[LTP] [PATCH] Hugetlb: Test to detect bug with freeing gigantic hugetlb pages
Cyril Hrubis
chrubis@suse.cz
Thu Apr 13 11:22:21 CEST 2023
Hi!
> +/*\
> + * [Description]
> + *
> + * Before kernel version 5.10-rc7, there was a bug that resulted in a "Bad Page
> + * State" error when freeing gigantic hugepages. This happened because the
> + * struct page entry compound_nr, which overlapped with page->mapping in the
> + * first tail page, was not cleared, causing the error. To ensure that this
> + * issue does not reoccur as struct page keeps changing and some fields are
> + * managed by folio, this test checks that freeing gigantic hugepages does not
> + * produce the above-mentioned error.
> + */
> +
> +#define _GNU_SOURCE
> +#include <dirent.h>
> +
> +#include <stdio.h>
> +
> +#include "hugetlb.h"
> +
> +#define PATH_GIGANTIC_HUGEPAGE "/sys/kernel/mm/hugepages"
> +#define GIGANTIC_MIN_ORDER 10
> +
> +static int org_g_hpages;
> +static char g_hpage_path[4096];
> +
> +static void run_test(void)
> +{
> + SAFE_FILE_PRINTF(g_hpage_path, "%d", 1);
I suppose this may still fail if there is not enough memory or the
memory is fragmented, right? I suppose that SAFE_FILE_PRINTF() will
cause TBROK here, right?
Maybe we should just use FILE_PRINTF() and ignore the errors.
> + SAFE_FILE_PRINTF(g_hpage_path, "%d", 0);
> + if (tst_taint_check())
> + tst_res(TFAIL, "Freeing Gigantic pages resulted in Bad Page State bug.");
> + else
> + tst_res(TPASS, "Successfully freed the gigantic hugepages");
> +}
> +
> +static void setup(void)
> +{
> + DIR *dir;
> + struct dirent *ent;
> + char *hpage_size_str;
> + unsigned long hpage_size;
> +
> + dir = SAFE_OPENDIR(PATH_GIGANTIC_HUGEPAGE);
> + while ((ent = SAFE_READDIR(dir)) != NULL) {
^
The != NULL is redundant
> + if (strstr(ent->d_name, "hugepages-") != NULL) {
> + hpage_size_str = ent->d_name + strlen("hugepages-");
> + hpage_size = atoi(hpage_size_str) * 1024;
Can we just do:
if ((sscanf(ent->d_name, "hugepages-%lu", &hpage_size) == 1) &&
is_hugetlb_gigantic(hpage_size * 1024)) {
sprintf(...);
}
> + if (is_hugetlb_gigantic(hpage_size)) {
> + sprintf(g_hpage_path, "%s/%s/%s",
> + PATH_GIGANTIC_HUGEPAGE, ent->d_name, "nr_hugepages");
> + break;
> + }
> + }
> + }
> + SAFE_CLOSEDIR(dir);
We should handle the case that there were no large hugepages found on
the system. On one of my machines:
# ls /sys/kernel/mm/hugepages/
hugepages-2048kB
#
I suppose:
if (!g_hpage_path[0])
tst_brk(TCONF, "Gigantic hugepages not supported");
> + SAFE_FILE_LINES_SCANF(g_hpage_path, "%d", &org_g_hpages);
> +}
> +
> +static void cleanup(void)
> +{
> + SAFE_FILE_PRINTF(g_hpage_path, "%d", org_g_hpages);
> +}
> +
> +static struct tst_test test = {
> + .tags = (struct tst_tag[]) {
> + {"linux-git", "ba9c1201beaa"},
> + {"linux-git", "7118fc2906e9"},
> + {}
> + },
> + .needs_root = 1,
> + .setup = setup,
> + .cleanup = cleanup,
> + .test_all = run_test,
> + .taint_check = TST_TAINT_B,
> +};
> diff --git a/testcases/kernel/mem/hugetlb/lib/hugetlb.h b/testcases/kernel/mem/hugetlb/lib/hugetlb.h
> index 241dab708..34fe08c24 100644
> --- a/testcases/kernel/mem/hugetlb/lib/hugetlb.h
> +++ b/testcases/kernel/mem/hugetlb/lib/hugetlb.h
> @@ -39,6 +39,12 @@
> # endif
> #endif
>
> +/* Check if hugetlb page is gigantic */
> +static inline int is_hugetlb_gigantic(unsigned long hpage_size)
> +{
> + return (hpage_size / getpagesize()) >> 11;
> +}
> +
> /*
> * to get the lower nine permission bits
> * from shmid_ds.ipc_perm.mode
> --
> 2.31.1
>
--
Cyril Hrubis
chrubis@suse.cz
More information about the ltp
mailing list