[LTP] [PATCH v6 1/2] cpuset02: Convert the test6 from cpuset_memory_testset.sh to C code
Petr Vorel
pvorel@suse.cz
Tue Mar 25 14:36:11 CET 2025
> Hi!
> > diff --git a/lib/tst_cgroup.c b/lib/tst_cgroup.c
> > index 73b696c58..545c779e7 100644
> > --- a/lib/tst_cgroup.c
> > +++ b/lib/tst_cgroup.c
> > @@ -204,6 +204,7 @@ static const struct cgroup_file cpuset_ctrl_files[] = {
> > { "cpuset.cpus", "cpuset.cpus", CTRL_CPUSET },
> > { "cpuset.mems", "cpuset.mems", CTRL_CPUSET },
> > { "cpuset.memory_migrate", "cpuset.memory_migrate", CTRL_CPUSET },
> > + { "cpuset.sched_load_balance", "cpuset.sched_load_balance", CTRL_CPUSET },
> > { }
> > };
> > diff --git a/runtest/mm b/runtest/mm
> > index d8e62af81..5af29b0ea 100644
> > --- a/runtest/mm
> > +++ b/runtest/mm
> > @@ -75,6 +75,7 @@ ksm06_2 ksm06 -n 8000
> > ksm07 ksm07
> > cpuset01 cpuset01
> > +cpuset02 cpuset02
> > oom01 oom01
> > oom02 oom02
> > diff --git a/testcases/kernel/mem/.gitignore b/testcases/kernel/mem/.gitignore
> > index 699e022fb..e24e96001 100644
> > --- a/testcases/kernel/mem/.gitignore
> > +++ b/testcases/kernel/mem/.gitignore
> > @@ -1,4 +1,5 @@
> > /cpuset/cpuset01
> > +/cpuset/cpuset02
> > /hugetlb/hugefallocate/hugefallocate01
> > /hugetlb/hugefallocate/hugefallocate02
> > /hugetlb/hugefork/hugefork01
> > diff --git a/testcases/kernel/mem/cpuset/Makefile b/testcases/kernel/mem/cpuset/Makefile
> > index bac13e02b..7010c7be4 100644
> > --- a/testcases/kernel/mem/cpuset/Makefile
> > +++ b/testcases/kernel/mem/cpuset/Makefile
> > @@ -19,6 +19,11 @@
> > top_srcdir ?= ../../../..
> > +LTPLIBS = numa
> > +
> > include $(top_srcdir)/include/mk/testcases.mk
> > include $(top_srcdir)/testcases/kernel/include/lib.mk
> > +
> > +cpuset02: LTPLDLIBS = -lltpnuma
> > +
> > include $(top_srcdir)/include/mk/generic_leaf_target.mk
> > diff --git a/testcases/kernel/mem/cpuset/cpuset02.c b/testcases/kernel/mem/cpuset/cpuset02.c
> > new file mode 100644
> > index 000000000..f43d8e98a
> > --- /dev/null
> > +++ b/testcases/kernel/mem/cpuset/cpuset02.c
> > @@ -0,0 +1,139 @@
> > +// SPDX-License-Identifier: LGPL-2.1-or-later
> > +/*
> > + * Copyright (c) 2009 FUJITSU LIMITED Miao Xie <miaox@cn.fujitsu.com>
> > + * Copyright (c) 2023 SUSE LLC <wegao@suse.com>
> > + */
> > +
> > +/*\
> > + * Test checks cpuset.mems works with hugepage file.
> > + */
> > +
> > +#define _GNU_SOURCE
> > +#include <stdio.h>
> > +#include <sys/mount.h>
> > +#include <limits.h>
> > +#include <sys/param.h>
> > +#include <sys/types.h>
> > +#include "tst_test.h"
> > +
> > +#ifdef HAVE_NUMA_V2
> > +#include <numa.h>
> > +#include <numaif.h>
> > +#include <errno.h>
> > +#include "tst_numa.h"
> > +#include "tst_safe_stdio.h"
> > +#include "numa_helper.h"
> > +
> > +#define MNTPOINT "hugetlbfs/"
> > +#define HUGE_PAGE_FILE MNTPOINT "hugepagefile"
> > +#define PAGES_ALLOCATED 16u
> > +
> > +static long hpage_size;
> > +static struct tst_nodemap *node;
> > +static struct tst_cg_group *cg_cpuset_0;
> > +
> > +static void touch_memory_and_check_node(char *p, int size)
> > +{
> > + int i;
> > + int node = -1;
> > + long ret;
> > + int pagesize = sysconf(_SC_PAGESIZE);
> > +
> > + for (i = 0; i < size; i += pagesize)
> > + p[i] = 0xef;
> > +
> > + ret = get_mempolicy(&node, NULL, 0, p, MPOL_F_NODE | MPOL_F_ADDR);
> > + if (ret < 0)
> > + tst_brk(TBROK | TERRNO, "get_mempolicy() failed");
> > +
> > + if (node == 0)
> > + tst_res(TPASS, "check node pass");
> > + else
> > + tst_res(TFAIL, "check node failed");
> > +}
> > +
> > +static void child(void)
> > +{
> > + char *p;
> > + int fd_hugepage;
> > +
> > + TST_CHECKPOINT_WAIT(0);
> > +
> > + fd_hugepage = SAFE_OPEN(HUGE_PAGE_FILE, O_CREAT | O_RDWR, 0755);
> > + p = SAFE_MMAP(NULL, hpage_size, PROT_WRITE | PROT_READ,
> > + MAP_SHARED, fd_hugepage, 0);
> > +
> > + touch_memory_and_check_node(p, hpage_size);
> > +
> > + SAFE_MUNMAP(p, hpage_size);
> > + SAFE_CLOSE(fd_hugepage);
> > +
> > + TST_CHECKPOINT_WAKE(1);
> > +}
> > +
> > +static void run_test(void)
> > +{
> > + int pid;
> > +
> > + cg_cpuset_0 = tst_cg_group_mk(tst_cg, "0");
> > +
> > + SAFE_CG_PRINT(cg_cpuset_0, "cpuset.cpus", "0");
> > + SAFE_CG_PRINT(cg_cpuset_0, "cpuset.mems", "0");
> > + SAFE_CG_PRINT(cg_cpuset_0, "cpuset.sched_load_balance", "0");
> Here you are locking the process to node0 but that does not necessarily
> means that the node0 has enough memory or CPUs. You have to choose a
> node from the nodemap that you got in the setup.
> Also if you want to have a node with CPU, so that you can lock the
> process there, you need to add TST_NUMA_CPU flag to the
> tst_get_nodemap(). And yes, there are machines that can have node with
> just memory or just CPU.
> However in the touch_memory_and_check_node() we just check a node the
> memory was allocated on. So I do not think that we need to lock the
> process on a particular CPU. And I do not think that we need the se the
> sched_load_balance either.
[ commenting old version, there is v7 already ]
Thanks for catching these. For a record original shell test does not set
cpuset.cpus, only cpuset.mems.
> > + SAFE_FILE_PRINTF("/proc/sys/vm/nr_hugepages", "%d", 2 * node->cnt);
> What is this needed for and why is there the 2*node->cnt useful for? As
> far as I can tell, we allocate a single page on a given node.
For a record, shell test does:
echo $((2*$nr_mems)) > /proc/sys/vm/nr_hugepages
I don't have deep understanding of hugetlbpage and NUMA, but kernel hugetlbpage
allows more huge pages in the kernel's huge page pool:
https://www.kernel.org/doc/Documentation/vm/hugetlbpage.txt
echo 20 > /proc/sys/vm/nr_hugepages
Kind regards,
Petr
> > + pid = SAFE_FORK();
> > +
> > + if (!pid) {
> > + child();
> > + return;
> > + }
> > +
> > + SAFE_CG_PRINTF(cg_cpuset_0, "cgroup.procs", "%d", pid);
> We can just get rid of all the checkpoints if this is done by the child
> process before it calls the child() function.
> > + TST_CHECKPOINT_WAKE(0);
> > + TST_CHECKPOINT_WAIT(1);
> > +
> > + SAFE_WAITPID(pid, NULL, 0);
> We should not wait processes if we are not interested in their return
> value. Such children are collected automatically by the test library.
> > + cg_cpuset_0 = tst_cg_group_rm(cg_cpuset_0);
> > +}
> > +
> > +static void setup(void)
> > +{
> > + node = tst_get_nodemap(TST_NUMA_MEM, PAGES_ALLOCATED * getpagesize() / 1024);
> ^
> Why is this here? The
> child in the test
> allocates just a single
> page.
> > + if (node->cnt <= 1)
> > + tst_brk(TCONF, "test requires NUMA system");
> > +
> > + hpage_size = SAFE_READ_MEMINFO(MEMINFO_HPAGE_SIZE)*1024;
> > +}
> > +
> > +static void cleanup(void)
> > +{
> > + if (cg_cpuset_0)
> > + cg_cpuset_0 = tst_cg_group_rm(cg_cpuset_0);
> > +}
> > +
> > +static struct tst_test test = {
> > + .needs_root = 1,
> > + .runs_script = 1,
> > + .mntpoint = MNTPOINT,
> > + .needs_hugetlbfs = 1,
> > + .setup = setup,
> > + .forks_child = 1,
> > + .cleanup = cleanup,
> > + .test_all = run_test,
> > + .hugepages = {3, TST_NEEDS},
> > + .needs_checkpoints = 1,
> > + .needs_cgroup_ver = TST_CG_V1,
> > + .needs_cgroup_ctrls = (const char *const []){ "cpuset", NULL },
> > + .save_restore = (const struct tst_path_val[]) {
> > + {"/proc/sys/vm/nr_hugepages", NULL, TST_SR_TBROK},
> > + {}
> > + },
> > +};
> > +
> > +#else
> > +TST_TEST_TCONF(NUMA_ERROR_MSG);
> > +#endif
> > --
> > 2.35.3
> > --
> > Mailing list info: https://lists.linux.it/listinfo/ltp
More information about the ltp
mailing list