[LTP] [PATCH 6/6] sched/cgroup: Add cfs_bandwidth01

Cyril Hrubis chrubis@suse.cz
Wed May 19 13:51:33 CEST 2021


On Thu, May 13, 2021 at 04:21:25PM +0100, Richard Palethorpe via ltp wrote:
> Signed-off-by: Richard Palethorpe <rpalethorpe@suse.com>
> ---
>  runtest/sched                                 |   1 +
>  .../kernel/sched/cfs-scheduler/.gitignore     |   1 +
>  testcases/kernel/sched/cfs-scheduler/Makefile |   4 +-
>  .../sched/cfs-scheduler/cfs_bandwidth01.c     | 175 ++++++++++++++++++
>  4 files changed, 179 insertions(+), 2 deletions(-)
>  create mode 100644 testcases/kernel/sched/cfs-scheduler/cfs_bandwidth01.c
> 
> diff --git a/runtest/sched b/runtest/sched
> index bfc4f2711..592898723 100644
> --- a/runtest/sched
> +++ b/runtest/sched
> @@ -6,6 +6,7 @@ pth_str03 pth_str03
>  time-schedule01		time-schedule
>  trace_sched01		trace_sched -c 1
>  
> +cfs_bandwidth01 cfs_bandwidth01 -i 5
>  hackbench01 hackbench 50 process 1000
>  hackbench02 hackbench 20 thread 1000
>  
> diff --git a/testcases/kernel/sched/cfs-scheduler/.gitignore b/testcases/kernel/sched/cfs-scheduler/.gitignore
> index db2759e4f..c5dacd6ef 100644
> --- a/testcases/kernel/sched/cfs-scheduler/.gitignore
> +++ b/testcases/kernel/sched/cfs-scheduler/.gitignore
> @@ -1 +1,2 @@
>  /hackbench
> +cfs_bandwidth01
> diff --git a/testcases/kernel/sched/cfs-scheduler/Makefile b/testcases/kernel/sched/cfs-scheduler/Makefile
> index aa3bf8459..2ffe1f7f9 100644
> --- a/testcases/kernel/sched/cfs-scheduler/Makefile
> +++ b/testcases/kernel/sched/cfs-scheduler/Makefile
> @@ -18,8 +18,8 @@
>  
>  top_srcdir		?= ../../../..
>  
> -include $(top_srcdir)/include/mk/env_pre.mk
> +include $(top_srcdir)/include/mk/testcases.mk
>  
> -LDLIBS			+= -lpthread
> +hackbench: LDLIBS			+= -lpthread
>  
>  include $(top_srcdir)/include/mk/generic_leaf_target.mk
> diff --git a/testcases/kernel/sched/cfs-scheduler/cfs_bandwidth01.c b/testcases/kernel/sched/cfs-scheduler/cfs_bandwidth01.c
> new file mode 100644
> index 000000000..b1f98d50f
> --- /dev/null
> +++ b/testcases/kernel/sched/cfs-scheduler/cfs_bandwidth01.c
> @@ -0,0 +1,175 @@
> +// SPDX-License-Identifier: GPL-2.0-or-later
> +/* Copyright (c) 2021 SUSE LLC <rpalethorpe@suse.com> */
> +/*\
> + *
> + * [Description]
> + *
> + * Creates a multi-level CGroup hierarchy with the cpu controller
> + * enabled. The leaf groups are populated with "busy" processes which
> + * simulate intermittent cpu load. They spin for some time then sleep
> + * then repeat.
> + *
> + * Both the trunk and leaf groups are set cpu bandwidth limits. The
> + * busy processes will intermittently exceed these limits. Causing
> + * them to be throttled. When they begin sleeping this will then cause
> + * them to be unthrottle.
> + *
> + * The test is known to reproduce an issue with an update to
> + * SLE-15-SP1 (kernel 4.12.14-197.64, bsc#1179093).
> + */
> +
> +#include <stdlib.h>
> +
> +#include "tst_test.h"
> +#include "tst_cgroup.h"
> +#include "tst_timer.h"
> +
> +static const struct tst_cgroup_group *cg_test;
> +static struct tst_cgroup_group *cg_level2, *cg_level3a, *cg_level3b;
> +static struct tst_cgroup_group *cg_workers[3];
> +
> +static void set_cpu_quota(const struct tst_cgroup_group *const cg,
> +			  const float quota_percent)
> +{
> +	const unsigned int period_us = 10000;
> +	const unsigned int quota_us = (quota_percent / 100) * (float)period_us;
> +
> +	if (TST_CGROUP_VER(cg, "cpu") != TST_CGROUP_V1) {
> +		SAFE_CGROUP_PRINTF(cg, "cpu.max",
> +				   "%u %u", quota_us, period_us);
> +	} else {
> +		SAFE_CGROUP_PRINTF(cg, "cpu.max",
> +				   "%u", quota_us);
> +		SAFE_CGROUP_PRINTF(cg, "cpu.cfs_period_us",
> +				  "%u", period_us);
> +	}
> +
> +	tst_res(TINFO, "Set '%s/cpu.max' = '%d %d'",
> +		tst_cgroup_group_name(cg), quota_us, period_us);
> +}
> +
> +static struct tst_cgroup_group *
> +mk_cpu_cgroup(const struct tst_cgroup_group *const cg_parent,
> +	      const char *const cg_child_name,
> +	      const float quota_percent)
> +{
> +	struct tst_cgroup_group *const cg =
> +		tst_cgroup_group_mk(cg_parent, cg_child_name);
> +
> +	set_cpu_quota(cg, quota_percent);
> +
> +	return cg;
> +}
> +
> +static void busy_loop(const unsigned int sleep_ms)
> +{
> +	for (;;) {
> +		tst_timer_start(CLOCK_MONOTONIC_RAW);
> +		while (!tst_timer_expired_ms(20))
> +			;
> +
> +		const int ret = tst_checkpoint_wait(0, sleep_ms);
> +
> +		if (!ret)
> +			exit(0);
> +
> +		if (errno != ETIMEDOUT)
> +			tst_brk(TBROK | TERRNO, "tst_checkpoint_wait");
> +	}
> +}
> +
> +static void fork_busy_procs_in_cgroup(const struct tst_cgroup_group *const cg)
> +{
> +	const unsigned int sleeps_ms[] = {3000, 1000, 10};
> +	const pid_t worker_pid = SAFE_FORK();
> +	size_t i;
> +
> +	if (worker_pid)
> +		return;
> +
> +	for (i = 0; i < ARRAY_SIZE(sleeps_ms); i++) {
> +		const pid_t busy_pid = SAFE_FORK();
> +
> +		if (!busy_pid)
> +			busy_loop(sleeps_ms[i]);
> +
> +		SAFE_CGROUP_PRINTF(cg, "cgroup.procs", "%d", busy_pid);
> +	}
> +
> +	tst_reap_children();
> +
> +	exit(0);
> +}
> +
> +static void do_test(void)
> +{
> +	size_t i;
> +
> +	cg_level2 = tst_cgroup_group_mk(cg_test, "level2");
> +
> +	cg_level3a = tst_cgroup_group_mk(cg_level2, "level3a");
> +	cg_workers[0] = mk_cpu_cgroup(cg_level3a, "worker1", 30);
> +	cg_workers[1] = mk_cpu_cgroup(cg_level3a, "worker2", 20);
> +
> +	cg_level3b = tst_cgroup_group_mk(cg_level2, "level3b");
> +	cg_workers[2] = mk_cpu_cgroup(cg_level3b, "worker3", 30);
> +
> +	for (i = 0; i < ARRAY_SIZE(cg_workers); i++)
> +		fork_busy_procs_in_cgroup(cg_workers[i]);
> +
> +	tst_res(TPASS, "Scheduled bandwidth constrained workers");
> +
> +	sleep(1);
> +
> +	set_cpu_quota(cg_level2, 50);
> +
> +	sleep(2);
> +
> +	TST_CHECKPOINT_WAKE2(0, 3 * 3);
> +	tst_reap_children();
> +
> +	tst_res(TPASS, "Workers exited");
> +}
> +
> +static void setup(void)
> +{
> +	tst_cgroup_require("cpu", NULL);
> +
> +	cg_test = tst_cgroup_get_test_group();
> +}
> +
> +static void cleanup(void)
> +{
> +	size_t i;
> +
> +	for (i = 0; i < ARRAY_SIZE(cg_workers); i++) {
> +		if (cg_workers[i])
> +			cg_workers[i] = tst_cgroup_group_rm(cg_workers[i]);
> +	}
> +
> +	if (cg_level3a)
> +		cg_level3a = tst_cgroup_group_rm(cg_level3a);
> +	if (cg_level3b)
> +		cg_level3b = tst_cgroup_group_rm(cg_level3b);
> +	if (cg_level2)
> +		cg_level2 = tst_cgroup_group_rm(cg_level2);

Hmm, I wonder if we can move this part of the cleanup to the test
library as well. If we add all cgroups the user has created into a FIFO
linked list then this could be implemented as a single loop in the
tst_cgroup_clean().

We would have to loop over the list in the tst_cgroup_group_rm() in
order to remove the about to be removed group from the list as well, but
I guess that this is still worth the trouble.

Other than that the test looks nice and clean.

> +	tst_cgroup_cleanup();
> +}
> +
> +static struct tst_test test = {
> +	.test_all = do_test,
> +	.setup = setup,
> +	.cleanup = cleanup,
> +	.forks_child = 1,
> +	.needs_checkpoints = 1,
> +	.taint_check = TST_TAINT_W | TST_TAINT_D,
> +	.tags = (const struct tst_tag[]) {
> +		{"linux-git", "39f23ce07b93"},
> +		{"linux-git", "b34cb07dde7c"},
> +		{"linux-git", "fe61468b2cbc"},
> +		{"linux-git", "5ab297bab984"},
> +		{"linux-git", "6d4d22468dae"},
> +		{ }
> +	}
> +};
> -- 
> 2.31.1
> 
> 
> -- 
> Mailing list info: https://lists.linux.it/listinfo/ltp

-- 
Cyril Hrubis
chrubis@suse.cz


More information about the ltp mailing list