[LTP] [PATCH v3] This test stress tests fallocate. This test starts three threads. First thread will continually punch/fill holes via falloc. Second thread will continually fault in those same pages. Third thread will continually mmap/munmap that page range.
Pavithra
pavrampu@linux.vnet.ibm.com
Mon Nov 25 15:29:32 CET 2024
Changes in v3:
- Used MIN() macro instead of defining min function.
- struch ts initialization done at the declaration.
- Added memory unmap before pthread_cleanup_pop.
- Removed if (falloc_th_ret)
Signed-off-by: Pavithra <pavrampu@linux.vnet.ibm.com>
---
runtest/hugetlb | 1 +
testcases/kernel/mem/.gitignore | 1 +
.../hugetlb/hugefallocate/hugefallocate03.c | 204 ++++++++++++++++++
3 files changed, 206 insertions(+)
create mode 100644 testcases/kernel/mem/hugetlb/hugefallocate/hugefallocate03.c
diff --git a/runtest/hugetlb b/runtest/hugetlb
index f294e9aaa..bba2460ff 100644
--- a/runtest/hugetlb
+++ b/runtest/hugetlb
@@ -1,5 +1,6 @@
hugefallocate01 hugefallocate01
hugefallocate02 hugefallocate02
+hugefallocate03 hugefallocate03
hugefork01 hugefork01
hugefork02 hugefork02
diff --git a/testcases/kernel/mem/.gitignore b/testcases/kernel/mem/.gitignore
index d88484fa1..2b84ac3b9 100644
--- a/testcases/kernel/mem/.gitignore
+++ b/testcases/kernel/mem/.gitignore
@@ -1,6 +1,7 @@
/cpuset/cpuset01
/hugetlb/hugefallocate/hugefallocate01
/hugetlb/hugefallocate/hugefallocate02
+/hugetlb/hugefallocate/hugefallocate03
/hugetlb/hugefork/hugefork01
/hugetlb/hugefork/hugefork02
/hugetlb/hugemmap/hugemmap01
diff --git a/testcases/kernel/mem/hugetlb/hugefallocate/hugefallocate03.c b/testcases/kernel/mem/hugetlb/hugefallocate/hugefallocate03.c
new file mode 100644
index 000000000..824816161
--- /dev/null
+++ b/testcases/kernel/mem/hugetlb/hugefallocate/hugefallocate03.c
@@ -0,0 +1,204 @@
+// SPDX-License-Identifier: LGPL-2.1-or-later
+/*
+ * Copyright (C) 2015 Oracle Corporation
+ * Author: Mike Kravetz
+ */
+
+/*\
+ * [Description]
+ *
+ * Stress test fallocate. This test starts three threads.
+ * Thread one will continually punch/fill holes via falloc.
+ * Thread two will continually fault in those same pages.
+ * Thread three will continually mmap/munmap that page range.
+ *
+ */
+
+#define _GNU_SOURCE
+#include <stdio.h>
+#include <sys/mount.h>
+#include <limits.h>
+#include <sys/param.h>
+#include <sys/types.h>
+#include <pthread.h>
+
+#include "hugetlb.h"
+#include "lapi/fallocate.h"
+#include "tst_safe_pthread.h"
+
+#define MNTPOINT "hugetlbfs/"
+#define MAX_PAGES_TO_USE 100
+#define FALLOCATE_ITERATIONS 100000
+
+static int fd = -1;
+static long nr_hpages_free;
+static unsigned long max_hpages;
+static int err;
+static long hpage_size;
+static unsigned long free_before, free_after;
+static unsigned long rsvd_before, rsvd_after;
+
+static void *thread_fallocate(void *)
+{
+ int i, err;
+ long tpage;
+
+ for (i = 0; i < FALLOCATE_ITERATIONS; i++) {
+ tpage = ((long long)random()) % (max_hpages);
+ err = fallocate(fd,
+ FALLOC_FL_PUNCH_HOLE | FALLOC_FL_KEEP_SIZE,
+ tpage * hpage_size, hpage_size);
+ if (err)
+ tst_res(TFAIL|TERRNO, "fallocate():");
+ err = fallocate(fd, 0, tpage * hpage_size, hpage_size);
+ if (err)
+ tst_res(TFAIL|TERRNO, "fallocate():");
+ }
+ return NULL;
+}
+
+static void *fault_mmap_addr;
+
+static void thread_fault_cleanup(void *)
+{
+ if (fault_mmap_addr)
+ munmap(fault_mmap_addr, max_hpages * hpage_size);
+}
+
+static void *thread_fault(void *)
+{
+ long tpage;
+ char foo;
+ struct timespec ts = {};
+
+ fault_mmap_addr = SAFE_MMAP(NULL, max_hpages * hpage_size,
+ PROT_READ | PROT_WRITE, MAP_SHARED,
+ fd, 0);
+
+ pthread_cleanup_push(thread_fault_cleanup, NULL);
+
+ while (1) {
+ tpage = ((long long)random()) % (max_hpages);
+ foo = *((char *)(fault_mmap_addr + (tpage * hpage_size)));
+ *((char *)(fault_mmap_addr + (tpage * hpage_size))) = foo;
+
+ nanosleep(&ts, NULL); /* thread cancellation point */
+ }
+
+ thread_fault_cleanup(NULL);
+
+ pthread_cleanup_pop(1);
+
+ return NULL;
+}
+
+static void *mmap_munmap_addr;
+
+static void thread_mmap_munmap_cleanup(void *)
+{
+ if (mmap_munmap_addr)
+ munmap(mmap_munmap_addr, max_hpages * hpage_size);
+}
+
+static void *thread_mmap_munmap(void *)
+{
+ struct timespec ts = {};
+
+ pthread_cleanup_push(thread_mmap_munmap_cleanup, NULL);
+
+ while (1) {
+ mmap_munmap_addr = SAFE_MMAP(NULL, max_hpages * hpage_size,
+ PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
+ SAFE_MUNMAP(mmap_munmap_addr, max_hpages * hpage_size);
+ mmap_munmap_addr = NULL;
+
+ nanosleep(&ts, NULL); /* thread cancellation point */
+ }
+
+ thread_mmap_munmap_cleanup(NULL);
+
+ pthread_cleanup_pop(1);
+ return NULL;
+}
+
+static void run_test(void)
+{
+ fd = tst_creat_unlinked(MNTPOINT, 0);
+ pthread_t falloc_th, fault_th, mmap_munmap_th;
+ void *falloc_th_ret, *fault_th_ret, *mmap_munmap_th_ret;
+
+ unsigned int seed = (int)getpid() * time(NULL);
+
+ srandom(seed);
+ tst_res(TINFO, "Seed = %d", seed);
+ nr_hpages_free = SAFE_READ_MEMINFO(MEMINFO_HPAGE_FREE);
+ max_hpages = MIN(nr_hpages_free, MAX_PAGES_TO_USE);
+ free_before = SAFE_READ_MEMINFO(MEMINFO_HPAGE_FREE);
+ rsvd_before = SAFE_READ_MEMINFO(MEMINFO_HPAGE_RSVD);
+
+ /* First preallocate file with max_hpages pages */
+ err = fallocate(fd, 0, 0, hpage_size * max_hpages);
+ if (err) {
+ if (errno == EOPNOTSUPP)
+ tst_brk(TCONF, "fallocate() Operation is not supported");
+ if (err) {
+ tst_res(TFAIL|TERRNO, "fallocate():");
+ goto windup;
+ }
+ }
+
+ free_after = SAFE_READ_MEMINFO(MEMINFO_HPAGE_FREE);
+ if (free_before - free_after != max_hpages) {
+ tst_res(TFAIL, "fallocate did not preallocate %ld huge pages\n",
+ max_hpages);
+ goto windup;
+ }
+
+ SAFE_PTHREAD_CREATE(&falloc_th, NULL, thread_fallocate, NULL);
+
+ SAFE_PTHREAD_CREATE(&fault_th, NULL, thread_fault, NULL);
+
+ SAFE_PTHREAD_CREATE(&mmap_munmap_th, NULL, thread_mmap_munmap, NULL);
+
+ SAFE_PTHREAD_JOIN(falloc_th, &falloc_th_ret);
+
+ SAFE_PTHREAD_CANCEL(fault_th);
+
+ SAFE_PTHREAD_JOIN(fault_th, &fault_th_ret);
+
+ SAFE_PTHREAD_CANCEL(mmap_munmap_th);
+
+ SAFE_PTHREAD_JOIN(mmap_munmap_th, &mmap_munmap_th_ret);
+
+windup:
+ SAFE_CLOSE(fd);
+
+ free_after = SAFE_READ_MEMINFO(MEMINFO_HPAGE_FREE);
+ rsvd_after = SAFE_READ_MEMINFO(MEMINFO_HPAGE_RSVD);
+ if (free_after != free_before || rsvd_after != rsvd_before)
+ tst_res(TFAIL, "free or reserve counts incorrect after fallocate stress test");
+ else
+ tst_res(TPASS, "fallocate stress test passed");
+}
+
+static void setup(void)
+{
+ hpage_size = tst_get_hugepage_size();
+}
+
+static void cleanup(void)
+{
+ if (fd > 0)
+ SAFE_CLOSE(fd);
+}
+
+static struct tst_test test = {
+ .needs_root = 1,
+ .mntpoint = MNTPOINT,
+ .needs_hugetlbfs = 1,
+ .needs_tmpdir = 1,
+ .setup = setup,
+ .cleanup = cleanup,
+ .test_all = run_test,
+ .hugepages = {2, TST_NEEDS},
+};
--
2.43.5
More information about the ltp
mailing list