* [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
@ 2025-07-16 3:05 ` Suren Baghdasaryan
2025-07-16 10:44 ` David Hildenbrand
2025-07-16 3:05 ` [PATCH v7 2/7] selftests/proc: extend /proc/pid/maps tearing test to include vma resizing Suren Baghdasaryan
` (6 subsequent siblings)
7 siblings, 1 reply; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 3:05 UTC (permalink / raw)
To: akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest, surenb
The /proc/pid/maps file is generated page by page, with the mmap_lock
released between pages. This can lead to inconsistent reads if the
underlying vmas are concurrently modified. For instance, if a vma split
or merge occurs at a page boundary while /proc/pid/maps is being read,
the same vma might be seen twice: once before and once after the change.
This duplication is considered acceptable for userspace handling.
However, observing a "hole" where a vma should be (e.g., due to a vma
being replaced and the space temporarily being empty) is unacceptable.
Implement a test that:
1. Forks a child process which continuously modifies its address space,
specifically targeting a vma at the boundary between two pages.
2. The parent process repeatedly reads the child's /proc/pid/maps.
3. The parent process checks the last vma of the first page and
the first vma of the second page for consistency, looking for the
effects of vma splits or merges.
The test duration is configurable via the -d command-line parameter
in seconds to increase the likelihood of catching the race condition.
The default test duration is 5 seconds.
Example Command: proc-maps-race -d 10
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
---
tools/testing/selftests/proc/.gitignore | 1 +
tools/testing/selftests/proc/Makefile | 1 +
tools/testing/selftests/proc/proc-maps-race.c | 459 ++++++++++++++++++
3 files changed, 461 insertions(+)
create mode 100644 tools/testing/selftests/proc/proc-maps-race.c
diff --git a/tools/testing/selftests/proc/.gitignore b/tools/testing/selftests/proc/.gitignore
index 973968f45bba..19bb333e2485 100644
--- a/tools/testing/selftests/proc/.gitignore
+++ b/tools/testing/selftests/proc/.gitignore
@@ -5,6 +5,7 @@
/proc-2-is-kthread
/proc-fsconfig-hidepid
/proc-loadavg-001
+/proc-maps-race
/proc-multiple-procfs
/proc-empty-vm
/proc-pid-vm
diff --git a/tools/testing/selftests/proc/Makefile b/tools/testing/selftests/proc/Makefile
index b12921b9794b..50aba102201a 100644
--- a/tools/testing/selftests/proc/Makefile
+++ b/tools/testing/selftests/proc/Makefile
@@ -9,6 +9,7 @@ TEST_GEN_PROGS += fd-002-posix-eq
TEST_GEN_PROGS += fd-003-kthread
TEST_GEN_PROGS += proc-2-is-kthread
TEST_GEN_PROGS += proc-loadavg-001
+TEST_GEN_PROGS += proc-maps-race
TEST_GEN_PROGS += proc-empty-vm
TEST_GEN_PROGS += proc-pid-vm
TEST_GEN_PROGS += proc-self-map-files-001
diff --git a/tools/testing/selftests/proc/proc-maps-race.c b/tools/testing/selftests/proc/proc-maps-race.c
new file mode 100644
index 000000000000..523afd83d34f
--- /dev/null
+++ b/tools/testing/selftests/proc/proc-maps-race.c
@@ -0,0 +1,459 @@
+/*
+ * Copyright (c) 2025 Suren Baghdasaryan <surenb@google.com>
+ *
+ * Permission to use, copy, modify, and distribute this software for any
+ * purpose with or without fee is hereby granted, provided that the above
+ * copyright notice and this permission notice appear in all copies.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
+ * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
+ * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
+ * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
+ * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
+ * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
+ * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
+ */
+/*
+ * Fork a child that concurrently modifies address space while the main
+ * process is reading /proc/$PID/maps and verifying the results. Address
+ * space modifications include:
+ * VMA splitting and merging
+ *
+ */
+#undef NDEBUG
+#include <assert.h>
+#include <errno.h>
+#include <fcntl.h>
+#include <pthread.h>
+#include <stdbool.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+#include <unistd.h>
+#include <sys/mman.h>
+#include <sys/stat.h>
+#include <sys/types.h>
+#include <sys/wait.h>
+
+static unsigned long test_duration_sec = 5UL;
+static int page_size;
+
+/* /proc/pid/maps parsing routines */
+struct page_content {
+ char *data;
+ ssize_t size;
+};
+
+#define LINE_MAX_SIZE 256
+
+struct line_content {
+ char text[LINE_MAX_SIZE];
+ unsigned long start_addr;
+ unsigned long end_addr;
+};
+
+static void read_two_pages(int maps_fd, struct page_content *page1,
+ struct page_content *page2)
+{
+ ssize_t bytes_read;
+
+ assert(lseek(maps_fd, 0, SEEK_SET) >= 0);
+ bytes_read = read(maps_fd, page1->data, page_size);
+ assert(bytes_read > 0 && bytes_read < page_size);
+ page1->size = bytes_read;
+
+ bytes_read = read(maps_fd, page2->data, page_size);
+ assert(bytes_read > 0 && bytes_read < page_size);
+ page2->size = bytes_read;
+}
+
+static void copy_first_line(struct page_content *page, char *first_line)
+{
+ char *pos = strchr(page->data, '\n');
+
+ strncpy(first_line, page->data, pos - page->data);
+ first_line[pos - page->data] = '\0';
+}
+
+static void copy_last_line(struct page_content *page, char *last_line)
+{
+ /* Get the last line in the first page */
+ const char *end = page->data + page->size - 1;
+ /* skip last newline */
+ const char *pos = end - 1;
+
+ /* search previous newline */
+ while (pos[-1] != '\n')
+ pos--;
+ strncpy(last_line, pos, end - pos);
+ last_line[end - pos] = '\0';
+}
+
+/* Read the last line of the first page and the first line of the second page */
+static void read_boundary_lines(int maps_fd, struct page_content *page1,
+ struct page_content *page2,
+ struct line_content *last_line,
+ struct line_content *first_line)
+{
+ read_two_pages(maps_fd, page1, page2);
+
+ copy_last_line(page1, last_line->text);
+ copy_first_line(page2, first_line->text);
+
+ assert(sscanf(last_line->text, "%lx-%lx", &last_line->start_addr,
+ &last_line->end_addr) == 2);
+ assert(sscanf(first_line->text, "%lx-%lx", &first_line->start_addr,
+ &first_line->end_addr) == 2);
+}
+
+/* Thread synchronization routines */
+enum test_state {
+ INIT,
+ CHILD_READY,
+ PARENT_READY,
+ SETUP_READY,
+ SETUP_MODIFY_MAPS,
+ SETUP_MAPS_MODIFIED,
+ SETUP_RESTORE_MAPS,
+ SETUP_MAPS_RESTORED,
+ TEST_READY,
+ TEST_DONE,
+};
+
+struct vma_modifier_info;
+
+typedef void (*vma_modifier_op)(const struct vma_modifier_info *mod_info);
+typedef void (*vma_mod_result_check_op)(struct line_content *mod_last_line,
+ struct line_content *mod_first_line,
+ struct line_content *restored_last_line,
+ struct line_content *restored_first_line);
+
+struct vma_modifier_info {
+ int vma_count;
+ void *addr;
+ int prot;
+ void *next_addr;
+ vma_modifier_op vma_modify;
+ vma_modifier_op vma_restore;
+ vma_mod_result_check_op vma_mod_check;
+ pthread_mutex_t sync_lock;
+ pthread_cond_t sync_cond;
+ enum test_state curr_state;
+ bool exit;
+ void *child_mapped_addr[];
+};
+
+static void wait_for_state(struct vma_modifier_info *mod_info, enum test_state state)
+{
+ pthread_mutex_lock(&mod_info->sync_lock);
+ while (mod_info->curr_state != state)
+ pthread_cond_wait(&mod_info->sync_cond, &mod_info->sync_lock);
+ pthread_mutex_unlock(&mod_info->sync_lock);
+}
+
+static void signal_state(struct vma_modifier_info *mod_info, enum test_state state)
+{
+ pthread_mutex_lock(&mod_info->sync_lock);
+ mod_info->curr_state = state;
+ pthread_cond_signal(&mod_info->sync_cond);
+ pthread_mutex_unlock(&mod_info->sync_lock);
+}
+
+/* VMA modification routines */
+static void *child_vma_modifier(struct vma_modifier_info *mod_info)
+{
+ int prot = PROT_READ | PROT_WRITE;
+ int i;
+
+ for (i = 0; i < mod_info->vma_count; i++) {
+ mod_info->child_mapped_addr[i] = mmap(NULL, page_size * 3, prot,
+ MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
+ assert(mod_info->child_mapped_addr[i] != MAP_FAILED);
+ /* change protection in adjacent maps to prevent merging */
+ prot ^= PROT_WRITE;
+ }
+ signal_state(mod_info, CHILD_READY);
+ wait_for_state(mod_info, PARENT_READY);
+ while (true) {
+ signal_state(mod_info, SETUP_READY);
+ wait_for_state(mod_info, SETUP_MODIFY_MAPS);
+ if (mod_info->exit)
+ break;
+
+ mod_info->vma_modify(mod_info);
+ signal_state(mod_info, SETUP_MAPS_MODIFIED);
+ wait_for_state(mod_info, SETUP_RESTORE_MAPS);
+ mod_info->vma_restore(mod_info);
+ signal_state(mod_info, SETUP_MAPS_RESTORED);
+
+ wait_for_state(mod_info, TEST_READY);
+ while (mod_info->curr_state != TEST_DONE) {
+ mod_info->vma_modify(mod_info);
+ mod_info->vma_restore(mod_info);
+ }
+ }
+ for (i = 0; i < mod_info->vma_count; i++)
+ munmap(mod_info->child_mapped_addr[i], page_size * 3);
+
+ return NULL;
+}
+
+static void stop_vma_modifier(struct vma_modifier_info *mod_info)
+{
+ wait_for_state(mod_info, SETUP_READY);
+ mod_info->exit = true;
+ signal_state(mod_info, SETUP_MODIFY_MAPS);
+}
+
+static void capture_mod_pattern(int maps_fd,
+ struct vma_modifier_info *mod_info,
+ struct page_content *page1,
+ struct page_content *page2,
+ struct line_content *last_line,
+ struct line_content *first_line,
+ struct line_content *mod_last_line,
+ struct line_content *mod_first_line,
+ struct line_content *restored_last_line,
+ struct line_content *restored_first_line)
+{
+ signal_state(mod_info, SETUP_MODIFY_MAPS);
+ wait_for_state(mod_info, SETUP_MAPS_MODIFIED);
+
+ /* Copy last line of the first page and first line of the last page */
+ read_boundary_lines(maps_fd, page1, page2, mod_last_line, mod_first_line);
+
+ signal_state(mod_info, SETUP_RESTORE_MAPS);
+ wait_for_state(mod_info, SETUP_MAPS_RESTORED);
+
+ /* Copy last line of the first page and first line of the last page */
+ read_boundary_lines(maps_fd, page1, page2, restored_last_line, restored_first_line);
+
+ mod_info->vma_mod_check(mod_last_line, mod_first_line,
+ restored_last_line, restored_first_line);
+
+ /*
+ * The content of these lines after modify+resore should be the same
+ * as the original.
+ */
+ assert(strcmp(restored_last_line->text, last_line->text) == 0);
+ assert(strcmp(restored_first_line->text, first_line->text) == 0);
+}
+
+static inline void split_vma(const struct vma_modifier_info *mod_info)
+{
+ assert(mmap(mod_info->addr, page_size, mod_info->prot | PROT_EXEC,
+ MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED,
+ -1, 0) != MAP_FAILED);
+}
+
+static inline void merge_vma(const struct vma_modifier_info *mod_info)
+{
+ assert(mmap(mod_info->addr, page_size, mod_info->prot,
+ MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED,
+ -1, 0) != MAP_FAILED);
+}
+
+static inline void check_split_result(struct line_content *mod_last_line,
+ struct line_content *mod_first_line,
+ struct line_content *restored_last_line,
+ struct line_content *restored_first_line)
+{
+ /* Make sure vmas at the boundaries are changing */
+ assert(strcmp(mod_last_line->text, restored_last_line->text) != 0);
+ assert(strcmp(mod_first_line->text, restored_first_line->text) != 0);
+}
+
+static void test_maps_tearing_from_split(int maps_fd,
+ struct vma_modifier_info *mod_info,
+ struct page_content *page1,
+ struct page_content *page2,
+ struct line_content *last_line,
+ struct line_content *first_line)
+{
+ struct line_content split_last_line;
+ struct line_content split_first_line;
+ struct line_content restored_last_line;
+ struct line_content restored_first_line;
+
+ wait_for_state(mod_info, SETUP_READY);
+
+ /* re-read the file to avoid using stale data from previous test */
+ read_boundary_lines(maps_fd, page1, page2, last_line, first_line);
+
+ mod_info->vma_modify = split_vma;
+ mod_info->vma_restore = merge_vma;
+ mod_info->vma_mod_check = check_split_result;
+
+ capture_mod_pattern(maps_fd, mod_info, page1, page2, last_line, first_line,
+ &split_last_line, &split_first_line,
+ &restored_last_line, &restored_first_line);
+
+ /* Now start concurrent modifications for test_duration_sec */
+ signal_state(mod_info, TEST_READY);
+
+ struct line_content new_last_line;
+ struct line_content new_first_line;
+ struct timespec start_ts, end_ts;
+
+ clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
+ do {
+ bool last_line_changed;
+ bool first_line_changed;
+
+ read_boundary_lines(maps_fd, page1, page2, &new_last_line, &new_first_line);
+
+ /* Check if we read vmas after split */
+ if (!strcmp(new_last_line.text, split_last_line.text)) {
+ /*
+ * The vmas should be consistent with split results,
+ * however if vma was concurrently restored after a
+ * split, it can be reported twice (first the original
+ * split one, then the same vma but extended after the
+ * merge) because we found it as the next vma again.
+ * In that case new first line will be the same as the
+ * last restored line.
+ */
+ assert(!strcmp(new_first_line.text, split_first_line.text) ||
+ !strcmp(new_first_line.text, restored_last_line.text));
+ } else {
+ /* The vmas should be consistent with merge results */
+ assert(!strcmp(new_last_line.text, restored_last_line.text) &&
+ !strcmp(new_first_line.text, restored_first_line.text));
+ }
+ /*
+ * First and last lines should change in unison. If the last
+ * line changed then the first line should change as well and
+ * vice versa.
+ */
+ last_line_changed = strcmp(new_last_line.text, last_line->text) != 0;
+ first_line_changed = strcmp(new_first_line.text, first_line->text) != 0;
+ assert(last_line_changed == first_line_changed);
+
+ clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
+ } while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
+
+ /* Signal the modifyer thread to stop and wait until it exits */
+ signal_state(mod_info, TEST_DONE);
+}
+
+int usage(void)
+{
+ fprintf(stderr, "Userland /proc/pid/{s}maps race test cases\n");
+ fprintf(stderr, " -d: Duration for time-consuming tests\n");
+ fprintf(stderr, " -h: Help screen\n");
+ exit(-1);
+}
+
+int main(int argc, char **argv)
+{
+ struct vma_modifier_info *mod_info;
+ pthread_mutexattr_t mutex_attr;
+ pthread_condattr_t cond_attr;
+ int shared_mem_size;
+ char fname[32];
+ int vma_count;
+ int maps_fd;
+ int status;
+ pid_t pid;
+ int opt;
+
+ while ((opt = getopt(argc, argv, "d:h")) != -1) {
+ if (opt == 'd')
+ test_duration_sec = strtoul(optarg, NULL, 0);
+ else if (opt == 'h')
+ usage();
+ }
+
+ page_size = sysconf(_SC_PAGESIZE);
+ /*
+ * Have to map enough vmas for /proc/pid/maps to contain more than one
+ * page worth of vmas. Assume at least 32 bytes per line in maps output
+ */
+ vma_count = page_size / 32 + 1;
+ shared_mem_size = sizeof(struct vma_modifier_info) + vma_count * sizeof(void *);
+
+ /* map shared memory for communication with the child process */
+ mod_info = (struct vma_modifier_info *)mmap(NULL, shared_mem_size,
+ PROT_READ | PROT_WRITE, MAP_SHARED | MAP_ANONYMOUS, -1, 0);
+
+ assert(mod_info != MAP_FAILED);
+
+ /* Initialize shared members */
+ pthread_mutexattr_init(&mutex_attr);
+ pthread_mutexattr_setpshared(&mutex_attr, PTHREAD_PROCESS_SHARED);
+ assert(!pthread_mutex_init(&mod_info->sync_lock, &mutex_attr));
+ pthread_condattr_init(&cond_attr);
+ pthread_condattr_setpshared(&cond_attr, PTHREAD_PROCESS_SHARED);
+ assert(!pthread_cond_init(&mod_info->sync_cond, &cond_attr));
+ mod_info->vma_count = vma_count;
+ mod_info->curr_state = INIT;
+ mod_info->exit = false;
+
+ pid = fork();
+ if (!pid) {
+ /* Child process */
+ child_vma_modifier(mod_info);
+ return 0;
+ }
+
+ sprintf(fname, "/proc/%d/maps", pid);
+ maps_fd = open(fname, O_RDONLY);
+ assert(maps_fd != -1);
+
+ /* Wait for the child to map the VMAs */
+ wait_for_state(mod_info, CHILD_READY);
+
+ /* Read first two pages */
+ struct page_content page1;
+ struct page_content page2;
+
+ page1.data = malloc(page_size);
+ assert(page1.data);
+ page2.data = malloc(page_size);
+ assert(page2.data);
+
+ struct line_content last_line;
+ struct line_content first_line;
+
+ read_boundary_lines(maps_fd, &page1, &page2, &last_line, &first_line);
+
+ /*
+ * Find the addresses corresponding to the last line in the first page
+ * and the first line in the last page.
+ */
+ mod_info->addr = NULL;
+ mod_info->next_addr = NULL;
+ for (int i = 0; i < mod_info->vma_count; i++) {
+ if (mod_info->child_mapped_addr[i] == (void *)last_line.start_addr) {
+ mod_info->addr = mod_info->child_mapped_addr[i];
+ mod_info->prot = PROT_READ;
+ /* Even VMAs have write permission */
+ if ((i % 2) == 0)
+ mod_info->prot |= PROT_WRITE;
+ } else if (mod_info->child_mapped_addr[i] == (void *)first_line.start_addr) {
+ mod_info->next_addr = mod_info->child_mapped_addr[i];
+ }
+
+ if (mod_info->addr && mod_info->next_addr)
+ break;
+ }
+ assert(mod_info->addr && mod_info->next_addr);
+
+ signal_state(mod_info, PARENT_READY);
+
+ test_maps_tearing_from_split(maps_fd, mod_info, &page1, &page2,
+ &last_line, &first_line);
+
+ stop_vma_modifier(mod_info);
+
+ free(page2.data);
+ free(page1.data);
+
+ for (int i = 0; i < vma_count; i++)
+ munmap(mod_info->child_mapped_addr[i], page_size);
+ close(maps_fd);
+ waitpid(pid, &status, 0);
+ munmap(mod_info, shared_mem_size);
+
+ return 0;
+}
--
2.50.0.727.gbf7dc18ff4-goog
^ permalink raw reply related [flat|nested] 20+ messages in thread
* Re: [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test
2025-07-16 3:05 ` [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test Suren Baghdasaryan
@ 2025-07-16 10:44 ` David Hildenbrand
2025-07-16 10:50 ` Lorenzo Stoakes
0 siblings, 1 reply; 20+ messages in thread
From: David Hildenbrand @ 2025-07-16 10:44 UTC (permalink / raw)
To: Suren Baghdasaryan, akpm
Cc: Liam.Howlett, lorenzo.stoakes, vbabka, peterx, jannh, hannes,
mhocko, paulmck, shuah, adobriyan, brauner, josef, yebin10, linux,
willy, osalvador, andrii, ryan.roberts, christophe.leroy,
tjmercier, kaleshsingh, aha310510, linux-kernel, linux-fsdevel,
linux-mm, linux-kselftest
On 16.07.25 05:05, Suren Baghdasaryan wrote:
> The /proc/pid/maps file is generated page by page, with the mmap_lock
> released between pages. This can lead to inconsistent reads if the
> underlying vmas are concurrently modified. For instance, if a vma split
> or merge occurs at a page boundary while /proc/pid/maps is being read,
> the same vma might be seen twice: once before and once after the change.
> This duplication is considered acceptable for userspace handling.
> However, observing a "hole" where a vma should be (e.g., due to a vma
> being replaced and the space temporarily being empty) is unacceptable.
>
> Implement a test that:
> 1. Forks a child process which continuously modifies its address space,
> specifically targeting a vma at the boundary between two pages.
> 2. The parent process repeatedly reads the child's /proc/pid/maps.
> 3. The parent process checks the last vma of the first page and
> the first vma of the second page for consistency, looking for the
> effects of vma splits or merges.
>
> The test duration is configurable via the -d command-line parameter
> in seconds to increase the likelihood of catching the race condition.
> The default test duration is 5 seconds.
>
> Example Command: proc-maps-race -d 10
>
> Signed-off-by: Suren Baghdasaryan <surenb@google.com>
Why is this selftest not making use of any kselftest framework?
I'm sure there is a very good reason :)
Reading assert() feels very weird compared to other selftests.
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test
2025-07-16 10:44 ` David Hildenbrand
@ 2025-07-16 10:50 ` Lorenzo Stoakes
2025-07-16 14:20 ` Suren Baghdasaryan
0 siblings, 1 reply; 20+ messages in thread
From: Lorenzo Stoakes @ 2025-07-16 10:50 UTC (permalink / raw)
To: David Hildenbrand
Cc: Suren Baghdasaryan, akpm, Liam.Howlett, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest
On Wed, Jul 16, 2025 at 12:44:23PM +0200, David Hildenbrand wrote:
> On 16.07.25 05:05, Suren Baghdasaryan wrote:
> > The /proc/pid/maps file is generated page by page, with the mmap_lock
> > released between pages. This can lead to inconsistent reads if the
> > underlying vmas are concurrently modified. For instance, if a vma split
> > or merge occurs at a page boundary while /proc/pid/maps is being read,
> > the same vma might be seen twice: once before and once after the change.
> > This duplication is considered acceptable for userspace handling.
> > However, observing a "hole" where a vma should be (e.g., due to a vma
> > being replaced and the space temporarily being empty) is unacceptable.
> >
> > Implement a test that:
> > 1. Forks a child process which continuously modifies its address space,
> > specifically targeting a vma at the boundary between two pages.
> > 2. The parent process repeatedly reads the child's /proc/pid/maps.
> > 3. The parent process checks the last vma of the first page and
> > the first vma of the second page for consistency, looking for the
> > effects of vma splits or merges.
> >
> > The test duration is configurable via the -d command-line parameter
> > in seconds to increase the likelihood of catching the race condition.
> > The default test duration is 5 seconds.
> >
> > Example Command: proc-maps-race -d 10
> >
> > Signed-off-by: Suren Baghdasaryan <surenb@google.com>
>
> Why is this selftest not making use of any kselftest framework?
>
> I'm sure there is a very good reason :)
>
> Reading assert() feels very weird compared to other selftests.
Sorry to meta-review via your review again David :P
But just to say tools/testing/selftests/kselftest_harness.h is really good, and
makes life simple. See tools/testing/selftests/mm/guard-regions.c for an example
of how they can be used - pretty straightforward and avoids a lot of kselftest
boilerplate.
>
> --
> Cheers,
>
> David / dhildenb
>
>
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test
2025-07-16 10:50 ` Lorenzo Stoakes
@ 2025-07-16 14:20 ` Suren Baghdasaryan
2025-07-16 16:44 ` Suren Baghdasaryan
0 siblings, 1 reply; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 14:20 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: David Hildenbrand, akpm, Liam.Howlett, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest
On Wed, Jul 16, 2025 at 3:50 AM Lorenzo Stoakes
<lorenzo.stoakes@oracle.com> wrote:
>
> On Wed, Jul 16, 2025 at 12:44:23PM +0200, David Hildenbrand wrote:
> > On 16.07.25 05:05, Suren Baghdasaryan wrote:
> > > The /proc/pid/maps file is generated page by page, with the mmap_lock
> > > released between pages. This can lead to inconsistent reads if the
> > > underlying vmas are concurrently modified. For instance, if a vma split
> > > or merge occurs at a page boundary while /proc/pid/maps is being read,
> > > the same vma might be seen twice: once before and once after the change.
> > > This duplication is considered acceptable for userspace handling.
> > > However, observing a "hole" where a vma should be (e.g., due to a vma
> > > being replaced and the space temporarily being empty) is unacceptable.
> > >
> > > Implement a test that:
> > > 1. Forks a child process which continuously modifies its address space,
> > > specifically targeting a vma at the boundary between two pages.
> > > 2. The parent process repeatedly reads the child's /proc/pid/maps.
> > > 3. The parent process checks the last vma of the first page and
> > > the first vma of the second page for consistency, looking for the
> > > effects of vma splits or merges.
> > >
> > > The test duration is configurable via the -d command-line parameter
> > > in seconds to increase the likelihood of catching the race condition.
> > > The default test duration is 5 seconds.
> > >
> > > Example Command: proc-maps-race -d 10
> > >
> > > Signed-off-by: Suren Baghdasaryan <surenb@google.com>
> >
> > Why is this selftest not making use of any kselftest framework?
> >
> > I'm sure there is a very good reason :)
It used to be a part of proc-pid-vm.c and after the split I kept its
overall structure. I'll look into using the kselftest framework.
Thanks!
> >
> > Reading assert() feels very weird compared to other selftests.
>
> Sorry to meta-review via your review again David :P
>
> But just to say tools/testing/selftests/kselftest_harness.h is really good, and
> makes life simple. See tools/testing/selftests/mm/guard-regions.c for an example
> of how they can be used - pretty straightforward and avoids a lot of kselftest
> boilerplate.
Thanks for the pointers. I need to figure out a way to pass
command-line parameters to my test. Maybe I can use fixtures for
that... Let me read more about it.
>
> >
> > --
> > Cheers,
> >
> > David / dhildenb
> >
> >
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test
2025-07-16 14:20 ` Suren Baghdasaryan
@ 2025-07-16 16:44 ` Suren Baghdasaryan
0 siblings, 0 replies; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 16:44 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: David Hildenbrand, akpm, Liam.Howlett, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest
On Wed, Jul 16, 2025 at 7:20 AM Suren Baghdasaryan <surenb@google.com> wrote:
>
> On Wed, Jul 16, 2025 at 3:50 AM Lorenzo Stoakes
> <lorenzo.stoakes@oracle.com> wrote:
> >
> > On Wed, Jul 16, 2025 at 12:44:23PM +0200, David Hildenbrand wrote:
> > > On 16.07.25 05:05, Suren Baghdasaryan wrote:
> > > > The /proc/pid/maps file is generated page by page, with the mmap_lock
> > > > released between pages. This can lead to inconsistent reads if the
> > > > underlying vmas are concurrently modified. For instance, if a vma split
> > > > or merge occurs at a page boundary while /proc/pid/maps is being read,
> > > > the same vma might be seen twice: once before and once after the change.
> > > > This duplication is considered acceptable for userspace handling.
> > > > However, observing a "hole" where a vma should be (e.g., due to a vma
> > > > being replaced and the space temporarily being empty) is unacceptable.
> > > >
> > > > Implement a test that:
> > > > 1. Forks a child process which continuously modifies its address space,
> > > > specifically targeting a vma at the boundary between two pages.
> > > > 2. The parent process repeatedly reads the child's /proc/pid/maps.
> > > > 3. The parent process checks the last vma of the first page and
> > > > the first vma of the second page for consistency, looking for the
> > > > effects of vma splits or merges.
> > > >
> > > > The test duration is configurable via the -d command-line parameter
> > > > in seconds to increase the likelihood of catching the race condition.
> > > > The default test duration is 5 seconds.
> > > >
> > > > Example Command: proc-maps-race -d 10
> > > >
> > > > Signed-off-by: Suren Baghdasaryan <surenb@google.com>
> > >
> > > Why is this selftest not making use of any kselftest framework?
> > >
> > > I'm sure there is a very good reason :)
>
> It used to be a part of proc-pid-vm.c and after the split I kept its
> overall structure. I'll look into using the kselftest framework.
> Thanks!
>
> > >
> > > Reading assert() feels very weird compared to other selftests.
> >
> > Sorry to meta-review via your review again David :P
> >
> > But just to say tools/testing/selftests/kselftest_harness.h is really good, and
> > makes life simple. See tools/testing/selftests/mm/guard-regions.c for an example
> > of how they can be used - pretty straightforward and avoids a lot of kselftest
> > boilerplate.
>
> Thanks for the pointers. I need to figure out a way to pass
> command-line parameters to my test. Maybe I can use fixtures for
> that... Let me read more about it.
Ok, I think I'll use environment variables to set verbosity and test
duration. That seems like the cleanest approach.
>
> >
> > >
> > > --
> > > Cheers,
> > >
> > > David / dhildenb
> > >
> > >
^ permalink raw reply [flat|nested] 20+ messages in thread
* [PATCH v7 2/7] selftests/proc: extend /proc/pid/maps tearing test to include vma resizing
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
2025-07-16 3:05 ` [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test Suren Baghdasaryan
@ 2025-07-16 3:05 ` Suren Baghdasaryan
2025-07-16 3:05 ` [PATCH v7 3/7] selftests/proc: extend /proc/pid/maps tearing test to include vma remapping Suren Baghdasaryan
` (5 subsequent siblings)
7 siblings, 0 replies; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 3:05 UTC (permalink / raw)
To: akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest, surenb
Test that /proc/pid/maps does not report unexpected holes in the address
space when a vma at the edge of the page is being concurrently remapped.
This remapping results in the vma shrinking and expanding from under the
reader. We should always see either shrunk or expanded (original) version
of the vma.
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
---
tools/testing/selftests/proc/proc-maps-race.c | 83 +++++++++++++++++++
1 file changed, 83 insertions(+)
diff --git a/tools/testing/selftests/proc/proc-maps-race.c b/tools/testing/selftests/proc/proc-maps-race.c
index 523afd83d34f..10365b4e68e1 100644
--- a/tools/testing/selftests/proc/proc-maps-race.c
+++ b/tools/testing/selftests/proc/proc-maps-race.c
@@ -336,6 +336,86 @@ static void test_maps_tearing_from_split(int maps_fd,
signal_state(mod_info, TEST_DONE);
}
+static inline void shrink_vma(const struct vma_modifier_info *mod_info)
+{
+ assert(mremap(mod_info->addr, page_size * 3, page_size, 0) != MAP_FAILED);
+}
+
+static inline void expand_vma(const struct vma_modifier_info *mod_info)
+{
+ assert(mremap(mod_info->addr, page_size, page_size * 3, 0) != MAP_FAILED);
+}
+
+static inline void check_shrink_result(struct line_content *mod_last_line,
+ struct line_content *mod_first_line,
+ struct line_content *restored_last_line,
+ struct line_content *restored_first_line)
+{
+ /* Make sure only the last vma of the first page is changing */
+ assert(strcmp(mod_last_line->text, restored_last_line->text) != 0);
+ assert(strcmp(mod_first_line->text, restored_first_line->text) == 0);
+}
+
+static void test_maps_tearing_from_resize(int maps_fd,
+ struct vma_modifier_info *mod_info,
+ struct page_content *page1,
+ struct page_content *page2,
+ struct line_content *last_line,
+ struct line_content *first_line)
+{
+ struct line_content shrunk_last_line;
+ struct line_content shrunk_first_line;
+ struct line_content restored_last_line;
+ struct line_content restored_first_line;
+
+ wait_for_state(mod_info, SETUP_READY);
+
+ /* re-read the file to avoid using stale data from previous test */
+ read_boundary_lines(maps_fd, page1, page2, last_line, first_line);
+
+ mod_info->vma_modify = shrink_vma;
+ mod_info->vma_restore = expand_vma;
+ mod_info->vma_mod_check = check_shrink_result;
+
+ capture_mod_pattern(maps_fd, mod_info, page1, page2, last_line, first_line,
+ &shrunk_last_line, &shrunk_first_line,
+ &restored_last_line, &restored_first_line);
+
+ /* Now start concurrent modifications for test_duration_sec */
+ signal_state(mod_info, TEST_READY);
+
+ struct line_content new_last_line;
+ struct line_content new_first_line;
+ struct timespec start_ts, end_ts;
+
+ clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
+ do {
+ read_boundary_lines(maps_fd, page1, page2, &new_last_line, &new_first_line);
+
+ /* Check if we read vmas after shrinking it */
+ if (!strcmp(new_last_line.text, shrunk_last_line.text)) {
+ /*
+ * The vmas should be consistent with shrunk results,
+ * however if the vma was concurrently restored, it
+ * can be reported twice (first as shrunk one, then
+ * as restored one) because we found it as the next vma
+ * again. In that case new first line will be the same
+ * as the last restored line.
+ */
+ assert(!strcmp(new_first_line.text, shrunk_first_line.text) ||
+ !strcmp(new_first_line.text, restored_last_line.text));
+ } else {
+ /* The vmas should be consistent with the original/resored state */
+ assert(!strcmp(new_last_line.text, restored_last_line.text) &&
+ !strcmp(new_first_line.text, restored_first_line.text));
+ }
+ clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
+ } while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
+
+ /* Signal the modifyer thread to stop and wait until it exits */
+ signal_state(mod_info, TEST_DONE);
+}
+
int usage(void)
{
fprintf(stderr, "Userland /proc/pid/{s}maps race test cases\n");
@@ -444,6 +524,9 @@ int main(int argc, char **argv)
test_maps_tearing_from_split(maps_fd, mod_info, &page1, &page2,
&last_line, &first_line);
+ test_maps_tearing_from_resize(maps_fd, mod_info, &page1, &page2,
+ &last_line, &first_line);
+
stop_vma_modifier(mod_info);
free(page2.data);
--
2.50.0.727.gbf7dc18ff4-goog
^ permalink raw reply related [flat|nested] 20+ messages in thread
* [PATCH v7 3/7] selftests/proc: extend /proc/pid/maps tearing test to include vma remapping
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
2025-07-16 3:05 ` [PATCH v7 1/7] selftests/proc: add /proc/pid/maps tearing from vma split test Suren Baghdasaryan
2025-07-16 3:05 ` [PATCH v7 2/7] selftests/proc: extend /proc/pid/maps tearing test to include vma resizing Suren Baghdasaryan
@ 2025-07-16 3:05 ` Suren Baghdasaryan
2025-07-16 3:05 ` [PATCH v7 4/7] selftests/proc: test PROCMAP_QUERY ioctl while vma is concurrently modified Suren Baghdasaryan
` (4 subsequent siblings)
7 siblings, 0 replies; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 3:05 UTC (permalink / raw)
To: akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest, surenb
Test that /proc/pid/maps does not report unexpected holes in the address
space when we concurrently remap a part of a vma into the middle of
another vma. This remapping results in the destination vma being split
into three parts and the part in the middle being patched back from,
all done concurrently from under the reader. We should always see either
original vma or the split one with no holes.
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
---
tools/testing/selftests/proc/proc-maps-race.c | 92 +++++++++++++++++++
1 file changed, 92 insertions(+)
diff --git a/tools/testing/selftests/proc/proc-maps-race.c b/tools/testing/selftests/proc/proc-maps-race.c
index 10365b4e68e1..764821ffd63d 100644
--- a/tools/testing/selftests/proc/proc-maps-race.c
+++ b/tools/testing/selftests/proc/proc-maps-race.c
@@ -416,6 +416,95 @@ static void test_maps_tearing_from_resize(int maps_fd,
signal_state(mod_info, TEST_DONE);
}
+static inline void remap_vma(const struct vma_modifier_info *mod_info)
+{
+ /*
+ * Remap the last page of the next vma into the middle of the vma.
+ * This splits the current vma and the first and middle parts (the
+ * parts at lower addresses) become the last vma objserved in the
+ * first page and the first vma observed in the last page.
+ */
+ assert(mremap(mod_info->next_addr + page_size * 2, page_size,
+ page_size, MREMAP_FIXED | MREMAP_MAYMOVE | MREMAP_DONTUNMAP,
+ mod_info->addr + page_size) != MAP_FAILED);
+}
+
+static inline void patch_vma(const struct vma_modifier_info *mod_info)
+{
+ assert(!mprotect(mod_info->addr + page_size, page_size,
+ mod_info->prot));
+}
+
+static inline void check_remap_result(struct line_content *mod_last_line,
+ struct line_content *mod_first_line,
+ struct line_content *restored_last_line,
+ struct line_content *restored_first_line)
+{
+ /* Make sure vmas at the boundaries are changing */
+ assert(strcmp(mod_last_line->text, restored_last_line->text) != 0);
+ assert(strcmp(mod_first_line->text, restored_first_line->text) != 0);
+}
+
+static void test_maps_tearing_from_remap(int maps_fd,
+ struct vma_modifier_info *mod_info,
+ struct page_content *page1,
+ struct page_content *page2,
+ struct line_content *last_line,
+ struct line_content *first_line)
+{
+ struct line_content remapped_last_line;
+ struct line_content remapped_first_line;
+ struct line_content restored_last_line;
+ struct line_content restored_first_line;
+
+ wait_for_state(mod_info, SETUP_READY);
+
+ /* re-read the file to avoid using stale data from previous test */
+ read_boundary_lines(maps_fd, page1, page2, last_line, first_line);
+
+ mod_info->vma_modify = remap_vma;
+ mod_info->vma_restore = patch_vma;
+ mod_info->vma_mod_check = check_remap_result;
+
+ capture_mod_pattern(maps_fd, mod_info, page1, page2, last_line, first_line,
+ &remapped_last_line, &remapped_first_line,
+ &restored_last_line, &restored_first_line);
+
+ /* Now start concurrent modifications for test_duration_sec */
+ signal_state(mod_info, TEST_READY);
+
+ struct line_content new_last_line;
+ struct line_content new_first_line;
+ struct timespec start_ts, end_ts;
+
+ clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
+ do {
+ read_boundary_lines(maps_fd, page1, page2, &new_last_line, &new_first_line);
+
+ /* Check if we read vmas after remapping it */
+ if (!strcmp(new_last_line.text, remapped_last_line.text)) {
+ /*
+ * The vmas should be consistent with remap results,
+ * however if the vma was concurrently restored, it
+ * can be reported twice (first as split one, then
+ * as restored one) because we found it as the next vma
+ * again. In that case new first line will be the same
+ * as the last restored line.
+ */
+ assert(!strcmp(new_first_line.text, remapped_first_line.text) ||
+ !strcmp(new_first_line.text, restored_last_line.text));
+ } else {
+ /* The vmas should be consistent with the original/resored state */
+ assert(!strcmp(new_last_line.text, restored_last_line.text) &&
+ !strcmp(new_first_line.text, restored_first_line.text));
+ }
+ clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
+ } while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
+
+ /* Signal the modifyer thread to stop and wait until it exits */
+ signal_state(mod_info, TEST_DONE);
+}
+
int usage(void)
{
fprintf(stderr, "Userland /proc/pid/{s}maps race test cases\n");
@@ -527,6 +616,9 @@ int main(int argc, char **argv)
test_maps_tearing_from_resize(maps_fd, mod_info, &page1, &page2,
&last_line, &first_line);
+ test_maps_tearing_from_remap(maps_fd, mod_info, &page1, &page2,
+ &last_line, &first_line);
+
stop_vma_modifier(mod_info);
free(page2.data);
--
2.50.0.727.gbf7dc18ff4-goog
^ permalink raw reply related [flat|nested] 20+ messages in thread
* [PATCH v7 4/7] selftests/proc: test PROCMAP_QUERY ioctl while vma is concurrently modified
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
` (2 preceding siblings ...)
2025-07-16 3:05 ` [PATCH v7 3/7] selftests/proc: extend /proc/pid/maps tearing test to include vma remapping Suren Baghdasaryan
@ 2025-07-16 3:05 ` Suren Baghdasaryan
2025-07-16 10:04 ` David Hildenbrand
2025-07-16 3:05 ` [PATCH v7 5/7] selftests/proc: add verbose more for tests to facilitate debugging Suren Baghdasaryan
` (3 subsequent siblings)
7 siblings, 1 reply; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 3:05 UTC (permalink / raw)
To: akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest, surenb
Extend /proc/pid/maps tearing test to verify PROCMAP_QUERY ioctl operation
correctness while the vma is being concurrently modified.
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
---
tools/testing/selftests/proc/proc-maps-race.c | 62 +++++++++++++++++++
1 file changed, 62 insertions(+)
diff --git a/tools/testing/selftests/proc/proc-maps-race.c b/tools/testing/selftests/proc/proc-maps-race.c
index 764821ffd63d..6acdafdac9db 100644
--- a/tools/testing/selftests/proc/proc-maps-race.c
+++ b/tools/testing/selftests/proc/proc-maps-race.c
@@ -30,6 +30,8 @@
#include <stdlib.h>
#include <string.h>
#include <unistd.h>
+#include <linux/fs.h>
+#include <sys/ioctl.h>
#include <sys/mman.h>
#include <sys/stat.h>
#include <sys/types.h>
@@ -239,6 +241,21 @@ static void capture_mod_pattern(int maps_fd,
assert(strcmp(restored_first_line->text, first_line->text) == 0);
}
+static void query_addr_at(int maps_fd, void *addr,
+ unsigned long *vma_start, unsigned long *vma_end)
+{
+ struct procmap_query q;
+
+ memset(&q, 0, sizeof(q));
+ q.size = sizeof(q);
+ /* Find the VMA at the split address */
+ q.query_addr = (unsigned long long)addr;
+ q.query_flags = 0;
+ assert(!ioctl(maps_fd, PROCMAP_QUERY, &q));
+ *vma_start = q.vma_start;
+ *vma_end = q.vma_end;
+}
+
static inline void split_vma(const struct vma_modifier_info *mod_info)
{
assert(mmap(mod_info->addr, page_size, mod_info->prot | PROT_EXEC,
@@ -299,6 +316,8 @@ static void test_maps_tearing_from_split(int maps_fd,
do {
bool last_line_changed;
bool first_line_changed;
+ unsigned long vma_start;
+ unsigned long vma_end;
read_boundary_lines(maps_fd, page1, page2, &new_last_line, &new_first_line);
@@ -329,6 +348,19 @@ static void test_maps_tearing_from_split(int maps_fd,
first_line_changed = strcmp(new_first_line.text, first_line->text) != 0;
assert(last_line_changed == first_line_changed);
+ /* Check if PROCMAP_QUERY ioclt() finds the right VMA */
+ query_addr_at(maps_fd, mod_info->addr + page_size,
+ &vma_start, &vma_end);
+ /*
+ * The vma at the split address can be either the same as
+ * original one (if read before the split) or the same as the
+ * first line in the second page (if read after the split).
+ */
+ assert((vma_start == last_line->start_addr &&
+ vma_end == last_line->end_addr) ||
+ (vma_start == split_first_line.start_addr &&
+ vma_end == split_first_line.end_addr));
+
clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
} while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
@@ -390,6 +422,9 @@ static void test_maps_tearing_from_resize(int maps_fd,
clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
do {
+ unsigned long vma_start;
+ unsigned long vma_end;
+
read_boundary_lines(maps_fd, page1, page2, &new_last_line, &new_first_line);
/* Check if we read vmas after shrinking it */
@@ -409,6 +444,17 @@ static void test_maps_tearing_from_resize(int maps_fd,
assert(!strcmp(new_last_line.text, restored_last_line.text) &&
!strcmp(new_first_line.text, restored_first_line.text));
}
+
+ /* Check if PROCMAP_QUERY ioclt() finds the right VMA */
+ query_addr_at(maps_fd, mod_info->addr, &vma_start, &vma_end);
+ /*
+ * The vma should stay at the same address and have either the
+ * original size of 3 pages or 1 page if read after shrinking.
+ */
+ assert(vma_start == last_line->start_addr &&
+ (vma_end - vma_start == page_size * 3 ||
+ vma_end - vma_start == page_size));
+
clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
} while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
@@ -479,6 +525,9 @@ static void test_maps_tearing_from_remap(int maps_fd,
clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
do {
+ unsigned long vma_start;
+ unsigned long vma_end;
+
read_boundary_lines(maps_fd, page1, page2, &new_last_line, &new_first_line);
/* Check if we read vmas after remapping it */
@@ -498,6 +547,19 @@ static void test_maps_tearing_from_remap(int maps_fd,
assert(!strcmp(new_last_line.text, restored_last_line.text) &&
!strcmp(new_first_line.text, restored_first_line.text));
}
+
+ /* Check if PROCMAP_QUERY ioclt() finds the right VMA */
+ query_addr_at(maps_fd, mod_info->addr + page_size, &vma_start, &vma_end);
+ /*
+ * The vma should either stay at the same address and have the
+ * original size of 3 pages or we should find the remapped vma
+ * at the remap destination address with size of 1 page.
+ */
+ assert((vma_start == last_line->start_addr &&
+ vma_end - vma_start == page_size * 3) ||
+ (vma_start == last_line->start_addr + page_size &&
+ vma_end - vma_start == page_size));
+
clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
} while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
--
2.50.0.727.gbf7dc18ff4-goog
^ permalink raw reply related [flat|nested] 20+ messages in thread
* Re: [PATCH v7 4/7] selftests/proc: test PROCMAP_QUERY ioctl while vma is concurrently modified
2025-07-16 3:05 ` [PATCH v7 4/7] selftests/proc: test PROCMAP_QUERY ioctl while vma is concurrently modified Suren Baghdasaryan
@ 2025-07-16 10:04 ` David Hildenbrand
2025-07-16 10:39 ` Lorenzo Stoakes
0 siblings, 1 reply; 20+ messages in thread
From: David Hildenbrand @ 2025-07-16 10:04 UTC (permalink / raw)
To: Suren Baghdasaryan, akpm
Cc: Liam.Howlett, lorenzo.stoakes, vbabka, peterx, jannh, hannes,
mhocko, paulmck, shuah, adobriyan, brauner, josef, yebin10, linux,
willy, osalvador, andrii, ryan.roberts, christophe.leroy,
tjmercier, kaleshsingh, aha310510, linux-kernel, linux-fsdevel,
linux-mm, linux-kselftest
On 16.07.25 05:05, Suren Baghdasaryan wrote:
> Extend /proc/pid/maps tearing test to verify PROCMAP_QUERY ioctl operation
> correctness while the vma is being concurrently modified.
>
Wonder if that should be moved out of this series as well. Of course, it
doesn't hurt to have this test already in.
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 4/7] selftests/proc: test PROCMAP_QUERY ioctl while vma is concurrently modified
2025-07-16 10:04 ` David Hildenbrand
@ 2025-07-16 10:39 ` Lorenzo Stoakes
2025-07-16 16:43 ` Suren Baghdasaryan
0 siblings, 1 reply; 20+ messages in thread
From: Lorenzo Stoakes @ 2025-07-16 10:39 UTC (permalink / raw)
To: David Hildenbrand
Cc: Suren Baghdasaryan, akpm, Liam.Howlett, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest
On Wed, Jul 16, 2025 at 12:04:25PM +0200, David Hildenbrand wrote:
> On 16.07.25 05:05, Suren Baghdasaryan wrote:
> > Extend /proc/pid/maps tearing test to verify PROCMAP_QUERY ioctl operation
> > correctness while the vma is being concurrently modified.
> >
>
> Wonder if that should be moved out of this series as well. Of course, it
> doesn't hurt to have this test already in.
Yeah that's move this out actually, in this series it's not actually
testing anything _pertinent_.
Though all the tests are designed to pass _before_ as well as after
obviously, so (as David says) no harm to have it here BUT - it's confusing
and a bit weird :P
So yeah let's just move this test over to the respun procmap query series.
BTW Suren - slightly unrelated but - the syzbot report I did the analysis
on is super super reproducable, so if you want to play around with
different solutions, using that should make it some quick iterations.
>
> --
> Cheers,
>
> David / dhildenb
>
Cheers, Lorenzo
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 4/7] selftests/proc: test PROCMAP_QUERY ioctl while vma is concurrently modified
2025-07-16 10:39 ` Lorenzo Stoakes
@ 2025-07-16 16:43 ` Suren Baghdasaryan
0 siblings, 0 replies; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 16:43 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: David Hildenbrand, akpm, Liam.Howlett, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest
On Wed, Jul 16, 2025 at 3:40 AM Lorenzo Stoakes
<lorenzo.stoakes@oracle.com> wrote:
>
> On Wed, Jul 16, 2025 at 12:04:25PM +0200, David Hildenbrand wrote:
> > On 16.07.25 05:05, Suren Baghdasaryan wrote:
> > > Extend /proc/pid/maps tearing test to verify PROCMAP_QUERY ioctl operation
> > > correctness while the vma is being concurrently modified.
> > >
> >
> > Wonder if that should be moved out of this series as well. Of course, it
> > doesn't hurt to have this test already in.
>
> Yeah that's move this out actually, in this series it's not actually
> testing anything _pertinent_.
>
> Though all the tests are designed to pass _before_ as well as after
> obviously, so (as David says) no harm to have it here BUT - it's confusing
> and a bit weird :P
>
> So yeah let's just move this test over to the respun procmap query series.
Ok.
>
> BTW Suren - slightly unrelated but - the syzbot report I did the analysis
> on is super super reproducable, so if you want to play around with
> different solutions, using that should make it some quick iterations.
Yes, I'm using that test routinely now.
>
> >
> > --
> > Cheers,
> >
> > David / dhildenb
> >
>
> Cheers, Lorenzo
^ permalink raw reply [flat|nested] 20+ messages in thread
* [PATCH v7 5/7] selftests/proc: add verbose more for tests to facilitate debugging
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
` (3 preceding siblings ...)
2025-07-16 3:05 ` [PATCH v7 4/7] selftests/proc: test PROCMAP_QUERY ioctl while vma is concurrently modified Suren Baghdasaryan
@ 2025-07-16 3:05 ` Suren Baghdasaryan
2025-07-16 3:05 ` [PATCH v7 6/7] fs/proc/task_mmu: remove conversion of seq_file position to unsigned Suren Baghdasaryan
` (2 subsequent siblings)
7 siblings, 0 replies; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 3:05 UTC (permalink / raw)
To: akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest, surenb
Add verbose mode to the proc tests to print debugging information.
Usage: proc-pid-vm -v
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
---
tools/testing/selftests/proc/proc-maps-race.c | 159 ++++++++++++++++--
1 file changed, 146 insertions(+), 13 deletions(-)
diff --git a/tools/testing/selftests/proc/proc-maps-race.c b/tools/testing/selftests/proc/proc-maps-race.c
index 6acdafdac9db..5f912fedd6cf 100644
--- a/tools/testing/selftests/proc/proc-maps-race.c
+++ b/tools/testing/selftests/proc/proc-maps-race.c
@@ -39,6 +39,7 @@
static unsigned long test_duration_sec = 5UL;
static int page_size;
+static bool verbose;
/* /proc/pid/maps parsing routines */
struct page_content {
@@ -207,6 +208,99 @@ static void stop_vma_modifier(struct vma_modifier_info *mod_info)
signal_state(mod_info, SETUP_MODIFY_MAPS);
}
+static void print_first_lines(char *text, int nr)
+{
+ const char *end = text;
+
+ while (nr && (end = strchr(end, '\n')) != NULL) {
+ nr--;
+ end++;
+ }
+
+ if (end) {
+ int offs = end - text;
+
+ text[offs] = '\0';
+ printf(text);
+ text[offs] = '\n';
+ printf("\n");
+ } else {
+ printf(text);
+ }
+}
+
+static void print_last_lines(char *text, int nr)
+{
+ const char *start = text + strlen(text);
+
+ nr++; /* to ignore the last newline */
+ while (nr) {
+ while (start > text && *start != '\n')
+ start--;
+ nr--;
+ start--;
+ }
+ printf(start);
+}
+
+static void print_boundaries(const char *title,
+ struct page_content *page1,
+ struct page_content *page2)
+{
+ if (!verbose)
+ return;
+
+ printf("%s", title);
+ /* Print 3 boundary lines from each page */
+ print_last_lines(page1->data, 3);
+ printf("-----------------page boundary-----------------\n");
+ print_first_lines(page2->data, 3);
+}
+
+static bool print_boundaries_on(bool condition, const char *title,
+ struct page_content *page1,
+ struct page_content *page2)
+{
+ if (verbose && condition)
+ print_boundaries(title, page1, page2);
+
+ return condition;
+}
+
+static void report_test_start(const char *name)
+{
+ if (verbose)
+ printf("==== %s ====\n", name);
+}
+
+static struct timespec print_ts;
+
+static void start_test_loop(struct timespec *ts)
+{
+ if (verbose)
+ print_ts.tv_sec = ts->tv_sec;
+}
+
+static void end_test_iteration(struct timespec *ts)
+{
+ if (!verbose)
+ return;
+
+ /* Update every second */
+ if (print_ts.tv_sec == ts->tv_sec)
+ return;
+
+ printf(".");
+ fflush(stdout);
+ print_ts.tv_sec = ts->tv_sec;
+}
+
+static void end_test_loop(void)
+{
+ if (verbose)
+ printf("\n");
+}
+
static void capture_mod_pattern(int maps_fd,
struct vma_modifier_info *mod_info,
struct page_content *page1,
@@ -218,18 +312,24 @@ static void capture_mod_pattern(int maps_fd,
struct line_content *restored_last_line,
struct line_content *restored_first_line)
{
+ print_boundaries("Before modification", page1, page2);
+
signal_state(mod_info, SETUP_MODIFY_MAPS);
wait_for_state(mod_info, SETUP_MAPS_MODIFIED);
/* Copy last line of the first page and first line of the last page */
read_boundary_lines(maps_fd, page1, page2, mod_last_line, mod_first_line);
+ print_boundaries("After modification", page1, page2);
+
signal_state(mod_info, SETUP_RESTORE_MAPS);
wait_for_state(mod_info, SETUP_MAPS_RESTORED);
/* Copy last line of the first page and first line of the last page */
read_boundary_lines(maps_fd, page1, page2, restored_last_line, restored_first_line);
+ print_boundaries("After restore", page1, page2);
+
mod_info->vma_mod_check(mod_last_line, mod_first_line,
restored_last_line, restored_first_line);
@@ -301,6 +401,7 @@ static void test_maps_tearing_from_split(int maps_fd,
mod_info->vma_restore = merge_vma;
mod_info->vma_mod_check = check_split_result;
+ report_test_start("Tearing from split");
capture_mod_pattern(maps_fd, mod_info, page1, page2, last_line, first_line,
&split_last_line, &split_first_line,
&restored_last_line, &restored_first_line);
@@ -313,6 +414,7 @@ static void test_maps_tearing_from_split(int maps_fd,
struct timespec start_ts, end_ts;
clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
+ start_test_loop(&start_ts);
do {
bool last_line_changed;
bool first_line_changed;
@@ -332,12 +434,18 @@ static void test_maps_tearing_from_split(int maps_fd,
* In that case new first line will be the same as the
* last restored line.
*/
- assert(!strcmp(new_first_line.text, split_first_line.text) ||
- !strcmp(new_first_line.text, restored_last_line.text));
+ assert(!print_boundaries_on(
+ strcmp(new_first_line.text, split_first_line.text) &&
+ strcmp(new_first_line.text, restored_last_line.text),
+ "Split result invalid", page1, page2));
} else {
/* The vmas should be consistent with merge results */
- assert(!strcmp(new_last_line.text, restored_last_line.text) &&
- !strcmp(new_first_line.text, restored_first_line.text));
+ assert(!print_boundaries_on(
+ strcmp(new_last_line.text, restored_last_line.text),
+ "Merge result invalid", page1, page2));
+ assert(!print_boundaries_on(
+ strcmp(new_first_line.text, restored_first_line.text),
+ "Merge result invalid", page1, page2));
}
/*
* First and last lines should change in unison. If the last
@@ -362,7 +470,9 @@ static void test_maps_tearing_from_split(int maps_fd,
vma_end == split_first_line.end_addr));
clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
+ end_test_iteration(&end_ts);
} while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
+ end_test_loop();
/* Signal the modifyer thread to stop and wait until it exits */
signal_state(mod_info, TEST_DONE);
@@ -409,6 +519,7 @@ static void test_maps_tearing_from_resize(int maps_fd,
mod_info->vma_restore = expand_vma;
mod_info->vma_mod_check = check_shrink_result;
+ report_test_start("Tearing from resize");
capture_mod_pattern(maps_fd, mod_info, page1, page2, last_line, first_line,
&shrunk_last_line, &shrunk_first_line,
&restored_last_line, &restored_first_line);
@@ -421,6 +532,7 @@ static void test_maps_tearing_from_resize(int maps_fd,
struct timespec start_ts, end_ts;
clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
+ start_test_loop(&start_ts);
do {
unsigned long vma_start;
unsigned long vma_end;
@@ -437,12 +549,18 @@ static void test_maps_tearing_from_resize(int maps_fd,
* again. In that case new first line will be the same
* as the last restored line.
*/
- assert(!strcmp(new_first_line.text, shrunk_first_line.text) ||
- !strcmp(new_first_line.text, restored_last_line.text));
+ assert(!print_boundaries_on(
+ strcmp(new_first_line.text, shrunk_first_line.text) &&
+ strcmp(new_first_line.text, restored_last_line.text),
+ "Shrink result invalid", page1, page2));
} else {
/* The vmas should be consistent with the original/resored state */
- assert(!strcmp(new_last_line.text, restored_last_line.text) &&
- !strcmp(new_first_line.text, restored_first_line.text));
+ assert(!print_boundaries_on(
+ strcmp(new_last_line.text, restored_last_line.text),
+ "Expand result invalid", page1, page2));
+ assert(!print_boundaries_on(
+ strcmp(new_first_line.text, restored_first_line.text),
+ "Expand result invalid", page1, page2));
}
/* Check if PROCMAP_QUERY ioclt() finds the right VMA */
@@ -456,7 +574,9 @@ static void test_maps_tearing_from_resize(int maps_fd,
vma_end - vma_start == page_size));
clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
+ end_test_iteration(&end_ts);
} while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
+ end_test_loop();
/* Signal the modifyer thread to stop and wait until it exits */
signal_state(mod_info, TEST_DONE);
@@ -512,6 +632,7 @@ static void test_maps_tearing_from_remap(int maps_fd,
mod_info->vma_restore = patch_vma;
mod_info->vma_mod_check = check_remap_result;
+ report_test_start("Tearing from remap");
capture_mod_pattern(maps_fd, mod_info, page1, page2, last_line, first_line,
&remapped_last_line, &remapped_first_line,
&restored_last_line, &restored_first_line);
@@ -524,6 +645,7 @@ static void test_maps_tearing_from_remap(int maps_fd,
struct timespec start_ts, end_ts;
clock_gettime(CLOCK_MONOTONIC_COARSE, &start_ts);
+ start_test_loop(&start_ts);
do {
unsigned long vma_start;
unsigned long vma_end;
@@ -540,12 +662,18 @@ static void test_maps_tearing_from_remap(int maps_fd,
* again. In that case new first line will be the same
* as the last restored line.
*/
- assert(!strcmp(new_first_line.text, remapped_first_line.text) ||
- !strcmp(new_first_line.text, restored_last_line.text));
+ assert(!print_boundaries_on(
+ strcmp(new_first_line.text, remapped_first_line.text) &&
+ strcmp(new_first_line.text, restored_last_line.text),
+ "Remap result invalid", page1, page2));
} else {
/* The vmas should be consistent with the original/resored state */
- assert(!strcmp(new_last_line.text, restored_last_line.text) &&
- !strcmp(new_first_line.text, restored_first_line.text));
+ assert(!print_boundaries_on(
+ strcmp(new_last_line.text, restored_last_line.text),
+ "Remap restore result invalid", page1, page2));
+ assert(!print_boundaries_on(
+ strcmp(new_first_line.text, restored_first_line.text),
+ "Remap restore result invalid", page1, page2));
}
/* Check if PROCMAP_QUERY ioclt() finds the right VMA */
@@ -561,7 +689,9 @@ static void test_maps_tearing_from_remap(int maps_fd,
vma_end - vma_start == page_size));
clock_gettime(CLOCK_MONOTONIC_COARSE, &end_ts);
+ end_test_iteration(&end_ts);
} while (end_ts.tv_sec - start_ts.tv_sec < test_duration_sec);
+ end_test_loop();
/* Signal the modifyer thread to stop and wait until it exits */
signal_state(mod_info, TEST_DONE);
@@ -571,6 +701,7 @@ int usage(void)
{
fprintf(stderr, "Userland /proc/pid/{s}maps race test cases\n");
fprintf(stderr, " -d: Duration for time-consuming tests\n");
+ fprintf(stderr, " -v: Verbose mode\n");
fprintf(stderr, " -h: Help screen\n");
exit(-1);
}
@@ -588,9 +719,11 @@ int main(int argc, char **argv)
pid_t pid;
int opt;
- while ((opt = getopt(argc, argv, "d:h")) != -1) {
+ while ((opt = getopt(argc, argv, "d:vh")) != -1) {
if (opt == 'd')
test_duration_sec = strtoul(optarg, NULL, 0);
+ else if (opt == 'v')
+ verbose = true;
else if (opt == 'h')
usage();
}
--
2.50.0.727.gbf7dc18ff4-goog
^ permalink raw reply related [flat|nested] 20+ messages in thread
* [PATCH v7 6/7] fs/proc/task_mmu: remove conversion of seq_file position to unsigned
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
` (4 preceding siblings ...)
2025-07-16 3:05 ` [PATCH v7 5/7] selftests/proc: add verbose more for tests to facilitate debugging Suren Baghdasaryan
@ 2025-07-16 3:05 ` Suren Baghdasaryan
2025-07-16 10:41 ` David Hildenbrand
2025-07-16 3:05 ` [PATCH v7 7/7] fs/proc/task_mmu: read proc/pid/maps under per-vma lock Suren Baghdasaryan
2025-07-16 22:55 ` [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Andrew Morton
7 siblings, 1 reply; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 3:05 UTC (permalink / raw)
To: akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest, surenb
Back in 2.6 era, last_addr used to be stored in seq_file->version
variable, which was unsigned long. As a result, sentinels to represent
gate vma and end of all vmas used unsigned values. In more recent
kernels we don't used seq_file->version anymore and therefore conversion
from loff_t into unsigned type is not needed. Similarly, sentinel values
don't need to be unsigned. Remove type conversion for set_file position
and change sentinel values to signed.
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
---
fs/proc/task_mmu.c | 14 +++++++-------
1 file changed, 7 insertions(+), 7 deletions(-)
diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c
index 751479eb128f..b8bc06d05a72 100644
--- a/fs/proc/task_mmu.c
+++ b/fs/proc/task_mmu.c
@@ -135,7 +135,7 @@ static struct vm_area_struct *proc_get_vma(struct proc_maps_private *priv,
if (vma) {
*ppos = vma->vm_start;
} else {
- *ppos = -2UL;
+ *ppos = -2;
vma = get_gate_vma(priv->mm);
}
@@ -145,11 +145,11 @@ static struct vm_area_struct *proc_get_vma(struct proc_maps_private *priv,
static void *m_start(struct seq_file *m, loff_t *ppos)
{
struct proc_maps_private *priv = m->private;
- unsigned long last_addr = *ppos;
+ loff_t last_addr = *ppos;
struct mm_struct *mm;
/* See m_next(). Zero at the start or after lseek. */
- if (last_addr == -1UL)
+ if (last_addr == -1)
return NULL;
priv->task = get_proc_task(priv->inode);
@@ -170,9 +170,9 @@ static void *m_start(struct seq_file *m, loff_t *ppos)
return ERR_PTR(-EINTR);
}
- vma_iter_init(&priv->iter, mm, last_addr);
+ vma_iter_init(&priv->iter, mm, (unsigned long)last_addr);
hold_task_mempolicy(priv);
- if (last_addr == -2UL)
+ if (last_addr == -2)
return get_gate_vma(mm);
return proc_get_vma(priv, ppos);
@@ -180,8 +180,8 @@ static void *m_start(struct seq_file *m, loff_t *ppos)
static void *m_next(struct seq_file *m, void *v, loff_t *ppos)
{
- if (*ppos == -2UL) {
- *ppos = -1UL;
+ if (*ppos == -2) {
+ *ppos = -1;
return NULL;
}
return proc_get_vma(m->private, ppos);
--
2.50.0.727.gbf7dc18ff4-goog
^ permalink raw reply related [flat|nested] 20+ messages in thread
* Re: [PATCH v7 6/7] fs/proc/task_mmu: remove conversion of seq_file position to unsigned
2025-07-16 3:05 ` [PATCH v7 6/7] fs/proc/task_mmu: remove conversion of seq_file position to unsigned Suren Baghdasaryan
@ 2025-07-16 10:41 ` David Hildenbrand
0 siblings, 0 replies; 20+ messages in thread
From: David Hildenbrand @ 2025-07-16 10:41 UTC (permalink / raw)
To: Suren Baghdasaryan, akpm
Cc: Liam.Howlett, lorenzo.stoakes, vbabka, peterx, jannh, hannes,
mhocko, paulmck, shuah, adobriyan, brauner, josef, yebin10, linux,
willy, osalvador, andrii, ryan.roberts, christophe.leroy,
tjmercier, kaleshsingh, aha310510, linux-kernel, linux-fsdevel,
linux-mm, linux-kselftest
On 16.07.25 05:05, Suren Baghdasaryan wrote:
> Back in 2.6 era, last_addr used to be stored in seq_file->version
> variable, which was unsigned long. As a result, sentinels to represent
> gate vma and end of all vmas used unsigned values. In more recent
> kernels we don't used seq_file->version anymore and therefore conversion
> from loff_t into unsigned type is not needed. Similarly, sentinel values
> don't need to be unsigned. Remove type conversion for set_file position
> and change sentinel values to signed.
>
> Signed-off-by: Suren Baghdasaryan <surenb@google.com>
> Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
> Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
> ---
Acked-by: David Hildenbrand <david@redhat.com>
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 20+ messages in thread
* [PATCH v7 7/7] fs/proc/task_mmu: read proc/pid/maps under per-vma lock
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
` (5 preceding siblings ...)
2025-07-16 3:05 ` [PATCH v7 6/7] fs/proc/task_mmu: remove conversion of seq_file position to unsigned Suren Baghdasaryan
@ 2025-07-16 3:05 ` Suren Baghdasaryan
2025-07-16 13:57 ` Vlastimil Babka
2025-07-16 22:55 ` [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Andrew Morton
7 siblings, 1 reply; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 3:05 UTC (permalink / raw)
To: akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest, surenb
With maple_tree supporting vma tree traversal under RCU and per-vma
locks, /proc/pid/maps can be read while holding individual vma locks
instead of locking the entire address space.
A completely lockless approach (walking vma tree under RCU) would be
quite complex with the main issue being get_vma_name() using callbacks
which might not work correctly with a stable vma copy, requiring
original (unstable) vma - see special_mapping_name() for example.
When per-vma lock acquisition fails, we take the mmap_lock for reading,
lock the vma, release the mmap_lock and continue. This fallback to mmap
read lock guarantees the reader to make forward progress even during
lock contention. This will interfere with the writer but for a very
short time while we are acquiring the per-vma lock and only when there
was contention on the vma reader is interested in.
We shouldn't see a repeated fallback to mmap read locks in practice, as
this require a very unlikely series of lock contentions (for instance
due to repeated vma split operations). However even if this did somehow
happen, we would still progress.
One case requiring special handling is when a vma changes between the
time it was found and the time it got locked. A problematic case would
be if a vma got shrunk so that its vm_start moved higher in the address
space and a new vma was installed at the beginning:
reader found: |--------VMA A--------|
VMA is modified: |-VMA B-|----VMA A----|
reader locks modified VMA A
reader reports VMA A: | gap |----VMA A----|
This would result in reporting a gap in the address space that does not
exist. To prevent this we retry the lookup after locking the vma, however
we do that only when we identify a gap and detect that the address space
was changed after we found the vma.
This change is designed to reduce mmap_lock contention and prevent a
process reading /proc/pid/maps files (often a low priority task, such
as monitoring/data collection services) from blocking address space
updates. Note that this change has a userspace visible disadvantage:
it allows for sub-page data tearing as opposed to the previous mechanism
where data tearing could happen only between pages of generated output
data. Since current userspace considers data tearing between pages to be
acceptable, we assume is will be able to handle sub-page data tearing
as well.
Signed-off-by: Suren Baghdasaryan <surenb@google.com>
---
fs/proc/internal.h | 5 ++
fs/proc/task_mmu.c | 145 +++++++++++++++++++++++++++++++++++---
include/linux/mmap_lock.h | 11 +++
mm/madvise.c | 3 +-
mm/mmap_lock.c | 93 ++++++++++++++++++++++++
5 files changed, 246 insertions(+), 11 deletions(-)
diff --git a/fs/proc/internal.h b/fs/proc/internal.h
index 3d48ffe72583..7c235451c5ea 100644
--- a/fs/proc/internal.h
+++ b/fs/proc/internal.h
@@ -384,6 +384,11 @@ struct proc_maps_private {
struct task_struct *task;
struct mm_struct *mm;
struct vma_iterator iter;
+ loff_t last_pos;
+#ifdef CONFIG_PER_VMA_LOCK
+ bool mmap_locked;
+ struct vm_area_struct *locked_vma;
+#endif
#ifdef CONFIG_NUMA
struct mempolicy *task_mempolicy;
#endif
diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c
index b8bc06d05a72..b15d0ef29896 100644
--- a/fs/proc/task_mmu.c
+++ b/fs/proc/task_mmu.c
@@ -127,15 +127,134 @@ static void release_task_mempolicy(struct proc_maps_private *priv)
}
#endif
-static struct vm_area_struct *proc_get_vma(struct proc_maps_private *priv,
- loff_t *ppos)
+#ifdef CONFIG_PER_VMA_LOCK
+
+static void unlock_vma(struct proc_maps_private *priv)
+{
+ if (priv->locked_vma) {
+ vma_end_read(priv->locked_vma);
+ priv->locked_vma = NULL;
+ }
+}
+
+static const struct seq_operations proc_pid_maps_op;
+
+static inline bool lock_vma_range(struct seq_file *m,
+ struct proc_maps_private *priv)
+{
+ /*
+ * smaps and numa_maps perform page table walk, therefore require
+ * mmap_lock but maps can be read with locking just the vma and
+ * walking the vma tree under rcu read protection.
+ */
+ if (m->op != &proc_pid_maps_op) {
+ if (mmap_read_lock_killable(priv->mm))
+ return false;
+
+ priv->mmap_locked = true;
+ } else {
+ rcu_read_lock();
+ priv->locked_vma = NULL;
+ priv->mmap_locked = false;
+ }
+
+ return true;
+}
+
+static inline void unlock_vma_range(struct proc_maps_private *priv)
+{
+ if (priv->mmap_locked) {
+ mmap_read_unlock(priv->mm);
+ } else {
+ unlock_vma(priv);
+ rcu_read_unlock();
+ }
+}
+
+static struct vm_area_struct *get_next_vma(struct proc_maps_private *priv,
+ loff_t last_pos)
+{
+ struct vm_area_struct *vma;
+
+ if (priv->mmap_locked)
+ return vma_next(&priv->iter);
+
+ unlock_vma(priv);
+ vma = lock_next_vma(priv->mm, &priv->iter, last_pos);
+ if (!IS_ERR_OR_NULL(vma))
+ priv->locked_vma = vma;
+
+ return vma;
+}
+
+static inline bool fallback_to_mmap_lock(struct proc_maps_private *priv,
+ loff_t pos)
{
- struct vm_area_struct *vma = vma_next(&priv->iter);
+ if (priv->mmap_locked)
+ return false;
+
+ rcu_read_unlock();
+ mmap_read_lock(priv->mm);
+ /* Reinitialize the iterator after taking mmap_lock */
+ vma_iter_set(&priv->iter, pos);
+ priv->mmap_locked = true;
+ return true;
+}
+
+#else /* CONFIG_PER_VMA_LOCK */
+
+static inline bool lock_vma_range(struct seq_file *m,
+ struct proc_maps_private *priv)
+{
+ return mmap_read_lock_killable(priv->mm) == 0;
+}
+
+static inline void unlock_vma_range(struct proc_maps_private *priv)
+{
+ mmap_read_unlock(priv->mm);
+}
+
+static struct vm_area_struct *get_next_vma(struct proc_maps_private *priv,
+ loff_t last_pos)
+{
+ return vma_next(&priv->iter);
+}
+
+static inline bool fallback_to_mmap_lock(struct proc_maps_private *priv,
+ loff_t pos)
+{
+ return false;
+}
+
+#endif /* CONFIG_PER_VMA_LOCK */
+
+static struct vm_area_struct *proc_get_vma(struct seq_file *m, loff_t *ppos)
+{
+ struct proc_maps_private *priv = m->private;
+ struct vm_area_struct *vma;
+
+retry:
+ vma = get_next_vma(priv, *ppos);
+ /* EINTR of EAGAIN is possible */
+ if (IS_ERR(vma)) {
+ if (PTR_ERR(vma) == -EAGAIN && fallback_to_mmap_lock(priv, *ppos))
+ goto retry;
+
+ return vma;
+ }
+
+ /* Store previous position to be able to restart if needed */
+ priv->last_pos = *ppos;
if (vma) {
- *ppos = vma->vm_start;
+ /*
+ * Track the end of the reported vma to ensure position changes
+ * even if previous vma was merged with the next vma and we
+ * found the extended vma with the same vm_start.
+ */
+ *ppos = vma->vm_end;
} else {
- *ppos = -2;
+ *ppos = -2; /* -2 indicates gate vma */
vma = get_gate_vma(priv->mm);
}
@@ -163,28 +282,34 @@ static void *m_start(struct seq_file *m, loff_t *ppos)
return NULL;
}
- if (mmap_read_lock_killable(mm)) {
+ if (!lock_vma_range(m, priv)) {
mmput(mm);
put_task_struct(priv->task);
priv->task = NULL;
return ERR_PTR(-EINTR);
}
+ /*
+ * Reset current position if last_addr was set before
+ * and it's not a sentinel.
+ */
+ if (last_addr > 0)
+ *ppos = last_addr = priv->last_pos;
vma_iter_init(&priv->iter, mm, (unsigned long)last_addr);
hold_task_mempolicy(priv);
if (last_addr == -2)
return get_gate_vma(mm);
- return proc_get_vma(priv, ppos);
+ return proc_get_vma(m, ppos);
}
static void *m_next(struct seq_file *m, void *v, loff_t *ppos)
{
if (*ppos == -2) {
- *ppos = -1;
+ *ppos = -1; /* -1 indicates no more vmas */
return NULL;
}
- return proc_get_vma(m->private, ppos);
+ return proc_get_vma(m, ppos);
}
static void m_stop(struct seq_file *m, void *v)
@@ -196,7 +321,7 @@ static void m_stop(struct seq_file *m, void *v)
return;
release_task_mempolicy(priv);
- mmap_read_unlock(mm);
+ unlock_vma_range(priv);
mmput(mm);
put_task_struct(priv->task);
priv->task = NULL;
diff --git a/include/linux/mmap_lock.h b/include/linux/mmap_lock.h
index 5da384bd0a26..1f4f44951abe 100644
--- a/include/linux/mmap_lock.h
+++ b/include/linux/mmap_lock.h
@@ -309,6 +309,17 @@ void vma_mark_detached(struct vm_area_struct *vma);
struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm,
unsigned long address);
+/*
+ * Locks next vma pointed by the iterator. Confirms the locked vma has not
+ * been modified and will retry under mmap_lock protection if modification
+ * was detected. Should be called from read RCU section.
+ * Returns either a valid locked VMA, NULL if no more VMAs or -EINTR if the
+ * process was interrupted.
+ */
+struct vm_area_struct *lock_next_vma(struct mm_struct *mm,
+ struct vma_iterator *iter,
+ unsigned long address);
+
#else /* CONFIG_PER_VMA_LOCK */
static inline void mm_lock_seqcount_init(struct mm_struct *mm) {}
diff --git a/mm/madvise.c b/mm/madvise.c
index 1c30031ab035..9de9b7c797c6 100644
--- a/mm/madvise.c
+++ b/mm/madvise.c
@@ -108,7 +108,8 @@ void anon_vma_name_free(struct kref *kref)
struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma)
{
- mmap_assert_locked(vma->vm_mm);
+ if (!rwsem_is_locked(&vma->vm_mm->mmap_lock))
+ vma_assert_locked(vma);
return vma->anon_name;
}
diff --git a/mm/mmap_lock.c b/mm/mmap_lock.c
index 5f725cc67334..729fb7d0dd59 100644
--- a/mm/mmap_lock.c
+++ b/mm/mmap_lock.c
@@ -178,6 +178,99 @@ struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm,
count_vm_vma_lock_event(VMA_LOCK_ABORT);
return NULL;
}
+
+static struct vm_area_struct *lock_next_vma_under_mmap_lock(struct mm_struct *mm,
+ struct vma_iterator *vmi,
+ unsigned long from_addr)
+{
+ struct vm_area_struct *vma;
+ int ret;
+
+ ret = mmap_read_lock_killable(mm);
+ if (ret)
+ return ERR_PTR(ret);
+
+ /* Lookup the vma at the last position again under mmap_read_lock */
+ vma_iter_set(vmi, from_addr);
+ vma = vma_next(vmi);
+ if (vma) {
+ /* Very unlikely vma->vm_refcnt overflow case */
+ if (unlikely(!vma_start_read_locked(vma)))
+ vma = ERR_PTR(-EAGAIN);
+ }
+
+ mmap_read_unlock(mm);
+
+ return vma;
+}
+
+struct vm_area_struct *lock_next_vma(struct mm_struct *mm,
+ struct vma_iterator *vmi,
+ unsigned long from_addr)
+{
+ struct vm_area_struct *vma;
+ unsigned int mm_wr_seq;
+ bool mmap_unlocked;
+
+ RCU_LOCKDEP_WARN(!rcu_read_lock_held(), "no rcu read lock held");
+retry:
+ /* Start mmap_lock speculation in case we need to verify the vma later */
+ mmap_unlocked = mmap_lock_speculate_try_begin(mm, &mm_wr_seq);
+ vma = vma_next(vmi);
+ if (!vma)
+ return NULL;
+
+ vma = vma_start_read(mm, vma);
+ if (IS_ERR_OR_NULL(vma)) {
+ /*
+ * Retry immediately if the vma gets detached from under us.
+ * Infinite loop should not happen because the vma we find will
+ * have to be constantly knocked out from under us.
+ */
+ if (PTR_ERR(vma) == -EAGAIN) {
+ /* reset to search from the last address */
+ vma_iter_set(vmi, from_addr);
+ goto retry;
+ }
+
+ goto fallback;
+ }
+
+ /*
+ * Verify the vma we locked belongs to the same address space and it's
+ * not behind of the last search position.
+ */
+ if (unlikely(vma->vm_mm != mm || from_addr >= vma->vm_end))
+ goto fallback_unlock;
+
+ /*
+ * vma can be ahead of the last search position but we need to verify
+ * it was not shrunk after we found it and another vma has not been
+ * installed ahead of it. Otherwise we might observe a gap that should
+ * not be there.
+ */
+ if (from_addr < vma->vm_start) {
+ /* Verify only if the address space might have changed since vma lookup. */
+ if (!mmap_unlocked || mmap_lock_speculate_retry(mm, mm_wr_seq)) {
+ vma_iter_set(vmi, from_addr);
+ if (vma != vma_next(vmi))
+ goto fallback_unlock;
+ }
+ }
+
+ return vma;
+
+fallback_unlock:
+ vma_end_read(vma);
+fallback:
+ rcu_read_unlock();
+ vma = lock_next_vma_under_mmap_lock(mm, vmi, from_addr);
+ rcu_read_lock();
+ /* Reinitialize the iterator after re-entering rcu read section */
+ vma_iter_set(vmi, IS_ERR_OR_NULL(vma) ? from_addr : vma->vm_end);
+
+ return vma;
+}
#endif /* CONFIG_PER_VMA_LOCK */
#ifdef CONFIG_LOCK_MM_AND_FIND_VMA
--
2.50.0.727.gbf7dc18ff4-goog
^ permalink raw reply related [flat|nested] 20+ messages in thread
* Re: [PATCH v7 7/7] fs/proc/task_mmu: read proc/pid/maps under per-vma lock
2025-07-16 3:05 ` [PATCH v7 7/7] fs/proc/task_mmu: read proc/pid/maps under per-vma lock Suren Baghdasaryan
@ 2025-07-16 13:57 ` Vlastimil Babka
2025-07-16 14:29 ` Suren Baghdasaryan
0 siblings, 1 reply; 20+ messages in thread
From: Vlastimil Babka @ 2025-07-16 13:57 UTC (permalink / raw)
To: Suren Baghdasaryan, akpm
Cc: Liam.Howlett, lorenzo.stoakes, david, peterx, jannh, hannes,
mhocko, paulmck, shuah, adobriyan, brauner, josef, yebin10, linux,
willy, osalvador, andrii, ryan.roberts, christophe.leroy,
tjmercier, kaleshsingh, aha310510, linux-kernel, linux-fsdevel,
linux-mm, linux-kselftest
On 7/16/25 05:05, Suren Baghdasaryan wrote:
> With maple_tree supporting vma tree traversal under RCU and per-vma
> locks, /proc/pid/maps can be read while holding individual vma locks
> instead of locking the entire address space.
> A completely lockless approach (walking vma tree under RCU) would be
> quite complex with the main issue being get_vma_name() using callbacks
> which might not work correctly with a stable vma copy, requiring
> original (unstable) vma - see special_mapping_name() for example.
>
> When per-vma lock acquisition fails, we take the mmap_lock for reading,
> lock the vma, release the mmap_lock and continue. This fallback to mmap
> read lock guarantees the reader to make forward progress even during
> lock contention. This will interfere with the writer but for a very
> short time while we are acquiring the per-vma lock and only when there
> was contention on the vma reader is interested in.
>
> We shouldn't see a repeated fallback to mmap read locks in practice, as
> this require a very unlikely series of lock contentions (for instance
> due to repeated vma split operations). However even if this did somehow
> happen, we would still progress.
>
> One case requiring special handling is when a vma changes between the
> time it was found and the time it got locked. A problematic case would
> be if a vma got shrunk so that its vm_start moved higher in the address
> space and a new vma was installed at the beginning:
>
> reader found: |--------VMA A--------|
> VMA is modified: |-VMA B-|----VMA A----|
> reader locks modified VMA A
> reader reports VMA A: | gap |----VMA A----|
>
> This would result in reporting a gap in the address space that does not
> exist. To prevent this we retry the lookup after locking the vma, however
> we do that only when we identify a gap and detect that the address space
> was changed after we found the vma.
>
> This change is designed to reduce mmap_lock contention and prevent a
> process reading /proc/pid/maps files (often a low priority task, such
> as monitoring/data collection services) from blocking address space
> updates. Note that this change has a userspace visible disadvantage:
> it allows for sub-page data tearing as opposed to the previous mechanism
> where data tearing could happen only between pages of generated output
> data. Since current userspace considers data tearing between pages to be
> acceptable, we assume is will be able to handle sub-page data tearing
> as well.
>
> Signed-off-by: Suren Baghdasaryan <surenb@google.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Nit: the previous patch changed lines with e.g. -2UL to -2 and this seems
changing the same lines to add a comment e.g. *ppos = -2; /* -2 indicates
gate vma */
That comment could have been added in the previous patch already. Also if
you feel the need to add the comments, maybe it's time to just name those
special values with a #define or something :)
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 7/7] fs/proc/task_mmu: read proc/pid/maps under per-vma lock
2025-07-16 13:57 ` Vlastimil Babka
@ 2025-07-16 14:29 ` Suren Baghdasaryan
0 siblings, 0 replies; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-16 14:29 UTC (permalink / raw)
To: Vlastimil Babka
Cc: akpm, Liam.Howlett, lorenzo.stoakes, david, peterx, jannh, hannes,
mhocko, paulmck, shuah, adobriyan, brauner, josef, yebin10, linux,
willy, osalvador, andrii, ryan.roberts, christophe.leroy,
tjmercier, kaleshsingh, aha310510, linux-kernel, linux-fsdevel,
linux-mm, linux-kselftest
On Wed, Jul 16, 2025 at 6:57 AM Vlastimil Babka <vbabka@suse.cz> wrote:
>
> On 7/16/25 05:05, Suren Baghdasaryan wrote:
> > With maple_tree supporting vma tree traversal under RCU and per-vma
> > locks, /proc/pid/maps can be read while holding individual vma locks
> > instead of locking the entire address space.
> > A completely lockless approach (walking vma tree under RCU) would be
> > quite complex with the main issue being get_vma_name() using callbacks
> > which might not work correctly with a stable vma copy, requiring
> > original (unstable) vma - see special_mapping_name() for example.
> >
> > When per-vma lock acquisition fails, we take the mmap_lock for reading,
> > lock the vma, release the mmap_lock and continue. This fallback to mmap
> > read lock guarantees the reader to make forward progress even during
> > lock contention. This will interfere with the writer but for a very
> > short time while we are acquiring the per-vma lock and only when there
> > was contention on the vma reader is interested in.
> >
> > We shouldn't see a repeated fallback to mmap read locks in practice, as
> > this require a very unlikely series of lock contentions (for instance
> > due to repeated vma split operations). However even if this did somehow
> > happen, we would still progress.
> >
> > One case requiring special handling is when a vma changes between the
> > time it was found and the time it got locked. A problematic case would
> > be if a vma got shrunk so that its vm_start moved higher in the address
> > space and a new vma was installed at the beginning:
> >
> > reader found: |--------VMA A--------|
> > VMA is modified: |-VMA B-|----VMA A----|
> > reader locks modified VMA A
> > reader reports VMA A: | gap |----VMA A----|
> >
> > This would result in reporting a gap in the address space that does not
> > exist. To prevent this we retry the lookup after locking the vma, however
> > we do that only when we identify a gap and detect that the address space
> > was changed after we found the vma.
> >
> > This change is designed to reduce mmap_lock contention and prevent a
> > process reading /proc/pid/maps files (often a low priority task, such
> > as monitoring/data collection services) from blocking address space
> > updates. Note that this change has a userspace visible disadvantage:
> > it allows for sub-page data tearing as opposed to the previous mechanism
> > where data tearing could happen only between pages of generated output
> > data. Since current userspace considers data tearing between pages to be
> > acceptable, we assume is will be able to handle sub-page data tearing
> > as well.
> >
> > Signed-off-by: Suren Baghdasaryan <surenb@google.com>
>
> Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
>
> Nit: the previous patch changed lines with e.g. -2UL to -2 and this seems
> changing the same lines to add a comment e.g. *ppos = -2; /* -2 indicates
> gate vma */
>
> That comment could have been added in the previous patch already. Also if
> you feel the need to add the comments, maybe it's time to just name those
> special values with a #define or something :)
Good point. I'll see if I can fit that into the next version.
>
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads
2025-07-16 3:05 [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Suren Baghdasaryan
` (6 preceding siblings ...)
2025-07-16 3:05 ` [PATCH v7 7/7] fs/proc/task_mmu: read proc/pid/maps under per-vma lock Suren Baghdasaryan
@ 2025-07-16 22:55 ` Andrew Morton
2025-07-17 1:38 ` Suren Baghdasaryan
7 siblings, 1 reply; 20+ messages in thread
From: Andrew Morton @ 2025-07-16 22:55 UTC (permalink / raw)
To: Suren Baghdasaryan
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest
On Tue, 15 Jul 2025 20:05:49 -0700 Suren Baghdasaryan <surenb@google.com> wrote:
> This patchset switches from holding mmap_lock while reading /proc/pid/maps
> to taking per-vma locks as we walk the vma tree.
Thanks, I added this v7 series to mm-new. Which I usually push out
mid-evening California time.
^ permalink raw reply [flat|nested] 20+ messages in thread
* Re: [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads
2025-07-16 22:55 ` [PATCH v7 0/7] use per-vma locks for /proc/pid/maps reads Andrew Morton
@ 2025-07-17 1:38 ` Suren Baghdasaryan
0 siblings, 0 replies; 20+ messages in thread
From: Suren Baghdasaryan @ 2025-07-17 1:38 UTC (permalink / raw)
To: Andrew Morton
Cc: Liam.Howlett, lorenzo.stoakes, david, vbabka, peterx, jannh,
hannes, mhocko, paulmck, shuah, adobriyan, brauner, josef,
yebin10, linux, willy, osalvador, andrii, ryan.roberts,
christophe.leroy, tjmercier, kaleshsingh, aha310510, linux-kernel,
linux-fsdevel, linux-mm, linux-kselftest
On Wed, Jul 16, 2025 at 3:55 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> On Tue, 15 Jul 2025 20:05:49 -0700 Suren Baghdasaryan <surenb@google.com> wrote:
>
> > This patchset switches from holding mmap_lock while reading /proc/pid/maps
> > to taking per-vma locks as we walk the vma tree.
>
> Thanks, I added this v7 series to mm-new. Which I usually push out
> mid-evening California time.
Thanks! There are some comments on the last version as well, so
unfortunately I'll have to respin and bother you again once that's
addressed but this update should at least remove the syzbot noise.
^ permalink raw reply [flat|nested] 20+ messages in thread