* [PATCH] mm: update stale locking comment in do_anonymous_page() @ 2026-03-31 14:29 Aditya Sharma 2026-04-01 10:52 ` David Hildenbrand (Arm) 2026-04-05 17:18 ` [PATCH v2] mm/memory: update stale locking comments for fault handlers Aditya Sharma 0 siblings, 2 replies; 7+ messages in thread From: Aditya Sharma @ 2026-03-31 14:29 UTC (permalink / raw) To: linux-mm Cc: akpm, david, ljs, Liam.Howlett, vbabka, rppt, surenb, mhocko, linux-kernel, Aditya Sharma The comment above do_anonymous_page() dates back to 2005 and describes the pre-per-VMA-lock world where mmap_lock was always held on entry. Since CONFIG_PER_VMA_LOCK was introduced (6.4), the fault handler now has a fast path that enters holding only a per-VMA read lock, with mmap_lock not held at all. Update the comment to describe both entry contexts accurately. Signed-off-by: Aditya Sharma <adi.sharma@zohomail.in> --- mm/memory.c | 22 +++++++++++++++++++--- 1 file changed, 19 insertions(+), 3 deletions(-) diff --git a/mm/memory.c b/mm/memory.c index c65e82c86..cc8dbbaea 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -5210,9 +5210,25 @@ static struct folio *alloc_anon_folio(struct vm_fault *vmf) } /* - * We enter with non-exclusive mmap_lock (to exclude vma changes, - * but allow concurrent faults), and pte mapped but not yet locked. - * We return with mmap_lock still held, but pte unmapped and unlocked. + * We enter in one of two locking contexts: + * + * 1) VMA lock path (FAULT_FLAG_VMA_LOCK set): + * Entered holding a read lock on the faulting VMA (vma_start_read), + * but NOT holding mmap_lock. This is the fast path introduced with + * per-VMA locking (CONFIG_PER_VMA_LOCK). If this function cannot + * complete the fault (e.g. needs to wait on I/O or encounters a + * condition requiring the mm lock), it must return VM_FAULT_RETRY + * and the caller will fall back to the mmap_lock path below. + * + * 2) mmap_lock path (FAULT_FLAG_VMA_LOCK not set): + * Entered holding a non-exclusive (read) lock on mmap_lock, which + * excludes VMA tree modifications but allows concurrent faults on + * other VMAs. No per-VMA lock is held. + * + * In both cases, on entry the pte is mapped but not yet locked. + * On return, the pte is unmapped and unlocked, and whichever of + * the above locks was held on entry is still held (mmap_lock is + * not dropped, VMA read lock is not dropped, rather, the caller releases it). */ static vm_fault_t do_anonymous_page(struct vm_fault *vmf) { -- 2.34.1 ^ permalink raw reply related [flat|nested] 7+ messages in thread
* Re: [PATCH] mm: update stale locking comment in do_anonymous_page() 2026-03-31 14:29 [PATCH] mm: update stale locking comment in do_anonymous_page() Aditya Sharma @ 2026-04-01 10:52 ` David Hildenbrand (Arm) 2026-04-01 16:42 ` Aditya Sharma 2026-04-05 17:18 ` [PATCH v2] mm/memory: update stale locking comments for fault handlers Aditya Sharma 1 sibling, 1 reply; 7+ messages in thread From: David Hildenbrand (Arm) @ 2026-04-01 10:52 UTC (permalink / raw) To: Aditya Sharma, linux-mm Cc: akpm, ljs, Liam.Howlett, vbabka, rppt, surenb, mhocko, linux-kernel On 3/31/26 16:29, Aditya Sharma wrote: > The comment above do_anonymous_page() dates back to 2005 and describes > the pre-per-VMA-lock world where mmap_lock was always held on entry. > Since CONFIG_PER_VMA_LOCK was introduced (6.4), the fault handler now > has a fast path that enters holding only a per-VMA read lock, with > mmap_lock not held at all. > > Update the comment to describe both entry contexts accurately. > > Signed-off-by: Aditya Sharma <adi.sharma@zohomail.in> > --- > mm/memory.c | 22 +++++++++++++++++++--- > 1 file changed, 19 insertions(+), 3 deletions(-) > > diff --git a/mm/memory.c b/mm/memory.c > index c65e82c86..cc8dbbaea 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -5210,9 +5210,25 @@ static struct folio *alloc_anon_folio(struct vm_fault *vmf) > } > > /* > - * We enter with non-exclusive mmap_lock (to exclude vma changes, > - * but allow concurrent faults), and pte mapped but not yet locked. > - * We return with mmap_lock still held, but pte unmapped and unlocked. > + * We enter in one of two locking contexts: > + * > + * 1) VMA lock path (FAULT_FLAG_VMA_LOCK set): > + * Entered holding a read lock on the faulting VMA (vma_start_read), > + * but NOT holding mmap_lock. This is the fast path introduced with > + * per-VMA locking (CONFIG_PER_VMA_LOCK). If this function cannot > + * complete the fault (e.g. needs to wait on I/O or encounters a > + * condition requiring the mm lock), it must return VM_FAULT_RETRY > + * and the caller will fall back to the mmap_lock path below. > + * > + * 2) mmap_lock path (FAULT_FLAG_VMA_LOCK not set): > + * Entered holding a non-exclusive (read) lock on mmap_lock, which > + * excludes VMA tree modifications but allows concurrent faults on > + * other VMAs. No per-VMA lock is held. > + * > + * In both cases, on entry the pte is mapped but not yet locked. > + * On return, the pte is unmapped and unlocked, and whichever of > + * the above locks was held on entry is still held (mmap_lock is > + * not dropped, VMA read lock is not dropped, rather, the caller releases it). > */ > static vm_fault_t do_anonymous_page(struct vm_fault *vmf) > { Was this AI generated? I don't think we want to have such a wall of text for each and every function that can be called with VMA lock or mmap lock in read mode. -- Cheers, David ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH] mm: update stale locking comment in do_anonymous_page() 2026-04-01 10:52 ` David Hildenbrand (Arm) @ 2026-04-01 16:42 ` Aditya Sharma 2026-04-01 18:47 ` David Hildenbrand (Arm) 0 siblings, 1 reply; 7+ messages in thread From: Aditya Sharma @ 2026-04-01 16:42 UTC (permalink / raw) To: david Cc: linux-mm, akpm, ljs, Liam.Howlett, vbabka, rppt, surenb, mhocko, linux-kernel [-- Attachment #1: Type: text/plain, Size: 2743 bytes --] Sorry about that. The comment was AI assisted and probably too long. I had this confusion myself spending half a day debugging only to realise that the comment was stale. Will send an updated patch with a concise comment Aditya ---- On Wed, 01 Apr 2026 16:22:17 +0530 David Hildenbrand (Arm) <david@kernel.org> wrote ---- On 3/31/26 16:29, Aditya Sharma wrote: > The comment above do_anonymous_page() dates back to 2005 and describes > the pre-per-VMA-lock world where mmap_lock was always held on entry. > Since CONFIG_PER_VMA_LOCK was introduced (6.4), the fault handler now > has a fast path that enters holding only a per-VMA read lock, with > mmap_lock not held at all. > > Update the comment to describe both entry contexts accurately. > > Signed-off-by: Aditya Sharma < mailto:adi.sharma@zohomail.in > > --- > mm/memory.c | 22 +++++++++++++++++++--- > 1 file changed, 19 insertions(+), 3 deletions(-) > > diff --git a/mm/memory.c b/mm/memory.c > index c65e82c86..cc8dbbaea 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -5210,9 +5210,25 @@ static struct folio *alloc_anon_folio(struct vm_fault *vmf) > } > > /* > - * We enter with non-exclusive mmap_lock (to exclude vma changes, > - * but allow concurrent faults), and pte mapped but not yet locked. > - * We return with mmap_lock still held, but pte unmapped and unlocked. > + * We enter in one of two locking contexts: > + * > + * 1) VMA lock path (FAULT_FLAG_VMA_LOCK set): > + * Entered holding a read lock on the faulting VMA (vma_start_read), > + * but NOT holding mmap_lock. This is the fast path introduced with > + * per-VMA locking (CONFIG_PER_VMA_LOCK). If this function cannot > + * complete the fault (e.g. needs to wait on I/O or encounters a > + * condition requiring the mm lock), it must return VM_FAULT_RETRY > + * and the caller will fall back to the mmap_lock path below. > + * > + * 2) mmap_lock path (FAULT_FLAG_VMA_LOCK not set): > + * Entered holding a non-exclusive (read) lock on mmap_lock, which > + * excludes VMA tree modifications but allows concurrent faults on > + * other VMAs. No per-VMA lock is held. > + * > + * In both cases, on entry the pte is mapped but not yet locked. > + * On return, the pte is unmapped and unlocked, and whichever of > + * the above locks was held on entry is still held (mmap_lock is > + * not dropped, VMA read lock is not dropped, rather, the caller releases it). > */ > static vm_fault_t do_anonymous_page(struct vm_fault *vmf) > { Was this AI generated? I don't think we want to have such a wall of text for each and every function that can be called with VMA lock or mmap lock in read mode. -- Cheers, David [-- Attachment #2: Type: text/html, Size: 3690 bytes --] ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH] mm: update stale locking comment in do_anonymous_page() 2026-04-01 16:42 ` Aditya Sharma @ 2026-04-01 18:47 ` David Hildenbrand (Arm) 0 siblings, 0 replies; 7+ messages in thread From: David Hildenbrand (Arm) @ 2026-04-01 18:47 UTC (permalink / raw) To: Aditya Sharma Cc: linux-mm, akpm, ljs, Liam.Howlett, vbabka, rppt, surenb, mhocko, linux-kernel On 4/1/26 18:42, Aditya Sharma wrote: > Sorry about that. The comment was AI assisted and probably too long. I > had this confusion myself spending half a day debugging only to realise > that the comment was stale. Will send an updated patch with a concise > comment Right "mmap lock or vma lock held", for example, is sufficient. While at it, please look for other cases in the file that need similar care. -- Cheers, David ^ permalink raw reply [flat|nested] 7+ messages in thread
* [PATCH v2] mm/memory: update stale locking comments for fault handlers 2026-03-31 14:29 [PATCH] mm: update stale locking comment in do_anonymous_page() Aditya Sharma 2026-04-01 10:52 ` David Hildenbrand (Arm) @ 2026-04-05 17:18 ` Aditya Sharma 2026-04-23 14:27 ` David Hildenbrand (Arm) 1 sibling, 1 reply; 7+ messages in thread From: Aditya Sharma @ 2026-04-05 17:18 UTC (permalink / raw) To: linux-mm Cc: akpm, david, ljs, Liam.Howlett, vbabka, rppt, surenb, mhocko, linux-kernel, Aditya Sharma Update the comments for wp_page_copy(), do_wp_page(), do_swap_page(), do_anonymous_page(), __do_fault(), do_fault(), handle_pte_fault(), __handle_mm_fault(), and handle_mm_fault() to concisely clarify that they can be entered holding either the mmap_lock or the VMA lock, and that the lock may be released upon returning VM_FAULT_RETRY. Additionally, make the following corrections: - In do_anonymous_page(), correct the outdated claim that the function is entered with the PTE "mapped but not yet locked". Since handle_pte_fault() unmaps the empty PTE before routing to do_pte_missing(), the comment now correctly states it is entered with the PTE unmapped and unlocked. - In __do_fault(), update the stale reference from __lock_page_retry() to __folio_lock_or_retry(). Signed-off-by: Aditya Sharma <adi.sharma@zohomail.in> --- v2: - Simplified the comment to concisely state "either the VMA lock or the mmap_lock" instead of a verbose explanation (per David Hildenbrand). - Expanded the scope to cover 8 other fault handlers in mm/memory.c that suffered from the same stale mmap_lock comments. - Fixed an additional historical inaccuracy in do_anonymous_page() regarding the PTE mapping state on entry. - Updated a stale reference in __do_fault() from __lock_page_retry() to __folio_lock_or_retry(). mm/memory.c | 49 ++++++++++++++++++++++++++----------------------- 1 file changed, 26 insertions(+), 23 deletions(-) diff --git a/mm/memory.c b/mm/memory.c index c65e82c86..2b407e3f9 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -3742,8 +3742,8 @@ vm_fault_t __vmf_anon_prepare(struct vm_fault *vmf) * Handle the case of a page which we actually need to copy to a new page, * either due to COW or unsharing. * - * Called with mmap_lock locked and the old page referenced, but - * without the ptl held. + * Called with either the VMA lock or the mmap_lock (FAULT_FLAG_VMA_LOCK + * tells you which) and the old page referenced, but without the ptl held. * * High level logic flow: * @@ -4142,9 +4142,9 @@ static bool wp_can_reuse_anon_folio(struct folio *folio, * though the page will change only once the write actually happens. This * avoids a few races, and potentially makes it more efficient. * - * We enter with non-exclusive mmap_lock (to exclude vma changes, - * but allow concurrent faults), with pte both mapped and locked. - * We return with mmap_lock still held, but pte unmapped and unlocked. + * We enter with either the VMA lock or the mmap_lock (FAULT_FLAG_VMA_LOCK + * tells you which), and pte both mapped and locked. We return with + * the same lock still held, but pte unmapped and unlocked. */ static vm_fault_t do_wp_page(struct vm_fault *vmf) __releases(vmf->ptl) @@ -4696,11 +4696,11 @@ static void check_swap_exclusive(struct folio *folio, swp_entry_t entry, } /* - * We enter with non-exclusive mmap_lock (to exclude vma changes, - * but allow concurrent faults), and pte mapped but not yet locked. + * We enter with either the VMA lock or the mmap_lock (FAULT_FLAG_VMA_LOCK + * tells you which), and pte mapped but not yet locked. * We return with pte unmapped and unlocked. * - * We return with the mmap_lock locked or unlocked in the same cases + * We return with the lock locked or unlocked in the same cases * as does filemap_fault(). */ vm_fault_t do_swap_page(struct vm_fault *vmf) @@ -5210,9 +5210,10 @@ static struct folio *alloc_anon_folio(struct vm_fault *vmf) } /* - * We enter with non-exclusive mmap_lock (to exclude vma changes, - * but allow concurrent faults), and pte mapped but not yet locked. - * We return with mmap_lock still held, but pte unmapped and unlocked. + * We enter with either the VMA lock or the mmap_lock (FAULT_FLAG_VMA_LOCK + * tells you which), and pte unmapped and unlocked. + * We return with the lock still held, but pte unmapped and unlocked. + * If VM_FAULT_RETRY is returned, the lock may have been released. */ static vm_fault_t do_anonymous_page(struct vm_fault *vmf) { @@ -5330,9 +5331,10 @@ static vm_fault_t do_anonymous_page(struct vm_fault *vmf) } /* - * The mmap_lock must have been held on entry, and may have been - * released depending on flags and vma->vm_ops->fault() return value. - * See filemap_fault() and __lock_page_retry(). + * Either the VMA lock or the mmap_lock must have been held on entry, + * and may have been released depending on flags and vma->vm_ops->fault() + * return value. + * See filemap_fault() and __folio_lock_or_retry(). */ static vm_fault_t __do_fault(struct vm_fault *vmf) { @@ -5893,11 +5895,11 @@ static vm_fault_t do_shared_fault(struct vm_fault *vmf) } /* - * We enter with non-exclusive mmap_lock (to exclude vma changes, - * but allow concurrent faults). - * The mmap_lock may have been released depending on flags and our + * We enter with either the VMA lock or the mmap_lock (FAULT_FLAG_VMA_LOCK + * tells you which). + * The lock may have been released depending on flags and our * return value. See filemap_fault() and __folio_lock_or_retry(). - * If mmap_lock is released, vma may become invalid (for example + * If the lock is released, vma may become invalid (for example * by other thread calling munmap()). */ static vm_fault_t do_fault(struct vm_fault *vmf) @@ -6264,10 +6266,11 @@ static void fix_spurious_fault(struct vm_fault *vmf, * with external mmu caches can use to update those (ie the Sparc or * PowerPC hashed page tables that act as extended TLBs). * - * We enter with non-exclusive mmap_lock (to exclude vma changes, but allow - * concurrent faults). + * On entry, we hold either the VMA lock or the mmap_lock + * (FAULT_FLAG_VMA_LOCK tells you which). * - * The mmap_lock may have been released depending on flags and our return value. + * The mmap_lock or VMA lock may have been released depending on flags + * and our return value. * See filemap_fault() and __folio_lock_or_retry(). */ static vm_fault_t handle_pte_fault(struct vm_fault *vmf) @@ -6349,7 +6352,7 @@ static vm_fault_t handle_pte_fault(struct vm_fault *vmf) /* * On entry, we hold either the VMA lock or the mmap_lock * (FAULT_FLAG_VMA_LOCK tells you which). If VM_FAULT_RETRY is set in - * the result, the mmap_lock is not held on exit. See filemap_fault() + * the result, the lock is not held on exit. See filemap_fault() * and __folio_lock_or_retry(). */ static vm_fault_t __handle_mm_fault(struct vm_area_struct *vma, @@ -6583,7 +6586,7 @@ static vm_fault_t sanitize_fault_flags(struct vm_area_struct *vma, * By the time we get here, we already hold either the VMA lock or the * mmap_lock (FAULT_FLAG_VMA_LOCK tells you which). * - * The mmap_lock may have been released depending on flags and our + * The lock may have been released depending on flags and our * return value. See filemap_fault() and __folio_lock_or_retry(). */ vm_fault_t handle_mm_fault(struct vm_area_struct *vma, unsigned long address, -- 2.34.1 ^ permalink raw reply related [flat|nested] 7+ messages in thread
* Re: [PATCH v2] mm/memory: update stale locking comments for fault handlers 2026-04-05 17:18 ` [PATCH v2] mm/memory: update stale locking comments for fault handlers Aditya Sharma @ 2026-04-23 14:27 ` David Hildenbrand (Arm) 2026-04-23 15:00 ` Aditya Sharma 0 siblings, 1 reply; 7+ messages in thread From: David Hildenbrand (Arm) @ 2026-04-23 14:27 UTC (permalink / raw) To: Aditya Sharma, linux-mm Cc: akpm, ljs, Liam.Howlett, vbabka, rppt, surenb, mhocko, linux-kernel On 4/5/26 19:18, Aditya Sharma wrote: > Update the comments for wp_page_copy(), do_wp_page(), do_swap_page(), > do_anonymous_page(), __do_fault(), do_fault(), handle_pte_fault(), > __handle_mm_fault(), and handle_mm_fault() to concisely clarify that > they can be entered holding either the mmap_lock or the VMA lock, > and that the lock may be released upon returning VM_FAULT_RETRY. > > Additionally, make the following corrections: > - In do_anonymous_page(), correct the outdated claim that the function > is entered with the PTE "mapped but not yet locked". Since > handle_pte_fault() unmaps the empty PTE before routing to > do_pte_missing(), the comment now correctly states it is entered > with the PTE unmapped and unlocked. > - In __do_fault(), update the stale reference from __lock_page_retry() > to __folio_lock_or_retry(). > > Signed-off-by: Aditya Sharma <adi.sharma@zohomail.in> Next time, please don't send as reply to your previous version > --- > v2: > - Simplified the comment to concisely state "either the VMA lock or > the mmap_lock" instead of a verbose explanation (per David Hildenbrand). > - Expanded the scope to cover 8 other fault handlers in mm/memory.c > that suffered from the same stale mmap_lock comments. > - Fixed an additional historical inaccuracy in do_anonymous_page() > regarding the PTE mapping state on entry. > - Updated a stale reference in __do_fault() from __lock_page_retry() > to __folio_lock_or_retry(). > > > mm/memory.c | 49 ++++++++++++++++++++++++++----------------------- > 1 file changed, 26 insertions(+), 23 deletions(-) > > diff --git a/mm/memory.c b/mm/memory.c > index c65e82c86..2b407e3f9 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -3742,8 +3742,8 @@ vm_fault_t __vmf_anon_prepare(struct vm_fault *vmf) > * Handle the case of a page which we actually need to copy to a new page, > * either due to COW or unsharing. > * > - * Called with mmap_lock locked and the old page referenced, but > - * without the ptl held. > + * Called with either the VMA lock or the mmap_lock (FAULT_FLAG_VMA_LOCK > + * tells you which) "Called with either the VMA lock or the mmap_lock *held*" Same for other cases. Can we shorten "FAULT_FLAG_VMA_LOCK tells you which" to "see FAULT_FLAG_VMA_LOCK" ? -- Cheers, David ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH v2] mm/memory: update stale locking comments for fault handlers 2026-04-23 14:27 ` David Hildenbrand (Arm) @ 2026-04-23 15:00 ` Aditya Sharma 0 siblings, 0 replies; 7+ messages in thread From: Aditya Sharma @ 2026-04-23 15:00 UTC (permalink / raw) To: David Hildenbrand (Arm) Cc: linux-mm, akpm, ljs, Liam.Howlett, vbabka, rppt, surenb, mhocko, linux-kernel Hi David, Understood. Will send v3 as a new thread, not a reply. I'll apply the wording changes ("held" and "see FAULT_FLAG_VMA_LOCK") and send out the v3 shortly Thanks, Aditya From: David Hildenbrand (Arm) <david@kernel.org> To: "Aditya Sharma"<adi.sharma@zohomail.in>, <linux-mm@kvack.org> Cc: <akpm@linux-foundation.org>, <ljs@kernel.org>, <Liam.Howlett@oracle.com>, <vbabka@kernel.org>, <rppt@kernel.org>, <surenb@google.com>, <mhocko@suse.com>, <linux-kernel@vger.kernel.org> Date: Thu, 23 Apr 2026 19:57:21 +0530 Subject: Re: [PATCH v2] mm/memory: update stale locking comments for fault handlers > On 4/5/26 19:18, Aditya Sharma wrote: > > Update the comments for wp_page_copy(), do_wp_page(), do_swap_page(), > > do_anonymous_page(), __do_fault(), do_fault(), handle_pte_fault(), > > __handle_mm_fault(), and handle_mm_fault() to concisely clarify that > > they can be entered holding either the mmap_lock or the VMA lock, > > and that the lock may be released upon returning VM_FAULT_RETRY. > > > > Additionally, make the following corrections: > > - In do_anonymous_page(), correct the outdated claim that the function > > is entered with the PTE "mapped but not yet locked". Since > > handle_pte_fault() unmaps the empty PTE before routing to > > do_pte_missing(), the comment now correctly states it is entered > > with the PTE unmapped and unlocked. > > - In __do_fault(), update the stale reference from __lock_page_retry() > > to __folio_lock_or_retry(). > > > > Signed-off-by: Aditya Sharma <adi.sharma@zohomail.in> > > Next time, please don't send as reply to your previous version > > > --- > > v2: > > - Simplified the comment to concisely state "either the VMA lock or > > the mmap_lock" instead of a verbose explanation (per David Hildenbrand). > > - Expanded the scope to cover 8 other fault handlers in mm/memory.c > > that suffered from the same stale mmap_lock comments. > > - Fixed an additional historical inaccuracy in do_anonymous_page() > > regarding the PTE mapping state on entry. > > - Updated a stale reference in __do_fault() from __lock_page_retry() > > to __folio_lock_or_retry(). > > > > > > mm/memory.c | 49 ++++++++++++++++++++++++++----------------------- > > 1 file changed, 26 insertions(+), 23 deletions(-) > > > > diff --git a/mm/memory.c b/mm/memory.c > > index c65e82c86..2b407e3f9 100644 > > --- a/mm/memory.c > > +++ b/mm/memory.c > > @@ -3742,8 +3742,8 @@ vm_fault_t __vmf_anon_prepare(struct vm_fault *vmf) > > * Handle the case of a page which we actually need to copy to a new page, > > * either due to COW or unsharing. > > * > > - * Called with mmap_lock locked and the old page referenced, but > > - * without the ptl held. > > + * Called with either the VMA lock or the mmap_lock (FAULT_FLAG_VMA_LOCK > > + * tells you which) > > "Called with either the VMA lock or the mmap_lock *held*" > > Same for other cases. > > Can we shorten "FAULT_FLAG_VMA_LOCK tells you which" to "see FAULT_FLAG_VMA_LOCK" > > ? > > -- > Cheers, > > David > ^ permalink raw reply [flat|nested] 7+ messages in thread
end of thread, other threads:[~2026-04-23 15:01 UTC | newest] Thread overview: 7+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2026-03-31 14:29 [PATCH] mm: update stale locking comment in do_anonymous_page() Aditya Sharma 2026-04-01 10:52 ` David Hildenbrand (Arm) 2026-04-01 16:42 ` Aditya Sharma 2026-04-01 18:47 ` David Hildenbrand (Arm) 2026-04-05 17:18 ` [PATCH v2] mm/memory: update stale locking comments for fault handlers Aditya Sharma 2026-04-23 14:27 ` David Hildenbrand (Arm) 2026-04-23 15:00 ` Aditya Sharma
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox