From: Dev Jain <dev.jain@arm.com>
To: "Lorenzo Stoakes (Oracle)" <ljs@kernel.org>
Cc: akpm@linux-foundation.org, axelrasmussen@google.com,
yuanchu@google.com, david@kernel.org, hughd@google.com,
chrisl@kernel.org, kasong@tencent.com, weixugc@google.com,
Liam.Howlett@oracle.com, vbabka@kernel.org, rppt@kernel.org,
surenb@google.com, mhocko@suse.com, riel@surriel.com,
harry.yoo@oracle.com, jannh@google.com, pfalcato@suse.de,
baolin.wang@linux.alibaba.com, shikemeng@huaweicloud.com,
nphamcs@gmail.com, bhe@redhat.com, baohua@kernel.org,
youngjun.park@lge.com, ziy@nvidia.com, kas@kernel.org,
willy@infradead.org, yuzhao@google.com, linux-mm@kvack.org,
linux-kernel@vger.kernel.org, ryan.roberts@arm.com,
anshuman.khandual@arm.com
Subject: Re: [PATCH 3/9] mm/rmap: refactor lazyfree unmap commit path to commit_ttu_lazyfree_folio()
Date: Tue, 10 Mar 2026 14:12:45 +0530 [thread overview]
Message-ID: <71e7df79-cfe1-4f3d-b00c-8ac53f886c04@arm.com> (raw)
In-Reply-To: <3915e3bc-c390-44bc-a463-bdd687598284@lucifer.local>
On 10/03/26 1:49 pm, Lorenzo Stoakes (Oracle) wrote:
> On Tue, Mar 10, 2026 at 01:00:07PM +0530, Dev Jain wrote:
>> Clean up the code by refactoring the post-pte-clearing path of lazyfree
>> folio unmapping, into commit_ttu_lazyfree_folio().
>>
>> No functional change is intended.
>>
>> Signed-off-by: Dev Jain <dev.jain@arm.com>
>
> This is a good idea, and we need more refactoring like this in the rmap code,
> but comments/nits below.
>
>> ---
>> mm/rmap.c | 93 ++++++++++++++++++++++++++++++++-----------------------
>> 1 file changed, 54 insertions(+), 39 deletions(-)
>>
>> diff --git a/mm/rmap.c b/mm/rmap.c
>> index 1fa020edd954a..a61978141ee3f 100644
>> --- a/mm/rmap.c
>> +++ b/mm/rmap.c
>> @@ -1966,6 +1966,57 @@ static inline unsigned int folio_unmap_pte_batch(struct folio *folio,
>> FPB_RESPECT_WRITE | FPB_RESPECT_SOFT_DIRTY);
>> }
>>
>> +static inline int commit_ttu_lazyfree_folio(struct vm_area_struct *vma,
>
> Strange name, maybe lazyfree_range()? Not sure what ttu has to do with
ttu means try_to_unmap, just like it is used in TTU_SYNC,
TTU_SPLIT_HUGE_PMD, etc. So personally I really like the name, it reads
"commit the try-to-unmap of a lazyfree folio". The "commit" comes because
the pte clearing has already happened, so now we are deciding if at all
to back-off and restore the ptes.
> anything...
>
>> + struct folio *folio, unsigned long address, pte_t *ptep,
>> + pte_t pteval, long nr_pages)
>
> That long nr_pages is really grating now...
Reading the discussion on patch 1, I'll convert this to unsigned long.
>
>> +{
>
> Come on Dev, it's 2026, why on earth are you returning an integer and not a
> bool?
>
> Also it would make sense for this to return false if something breaks, otherwise
> true.
Yes I was confused on which one of the options to choose :). Since the
function does a lot more than just test some functionality (which is what
boolean functions usually do) I was feeling weird when returning bool.
But yeah alright, I'll convert this to bool.
>
>> + struct mm_struct *mm = vma->vm_mm;
>> + int ref_count, map_count;
>> +
>> + /*
>> + * Synchronize with gup_pte_range():
>> + * - clear PTE; barrier; read refcount
>> + * - inc refcount; barrier; read PTE
>> + */
>> + smp_mb();
>> +
>> + ref_count = folio_ref_count(folio);
>> + map_count = folio_mapcount(folio);
>> +
>> + /*
>> + * Order reads for page refcount and dirty flag
>> + * (see comments in __remove_mapping()).
>> + */
>> + smp_rmb();
>> +
>> + if (folio_test_dirty(folio) && !(vma->vm_flags & VM_DROPPABLE)) {
>> + /*
>> + * redirtied either using the page table or a previously
>> + * obtained GUP reference.
>> + */
>> + set_ptes(mm, address, ptep, pteval, nr_pages);
>> + folio_set_swapbacked(folio);
>> + return 1;
>> + }
>> +
>> + if (ref_count != 1 + map_count) {
>> + /*
>> + * Additional reference. Could be a GUP reference or any
>> + * speculative reference. GUP users must mark the folio
>> + * dirty if there was a modification. This folio cannot be
>> + * reclaimed right now either way, so act just like nothing
>> + * happened.
>> + * We'll come back here later and detect if the folio was
>> + * dirtied when the additional reference is gone.
>> + */
>> + set_ptes(mm, address, ptep, pteval, nr_pages);
>> + return 1;
>> + }
>> +
>> + add_mm_counter(mm, MM_ANONPAGES, -nr_pages);
>> + return 0;
>> +}
>> +
>> /*
>> * @arg: enum ttu_flags will be passed to this argument
>> */
>> @@ -2227,46 +2278,10 @@ static bool try_to_unmap_one(struct folio *folio, struct vm_area_struct *vma,
>>
>> /* MADV_FREE page check */
>> if (!folio_test_swapbacked(folio)) {
>> - int ref_count, map_count;
>> -
>> - /*
>> - * Synchronize with gup_pte_range():
>> - * - clear PTE; barrier; read refcount
>> - * - inc refcount; barrier; read PTE
>> - */
>> - smp_mb();
>> -
>> - ref_count = folio_ref_count(folio);
>> - map_count = folio_mapcount(folio);
>> -
>> - /*
>> - * Order reads for page refcount and dirty flag
>> - * (see comments in __remove_mapping()).
>> - */
>> - smp_rmb();
>> -
>> - if (folio_test_dirty(folio) && !(vma->vm_flags & VM_DROPPABLE)) {
>> - /*
>> - * redirtied either using the page table or a previously
>> - * obtained GUP reference.
>> - */
>> - set_ptes(mm, address, pvmw.pte, pteval, nr_pages);
>> - folio_set_swapbacked(folio);
>> + if (commit_ttu_lazyfree_folio(vma, folio, address,
>> + pvmw.pte, pteval,
>> + nr_pages))
>
> With above corrections this would be:
>
> if (!lazyfree_range(vma, folio, address, pvme.pte, pteval, nr_pages))
> ...
>
>> goto walk_abort;
>> - } else if (ref_count != 1 + map_count) {
>> - /*
>> - * Additional reference. Could be a GUP reference or any
>> - * speculative reference. GUP users must mark the folio
>> - * dirty if there was a modification. This folio cannot be
>> - * reclaimed right now either way, so act just like nothing
>> - * happened.
>> - * We'll come back here later and detect if the folio was
>> - * dirtied when the additional reference is gone.
>> - */
>> - set_ptes(mm, address, pvmw.pte, pteval, nr_pages);
>> - goto walk_abort;
>> - }
>> - add_mm_counter(mm, MM_ANONPAGES, -nr_pages);
>> goto discard;
>> }
>>
>> --
>> 2.34.1
>>
>
> Thanks, Lorenzo
>
next prev parent reply other threads:[~2026-03-10 8:43 UTC|newest]
Thread overview: 47+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-03-10 7:30 [PATCH 0/9] mm/rmap: Optimize anonymous large folio unmapping Dev Jain
2026-03-10 7:30 ` [PATCH 1/9] mm/rmap: make nr_pages signed in try_to_unmap_one Dev Jain
2026-03-10 7:56 ` Lorenzo Stoakes (Oracle)
2026-03-10 8:06 ` David Hildenbrand (Arm)
2026-03-10 8:23 ` Dev Jain
2026-03-10 12:40 ` Matthew Wilcox
2026-03-11 4:54 ` Dev Jain
2026-03-10 7:30 ` [PATCH 2/9] mm/rmap: initialize nr_pages to 1 at loop start " Dev Jain
2026-03-10 8:10 ` Lorenzo Stoakes (Oracle)
2026-03-10 8:31 ` Dev Jain
2026-03-10 8:39 ` Lorenzo Stoakes (Oracle)
2026-03-10 8:43 ` Dev Jain
2026-03-10 7:30 ` [PATCH 3/9] mm/rmap: refactor lazyfree unmap commit path to commit_ttu_lazyfree_folio() Dev Jain
2026-03-10 8:19 ` Lorenzo Stoakes (Oracle)
2026-03-10 8:42 ` Dev Jain [this message]
2026-03-19 15:53 ` Lorenzo Stoakes (Oracle)
2026-03-10 7:30 ` [PATCH 4/9] mm/memory: Batch set uffd-wp markers during zapping Dev Jain
2026-03-10 7:30 ` [PATCH 5/9] mm/rmap: batch unmap folios belonging to uffd-wp VMAs Dev Jain
2026-03-10 8:34 ` Lorenzo Stoakes (Oracle)
2026-03-10 23:32 ` Barry Song
2026-03-11 4:14 ` Barry Song
2026-03-11 4:52 ` Dev Jain
2026-03-11 4:56 ` Dev Jain
2026-03-10 7:30 ` [PATCH 6/9] mm/swapfile: Make folio_dup_swap batchable Dev Jain
2026-03-10 8:27 ` Kairui Song
2026-03-10 8:46 ` Dev Jain
2026-03-10 8:49 ` Lorenzo Stoakes (Oracle)
2026-03-11 5:42 ` Dev Jain
2026-03-19 15:26 ` Lorenzo Stoakes (Oracle)
2026-03-19 16:47 ` Matthew Wilcox
2026-03-18 0:20 ` kernel test robot
2026-03-10 7:30 ` [PATCH 7/9] mm/swapfile: Make folio_put_swap batchable Dev Jain
2026-03-10 8:29 ` Kairui Song
2026-03-10 8:50 ` Dev Jain
2026-03-10 8:55 ` Lorenzo Stoakes (Oracle)
2026-03-18 1:04 ` kernel test robot
2026-03-10 7:30 ` [PATCH 8/9] mm/rmap: introduce folio_try_share_anon_rmap_ptes Dev Jain
2026-03-10 9:38 ` Lorenzo Stoakes (Oracle)
2026-03-11 8:09 ` Dev Jain
2026-03-12 8:19 ` Wei Yang
2026-03-19 15:47 ` Lorenzo Stoakes (Oracle)
2026-04-08 7:14 ` Dev Jain
2026-03-10 7:30 ` [PATCH 9/9] mm/rmap: enable batch unmapping of anonymous folios Dev Jain
2026-03-10 8:02 ` [PATCH 0/9] mm/rmap: Optimize anonymous large folio unmapping Lorenzo Stoakes (Oracle)
2026-03-10 9:28 ` Dev Jain
2026-03-10 12:59 ` Lance Yang
2026-03-11 8:11 ` Dev Jain
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=71e7df79-cfe1-4f3d-b00c-8ac53f886c04@arm.com \
--to=dev.jain@arm.com \
--cc=Liam.Howlett@oracle.com \
--cc=akpm@linux-foundation.org \
--cc=anshuman.khandual@arm.com \
--cc=axelrasmussen@google.com \
--cc=baohua@kernel.org \
--cc=baolin.wang@linux.alibaba.com \
--cc=bhe@redhat.com \
--cc=chrisl@kernel.org \
--cc=david@kernel.org \
--cc=harry.yoo@oracle.com \
--cc=hughd@google.com \
--cc=jannh@google.com \
--cc=kas@kernel.org \
--cc=kasong@tencent.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=ljs@kernel.org \
--cc=mhocko@suse.com \
--cc=nphamcs@gmail.com \
--cc=pfalcato@suse.de \
--cc=riel@surriel.com \
--cc=rppt@kernel.org \
--cc=ryan.roberts@arm.com \
--cc=shikemeng@huaweicloud.com \
--cc=surenb@google.com \
--cc=vbabka@kernel.org \
--cc=weixugc@google.com \
--cc=willy@infradead.org \
--cc=youngjun.park@lge.com \
--cc=yuanchu@google.com \
--cc=yuzhao@google.com \
--cc=ziy@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.