* [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page
@ 2025-08-06 14:56 Dev Jain
2025-08-06 16:40 ` Lorenzo Stoakes
2025-08-06 20:21 ` David Hildenbrand
0 siblings, 2 replies; 5+ messages in thread
From: Dev Jain @ 2025-08-06 14:56 UTC (permalink / raw)
To: akpm
Cc: ryan.roberts, david, willy, linux-mm, linux-kernel,
catalin.marinas, will, Liam.Howlett, lorenzo.stoakes, vbabka,
jannh, anshuman.khandual, peterx, joey.gouly, ioworker0, baohua,
kevin.brodsky, quic_zhenhuah, christophe.leroy, yangyicong,
linux-arm-kernel, hughd, yang, ziy, Dev Jain,
syzbot+57bcc752f0df8bb1365c
In commit_anon_folio_batch(), we iterate over all pages pointed to by the
PTE batch. Therefore we need to know the first page of the batch;
currently we derive that via folio_page(folio, 0), but, that takes us
to the first (head) page of the folio instead - our PTE batch may lie
in the middle of the folio, leading to incorrectness.
Bite the bullet and throw away the micro-optimization of reusing the
folio in favour of code simplicity. Derive the page and the folio in
change_pte_range, and pass the page too to commit_anon_folio_batch to
fix the aforementioned issue.
Reported-by: syzbot+57bcc752f0df8bb1365c@syzkaller.appspotmail.com
Fixes: cac1db8c3aad ("mm: optimize mprotect() by PTE batching")
Signed-off-by: Dev Jain <dev.jain@arm.com>
---
mm/mprotect.c | 23 ++++++++++-------------
1 file changed, 10 insertions(+), 13 deletions(-)
diff --git a/mm/mprotect.c b/mm/mprotect.c
index 78bded7acf79..113b48985834 100644
--- a/mm/mprotect.c
+++ b/mm/mprotect.c
@@ -120,9 +120,8 @@ static int mprotect_folio_pte_batch(struct folio *folio, pte_t *ptep,
static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
pte_t oldpte, pte_t *pte, int target_node,
- struct folio **foliop)
+ struct folio *folio)
{
- struct folio *folio = NULL;
bool ret = true;
bool toptier;
int nid;
@@ -131,7 +130,6 @@ static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
if (pte_protnone(oldpte))
goto skip;
- folio = vm_normal_folio(vma, addr, oldpte);
if (!folio)
goto skip;
@@ -173,7 +171,6 @@ static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
folio_xchg_access_time(folio, jiffies_to_msecs(jiffies));
skip:
- *foliop = folio;
return ret;
}
@@ -231,10 +228,9 @@ static int page_anon_exclusive_sub_batch(int start_idx, int max_len,
* retrieve sub-batches.
*/
static void commit_anon_folio_batch(struct vm_area_struct *vma,
- struct folio *folio, unsigned long addr, pte_t *ptep,
+ struct folio *folio, struct page *first_page, unsigned long addr, pte_t *ptep,
pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb)
{
- struct page *first_page = folio_page(folio, 0);
bool expected_anon_exclusive;
int sub_batch_idx = 0;
int len;
@@ -251,7 +247,7 @@ static void commit_anon_folio_batch(struct vm_area_struct *vma,
}
static void set_write_prot_commit_flush_ptes(struct vm_area_struct *vma,
- struct folio *folio, unsigned long addr, pte_t *ptep,
+ struct folio *folio, struct page *page, unsigned long addr, pte_t *ptep,
pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb)
{
bool set_write;
@@ -270,7 +266,7 @@ static void set_write_prot_commit_flush_ptes(struct vm_area_struct *vma,
/* idx = */ 0, set_write, tlb);
return;
}
- commit_anon_folio_batch(vma, folio, addr, ptep, oldpte, ptent, nr_ptes, tlb);
+ commit_anon_folio_batch(vma, folio, page, addr, ptep, oldpte, ptent, nr_ptes, tlb);
}
static long change_pte_range(struct mmu_gather *tlb,
@@ -305,15 +301,19 @@ static long change_pte_range(struct mmu_gather *tlb,
const fpb_t flags = FPB_RESPECT_SOFT_DIRTY | FPB_RESPECT_WRITE;
int max_nr_ptes = (end - addr) >> PAGE_SHIFT;
struct folio *folio = NULL;
+ struct page *page;
pte_t ptent;
+ page = vm_normal_page(vma, addr, oldpte);
+ if (page)
+ folio = page_folio(page);
/*
* Avoid trapping faults against the zero or KSM
* pages. See similar comment in change_huge_pmd.
*/
if (prot_numa) {
int ret = prot_numa_skip(vma, addr, oldpte, pte,
- target_node, &folio);
+ target_node, folio);
if (ret) {
/* determine batch to skip */
@@ -323,9 +323,6 @@ static long change_pte_range(struct mmu_gather *tlb,
}
}
- if (!folio)
- folio = vm_normal_folio(vma, addr, oldpte);
-
nr_ptes = mprotect_folio_pte_batch(folio, pte, oldpte, max_nr_ptes, flags);
oldpte = modify_prot_start_ptes(vma, addr, pte, nr_ptes);
@@ -351,7 +348,7 @@ static long change_pte_range(struct mmu_gather *tlb,
*/
if ((cp_flags & MM_CP_TRY_CHANGE_WRITABLE) &&
!pte_write(ptent))
- set_write_prot_commit_flush_ptes(vma, folio,
+ set_write_prot_commit_flush_ptes(vma, folio, page,
addr, pte, oldpte, ptent, nr_ptes, tlb);
else
prot_commit_flush_ptes(vma, addr, pte, oldpte, ptent,
--
2.30.2
^ permalink raw reply related [flat|nested] 5+ messages in thread
* Re: [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page
2025-08-06 14:56 [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page Dev Jain
@ 2025-08-06 16:40 ` Lorenzo Stoakes
2025-08-06 17:19 ` Dev Jain
2025-08-06 20:21 ` David Hildenbrand
1 sibling, 1 reply; 5+ messages in thread
From: Lorenzo Stoakes @ 2025-08-06 16:40 UTC (permalink / raw)
To: Dev Jain
Cc: akpm, ryan.roberts, david, willy, linux-mm, linux-kernel,
catalin.marinas, will, Liam.Howlett, vbabka, jannh,
anshuman.khandual, peterx, joey.gouly, ioworker0, baohua,
kevin.brodsky, quic_zhenhuah, christophe.leroy, yangyicong,
linux-arm-kernel, hughd, yang, ziy, syzbot+57bcc752f0df8bb1365c
On Wed, Aug 06, 2025 at 08:26:11PM +0530, Dev Jain wrote:
> In commit_anon_folio_batch(), we iterate over all pages pointed to by the
> PTE batch. Therefore we need to know the first page of the batch;
> currently we derive that via folio_page(folio, 0), but, that takes us
> to the first (head) page of the folio instead - our PTE batch may lie
> in the middle of the folio, leading to incorrectness.
>
> Bite the bullet and throw away the micro-optimization of reusing the
> folio in favour of code simplicity. Derive the page and the folio in
> change_pte_range, and pass the page too to commit_anon_folio_batch to
> fix the aforementioned issue.
>
> Reported-by: syzbot+57bcc752f0df8bb1365c@syzkaller.appspotmail.com
> Fixes: cac1db8c3aad ("mm: optimize mprotect() by PTE batching")
> Signed-off-by: Dev Jain <dev.jain@arm.com>
This looks reasonable, fixes the problem and compiles/works on my machine so:
Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
This badly needs refactoring as 13 parameters being passed to a function is
ridiculous, but we can do that later.
Let's get this in as a hotfix asap.
> ---
> mm/mprotect.c | 23 ++++++++++-------------
> 1 file changed, 10 insertions(+), 13 deletions(-)
>
> diff --git a/mm/mprotect.c b/mm/mprotect.c
> index 78bded7acf79..113b48985834 100644
> --- a/mm/mprotect.c
> +++ b/mm/mprotect.c
> @@ -120,9 +120,8 @@ static int mprotect_folio_pte_batch(struct folio *folio, pte_t *ptep,
>
> static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
> pte_t oldpte, pte_t *pte, int target_node,
> - struct folio **foliop)
> + struct folio *folio)
> {
> - struct folio *folio = NULL;
> bool ret = true;
> bool toptier;
> int nid;
> @@ -131,7 +130,6 @@ static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
> if (pte_protnone(oldpte))
> goto skip;
>
> - folio = vm_normal_folio(vma, addr, oldpte);
> if (!folio)
> goto skip;
>
> @@ -173,7 +171,6 @@ static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
> folio_xchg_access_time(folio, jiffies_to_msecs(jiffies));
>
> skip:
> - *foliop = folio;
> return ret;
> }
>
> @@ -231,10 +228,9 @@ static int page_anon_exclusive_sub_batch(int start_idx, int max_len,
> * retrieve sub-batches.
> */
> static void commit_anon_folio_batch(struct vm_area_struct *vma,
> - struct folio *folio, unsigned long addr, pte_t *ptep,
> + struct folio *folio, struct page *first_page, unsigned long addr, pte_t *ptep,
> pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb)
> {
> - struct page *first_page = folio_page(folio, 0);
> bool expected_anon_exclusive;
> int sub_batch_idx = 0;
> int len;
> @@ -251,7 +247,7 @@ static void commit_anon_folio_batch(struct vm_area_struct *vma,
> }
>
> static void set_write_prot_commit_flush_ptes(struct vm_area_struct *vma,
> - struct folio *folio, unsigned long addr, pte_t *ptep,
> + struct folio *folio, struct page *page, unsigned long addr, pte_t *ptep,
> pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb)
> {
> bool set_write;
> @@ -270,7 +266,7 @@ static void set_write_prot_commit_flush_ptes(struct vm_area_struct *vma,
> /* idx = */ 0, set_write, tlb);
> return;
> }
> - commit_anon_folio_batch(vma, folio, addr, ptep, oldpte, ptent, nr_ptes, tlb);
> + commit_anon_folio_batch(vma, folio, page, addr, ptep, oldpte, ptent, nr_ptes, tlb);
> }
>
> static long change_pte_range(struct mmu_gather *tlb,
> @@ -305,15 +301,19 @@ static long change_pte_range(struct mmu_gather *tlb,
> const fpb_t flags = FPB_RESPECT_SOFT_DIRTY | FPB_RESPECT_WRITE;
> int max_nr_ptes = (end - addr) >> PAGE_SHIFT;
> struct folio *folio = NULL;
> + struct page *page;
> pte_t ptent;
>
> + page = vm_normal_page(vma, addr, oldpte);
> + if (page)
> + folio = page_folio(page);
> /*
> * Avoid trapping faults against the zero or KSM
> * pages. See similar comment in change_huge_pmd.
> */
> if (prot_numa) {
> int ret = prot_numa_skip(vma, addr, oldpte, pte,
> - target_node, &folio);
> + target_node, folio);
> if (ret) {
>
> /* determine batch to skip */
> @@ -323,9 +323,6 @@ static long change_pte_range(struct mmu_gather *tlb,
> }
> }
>
> - if (!folio)
> - folio = vm_normal_folio(vma, addr, oldpte);
> -
> nr_ptes = mprotect_folio_pte_batch(folio, pte, oldpte, max_nr_ptes, flags);
>
> oldpte = modify_prot_start_ptes(vma, addr, pte, nr_ptes);
> @@ -351,7 +348,7 @@ static long change_pte_range(struct mmu_gather *tlb,
> */
> if ((cp_flags & MM_CP_TRY_CHANGE_WRITABLE) &&
> !pte_write(ptent))
> - set_write_prot_commit_flush_ptes(vma, folio,
> + set_write_prot_commit_flush_ptes(vma, folio, page,
> addr, pte, oldpte, ptent, nr_ptes, tlb);
> else
> prot_commit_flush_ptes(vma, addr, pte, oldpte, ptent,
> --
> 2.30.2
>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page
2025-08-06 16:40 ` Lorenzo Stoakes
@ 2025-08-06 17:19 ` Dev Jain
2025-08-06 17:23 ` Lorenzo Stoakes
0 siblings, 1 reply; 5+ messages in thread
From: Dev Jain @ 2025-08-06 17:19 UTC (permalink / raw)
To: Lorenzo Stoakes
Cc: akpm, ryan.roberts, david, willy, linux-mm, linux-kernel,
catalin.marinas, will, Liam.Howlett, vbabka, jannh,
anshuman.khandual, peterx, joey.gouly, ioworker0, baohua,
kevin.brodsky, quic_zhenhuah, christophe.leroy, yangyicong,
linux-arm-kernel, hughd, yang, ziy, syzbot+57bcc752f0df8bb1365c
On 06/08/25 10:10 pm, Lorenzo Stoakes wrote:
> On Wed, Aug 06, 2025 at 08:26:11PM +0530, Dev Jain wrote:
>> In commit_anon_folio_batch(), we iterate over all pages pointed to by the
>> PTE batch. Therefore we need to know the first page of the batch;
>> currently we derive that via folio_page(folio, 0), but, that takes us
>> to the first (head) page of the folio instead - our PTE batch may lie
>> in the middle of the folio, leading to incorrectness.
>>
>> Bite the bullet and throw away the micro-optimization of reusing the
>> folio in favour of code simplicity. Derive the page and the folio in
>> change_pte_range, and pass the page too to commit_anon_folio_batch to
>> fix the aforementioned issue.
>>
>> Reported-by: syzbot+57bcc752f0df8bb1365c@syzkaller.appspotmail.com
>> Fixes: cac1db8c3aad ("mm: optimize mprotect() by PTE batching")
>> Signed-off-by: Dev Jain <dev.jain@arm.com>
> This looks reasonable, fixes the problem and compiles/works on my machine so:
>
> Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
Thanks.
>
> This badly needs refactoring as 13 parameters being passed to a function is
Which function are you talking about?
> ridiculous, but we can do that later.
>
> Let's get this in as a hotfix asap.
>
>> ---
>> mm/mprotect.c | 23 ++++++++++-------------
>> 1 file changed, 10 insertions(+), 13 deletions(-)
>>
>> diff --git a/mm/mprotect.c b/mm/mprotect.c
>> index 78bded7acf79..113b48985834 100644
>> --- a/mm/mprotect.c
>> +++ b/mm/mprotect.c
>> @@ -120,9 +120,8 @@ static int mprotect_folio_pte_batch(struct folio *folio, pte_t *ptep,
>>
>> static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
>> pte_t oldpte, pte_t *pte, int target_node,
>> - struct folio **foliop)
>> + struct folio *folio)
>> {
>> - struct folio *folio = NULL;
>> bool ret = true;
>> bool toptier;
>> int nid;
>> @@ -131,7 +130,6 @@ static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
>> if (pte_protnone(oldpte))
>> goto skip;
>>
>> - folio = vm_normal_folio(vma, addr, oldpte);
>> if (!folio)
>> goto skip;
>>
>> @@ -173,7 +171,6 @@ static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
>> folio_xchg_access_time(folio, jiffies_to_msecs(jiffies));
>>
>> skip:
>> - *foliop = folio;
>> return ret;
>> }
>>
>> @@ -231,10 +228,9 @@ static int page_anon_exclusive_sub_batch(int start_idx, int max_len,
>> * retrieve sub-batches.
>> */
>> static void commit_anon_folio_batch(struct vm_area_struct *vma,
>> - struct folio *folio, unsigned long addr, pte_t *ptep,
>> + struct folio *folio, struct page *first_page, unsigned long addr, pte_t *ptep,
>> pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb)
>> {
>> - struct page *first_page = folio_page(folio, 0);
>> bool expected_anon_exclusive;
>> int sub_batch_idx = 0;
>> int len;
>> @@ -251,7 +247,7 @@ static void commit_anon_folio_batch(struct vm_area_struct *vma,
>> }
>>
>> static void set_write_prot_commit_flush_ptes(struct vm_area_struct *vma,
>> - struct folio *folio, unsigned long addr, pte_t *ptep,
>> + struct folio *folio, struct page *page, unsigned long addr, pte_t *ptep,
>> pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb)
>> {
>> bool set_write;
>> @@ -270,7 +266,7 @@ static void set_write_prot_commit_flush_ptes(struct vm_area_struct *vma,
>> /* idx = */ 0, set_write, tlb);
>> return;
>> }
>> - commit_anon_folio_batch(vma, folio, addr, ptep, oldpte, ptent, nr_ptes, tlb);
>> + commit_anon_folio_batch(vma, folio, page, addr, ptep, oldpte, ptent, nr_ptes, tlb);
>> }
>>
>> static long change_pte_range(struct mmu_gather *tlb,
>> @@ -305,15 +301,19 @@ static long change_pte_range(struct mmu_gather *tlb,
>> const fpb_t flags = FPB_RESPECT_SOFT_DIRTY | FPB_RESPECT_WRITE;
>> int max_nr_ptes = (end - addr) >> PAGE_SHIFT;
>> struct folio *folio = NULL;
>> + struct page *page;
>> pte_t ptent;
>>
>> + page = vm_normal_page(vma, addr, oldpte);
>> + if (page)
>> + folio = page_folio(page);
>> /*
>> * Avoid trapping faults against the zero or KSM
>> * pages. See similar comment in change_huge_pmd.
>> */
>> if (prot_numa) {
>> int ret = prot_numa_skip(vma, addr, oldpte, pte,
>> - target_node, &folio);
>> + target_node, folio);
>> if (ret) {
>>
>> /* determine batch to skip */
>> @@ -323,9 +323,6 @@ static long change_pte_range(struct mmu_gather *tlb,
>> }
>> }
>>
>> - if (!folio)
>> - folio = vm_normal_folio(vma, addr, oldpte);
>> -
>> nr_ptes = mprotect_folio_pte_batch(folio, pte, oldpte, max_nr_ptes, flags);
>>
>> oldpte = modify_prot_start_ptes(vma, addr, pte, nr_ptes);
>> @@ -351,7 +348,7 @@ static long change_pte_range(struct mmu_gather *tlb,
>> */
>> if ((cp_flags & MM_CP_TRY_CHANGE_WRITABLE) &&
>> !pte_write(ptent))
>> - set_write_prot_commit_flush_ptes(vma, folio,
>> + set_write_prot_commit_flush_ptes(vma, folio, page,
>> addr, pte, oldpte, ptent, nr_ptes, tlb);
>> else
>> prot_commit_flush_ptes(vma, addr, pte, oldpte, ptent,
>> --
>> 2.30.2
>>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page
2025-08-06 17:19 ` Dev Jain
@ 2025-08-06 17:23 ` Lorenzo Stoakes
0 siblings, 0 replies; 5+ messages in thread
From: Lorenzo Stoakes @ 2025-08-06 17:23 UTC (permalink / raw)
To: Dev Jain
Cc: akpm, ryan.roberts, david, willy, linux-mm, linux-kernel,
catalin.marinas, will, Liam.Howlett, vbabka, jannh,
anshuman.khandual, peterx, joey.gouly, ioworker0, baohua,
kevin.brodsky, quic_zhenhuah, christophe.leroy, yangyicong,
linux-arm-kernel, hughd, yang, ziy, syzbot+57bcc752f0df8bb1365c
On Wed, Aug 06, 2025 at 10:49:46PM +0530, Dev Jain wrote:
>
> On 06/08/25 10:10 pm, Lorenzo Stoakes wrote:
> > On Wed, Aug 06, 2025 at 08:26:11PM +0530, Dev Jain wrote:
> > > In commit_anon_folio_batch(), we iterate over all pages pointed to by the
> > > PTE batch. Therefore we need to know the first page of the batch;
> > > currently we derive that via folio_page(folio, 0), but, that takes us
> > > to the first (head) page of the folio instead - our PTE batch may lie
> > > in the middle of the folio, leading to incorrectness.
> > >
> > > Bite the bullet and throw away the micro-optimization of reusing the
> > > folio in favour of code simplicity. Derive the page and the folio in
> > > change_pte_range, and pass the page too to commit_anon_folio_batch to
> > > fix the aforementioned issue.
> > >
> > > Reported-by: syzbot+57bcc752f0df8bb1365c@syzkaller.appspotmail.com
> > > Fixes: cac1db8c3aad ("mm: optimize mprotect() by PTE batching")
> > > Signed-off-by: Dev Jain <dev.jain@arm.com>
> > This looks reasonable, fixes the problem and compiles/works on my machine so:
> >
> > Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
>
> Thanks.
>
> >
> > This badly needs refactoring as 13 parameters being passed to a function is
>
> Which function are you talking about?
Misread, 9, but it's still silly. mprotect in general needs rework. I may submit
a patch for this at some point...
Anyway let's just get this hotfix in asap.
>
> > ridiculous, but we can do that later.
> >
> > Let's get this in as a hotfix asap.
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page
2025-08-06 14:56 [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page Dev Jain
2025-08-06 16:40 ` Lorenzo Stoakes
@ 2025-08-06 20:21 ` David Hildenbrand
1 sibling, 0 replies; 5+ messages in thread
From: David Hildenbrand @ 2025-08-06 20:21 UTC (permalink / raw)
To: Dev Jain, akpm
Cc: ryan.roberts, willy, linux-mm, linux-kernel, catalin.marinas,
will, Liam.Howlett, lorenzo.stoakes, vbabka, jannh,
anshuman.khandual, peterx, joey.gouly, ioworker0, baohua,
kevin.brodsky, quic_zhenhuah, christophe.leroy, yangyicong,
linux-arm-kernel, hughd, yang, ziy, syzbot+57bcc752f0df8bb1365c
On 06.08.25 16:56, Dev Jain wrote:
> In commit_anon_folio_batch(), we iterate over all pages pointed to by the
> PTE batch. Therefore we need to know the first page of the batch;
> currently we derive that via folio_page(folio, 0), but, that takes us
> to the first (head) page of the folio instead - our PTE batch may lie
> in the middle of the folio, leading to incorrectness.
>
> Bite the bullet and throw away the micro-optimization of reusing the
> folio in favour of code simplicity.
Huh? We are still reusing the folio. There is a single
vm_normal_page()+page_folio() lookup in that code.
Maybe what you meant is "Simplify the code by moving the folio lookup
out of prot_numa_skip() such that we only have a single page+folio lookup."
> Derive the page and the folio in
> change_pte_range, and pass the page too to commit_anon_folio_batch to
> fix the aforementioned issue.
>
> Reported-by: syzbot+57bcc752f0df8bb1365c@syzkaller.appspotmail.com
> Fixes: cac1db8c3aad ("mm: optimize mprotect() by PTE batching")
Fell free to add a
Debugged-by: David Hildenbrand <david@redhat.com>
> Signed-off-by: Dev Jain <dev.jain@arm.com>
Acked-by: David Hildenbrand <david@redhat.com>
--
Cheers,
David / dhildenb
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2025-08-06 20:21 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-08-06 14:56 [PATCH mm-hotfixes-unstable] mm: Pass page directly instead of using folio_page Dev Jain
2025-08-06 16:40 ` Lorenzo Stoakes
2025-08-06 17:19 ` Dev Jain
2025-08-06 17:23 ` Lorenzo Stoakes
2025-08-06 20:21 ` David Hildenbrand
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).