* [PATCH v4] mm/huge_memory: fix folio isn't locked in softleaf_to_folio()
@ 2026-03-21 7:52 Jinjiang Tu
2026-03-21 18:07 ` Andrew Morton
0 siblings, 1 reply; 3+ messages in thread
From: Jinjiang Tu @ 2026-03-21 7:52 UTC (permalink / raw)
To: akpm, david, lorenzo.stoakes, Liam.Howlett, vbabka, rppt, surenb,
mhocko, baohua, ryan.roberts, linux-mm
Cc: wangkefeng.wang, sunnanyong, tujinjiang
On arm64 server, we found folio that get from migration entry isn't locked
in softleaf_to_folio(). This issue triggers when mTHP splitting and
zap_nonpresent_ptes() races, and the root cause is lack of memory barrier
in softleaf_to_folio(). The race is as follows:
CPU0 CPU1
deferred_split_scan() zap_nonpresent_ptes()
lock folio
split_folio()
unmap_folio()
change ptes to migration entries
__split_folio_to_order() softleaf_to_folio()
set flags(including PG_locked) for tail pages folio = pfn_folio(softleaf_to_pfn(entry))
smp_wmb() VM_WARN_ON_ONCE(!folio_test_locked(folio))
prep_compound_page() for tail pages
In __split_folio_to_order(), smp_wmb() guarantees page flags of tail pages
are visible before the tail page becomes non-compound. smp_wmb() should
be paired with smp_rmb() in softleaf_to_folio(), which is missed. As a
result, if zap_nonpresent_ptes() accesses migration entry that stores
tail pfn, softleaf_to_folio() may see the updated compound_head of tail
page before page->flags.
To fix it, add missing smp_rmb() if the softleaf entry is migration entry
in softleaf_to_folio() and softleaf_to_page().
Fixes: e9b61f19858a ("thp: reintroduce split_huge_page()")
Cc: stable@kernel.org
Acked-by: David Hildenbrand (Arm) <david@kernel.org>
Reviewed-by: Lorenzo Stoakes (Oracle) <ljs@kernel.org>
Signed-off-by: Jinjiang Tu <tujinjiang@huawei.com>
---
Changes in v4:
* update function name and comments.
* collect Acked-by and Reviewed-by.
include/linux/leafops.h | 32 +++++++++++++++++++++-----------
1 file changed, 21 insertions(+), 11 deletions(-)
diff --git a/include/linux/leafops.h b/include/linux/leafops.h
index a9ff94b744f2..05673d3529e7 100644
--- a/include/linux/leafops.h
+++ b/include/linux/leafops.h
@@ -363,6 +363,23 @@ static inline unsigned long softleaf_to_pfn(softleaf_t entry)
return swp_offset(entry) & SWP_PFN_MASK;
}
+static inline void softleaf_migration_sync(softleaf_t entry,
+ struct folio *folio)
+{
+ /*
+ * Ensure we do not race with split, which might alter tail pages into new
+ * folios and thus result in observing an unlocked folio.
+ * This matches the write barrier in __split_folio_to_order().
+ */
+ smp_rmb();
+
+ /*
+ * Any use of migration entries may only occur while the
+ * corresponding page is locked
+ */
+ VM_WARN_ON_ONCE(!folio_test_locked(folio));
+}
+
/**
* softleaf_to_page() - Obtains struct page for PFN encoded within leaf entry.
* @entry: Leaf entry, softleaf_has_pfn(@entry) must return true.
@@ -374,11 +391,8 @@ static inline struct page *softleaf_to_page(softleaf_t entry)
struct page *page = pfn_to_page(softleaf_to_pfn(entry));
VM_WARN_ON_ONCE(!softleaf_has_pfn(entry));
- /*
- * Any use of migration entries may only occur while the
- * corresponding page is locked
- */
- VM_WARN_ON_ONCE(softleaf_is_migration(entry) && !PageLocked(page));
+ if (softleaf_is_migration(entry))
+ softleaf_migration_sync(entry, page_folio(page));
return page;
}
@@ -394,12 +408,8 @@ static inline struct folio *softleaf_to_folio(softleaf_t entry)
struct folio *folio = pfn_folio(softleaf_to_pfn(entry));
VM_WARN_ON_ONCE(!softleaf_has_pfn(entry));
- /*
- * Any use of migration entries may only occur while the
- * corresponding folio is locked.
- */
- VM_WARN_ON_ONCE(softleaf_is_migration(entry) &&
- !folio_test_locked(folio));
+ if (softleaf_is_migration(entry))
+ softleaf_migration_sync(entry, folio);
return folio;
}
--
2.43.0
^ permalink raw reply related [flat|nested] 3+ messages in thread
* Re: [PATCH v4] mm/huge_memory: fix folio isn't locked in softleaf_to_folio()
2026-03-21 7:52 [PATCH v4] mm/huge_memory: fix folio isn't locked in softleaf_to_folio() Jinjiang Tu
@ 2026-03-21 18:07 ` Andrew Morton
2026-03-23 2:06 ` Jinjiang Tu
0 siblings, 1 reply; 3+ messages in thread
From: Andrew Morton @ 2026-03-21 18:07 UTC (permalink / raw)
To: Jinjiang Tu
Cc: david, lorenzo.stoakes, Liam.Howlett, vbabka, rppt, surenb,
mhocko, baohua, ryan.roberts, linux-mm, wangkefeng.wang,
sunnanyong
On Sat, 21 Mar 2026 15:52:14 +0800 Jinjiang Tu <tujinjiang@huawei.com> wrote:
> On arm64 server, we found folio that get from migration entry isn't locked
> in softleaf_to_folio(). This issue triggers when mTHP splitting and
> zap_nonpresent_ptes() races, and the root cause is lack of memory barrier
> in softleaf_to_folio(). The race is as follows:
>
> CPU0 CPU1
>
> deferred_split_scan() zap_nonpresent_ptes()
> lock folio
> split_folio()
> unmap_folio()
> change ptes to migration entries
> __split_folio_to_order() softleaf_to_folio()
> set flags(including PG_locked) for tail pages folio = pfn_folio(softleaf_to_pfn(entry))
> smp_wmb() VM_WARN_ON_ONCE(!folio_test_locked(folio))
> prep_compound_page() for tail pages
>
> In __split_folio_to_order(), smp_wmb() guarantees page flags of tail pages
> are visible before the tail page becomes non-compound. smp_wmb() should
> be paired with smp_rmb() in softleaf_to_folio(), which is missed. As a
> result, if zap_nonpresent_ptes() accesses migration entry that stores
> tail pfn, softleaf_to_folio() may see the updated compound_head of tail
> page before page->flags.
>
> To fix it, add missing smp_rmb() if the softleaf entry is migration entry
> in softleaf_to_folio() and softleaf_to_page().
AI review isn't entirely happy:
https://sashiko.dev/#/patchset/20260321075214.3305564-1-tujinjiang@huawei.com
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [PATCH v4] mm/huge_memory: fix folio isn't locked in softleaf_to_folio()
2026-03-21 18:07 ` Andrew Morton
@ 2026-03-23 2:06 ` Jinjiang Tu
0 siblings, 0 replies; 3+ messages in thread
From: Jinjiang Tu @ 2026-03-23 2:06 UTC (permalink / raw)
To: Andrew Morton
Cc: david, lorenzo.stoakes, Liam.Howlett, vbabka, rppt, surenb,
mhocko, baohua, ryan.roberts, linux-mm, wangkefeng.wang,
sunnanyong
在 2026/3/22 2:07, Andrew Morton 写道:
> On Sat, 21 Mar 2026 15:52:14 +0800 Jinjiang Tu <tujinjiang@huawei.com> wrote:
>
>> On arm64 server, we found folio that get from migration entry isn't locked
>> in softleaf_to_folio(). This issue triggers when mTHP splitting and
>> zap_nonpresent_ptes() races, and the root cause is lack of memory barrier
>> in softleaf_to_folio(). The race is as follows:
>>
>> CPU0 CPU1
>>
>> deferred_split_scan() zap_nonpresent_ptes()
>> lock folio
>> split_folio()
>> unmap_folio()
>> change ptes to migration entries
>> __split_folio_to_order() softleaf_to_folio()
>> set flags(including PG_locked) for tail pages folio = pfn_folio(softleaf_to_pfn(entry))
>> smp_wmb() VM_WARN_ON_ONCE(!folio_test_locked(folio))
>> prep_compound_page() for tail pages
>>
>> In __split_folio_to_order(), smp_wmb() guarantees page flags of tail pages
>> are visible before the tail page becomes non-compound. smp_wmb() should
>> be paired with smp_rmb() in softleaf_to_folio(), which is missed. As a
>> result, if zap_nonpresent_ptes() accesses migration entry that stores
>> tail pfn, softleaf_to_folio() may see the updated compound_head of tail
>> page before page->flags.
>>
>> To fix it, add missing smp_rmb() if the softleaf entry is migration entry
>> in softleaf_to_folio() and softleaf_to_page().
> AI review isn't entirely happy:
> https://sashiko.dev/#/patchset/20260321075214.3305564-1-tujinjiang@huawei.com
commit 54150f26b76fbe8cbf90c19d0052add5199851f0
Author: Jinjiang Tu <tujinjiang@huawei.com>
mm/huge_memory: fix folio isn't locked in softleaf_to_folio()
This commit adds a memory barrier to softleaf_to_folio() and
softleaf_to_page() to prevent a race condition during mTHP splitting
where the updated compound_head of a tail page might be observed before
its page flags.
> diff --git a/include/linux/leafops.h b/include/linux/leafops.h
> --- a/include/linux/leafops.h
> +++ b/include/linux/leafops.h
> @@ -363,6 +363,23 @@ static inline unsigned long softleaf_to_pfn(softleaf_t entry)
> return swp_offset(entry) & SWP_PFN_MASK;
> }
>
> +static inline void softleaf_migration_sync(softleaf_t entry,
> + struct folio *folio)
"This isn't a bug, but is the entry parameter needed here? It doesn't appear
to be used in the function body."
Indeed, the previous version move softleaf_is_migration() check out of
softleaf_migration_sync(), the parameter could be removed.
> +{
> + /*
> + * Ensure we do not race with split, which might alter tail pages into new
> + * folios and thus result in observing an unlocked folio.
> + * This matches the write barrier in __split_folio_to_order().
> + */
> + smp_rmb();
> +
> + /*
> + * Any use of migration entries may only occur while the
> + * corresponding page is locked
> + */
> + VM_WARN_ON_ONCE(!folio_test_locked(folio));
> +}
> +
> /**
> * softleaf_to_page() - Obtains struct page for PFN encoded within leaf entry.
> * @entry: Leaf entry, softleaf_has_pfn(@entry) must return true.
> @@ -374,11 +391,8 @@ static inline struct page *softleaf_to_page(softleaf_t entry)
> struct page *page = pfn_to_page(softleaf_to_pfn(entry));
>
> VM_WARN_ON_ONCE(!softleaf_has_pfn(entry));
> - /*
> - * Any use of migration entries may only occur while the
> - * corresponding page is locked
> - */
> - VM_WARN_ON_ONCE(softleaf_is_migration(entry) && !PageLocked(page));
> + if (softleaf_is_migration(entry))
> + softleaf_migration_sync(entry, page_folio(page));
>
> return page;
> }
"Can this still race if callers use softleaf_to_page() followed by
page_folio()?
If a caller does something similar to:
page = softleaf_to_page(entry);
folio = page_folio(page);
if (folio_test_locked(folio)) { ... }
The call to page_folio() performs a new read of compound_head. Because
this new read and the subsequent reads of folio->flags both execute after
the smp_rmb() inside softleaf_to_page(), there is no memory barrier
between them.
If a weakly-ordered CPU speculatively reads folio->flags before the new
read of compound_head evaluates as order-0, could the caller still see
stale flags and reproduce the crash this patch intends to fix?
Would it be safer to require vulnerable callers to use softleaf_to_folio()
directly?
[ ... ]
"
This issue only exist when the softleaf_entry is migration entry and is
pte-level. Check all callsites of softleaf_to_page(), the softleaf entries
are almost SOFTLEAF_DEVICE_* or pmd-level. The exceptions are smaps_pte_entry()
and pte_to_pagemap_entry(), they only collect information for userspace, and
I think such race doesn't matter. So, I think this AI review comment is false positive.
Thanks.
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2026-03-23 2:07 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2026-03-21 7:52 [PATCH v4] mm/huge_memory: fix folio isn't locked in softleaf_to_folio() Jinjiang Tu
2026-03-21 18:07 ` Andrew Morton
2026-03-23 2:06 ` Jinjiang Tu
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox