* Re: [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size()
2026-03-09 7:22 ` [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size() Harry Yoo
@ 2026-03-09 14:00 ` vbabka
2026-03-10 3:25 ` Harry Yoo
2026-03-10 3:29 ` Harry Yoo
2026-03-10 3:40 ` Zw Tang
2 siblings, 1 reply; 7+ messages in thread
From: vbabka @ 2026-03-09 14:00 UTC (permalink / raw)
To: Harry Yoo
Cc: adilger.kernel, akpm, cgroups, hannes, hao.li, linux-ext4,
linux-fsdevel, linux-kernel, linux-mm, shicenci, cl, rientjes,
roman.gushchin, viro, surenb, stable
On 3/9/26 08:22, Harry Yoo wrote:
> obj_exts_alloc_size() prevents recursive allocation of slabobj_ext
> array from the same cache, to avoid creating slabs that are never freed.
>
> There is one mistake that returns the original size when memory
> allocation profiling is disabled. The assumption was that
> memcg-triggered slabobj_ext allocation is always served from
> KMALLOC_CGROUP type. But this is wrong [1]: when the caller specifies
> both __GFP_RECLAIMABLE and __GFP_ACCOUNT with SLUB_TINY enabled, the
> allocation is served from normal kmalloc. This is because kmalloc_type()
> prioritizes __GFP_RECLAIMABLE over __GFP_ACCOUNT, and SLUB_TINY aliases
> KMALLOC_RECLAIM with KMALLOC_NORMAL.
Hm that's suboptimal (leads to sparsely used obj_exts in normal kmalloc
slabs) and maybe separately from this hotfix we could make sure that with
SLUB_TINY, __GFP_ACCOUNT is preferred going forward?
> As a result, the recursion guard is bypassed and the problematic slabs
> can be created. Fix this by removing the mem_alloc_profiling_enabled()
> check entirely. The remaining is_kmalloc_normal() check is still
> sufficient to detect whether the cache is of KMALLOC_NORMAL type and
> avoid bumping the size if it's not.
>
> Without SLUB_TINY, no functional change intended.
> With SLUB_TINY, allocations with __GFP_ACCOUNT|__GFP_RECLAIMABLE
> now allocate a larger array if the sizes equal.
>
> Reported-by: Zw Tang <shicenci@gmail.com>
> Fixes: 280ea9c3154b ("mm/slab: avoid allocating slabobj_ext array from its own slab")
> Closes: https://lore.kernel.org/linux-mm/CAPHJ_VKuMKSke8b11AZQw1PTSFN4n2C0gFxC6xGOG0ZLHgPmnA@mail.gmail.com [1]
> Cc: stable@vger.kernel.org
> Signed-off-by: Harry Yoo <harry.yoo@oracle.com>
Added to slab/for-next-fixes, thanks!
> ---
>
> Zw Tang, could you please confirm that the warning disappears
> on your test environment, with this patch applied?
>
> mm/slub.c | 7 -------
> 1 file changed, 7 deletions(-)
>
> diff --git a/mm/slub.c b/mm/slub.c
> index 20cb4f3b636d..6371838d2352 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -2119,13 +2119,6 @@ static inline size_t obj_exts_alloc_size(struct kmem_cache *s,
> size_t sz = sizeof(struct slabobj_ext) * slab->objects;
> struct kmem_cache *obj_exts_cache;
>
> - /*
> - * slabobj_ext array for KMALLOC_CGROUP allocations
> - * are served from KMALLOC_NORMAL caches.
> - */
> - if (!mem_alloc_profiling_enabled())
> - return sz;
> -
> if (sz > KMALLOC_MAX_CACHE_SIZE)
> return sz;
>
>
> base-commit: 6432f15c818cb30eec7c4ca378ecdebd9796f741
^ permalink raw reply [flat|nested] 7+ messages in thread* Re: [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size()
2026-03-09 14:00 ` vbabka
@ 2026-03-10 3:25 ` Harry Yoo
2026-03-10 10:06 ` vbabka
0 siblings, 1 reply; 7+ messages in thread
From: Harry Yoo @ 2026-03-10 3:25 UTC (permalink / raw)
To: vbabka
Cc: adilger.kernel, akpm, cgroups, hannes, hao.li, linux-ext4,
linux-fsdevel, linux-kernel, linux-mm, shicenci, cl, rientjes,
roman.gushchin, viro, surenb, stable
On Mon, Mar 09, 2026 at 03:00:17PM +0100, vbabka@kernel.org wrote:
> On 3/9/26 08:22, Harry Yoo wrote:
> > obj_exts_alloc_size() prevents recursive allocation of slabobj_ext
> > array from the same cache, to avoid creating slabs that are never freed.
> >
> > There is one mistake that returns the original size when memory
> > allocation profiling is disabled. The assumption was that
> > memcg-triggered slabobj_ext allocation is always served from
> > KMALLOC_CGROUP type. But this is wrong [1]: when the caller specifies
> > both __GFP_RECLAIMABLE and __GFP_ACCOUNT with SLUB_TINY enabled, the
> > allocation is served from normal kmalloc. This is because kmalloc_type()
> > prioritizes __GFP_RECLAIMABLE over __GFP_ACCOUNT, and SLUB_TINY aliases
> > KMALLOC_RECLAIM with KMALLOC_NORMAL.
>
> Hm that's suboptimal (leads to sparsely used obj_exts in normal kmalloc
> slabs) and maybe separately from this hotfix we could make sure that with
> SLUB_TINY, __GFP_ACCOUNT is preferred going forward?
To be honest, I don't a have strong opinion on that.
Is grouping by mobility (for anti-fragmentation less) important on
SLUB_TINY systems?
> > As a result, the recursion guard is bypassed and the problematic slabs
> > can be created. Fix this by removing the mem_alloc_profiling_enabled()
> > check entirely. The remaining is_kmalloc_normal() check is still
> > sufficient to detect whether the cache is of KMALLOC_NORMAL type and
> > avoid bumping the size if it's not.
> >
> > Without SLUB_TINY, no functional change intended.
> > With SLUB_TINY, allocations with __GFP_ACCOUNT|__GFP_RECLAIMABLE
> > now allocate a larger array if the sizes equal.
> >
> > Reported-by: Zw Tang <shicenci@gmail.com>
> > Fixes: 280ea9c3154b ("mm/slab: avoid allocating slabobj_ext array from its own slab")
> > Closes: https://lore.kernel.org/linux-mm/CAPHJ_VKuMKSke8b11AZQw1PTSFN4n2C0gFxC6xGOG0ZLHgPmnA@mail.gmail.com
> > Cc: stable@vger.kernel.org
> > Signed-off-by: Harry Yoo <harry.yoo@oracle.com>
>
> Added to slab/for-next-fixes, thanks!
Thanks!
--
Cheers,
Harry / Hyeonggon
^ permalink raw reply [flat|nested] 7+ messages in thread* Re: [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size()
2026-03-10 3:25 ` Harry Yoo
@ 2026-03-10 10:06 ` vbabka
0 siblings, 0 replies; 7+ messages in thread
From: vbabka @ 2026-03-10 10:06 UTC (permalink / raw)
To: Harry Yoo
Cc: adilger.kernel, akpm, cgroups, hannes, hao.li, linux-ext4,
linux-fsdevel, linux-kernel, linux-mm, shicenci, cl, rientjes,
roman.gushchin, viro, surenb, stable
On 3/10/26 04:25, Harry Yoo wrote:
> On Mon, Mar 09, 2026 at 03:00:17PM +0100, vbabka@kernel.org wrote:
>> On 3/9/26 08:22, Harry Yoo wrote:
>> > obj_exts_alloc_size() prevents recursive allocation of slabobj_ext
>> > array from the same cache, to avoid creating slabs that are never freed.
>> >
>> > There is one mistake that returns the original size when memory
>> > allocation profiling is disabled. The assumption was that
>> > memcg-triggered slabobj_ext allocation is always served from
>> > KMALLOC_CGROUP type. But this is wrong [1]: when the caller specifies
>> > both __GFP_RECLAIMABLE and __GFP_ACCOUNT with SLUB_TINY enabled, the
>> > allocation is served from normal kmalloc. This is because kmalloc_type()
>> > prioritizes __GFP_RECLAIMABLE over __GFP_ACCOUNT, and SLUB_TINY aliases
>> > KMALLOC_RECLAIM with KMALLOC_NORMAL.
>>
>> Hm that's suboptimal (leads to sparsely used obj_exts in normal kmalloc
>> slabs) and maybe separately from this hotfix we could make sure that with
>> SLUB_TINY, __GFP_ACCOUNT is preferred going forward?
>
> To be honest, I don't a have strong opinion on that.
>
> Is grouping by mobility (for anti-fragmentation less) important on
> SLUB_TINY systems?
Yeah, that's why "KMALLOC_RECLAIM = KMALLOC_NORMAL" there. So prioritizing
__GFP_RECLAIMABLE does nothing there, it goes to the same kmalloc_normal
cache. It only results in ignoring KMALLOC_CGROUP.
(I think in practice SLUB_TINY systems wouldn't enabled CONFIG_MEMCG either,
so it's a low priority, but still logical imho).
>> > As a result, the recursion guard is bypassed and the problematic slabs
>> > can be created. Fix this by removing the mem_alloc_profiling_enabled()
>> > check entirely. The remaining is_kmalloc_normal() check is still
>> > sufficient to detect whether the cache is of KMALLOC_NORMAL type and
>> > avoid bumping the size if it's not.
>> >
>> > Without SLUB_TINY, no functional change intended.
>> > With SLUB_TINY, allocations with __GFP_ACCOUNT|__GFP_RECLAIMABLE
>> > now allocate a larger array if the sizes equal.
>> >
>> > Reported-by: Zw Tang <shicenci@gmail.com>
>> > Fixes: 280ea9c3154b ("mm/slab: avoid allocating slabobj_ext array from its own slab")
>> > Closes: https://lore.kernel.org/linux-mm/CAPHJ_VKuMKSke8b11AZQw1PTSFN4n2C0gFxC6xGOG0ZLHgPmnA@mail.gmail.com
>> > Cc: stable@vger.kernel.org
>> > Signed-off-by: Harry Yoo <harry.yoo@oracle.com>
>>
>> Added to slab/for-next-fixes, thanks!
>
> Thanks!
>
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size()
2026-03-09 7:22 ` [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size() Harry Yoo
2026-03-09 14:00 ` vbabka
@ 2026-03-10 3:29 ` Harry Yoo
2026-03-10 3:40 ` Zw Tang
2 siblings, 0 replies; 7+ messages in thread
From: Harry Yoo @ 2026-03-10 3:29 UTC (permalink / raw)
To: adilger.kernel, akpm, cgroups, hannes, hao.li, linux-ext4,
linux-fsdevel, linux-kernel, linux-mm, shicenci, vbabka, cl,
rientjes, roman.gushchin, viro, surenb, stable
On Mon, Mar 09, 2026 at 04:22:19PM +0900, Harry Yoo wrote:
> obj_exts_alloc_size() prevents recursive allocation of slabobj_ext
> array from the same cache, to avoid creating slabs that are never freed.
>
> There is one mistake that returns the original size when memory
> allocation profiling is disabled. The assumption was that
> memcg-triggered slabobj_ext allocation is always served from
> KMALLOC_CGROUP type. But this is wrong [1]: when the caller specifies
> both __GFP_RECLAIMABLE and __GFP_ACCOUNT with SLUB_TINY enabled, the
> allocation is served from normal kmalloc. This is because kmalloc_type()
> prioritizes __GFP_RECLAIMABLE over __GFP_ACCOUNT, and SLUB_TINY aliases
> KMALLOC_RECLAIM with KMALLOC_NORMAL.
>
> As a result, the recursion guard is bypassed and the problematic slabs
> can be created. Fix this by removing the mem_alloc_profiling_enabled()
> check entirely. The remaining is_kmalloc_normal() check is still
> sufficient to detect whether the cache is of KMALLOC_NORMAL type and
> avoid bumping the size if it's not.
>
> Without SLUB_TINY, no functional change intended.
> With SLUB_TINY, allocations with __GFP_ACCOUNT|__GFP_RECLAIMABLE
> now allocate a larger array if the sizes equal.
>
> Reported-by: Zw Tang <shicenci@gmail.com>
> Fixes: 280ea9c3154b ("mm/slab: avoid allocating slabobj_ext array from its own slab")
> Closes: https://lore.kernel.org/linux-mm/CAPHJ_VKuMKSke8b11AZQw1PTSFN4n2C0gFxC6xGOG0ZLHgPmnA@mail.gmail.com [1]
> Cc: stable@vger.kernel.org
> Signed-off-by: Harry Yoo <harry.yoo@oracle.com>
> ---
>
> Zw Tang, could you please confirm that the warning disappears
> on your test environment, with this patch applied?
Oops, I think I saw Zw Tang's Tested-by: (thanks!), but appearently
it's not sent to linux-mm. Could you please add your Tested-by:
by replying to all, again?
--
Cheers,
Harry / Hyeonggon
^ permalink raw reply [flat|nested] 7+ messages in thread* Re: [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size()
2026-03-09 7:22 ` [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size() Harry Yoo
2026-03-09 14:00 ` vbabka
2026-03-10 3:29 ` Harry Yoo
@ 2026-03-10 3:40 ` Zw Tang
2026-03-10 10:02 ` vbabka
2 siblings, 1 reply; 7+ messages in thread
From: Zw Tang @ 2026-03-10 3:40 UTC (permalink / raw)
To: Harry Yoo
Cc: adilger.kernel, akpm, cgroups, hannes, hao.li, linux-ext4,
linux-fsdevel, linux-kernel, linux-mm, vbabka, cl, rientjes,
roman.gushchin, viro, surenb, stable
Hi Harry,
Thanks for the patch.
I tested it on my environment with the original syzkaller reproducer,
and the warning no longer reproduces after applying the patch.
Kernel version tested: v7.0-rc2
Tested-by: Zw Tang shicenci@gmail.com
Best regards,
Zw Tang
Harry Yoo <harry.yoo@oracle.com> 于2026年3月9日周一 15:22写道:
>
> obj_exts_alloc_size() prevents recursive allocation of slabobj_ext
> array from the same cache, to avoid creating slabs that are never freed.
>
> There is one mistake that returns the original size when memory
> allocation profiling is disabled. The assumption was that
> memcg-triggered slabobj_ext allocation is always served from
> KMALLOC_CGROUP type. But this is wrong [1]: when the caller specifies
> both __GFP_RECLAIMABLE and __GFP_ACCOUNT with SLUB_TINY enabled, the
> allocation is served from normal kmalloc. This is because kmalloc_type()
> prioritizes __GFP_RECLAIMABLE over __GFP_ACCOUNT, and SLUB_TINY aliases
> KMALLOC_RECLAIM with KMALLOC_NORMAL.
>
> As a result, the recursion guard is bypassed and the problematic slabs
> can be created. Fix this by removing the mem_alloc_profiling_enabled()
> check entirely. The remaining is_kmalloc_normal() check is still
> sufficient to detect whether the cache is of KMALLOC_NORMAL type and
> avoid bumping the size if it's not.
>
> Without SLUB_TINY, no functional change intended.
> With SLUB_TINY, allocations with __GFP_ACCOUNT|__GFP_RECLAIMABLE
> now allocate a larger array if the sizes equal.
>
> Reported-by: Zw Tang <shicenci@gmail.com>
> Fixes: 280ea9c3154b ("mm/slab: avoid allocating slabobj_ext array from its own slab")
> Closes: https://lore.kernel.org/linux-mm/CAPHJ_VKuMKSke8b11AZQw1PTSFN4n2C0gFxC6xGOG0ZLHgPmnA@mail.gmail.com [1]
> Cc: stable@vger.kernel.org
> Signed-off-by: Harry Yoo <harry.yoo@oracle.com>
> ---
>
> Zw Tang, could you please confirm that the warning disappears
> on your test environment, with this patch applied?
>
> mm/slub.c | 7 -------
> 1 file changed, 7 deletions(-)
>
> diff --git a/mm/slub.c b/mm/slub.c
> index 20cb4f3b636d..6371838d2352 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -2119,13 +2119,6 @@ static inline size_t obj_exts_alloc_size(struct kmem_cache *s,
> size_t sz = sizeof(struct slabobj_ext) * slab->objects;
> struct kmem_cache *obj_exts_cache;
>
> - /*
> - * slabobj_ext array for KMALLOC_CGROUP allocations
> - * are served from KMALLOC_NORMAL caches.
> - */
> - if (!mem_alloc_profiling_enabled())
> - return sz;
> -
> if (sz > KMALLOC_MAX_CACHE_SIZE)
> return sz;
>
>
> base-commit: 6432f15c818cb30eec7c4ca378ecdebd9796f741
> --
> 2.43.0
>
^ permalink raw reply [flat|nested] 7+ messages in thread* Re: [PATCH] mm/slab: fix an incorrect check in obj_exts_alloc_size()
2026-03-10 3:40 ` Zw Tang
@ 2026-03-10 10:02 ` vbabka
0 siblings, 0 replies; 7+ messages in thread
From: vbabka @ 2026-03-10 10:02 UTC (permalink / raw)
To: Zw Tang, Harry Yoo
Cc: adilger.kernel, akpm, cgroups, hannes, hao.li, linux-ext4,
linux-fsdevel, linux-kernel, linux-mm, cl, rientjes,
roman.gushchin, viro, surenb, stable
On 3/10/26 04:40, Zw Tang wrote:
> Hi Harry,
>
> Thanks for the patch.
>
> I tested it on my environment with the original syzkaller reproducer,
> and the warning no longer reproduces after applying the patch.
>
> Kernel version tested: v7.0-rc2
>
> Tested-by: Zw Tang shicenci@gmail.com
Thanks!
^ permalink raw reply [flat|nested] 7+ messages in thread