From: "Shakeel Butt" <shakeel.butt@linux.dev>
To: "Qi Zheng" <qi.zheng@linux.dev>,
"kernel test robot" <oliver.sang@intel.com>
Cc: oe-lkp@lists.linux.dev, lkp@intel.com,
linux-kernel@vger.kernel.org,
"Andrew Morton" <akpm@linux-foundation.org>,
"David Carlier" <devnexen@gmail.com>,
"Allen Pais" <apais@linux.microsoft.com>,
"Axel Rasmussen" <axelrasmussen@google.com>,
"Baoquan He" <bhe@redhat.com>,
"Chengming Zhou" <chengming.zhou@linux.dev>,
"Chen Ridong" <chenridong@huawei.com>,
"David Hildenbrand" <david@kernel.org>,
"Hamza Mahfooz" <hamzamahfooz@linux.microsoft.com>,
"Harry Yoo" <harry.yoo@oracle.com>,
"Hugh Dickins" <hughd@google.com>,
"Imran Khan" <imran.f.khan@oracle.com>,
"Johannes Weiner" <hannes@cmpxchg.org>,
"Kamalesh Babulal" <kamalesh.babulal@oracle.com>,
"Lance Yang" <lance.yang@linux.dev>,
"Liam Howlett" <Liam.Howlett@oracle.com>,
"Lorenzo Stoakes" <ljs@kernel.org>,
"Michal Hocko" <mhocko@suse.com>,
"Michal Koutný" <mkoutny@suse.com>,
"Mike Rapoport" <rppt@kernel.org>,
"Muchun Song" <muchun.song@linux.dev>,
"Muchun Song" <songmuchun@bytedance.com>,
"Nhat Pham" <nphamcs@gmail.com>,
"Roman Gushchin" <roman.gushchin@linux.dev>,
"Suren Baghdasaryan" <surenb@google.com>,
"Usama Arif" <usamaarif642@gmail.com>,
"Vlastimil Babka" <vbabka@kernel.org>,
"Wei Xu" <weixugc@google.com>, "Yosry Ahmed" <yosry@kernel.org>,
"Yuanchu Xie" <yuanchu@google.com>, "Zi Yan" <ziy@nvidia.com>,
"Usama Arif" <usama.arif@linux.dev>,
cgroups@vger.kernel.org, linux-mm@kvack.org
Subject: Re: [linus:master] [mm] 01b9da291c: stress-ng.switch.ops_per_sec 67.7% regression
Date: Thu, 14 May 2026 13:40:06 +0000 [thread overview]
Message-ID: <93b7c3f206f158e7387cbb5f0bf5845b59b93053@linux.dev> (raw)
In-Reply-To: <46e9f5cf-34cb-466d-a53a-5778768af4d9@linux.dev>
May 14, 2026 at 12:46 AM, "Qi Zheng" <qi.zheng@linux.dev mailto:qi.zheng@linux.dev?to=%22Qi%20Zheng%22%20%3Cqi.zheng%40linux.dev%3E > wrote:
>
> On 5/13/26 10:27 PM, Shakeel Butt wrote:
>
> >
> > On Wed, May 13, 2026 at 06:49:45AM -0700, Shakeel Butt wrote:
> >
> > >
> > > On Wed, May 13, 2026 at 10:10:34AM +0800, Qi Zheng wrote:
> > >
> > On 5/13/26 12:03 AM, Shakeel Butt wrote:
> > On Tue, May 12, 2026 at 08:56:52PM +0800, kernel test robot wrote:
> >
> > Hello,
> >
> > kernel test robot noticed a 67.7% regression of stress-ng.switch.ops_per_sec on:
> >
> > commit: 01b9da291c4969354807b52956f4aae1f41b4924 ("mm: memcontrol: convert objcg to be per-memcg per-node type")
> > https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
> >
> > This is most probably due to shuffling of struct mem_cgroup and struct
> > mem_cgroup_per_node members.
> >
> > Another possibility is that after objcg was split into per-node, the
> > slab accounting fast path is still designed assuming only one current
> > objcg per CPU:
> >
> > struct obj_stock_pcp {
> > struct obj_cgroup *cached_objcg;
> > };
> >
> > So it's may cause the following thrashing:
> >
> > CPU stock cached = memcg/node0 objcg
> > free object tagged = memcg/node1 objcg
> > => __refill_obj_stock --> objcg mismatch
> > => drain_obj_stock()
> > => cache switches to node1 objcg
> >
> > next local allocation tagged = node0 objcg
> > => mismatch again
> > => drain_obj_stock()
> >
> > >
> > > Actually I think this is the issue, we have ping pong threads running on
> > > different nodes where though theu are in same cgroup but their current->obcg is
> > > for local node and thus this ping pong is thrashing the per-cpu objcg stock.
> > >
> > > The easier fix would be to compare objcg->memcg instead of just objcg during
> > > draining and caching. In addition we can add support for multiple objcg per-cpu
> > > stock caching.
> > >
> > Something like the following:
> > From d756abe831a905d6fe32bad9a984fc619dafb7e0 Mon Sep 17 00:00:00 2001
> > From: Shakeel Butt <shakeel.butt@linux.dev>
> > Date: Wed, 13 May 2026 07:24:55 -0700
> > Subject: [PATCH] mm/memcontrol: skip obj_stock drain when refilled objcg
> > shares memcg
> > Signed-off-by: Shakeel Butt <shakeel.butt@linux.dev>
> > ---
> > mm/memcontrol.c | 14 +++++++++++++-
> > 1 file changed, 13 insertions(+), 1 deletion(-)
> > diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> > index d978e18b9b2d..01ed7a8e18ac 100644
> > --- a/mm/memcontrol.c
> > +++ b/mm/memcontrol.c
> > @@ -3318,6 +3318,7 @@ static void __refill_obj_stock(struct obj_cgroup *objcg,
> > unsigned int nr_bytes,
> > bool allow_uncharge)
> > {
> > + struct obj_cgroup *cached;
> > unsigned int nr_pages = 0;
> > > if (!stock) {
> > @@ -3327,7 +3328,18 @@ static void __refill_obj_stock(struct obj_cgroup *objcg,
> > goto out;
> > }
> > > - if (READ_ONCE(stock->cached_objcg) != objcg) { /* reset if necessary */
> > + cached = READ_ONCE(stock->cached_objcg);
> > + if (cached != objcg &&
> > + (!cached || obj_cgroup_memcg(cached) != obj_cgroup_memcg(objcg))) {
> > drain_obj_stock(stock);
> > obj_cgroup_get(objcg);
> > stock->nr_bytes = atomic_read(&objcg->nr_charged_bytes)
> >
> This change looks like it should be able to fix the ping-pong issue, but
> I stiil haven't reproduced the performance regression locally. I'll
> continue testing it.
Same here, couldn't reproduce locally. It seems like we had to craft a scenario
where the pair pingpong threads get their current->objcg from different nodes.
I will try that.
>
> Hi kernel-test-robot, could you help check if the patch above fixes the
> issue on your end?
>
In the meantime, Oliver, can you please help in testing this patch?
prev parent reply other threads:[~2026-05-14 13:40 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-05-12 12:56 [linus:master] [mm] 01b9da291c: stress-ng.switch.ops_per_sec 67.7% regression kernel test robot
2026-05-12 16:03 ` Shakeel Butt
2026-05-13 2:10 ` Qi Zheng
2026-05-13 13:49 ` Shakeel Butt
2026-05-13 14:27 ` Shakeel Butt
2026-05-14 7:46 ` Qi Zheng
2026-05-14 13:40 ` Shakeel Butt [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=93b7c3f206f158e7387cbb5f0bf5845b59b93053@linux.dev \
--to=shakeel.butt@linux.dev \
--cc=Liam.Howlett@oracle.com \
--cc=akpm@linux-foundation.org \
--cc=apais@linux.microsoft.com \
--cc=axelrasmussen@google.com \
--cc=bhe@redhat.com \
--cc=cgroups@vger.kernel.org \
--cc=chengming.zhou@linux.dev \
--cc=chenridong@huawei.com \
--cc=david@kernel.org \
--cc=devnexen@gmail.com \
--cc=hamzamahfooz@linux.microsoft.com \
--cc=hannes@cmpxchg.org \
--cc=harry.yoo@oracle.com \
--cc=hughd@google.com \
--cc=imran.f.khan@oracle.com \
--cc=kamalesh.babulal@oracle.com \
--cc=lance.yang@linux.dev \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=ljs@kernel.org \
--cc=lkp@intel.com \
--cc=mhocko@suse.com \
--cc=mkoutny@suse.com \
--cc=muchun.song@linux.dev \
--cc=nphamcs@gmail.com \
--cc=oe-lkp@lists.linux.dev \
--cc=oliver.sang@intel.com \
--cc=qi.zheng@linux.dev \
--cc=roman.gushchin@linux.dev \
--cc=rppt@kernel.org \
--cc=songmuchun@bytedance.com \
--cc=surenb@google.com \
--cc=usama.arif@linux.dev \
--cc=usamaarif642@gmail.com \
--cc=vbabka@kernel.org \
--cc=weixugc@google.com \
--cc=yosry@kernel.org \
--cc=yuanchu@google.com \
--cc=ziy@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox