public inbox for linux-nfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Qi Zheng <zhengqi.arch@bytedance.com>
To: Dave Chinner <david@fromorbit.com>
Cc: akpm@linux-foundation.org, tkhai@ya.ru, vbabka@suse.cz,
	roman.gushchin@linux.dev, djwong@kernel.org, brauner@kernel.org,
	paulmck@kernel.org, tytso@mit.edu, steven.price@arm.com,
	cel@kernel.org, senozhatsky@chromium.org, yujie.liu@intel.com,
	gregkh@linuxfoundation.org, muchun.song@linux.dev,
	simon.horman@corigine.com, dlemoal@kernel.org,
	linux-kernel@vger.kernel.org, linux-mm@kvack.org, x86@kernel.org,
	kvm@vger.kernel.org, xen-devel@lists.xenproject.org,
	linux-erofs@lists.ozlabs.org,
	linux-f2fs-devel@lists.sourceforge.net, cluster-devel@redhat.com,
	linux-nfs@vger.kernel.org, linux-mtd@lists.infradead.org,
	rcu@vger.kernel.org, netdev@vger.kernel.org,
	dri-devel@lists.freedesktop.org, linux-arm-msm@vger.kernel.org,
	dm-devel@redhat.com, linux-raid@vger.kernel.org,
	linux-bcache@vger.kernel.org,
	virtualization@lists.linux-foundation.org,
	linux-fsdevel@vger.kernel.org, linux-ext4@vger.kernel.org,
	linux-xfs@vger.kernel.org, linux-btrfs@vger.kernel.org
Subject: Re: [PATCH v4 46/48] mm: shrinker: make memcg slab shrink lockless
Date: Tue, 8 Aug 2023 15:50:18 +0800	[thread overview]
Message-ID: <5757e341-b261-14de-e052-46606d530460@bytedance.com> (raw)
In-Reply-To: <ZNGr+1orhHaBORJG@dread.disaster.area>

Hi Dave,

On 2023/8/8 10:44, Dave Chinner wrote:
> On Mon, Aug 07, 2023 at 07:09:34PM +0800, Qi Zheng wrote:
>> Like global slab shrink, this commit also uses refcount+RCU method to make
>> memcg slab shrink lockless.
> 
> This patch does random code cleanups amongst the actual RCU changes.
> Can you please move the cleanups to a spearate patch to reduce the
> noise in this one?

Sure, will do.

> 
>> diff --git a/mm/shrinker.c b/mm/shrinker.c
>> index d318f5621862..fee6f62904fb 100644
>> --- a/mm/shrinker.c
>> +++ b/mm/shrinker.c
>> @@ -107,6 +107,12 @@ static struct shrinker_info *shrinker_info_protected(struct mem_cgroup *memcg,
>>   					 lockdep_is_held(&shrinker_rwsem));
>>   }
>>   
>> +static struct shrinker_info *shrinker_info_rcu(struct mem_cgroup *memcg,
>> +					       int nid)
>> +{
>> +	return rcu_dereference(memcg->nodeinfo[nid]->shrinker_info);
>> +}
> 
> This helper doesn't add value. It doesn't tell me that
> rcu_read_lock() needs to be held when it is called, for one....

How about adding a comment or an assertion here?

> 
>>   static int expand_one_shrinker_info(struct mem_cgroup *memcg, int new_size,
>>   				    int old_size, int new_nr_max)
>>   {
>> @@ -198,7 +204,7 @@ void set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id)
>>   		struct shrinker_info_unit *unit;
>>   
>>   		rcu_read_lock();
>> -		info = rcu_dereference(memcg->nodeinfo[nid]->shrinker_info);
>> +		info = shrinker_info_rcu(memcg, nid);
> 
> ... whilst the original code here was obviously correct.
> 
>>   		unit = info->unit[shriner_id_to_index(shrinker_id)];
>>   		if (!WARN_ON_ONCE(shrinker_id >= info->map_nr_max)) {
>>   			/* Pairs with smp mb in shrink_slab() */
>> @@ -211,7 +217,7 @@ void set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id)
>>   
>>   static DEFINE_IDR(shrinker_idr);
>>   
>> -static int prealloc_memcg_shrinker(struct shrinker *shrinker)
>> +static int shrinker_memcg_alloc(struct shrinker *shrinker)
> 
> Cleanups in a separate patch.

OK.

> 
>> @@ -253,10 +258,15 @@ static long xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker,
>>   {
>>   	struct shrinker_info *info;
>>   	struct shrinker_info_unit *unit;
>> +	long nr_deferred;
>>   
>> -	info = shrinker_info_protected(memcg, nid);
>> +	rcu_read_lock();
>> +	info = shrinker_info_rcu(memcg, nid);
>>   	unit = info->unit[shriner_id_to_index(shrinker->id)];
>> -	return atomic_long_xchg(&unit->nr_deferred[shriner_id_to_offset(shrinker->id)], 0);
>> +	nr_deferred = atomic_long_xchg(&unit->nr_deferred[shriner_id_to_offset(shrinker->id)], 0);
>> +	rcu_read_unlock();
>> +
>> +	return nr_deferred;
>>   }
> 
> This adds two rcu_read_lock() sections to every call to
> do_shrink_slab(). It's not at all clear ifrom any of the other code
> that do_shrink_slab() now has internal rcu_read_lock() sections....

The xchg_nr_deferred_memcg() will only be called in shrink_slab_memcg(),
so other code doesn't need to know that information?

> 
>> @@ -464,18 +480,23 @@ static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid,
>>   	if (!mem_cgroup_online(memcg))
>>   		return 0;
>>   
>> -	if (!down_read_trylock(&shrinker_rwsem))
>> -		return 0;
>> -
>> -	info = shrinker_info_protected(memcg, nid);
>> +again:
>> +	rcu_read_lock();
>> +	info = shrinker_info_rcu(memcg, nid);
>>   	if (unlikely(!info))
>>   		goto unlock;
>>   
>> -	for (; index < shriner_id_to_index(info->map_nr_max); index++) {
>> +	if (index < shriner_id_to_index(info->map_nr_max)) {
>>   		struct shrinker_info_unit *unit;
>>   
>>   		unit = info->unit[index];
>>   
>> +		/*
>> +		 * The shrinker_info_unit will not be freed, so we can
>> +		 * safely release the RCU lock here.
>> +		 */
>> +		rcu_read_unlock();
> 
> Why - what guarantees that the shrinker_info_unit exists at this
> point? We hold no reference to it, we hold no reference to any
> shrinker, etc. What provides this existence guarantee?

The shrinker_info_unit is never freed unless the memcg is destroyed.
Here we hold the refcount of this memcg (mem_cgroup_iter() -->
css_tryget()), so the shrinker_info_unit will not be freed.

> 
>> +
>>   		for_each_set_bit(offset, unit->map, SHRINKER_UNIT_BITS) {
>>   			struct shrink_control sc = {
>>   				.gfp_mask = gfp_mask,
>> @@ -485,12 +506,14 @@ static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid,
>>   			struct shrinker *shrinker;
>>   			int shrinker_id = calc_shrinker_id(index, offset);
>>   
>> +			rcu_read_lock();
>>   			shrinker = idr_find(&shrinker_idr, shrinker_id);
>> -			if (unlikely(!shrinker || !(shrinker->flags & SHRINKER_REGISTERED))) {
>> -				if (!shrinker)
>> -					clear_bit(offset, unit->map);
>> +			if (unlikely(!shrinker || !shrinker_try_get(shrinker))) {
>> +				clear_bit(offset, unit->map);
>> +				rcu_read_unlock();
>>   				continue;
>>   			}
>> +			rcu_read_unlock();
>>   
>>   			/* Call non-slab shrinkers even though kmem is disabled */
>>   			if (!memcg_kmem_online() &&
>> @@ -523,15 +546,20 @@ static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid,
>>   					set_shrinker_bit(memcg, nid, shrinker_id);
>>   			}
>>   			freed += ret;
>> -
>> -			if (rwsem_is_contended(&shrinker_rwsem)) {
>> -				freed = freed ? : 1;
>> -				goto unlock;
>> -			}
>> +			shrinker_put(shrinker);
> 
> Ok, so why is this safe to call without holding the rcu read lock?
> The global shrinker has to hold the rcu_read_lock() whilst calling
> shrinker_put() to guarantee the validity of the list next pointer,
> but we don't hold off RCU here so what guarantees a racing global
> shrinker walk doesn't trip over this shrinker_put() call dropping
> the refcount to zero and freeing occuring in a different context...

This will not be a problem, even if shrinker::refcount is reduced to
0 here, the racing global shrinker walk already holds the rcu lock.

         shrink_slab            shrink_slab_memcg
         ===========            =================

         rcu_read_lock()
         shrinker_put()
                                shrinker_put()

And in shrink_slab_memcg(), the shrinker is not required to traverse the
next bit in the shrinker_info_unit::map, so there is no need to hold the
rcu lock to ensure the existence of this shrinker.

> 
> 
>> +		/*
>> +		 * We have already exited the read-side of rcu critical section
>> +		 * before calling do_shrink_slab(), the shrinker_info may be
>> +		 * released in expand_one_shrinker_info(), so reacquire the
>> +		 * shrinker_info.
>> +		 */
>> +		index++;
>> +		goto again;
> 
> With that, what makes the use of shrinker_info in
> xchg_nr_deferred_memcg() in do_shrink_slab() coherent and valid?

Holding rcu lock can ensure that the old shrinker_info will not be
freed, and the shrinker_info_unit::nr_deferred can also be indexed from
the old shrinker_info::unit[x], so the updated nr_deferred will not be
lost.

Thanks,
Qi

> 
> -Dave.

  reply	other threads:[~2023-08-08 17:10 UTC|newest]

Thread overview: 66+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-08-07 11:08 [PATCH v4 00/48] use refcount+RCU method to implement lockless slab shrink Qi Zheng
2023-08-07 11:08 ` [PATCH v4 01/48] mm: move some shrinker-related function declarations to mm/internal.h Qi Zheng
2023-08-15  8:36   ` Muchun Song
2023-08-15  9:14     ` Qi Zheng
2023-08-07 11:08 ` [PATCH v4 02/48] mm: vmscan: move shrinker-related code into a separate file Qi Zheng
2023-08-15  8:44   ` Muchun Song
2023-08-07 11:08 ` [PATCH v4 03/48] mm: shrinker: remove redundant shrinker_rwsem in debugfs operations Qi Zheng
2023-08-07 11:08 ` [PATCH v4 04/48] mm: shrinker: add infrastructure for dynamically allocating shrinker Qi Zheng
2023-08-07 11:08 ` [PATCH v4 05/48] kvm: mmu: dynamically allocate the x86-mmu shrinker Qi Zheng
2023-08-07 11:08 ` [PATCH v4 06/48] binder: dynamically allocate the android-binder shrinker Qi Zheng
2023-08-08  2:30   ` Muchun Song
2023-08-07 11:08 ` [PATCH v4 07/48] drm/ttm: dynamically allocate the drm-ttm_pool shrinker Qi Zheng
2023-08-07 11:08 ` [PATCH v4 08/48] xenbus/backend: dynamically allocate the xen-backend shrinker Qi Zheng
2023-08-07 11:08 ` [PATCH v4 09/48] erofs: dynamically allocate the erofs-shrinker Qi Zheng
2023-08-07 11:08 ` [PATCH v4 10/48] f2fs: dynamically allocate the f2fs-shrinker Qi Zheng
2023-08-07 11:08 ` [PATCH v4 11/48] gfs2: dynamically allocate the gfs2-glock shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 12/48] gfs2: dynamically allocate the gfs2-qd shrinker Qi Zheng
2023-08-15  8:56   ` Muchun Song
2023-08-07 11:09 ` [PATCH v4 13/48] NFSv4.2: dynamically allocate the nfs-xattr shrinkers Qi Zheng
2023-08-07 11:09 ` [PATCH v4 14/48] nfs: dynamically allocate the nfs-acl shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 15/48] nfsd: dynamically allocate the nfsd-filecache shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 16/48] quota: dynamically allocate the dquota-cache shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 17/48] ubifs: dynamically allocate the ubifs-slab shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 18/48] rcu: dynamically allocate the rcu-lazy shrinker Qi Zheng
2023-08-07 14:16   ` Joel Fernandes
2023-08-07 11:09 ` [PATCH v4 19/48] rcu: dynamically allocate the rcu-kfree shrinker Qi Zheng
2023-08-07 14:14   ` Joel Fernandes
2023-08-08  2:28   ` Muchun Song
2023-08-07 11:09 ` [PATCH v4 20/48] mm: thp: dynamically allocate the thp-related shrinkers Qi Zheng
2023-08-07 11:09 ` [PATCH v4 21/48] sunrpc: dynamically allocate the sunrpc_cred shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 22/48] mm: workingset: dynamically allocate the mm-shadow shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 23/48] drm/i915: dynamically allocate the i915_gem_mm shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 24/48] drm/msm: dynamically allocate the drm-msm_gem shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 25/48] drm/panfrost: dynamically allocate the drm-panfrost shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 26/48] dm: dynamically allocate the dm-bufio shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 27/48] dm zoned: dynamically allocate the dm-zoned-meta shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 28/48] md/raid5: dynamically allocate the md-raid5 shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 29/48] bcache: dynamically allocate the md-bcache shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 30/48] vmw_balloon: dynamically allocate the vmw-balloon shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 31/48] virtio_balloon: dynamically allocate the virtio-balloon shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 32/48] mbcache: dynamically allocate the mbcache shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 33/48] ext4: dynamically allocate the ext4-es shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 34/48] jbd2,ext4: dynamically allocate the jbd2-journal shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 35/48] nfsd: dynamically allocate the nfsd-client shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 36/48] nfsd: dynamically allocate the nfsd-reply shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 37/48] xfs: dynamically allocate the xfs-buf shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 38/48] xfs: dynamically allocate the xfs-inodegc shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 39/48] xfs: dynamically allocate the xfs-qm shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 40/48] zsmalloc: dynamically allocate the mm-zspool shrinker Qi Zheng
2023-08-07 11:09 ` [PATCH v4 41/48] fs: super: dynamically allocate the s_shrink Qi Zheng
2023-08-07 11:09 ` [PATCH v4 42/48] mm: shrinker: remove old APIs Qi Zheng
2023-08-07 11:09 ` [PATCH v4 43/48] drm/ttm: introduce pool_shrink_rwsem Qi Zheng
2023-08-22 13:56   ` Daniel Vetter
2023-08-23  2:59     ` Qi Zheng
2023-08-07 11:09 ` [PATCH v4 44/48] mm: shrinker: add a secondary array for shrinker_info::{map, nr_deferred} Qi Zheng
2023-08-08  2:12   ` Dave Chinner
2023-08-08  6:32     ` Qi Zheng
2023-08-07 11:09 ` [PATCH v4 45/48] mm: shrinker: make global slab shrink lockless Qi Zheng
2023-08-07 23:28   ` Dave Chinner
2023-08-08  2:24   ` Dave Chinner
2023-08-08  7:22     ` Qi Zheng
2023-08-07 11:09 ` [PATCH v4 46/48] mm: shrinker: make memcg " Qi Zheng
2023-08-08  2:44   ` Dave Chinner
2023-08-08  7:50     ` Qi Zheng [this message]
2023-08-07 11:09 ` [PATCH v4 47/48] mm: shrinker: hold write lock to reparent shrinker nr_deferred Qi Zheng
2023-08-07 11:09 ` [PATCH v4 48/48] mm: shrinker: convert shrinker_rwsem to mutex Qi Zheng

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=5757e341-b261-14de-e052-46606d530460@bytedance.com \
    --to=zhengqi.arch@bytedance.com \
    --cc=akpm@linux-foundation.org \
    --cc=brauner@kernel.org \
    --cc=cel@kernel.org \
    --cc=cluster-devel@redhat.com \
    --cc=david@fromorbit.com \
    --cc=djwong@kernel.org \
    --cc=dlemoal@kernel.org \
    --cc=dm-devel@redhat.com \
    --cc=dri-devel@lists.freedesktop.org \
    --cc=gregkh@linuxfoundation.org \
    --cc=kvm@vger.kernel.org \
    --cc=linux-arm-msm@vger.kernel.org \
    --cc=linux-bcache@vger.kernel.org \
    --cc=linux-btrfs@vger.kernel.org \
    --cc=linux-erofs@lists.ozlabs.org \
    --cc=linux-ext4@vger.kernel.org \
    --cc=linux-f2fs-devel@lists.sourceforge.net \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux-mtd@lists.infradead.org \
    --cc=linux-nfs@vger.kernel.org \
    --cc=linux-raid@vger.kernel.org \
    --cc=linux-xfs@vger.kernel.org \
    --cc=muchun.song@linux.dev \
    --cc=netdev@vger.kernel.org \
    --cc=paulmck@kernel.org \
    --cc=rcu@vger.kernel.org \
    --cc=roman.gushchin@linux.dev \
    --cc=senozhatsky@chromium.org \
    --cc=simon.horman@corigine.com \
    --cc=steven.price@arm.com \
    --cc=tkhai@ya.ru \
    --cc=tytso@mit.edu \
    --cc=vbabka@suse.cz \
    --cc=virtualization@lists.linux-foundation.org \
    --cc=x86@kernel.org \
    --cc=xen-devel@lists.xenproject.org \
    --cc=yujie.liu@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox