public inbox for linux-mm@kvack.org
 help / color / mirror / Atom feed
From: Baolin Wang <baolin.wang@linux.alibaba.com>
To: Kairui Song <ryncsn@gmail.com>
Cc: kasong@tencent.com, linux-mm@kvack.org,
	Andrew Morton <akpm@linux-foundation.org>,
	Axel Rasmussen <axelrasmussen@google.com>,
	Yuanchu Xie <yuanchu@google.com>, Wei Xu <weixugc@google.com>,
	Johannes Weiner <hannes@cmpxchg.org>,
	David Hildenbrand <david@kernel.org>,
	Michal Hocko <mhocko@kernel.org>,
	Qi Zheng <zhengqi.arch@bytedance.com>,
	Shakeel Butt <shakeel.butt@linux.dev>,
	Lorenzo Stoakes <ljs@kernel.org>, Barry Song <baohua@kernel.org>,
	David Stevens <stevensd@google.com>,
	Chen Ridong <chenridong@huaweicloud.com>,
	Leno Hou <lenohou@gmail.com>, Yafang Shao <laoar.shao@gmail.com>,
	Yu Zhao <yuzhao@google.com>, Zicheng Wang <wangzicheng@honor.com>,
	Kalesh Singh <kaleshsingh@google.com>,
	Suren Baghdasaryan <surenb@google.com>,
	Chris Li <chrisl@kernel.org>, Vernon Yang <vernon2gm@gmail.com>,
	linux-kernel@vger.kernel.org, Qi Zheng <qi.zheng@linux.dev>
Subject: Re: [PATCH v2 12/12] mm/vmscan: unify writeback reclaim statistic and throttling
Date: Tue, 31 Mar 2026 17:36:26 +0800	[thread overview]
Message-ID: <522f4898-78c3-453f-8367-29327e29290e@linux.alibaba.com> (raw)
In-Reply-To: <acuTMVSnGb0XRvTV@KASONG-MC4>



On 3/31/26 5:29 PM, Kairui Song wrote:
> On Tue, Mar 31, 2026 at 05:24:39PM +0800, Baolin Wang wrote:
>>
>>
>> On 3/29/26 3:52 AM, Kairui Song via B4 Relay wrote:
>>> From: Kairui Song <kasong@tencent.com>
>>>
>>> Currently MGLRU and non-MGLRU handle the reclaim statistic and
>>> writeback handling very differently, especially throttling.
>>> Basically MGLRU just ignored the throttling part.
>>>
>>> Let's just unify this part, use a helper to deduplicate the code
>>> so both setups will share the same behavior. Also remove the
>>> folio_clear_reclaim in isolate_folio which was actively invalidating
>>> the congestion control. PG_reclaim is now handled by shrink_folio_list,
>>> keeping it in isolate_folio is not helpful.
>>>
>>> Test using following reproducer using bash:
>>>
>>>     echo "Setup a slow device using dm delay"
>>>     dd if=/dev/zero of=/var/tmp/backing bs=1M count=2048
>>>     LOOP=$(losetup --show -f /var/tmp/backing)
>>>     mkfs.ext4 -q $LOOP
>>>     echo "0 $(blockdev --getsz $LOOP) delay $LOOP 0 0 $LOOP 0 1000" | \
>>>         dmsetup create slow_dev
>>>     mkdir -p /mnt/slow && mount /dev/mapper/slow_dev /mnt/slow
>>>
>>>     echo "Start writeback pressure"
>>>     sync && echo 3 > /proc/sys/vm/drop_caches
>>>     mkdir /sys/fs/cgroup/test_wb
>>>     echo 128M > /sys/fs/cgroup/test_wb/memory.max
>>>     (echo $BASHPID > /sys/fs/cgroup/test_wb/cgroup.procs && \
>>>         dd if=/dev/zero of=/mnt/slow/testfile bs=1M count=192)
>>>
>>>     echo "Clean up"
>>>     echo "0 $(blockdev --getsz $LOOP) error" | dmsetup load slow_dev
>>>     dmsetup resume slow_dev
>>>     umount -l /mnt/slow && sync
>>>     dmsetup remove slow_dev
>>>
>>> Before this commit, `dd` will get OOM killed immediately if
>>> MGLRU is enabled. Classic LRU is fine.
>>>
>>> After this commit, congestion control is now effective and no more
>>> spin on LRU or premature OOM.
>>>
>>> Stress test on other workloads also looking good.
>>>
>>> Suggested-by: Chen Ridong <chenridong@huaweicloud.com>
>>> Signed-off-by: Kairui Song <kasong@tencent.com>
>>> ---
>>>    mm/vmscan.c | 93 +++++++++++++++++++++++++++----------------------------------
>>>    1 file changed, 41 insertions(+), 52 deletions(-)
>>>
>>> diff --git a/mm/vmscan.c b/mm/vmscan.c
>>> index 1783da54ada1..83c8fdf8fdc4 100644
>>> --- a/mm/vmscan.c
>>> +++ b/mm/vmscan.c
>>> @@ -1942,6 +1942,44 @@ static int current_may_throttle(void)
>>>    	return !(current->flags & PF_LOCAL_THROTTLE);
>>>    }
>>> +static void handle_reclaim_writeback(unsigned long nr_taken,
>>> +				     struct pglist_data *pgdat,
>>> +				     struct scan_control *sc,
>>> +				     struct reclaim_stat *stat)
>>> +{
>>> +	/*
>>> +	 * If dirty folios are scanned that are not queued for IO, it
>>> +	 * implies that flushers are not doing their job. This can
>>> +	 * happen when memory pressure pushes dirty folios to the end of
>>> +	 * the LRU before the dirty limits are breached and the dirty
>>> +	 * data has expired. It can also happen when the proportion of
>>> +	 * dirty folios grows not through writes but through memory
>>> +	 * pressure reclaiming all the clean cache. And in some cases,
>>> +	 * the flushers simply cannot keep up with the allocation
>>> +	 * rate. Nudge the flusher threads in case they are asleep.
>>> +	 */
>>> +	if (stat->nr_unqueued_dirty == nr_taken && nr_taken) {
>>> +		wakeup_flusher_threads(WB_REASON_VMSCAN);
>>> +		/*
>>> +		 * For cgroupv1 dirty throttling is achieved by waking up
>>> +		 * the kernel flusher here and later waiting on folios
>>> +		 * which are in writeback to finish (see shrink_folio_list()).
>>> +		 *
>>> +		 * Flusher may not be able to issue writeback quickly
>>> +		 * enough for cgroupv1 writeback throttling to work
>>> +		 * on a large system.
>>> +		 */
>>> +		if (!writeback_throttling_sane(sc))
>>> +			reclaim_throttle(pgdat, VMSCAN_THROTTLE_WRITEBACK);
>>> +	}
>>> +
>>> +	sc->nr.dirty += stat->nr_dirty;
>>> +	sc->nr.congested += stat->nr_congested;
>>> +	sc->nr.writeback += stat->nr_writeback;
>>> +	sc->nr.immediate += stat->nr_immediate;
>>> +	sc->nr.taken += nr_taken;
>>> +}
>>> +
>>>    /*
>>>     * shrink_inactive_list() is a helper for shrink_node().  It returns the number
>>>     * of reclaimed pages
>>> @@ -2005,39 +2043,7 @@ static unsigned long shrink_inactive_list(unsigned long nr_to_scan,
>>>    	lruvec_lock_irq(lruvec);
>>>    	lru_note_cost_unlock_irq(lruvec, file, stat.nr_pageout,
>>>    					nr_scanned - nr_reclaimed);
>>> -
>>> -	/*
>>> -	 * If dirty folios are scanned that are not queued for IO, it
>>> -	 * implies that flushers are not doing their job. This can
>>> -	 * happen when memory pressure pushes dirty folios to the end of
>>> -	 * the LRU before the dirty limits are breached and the dirty
>>> -	 * data has expired. It can also happen when the proportion of
>>> -	 * dirty folios grows not through writes but through memory
>>> -	 * pressure reclaiming all the clean cache. And in some cases,
>>> -	 * the flushers simply cannot keep up with the allocation
>>> -	 * rate. Nudge the flusher threads in case they are asleep.
>>> -	 */
>>> -	if (stat.nr_unqueued_dirty == nr_taken) {
>>> -		wakeup_flusher_threads(WB_REASON_VMSCAN);
>>> -		/*
>>> -		 * For cgroupv1 dirty throttling is achieved by waking up
>>> -		 * the kernel flusher here and later waiting on folios
>>> -		 * which are in writeback to finish (see shrink_folio_list()).
>>> -		 *
>>> -		 * Flusher may not be able to issue writeback quickly
>>> -		 * enough for cgroupv1 writeback throttling to work
>>> -		 * on a large system.
>>> -		 */
>>> -		if (!writeback_throttling_sane(sc))
>>> -			reclaim_throttle(pgdat, VMSCAN_THROTTLE_WRITEBACK);
>>> -	}
>>> -
>>> -	sc->nr.dirty += stat.nr_dirty;
>>> -	sc->nr.congested += stat.nr_congested;
>>> -	sc->nr.writeback += stat.nr_writeback;
>>> -	sc->nr.immediate += stat.nr_immediate;
>>> -	sc->nr.taken += nr_taken;
>>> -
>>> +	handle_reclaim_writeback(nr_taken, pgdat, sc, &stat);
>>>    	trace_mm_vmscan_lru_shrink_inactive(pgdat->node_id,
>>>    			nr_scanned, nr_reclaimed, &stat, sc->priority, file);
>>>    	return nr_reclaimed;
>>> @@ -4651,9 +4657,6 @@ static bool isolate_folio(struct lruvec *lruvec, struct folio *folio, struct sca
>>>    	if (!folio_test_referenced(folio))
>>>    		set_mask_bits(&folio->flags.f, LRU_REFS_MASK, 0);
>>> -	/* for shrink_folio_list() */
>>> -	folio_clear_reclaim(folio);
>>
>> IMO, Moving this change into patch 8 would make more sense. Otherwise LGTM.
> 
> Thanks for the review! I made it a separate patch so we can better
> identify which part had the performance gain, and patch 8 can keep
> the review by. Patch 8 is still good without this, a few counters
> are updated with no user, kind of wasted but that's harmless.

I’m not referring to all the above changes. What I mean is that the 
'folio_clear_reclaim' removal should belong to patch 8. Since 
shrink_folio_list() in patch 8 will handle the writeback logic, 
folio_clear_reclaim() should also be removed in the same patch.


  reply	other threads:[~2026-03-31  9:36 UTC|newest]

Thread overview: 44+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-03-28 19:52 [PATCH v2 00/12] mm/mglru: improve reclaim loop and dirty folio handling Kairui Song via B4 Relay
2026-03-28 19:52 ` [PATCH v2 01/12] mm/mglru: consolidate common code for retrieving evitable size Kairui Song via B4 Relay
2026-03-28 19:52 ` [PATCH v2 02/12] mm/mglru: rename variables related to aging and rotation Kairui Song via B4 Relay
2026-03-30  1:57   ` Chen Ridong
2026-03-30  7:59   ` Baolin Wang
2026-04-01  0:00   ` Barry Song
2026-03-28 19:52 ` [PATCH v2 03/12] mm/mglru: relocate the LRU scan batch limit to callers Kairui Song via B4 Relay
2026-03-30  8:14   ` Baolin Wang
2026-04-01  0:20     ` Barry Song
2026-03-28 19:52 ` [PATCH v2 04/12] mm/mglru: restructure the reclaim loop Kairui Song via B4 Relay
2026-03-29  6:47   ` Kairui Song
2026-03-28 19:52 ` [PATCH v2 05/12] mm/mglru: scan and count the exact number of folios Kairui Song via B4 Relay
2026-03-31  8:04   ` Baolin Wang
2026-03-31  9:01     ` Kairui Song
2026-03-31  9:52       ` Baolin Wang
2026-03-28 19:52 ` [PATCH v2 06/12] mm/mglru: use a smaller batch for reclaim Kairui Song via B4 Relay
2026-03-31  8:08   ` Baolin Wang
2026-03-28 19:52 ` [PATCH v2 07/12] mm/mglru: don't abort scan immediately right after aging Kairui Song via B4 Relay
2026-03-28 19:52 ` [PATCH v2 08/12] mm/mglru: simplify and improve dirty writeback handling Kairui Song via B4 Relay
2026-03-29  8:21   ` Kairui Song
2026-03-29  8:46     ` Kairui Song
2026-03-31  8:42   ` Baolin Wang
2026-03-31  9:18     ` Kairui Song
2026-04-01  2:52       ` Baolin Wang
2026-04-01  4:57         ` Kairui Song
2026-04-02  0:11       ` Barry Song
2026-04-01 23:37   ` Shakeel Butt
2026-04-02 11:44     ` Kairui Song
2026-03-28 19:52 ` [PATCH v2 09/12] mm/mglru: remove no longer used reclaim argument for folio protection Kairui Song via B4 Relay
2026-03-28 19:52 ` [PATCH v2 10/12] mm/vmscan: remove sc->file_taken Kairui Song via B4 Relay
2026-03-31  8:49   ` Baolin Wang
2026-03-28 19:52 ` [PATCH v2 11/12] mm/vmscan: remove sc->unqueued_dirty Kairui Song via B4 Relay
2026-03-31  8:51   ` Baolin Wang
2026-03-28 19:52 ` [PATCH v2 12/12] mm/vmscan: unify writeback reclaim statistic and throttling Kairui Song via B4 Relay
2026-03-31  9:24   ` Baolin Wang
2026-03-31  9:29     ` Kairui Song
2026-03-31  9:36       ` Baolin Wang [this message]
2026-03-31  9:40         ` Kairui Song
2026-04-01  5:01   ` Leno Hou
2026-04-02  2:39   ` Shakeel Butt
2026-04-02  2:56     ` Kairui Song
2026-04-02  3:17       ` Shakeel Butt
2026-04-01  5:18 ` [PATCH v2 00/12] mm/mglru: improve reclaim loop and dirty folio handling Leno Hou
2026-04-01  7:36   ` Kairui Song

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=522f4898-78c3-453f-8367-29327e29290e@linux.alibaba.com \
    --to=baolin.wang@linux.alibaba.com \
    --cc=akpm@linux-foundation.org \
    --cc=axelrasmussen@google.com \
    --cc=baohua@kernel.org \
    --cc=chenridong@huaweicloud.com \
    --cc=chrisl@kernel.org \
    --cc=david@kernel.org \
    --cc=hannes@cmpxchg.org \
    --cc=kaleshsingh@google.com \
    --cc=kasong@tencent.com \
    --cc=laoar.shao@gmail.com \
    --cc=lenohou@gmail.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=ljs@kernel.org \
    --cc=mhocko@kernel.org \
    --cc=qi.zheng@linux.dev \
    --cc=ryncsn@gmail.com \
    --cc=shakeel.butt@linux.dev \
    --cc=stevensd@google.com \
    --cc=surenb@google.com \
    --cc=vernon2gm@gmail.com \
    --cc=wangzicheng@honor.com \
    --cc=weixugc@google.com \
    --cc=yuanchu@google.com \
    --cc=yuzhao@google.com \
    --cc=zhengqi.arch@bytedance.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox