From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id C777634DCF2 for ; Thu, 15 Jan 2026 09:26:57 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=170.10.133.124 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1768469220; cv=none; b=eIH6bsa02wldGSZdS6LYK+wFuY6EUZ+GDWOGh8T6GEGye0X6UC09J3YXHiuTlisI1nvcuilnEgtKiLUzla4QIhKLex8yJdn5Ii1d5KdMrd1DvBAo2myHCTLQfz8bwPPGGDtmjynscscHZO2hpkwQ3zqudmfkhv3gOqRIEtOnZLM= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1768469220; c=relaxed/simple; bh=+HANNRnvJo+yi8LbLlChnIfO2AA43c+quMMVkxCprM0=; h=Date:From:To:Cc:Subject:Message-ID:References:MIME-Version: Content-Type:Content-Disposition:In-Reply-To; b=DdMO4U6P9bvdPzexXcQBWEDGsKW8as261vMmFOMMp06K2D6877gyI2Hmy0qoQpGTTaoNJD5S13+hANeeUUWCqLV+IWBsxmw76YCcxrgMq12+CJzAmIOIlF7lHZtcXNVt2u/epTsqnfNqUGLxuvPtZkF5wNz5s9qCbXd9vrOt0+8= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=redhat.com; spf=pass smtp.mailfrom=redhat.com; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b=GTPbEKNv; arc=none smtp.client-ip=170.10.133.124 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=redhat.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=redhat.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b="GTPbEKNv" DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1768469216; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=o3P5TXz9n5urVuMSto49THOUDq4k6hsZfwksZf+Xzdg=; b=GTPbEKNv/izx7XIIDaEzfYfGIzl4SYts/eHc9xJCuCXH1GMBMPBnuiSCWeJbVBvubANgdN YdDI16ptomUcDvL6zVsgsYeQF+poLd+Th6q5yNmsLSv8KTXoSrdzM2JBQK8scMJEFUQxuO LnWDZPAeL3IOHVDM8vlWk5VsER636Xs= Received: from mx-prod-mc-08.mail-002.prod.us-west-2.aws.redhat.com (ec2-35-165-154-97.us-west-2.compute.amazonaws.com [35.165.154.97]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-628-3KKHiIK4PEicFgHIz8RiFw-1; Thu, 15 Jan 2026 04:26:52 -0500 X-MC-Unique: 3KKHiIK4PEicFgHIz8RiFw-1 X-Mimecast-MFC-AGG-ID: 3KKHiIK4PEicFgHIz8RiFw_1768469209 Received: from mx-prod-int-08.mail-002.prod.us-west-2.aws.redhat.com (mx-prod-int-08.mail-002.prod.us-west-2.aws.redhat.com [10.30.177.111]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits) server-digest SHA256) (No client certificate requested) by mx-prod-mc-08.mail-002.prod.us-west-2.aws.redhat.com (Postfix) with ESMTPS id B9EE51800473; Thu, 15 Jan 2026 09:26:48 +0000 (UTC) Received: from localhost (unknown [10.72.112.110]) by mx-prod-int-08.mail-002.prod.us-west-2.aws.redhat.com (Postfix) with ESMTPS id BA338180066A; Thu, 15 Jan 2026 09:26:46 +0000 (UTC) Date: Thu, 15 Jan 2026 17:26:42 +0800 From: Baoquan He To: Qi Zheng Cc: hannes@cmpxchg.org, hughd@google.com, mhocko@suse.com, roman.gushchin@linux.dev, shakeel.butt@linux.dev, muchun.song@linux.dev, david@kernel.org, lorenzo.stoakes@oracle.com, ziy@nvidia.com, harry.yoo@oracle.com, yosry.ahmed@linux.dev, imran.f.khan@oracle.com, kamalesh.babulal@oracle.com, axelrasmussen@google.com, yuanchu@google.com, weixugc@google.com, chenridong@huaweicloud.com, mkoutny@suse.com, akpm@linux-foundation.org, hamzamahfooz@linux.microsoft.com, apais@linux.microsoft.com, lance.yang@linux.dev, linux-mm@kvack.org, linux-kernel@vger.kernel.org, cgroups@vger.kernel.org, Qi Zheng Subject: Re: [PATCH v3 23/30] mm: do not open-code lruvec lock Message-ID: References: <33fef62fd821f669fcdc999e54c4035a4e91b47d.1768389889.git.zhengqi.arch@bytedance.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <33fef62fd821f669fcdc999e54c4035a4e91b47d.1768389889.git.zhengqi.arch@bytedance.com> X-Scanned-By: MIMEDefang 3.4.1 on 10.30.177.111 On 01/14/26 at 07:32pm, Qi Zheng wrote: > From: Qi Zheng > > Now we have lruvec_unlock(), lruvec_unlock_irq() and > lruvec_unlock_irqrestore(), but not the paired lruvec_lock(), ~~ typo, 'no'? > lruvec_lock_irq() and lruvec_lock_irqsave(). > > There is currently no use case for lruvec_lock_irqsave(), so only > introduce lruvec_lock() and lruvec_lock_irq(), and change all open-code I didn't see lruvec_lock() is introduced in this patch, do I miss anthing? > places to use these helper function. This looks cleaner and prepares for > reparenting LRU pages, preventing user from missing RCU lock calls due to > open-code lruvec lock. > > Signed-off-by: Qi Zheng > --- > include/linux/memcontrol.h | 5 +++++ > mm/vmscan.c | 38 +++++++++++++++++++------------------- > 2 files changed, 24 insertions(+), 19 deletions(-) > > diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h > index f1556759d0d3f..4b6f20dc694ba 100644 > --- a/include/linux/memcontrol.h > +++ b/include/linux/memcontrol.h > @@ -1499,6 +1499,11 @@ static inline struct lruvec *parent_lruvec(struct lruvec *lruvec) > return mem_cgroup_lruvec(memcg, lruvec_pgdat(lruvec)); > } > > +static inline void lruvec_lock_irq(struct lruvec *lruvec) > +{ > + spin_lock_irq(&lruvec->lru_lock); > +} > + > static inline void lruvec_unlock(struct lruvec *lruvec) > { > spin_unlock(&lruvec->lru_lock); > diff --git a/mm/vmscan.c b/mm/vmscan.c > index f206d4dac9e77..c48ff6e05e004 100644 > --- a/mm/vmscan.c > +++ b/mm/vmscan.c > @@ -2020,7 +2020,7 @@ static unsigned long shrink_inactive_list(unsigned long nr_to_scan, > > lru_add_drain(); > > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > > nr_taken = isolate_lru_folios(nr_to_scan, lruvec, &folio_list, > &nr_scanned, sc, lru); > @@ -2032,7 +2032,7 @@ static unsigned long shrink_inactive_list(unsigned long nr_to_scan, > count_memcg_events(lruvec_memcg(lruvec), item, nr_scanned); > __count_vm_events(PGSCAN_ANON + file, nr_scanned); > > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > > if (nr_taken == 0) > return 0; > @@ -2051,7 +2051,7 @@ static unsigned long shrink_inactive_list(unsigned long nr_to_scan, > count_memcg_events(lruvec_memcg(lruvec), item, nr_reclaimed); > count_vm_events(PGSTEAL_ANON + file, nr_reclaimed); > > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > lru_note_cost_unlock_irq(lruvec, file, stat.nr_pageout, > nr_scanned - nr_reclaimed); > > @@ -2130,7 +2130,7 @@ static void shrink_active_list(unsigned long nr_to_scan, > > lru_add_drain(); > > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > > nr_taken = isolate_lru_folios(nr_to_scan, lruvec, &l_hold, > &nr_scanned, sc, lru); > @@ -2141,7 +2141,7 @@ static void shrink_active_list(unsigned long nr_to_scan, > __count_vm_events(PGREFILL, nr_scanned); > count_memcg_events(lruvec_memcg(lruvec), PGREFILL, nr_scanned); > > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > > while (!list_empty(&l_hold)) { > struct folio *folio; > @@ -2197,7 +2197,7 @@ static void shrink_active_list(unsigned long nr_to_scan, > count_memcg_events(lruvec_memcg(lruvec), PGDEACTIVATE, nr_deactivate); > mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, -nr_taken); > > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > lru_note_cost_unlock_irq(lruvec, file, 0, nr_rotated); > trace_mm_vmscan_lru_shrink_active(pgdat->node_id, nr_taken, nr_activate, > nr_deactivate, nr_rotated, sc->priority, file); > @@ -3832,9 +3832,9 @@ static void walk_mm(struct mm_struct *mm, struct lru_gen_mm_walk *walk) > } > > if (walk->batched) { > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > reset_batch_size(walk); > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > } > > cond_resched(); > @@ -3993,7 +3993,7 @@ static bool inc_max_seq(struct lruvec *lruvec, unsigned long seq, int swappiness > if (seq < READ_ONCE(lrugen->max_seq)) > return false; > > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > > VM_WARN_ON_ONCE(!seq_is_valid(lruvec)); > > @@ -4008,7 +4008,7 @@ static bool inc_max_seq(struct lruvec *lruvec, unsigned long seq, int swappiness > if (inc_min_seq(lruvec, type, swappiness)) > continue; > > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > cond_resched(); > goto restart; > } > @@ -4043,7 +4043,7 @@ static bool inc_max_seq(struct lruvec *lruvec, unsigned long seq, int swappiness > /* make sure preceding modifications appear */ > smp_store_release(&lrugen->max_seq, lrugen->max_seq + 1); > unlock: > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > > return success; > } > @@ -4739,7 +4739,7 @@ static int evict_folios(unsigned long nr_to_scan, struct lruvec *lruvec, > struct mem_cgroup *memcg = lruvec_memcg(lruvec); > struct pglist_data *pgdat = lruvec_pgdat(lruvec); > > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > > scanned = isolate_folios(nr_to_scan, lruvec, sc, swappiness, &type, &list); > > @@ -4748,7 +4748,7 @@ static int evict_folios(unsigned long nr_to_scan, struct lruvec *lruvec, > if (evictable_min_seq(lrugen->min_seq, swappiness) + MIN_NR_GENS > lrugen->max_seq) > scanned = 0; > > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > > if (list_empty(&list)) > return scanned; > @@ -4786,9 +4786,9 @@ static int evict_folios(unsigned long nr_to_scan, struct lruvec *lruvec, > walk = current->reclaim_state->mm_walk; > if (walk && walk->batched) { > walk->lruvec = lruvec; > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > reset_batch_size(walk); > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > } > > mod_lruvec_state(lruvec, PGDEMOTE_KSWAPD + reclaimer_offset(sc), > @@ -5226,7 +5226,7 @@ static void lru_gen_change_state(bool enabled) > for_each_node(nid) { > struct lruvec *lruvec = get_lruvec(memcg, nid); > > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > > VM_WARN_ON_ONCE(!seq_is_valid(lruvec)); > VM_WARN_ON_ONCE(!state_is_valid(lruvec)); > @@ -5234,12 +5234,12 @@ static void lru_gen_change_state(bool enabled) > lruvec->lrugen.enabled = enabled; > > while (!(enabled ? fill_evictable(lruvec) : drain_evictable(lruvec))) { > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > cond_resched(); > - spin_lock_irq(&lruvec->lru_lock); > + lruvec_lock_irq(lruvec); > } > > - spin_unlock_irq(&lruvec->lru_lock); > + lruvec_unlock_irq(lruvec); > } > > cond_resched(); > -- > 2.20.1 > >