From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 9B201C433EF for ; Tue, 22 Mar 2022 10:32:36 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232558AbiCVKeC (ORCPT ); Tue, 22 Mar 2022 06:34:02 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:40826 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232517AbiCVKd5 (ORCPT ); Tue, 22 Mar 2022 06:33:57 -0400 Received: from ams.source.kernel.org (ams.source.kernel.org [145.40.68.75]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id DDC515C358 for ; Tue, 22 Mar 2022 03:32:29 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ams.source.kernel.org (Postfix) with ESMTPS id 91FE4B81B67 for ; Tue, 22 Mar 2022 10:32:28 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 0BF1AC340EC; Tue, 22 Mar 2022 10:32:26 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1647945147; bh=4vRfTKQYtgyUB3FKZacyM7oyLXrvgTFjMYsTjUSXKDg=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=PxmMRf3HTtDzvtAo/AKdcGvlPJFvhun1kKsI1kashCqdLg+Gop2Y4rfloWtVop9AH P3XXAtCs+EEa64rMv4eLb/JhDRF+f0S4dQwnrpOvg/GsEssDOPH4ZWym/u3jYGtRGO 2R4cC0lyhzb7KLzdjR2GGE664kphyodARO5nYuj3alZMD5RB3WWcd6n1IR1eXTAJB+ alwYwpsJQyMGAgkD45/8ev02f44hySjxrDSkqjG3Psl8B4wL5ysbgtpmJGguWgxl68 oq8WbU07i9zFRgIZ08A1hf7Vi/kv/Jcd7j+Dyb5CeIXuDLd1f4snxPpqOZsaAzBVKg bbWoGtnKs/FXA== Date: Tue, 22 Mar 2022 11:32:24 +0100 From: Frederic Weisbecker To: "Paul E. McKenney" Cc: LKML , Uladzislau Rezki , Boqun Feng , Neeraj Upadhyay , Joel Fernandes Subject: Re: [PATCH 4/4] rcu: Name internal polling flag Message-ID: <20220322103224.GA701946@lothringen> References: <20220316144255.336021-1-frederic@kernel.org> <20220316144255.336021-5-frederic@kernel.org> <20220322021107.GP4285@paulmck-ThinkPad-P17-Gen-1> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20220322021107.GP4285@paulmck-ThinkPad-P17-Gen-1> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Mar 21, 2022 at 07:11:07PM -0700, Paul E. McKenney wrote: > On Wed, Mar 16, 2022 at 03:42:55PM +0100, Frederic Weisbecker wrote: > > Give a proper self-explanatory name to the expedited grace period > > internal polling flag. > > > > Signed-off-by: Frederic Weisbecker > > Cc: Neeraj Upadhyay > > Cc: Boqun Feng > > Cc: Uladzislau Rezki > > Cc: Joel Fernandes > > --- > > kernel/rcu/rcu.h | 5 +++++ > > kernel/rcu/tree.c | 2 +- > > kernel/rcu/tree_exp.h | 9 +++++---- > > 3 files changed, 11 insertions(+), 5 deletions(-) > > > > diff --git a/kernel/rcu/rcu.h b/kernel/rcu/rcu.h > > index eccbdbdaa02e..8a62bb416ba4 100644 > > --- a/kernel/rcu/rcu.h > > +++ b/kernel/rcu/rcu.h > > @@ -30,6 +30,11 @@ > > #define RCU_GET_STATE_USE_NORMAL 0x2 > > #define RCU_GET_STATE_BAD_FOR_NORMAL (RCU_GET_STATE_FROM_EXPEDITED | RCU_GET_STATE_USE_NORMAL) > > > > +/* > > + * Low-order bit definitions for polled grace-period internals. > > + */ > > +#define RCU_EXP_SEQ_POLL_DONE 0x1 > > + > > /* > > * Return the counter portion of a sequence number previously returned > > * by rcu_seq_snap() or rcu_seq_current(). > > diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c > > index 5da381a3cbe5..b3223b365f9f 100644 > > --- a/kernel/rcu/tree.c > > +++ b/kernel/rcu/tree.c > > @@ -4679,7 +4679,7 @@ static void __init rcu_init_one(void) > > spin_lock_init(&rnp->exp_lock); > > mutex_init(&rnp->boost_kthread_mutex); > > raw_spin_lock_init(&rnp->exp_poll_lock); > > - rnp->exp_seq_poll_rq = 0x1; > > + rnp->exp_seq_poll_rq = RCU_EXP_SEQ_POLL_DONE; > > INIT_WORK(&rnp->exp_poll_wq, sync_rcu_do_polled_gp); > > } > > } > > diff --git a/kernel/rcu/tree_exp.h b/kernel/rcu/tree_exp.h > > index c4a19c6a83cf..7ccb909d6355 100644 > > --- a/kernel/rcu/tree_exp.h > > +++ b/kernel/rcu/tree_exp.h > > @@ -910,14 +910,14 @@ static void sync_rcu_do_polled_gp(struct work_struct *wp) > > unsigned long s; > > > > s = READ_ONCE(rnp->exp_seq_poll_rq); > > - if (s & 0x1) > > + if (s & RCU_EXP_SEQ_POLL_DONE) > > return; > > while (!sync_exp_work_done(s)) > > __synchronize_rcu_expedited(true); > > One additional question. If we re-read rnp->exp_seq_poll_rq on each pass > through the loop, wouldn't we have less trouble with counter wrap? We can indeed do that, though it won't eliminate the possibility of wrapping. > > Thanx, Paul > > > raw_spin_lock_irqsave(&rnp->exp_poll_lock, flags); > > s = rnp->exp_seq_poll_rq; > > - if (!(s & 0x1) && sync_exp_work_done(s)) > > - WRITE_ONCE(rnp->exp_seq_poll_rq, s | 0x1); > > + if (!(s & RCU_EXP_SEQ_POLL_DONE) && sync_exp_work_done(s)) > > + WRITE_ONCE(rnp->exp_seq_poll_rq, s | RCU_EXP_SEQ_POLL_DONE); > > raw_spin_unlock_irqrestore(&rnp->exp_poll_lock, flags); > > } > > > > @@ -946,7 +946,8 @@ unsigned long start_poll_synchronize_rcu_expedited(void) > > rnp = rdp->mynode; > > if (rcu_init_invoked()) > > raw_spin_lock_irqsave(&rnp->exp_poll_lock, flags); > > - if ((rnp->exp_seq_poll_rq & 0x1) || ULONG_CMP_LT(rnp->exp_seq_poll_rq, s)) { > > + if ((rnp->exp_seq_poll_rq & RCU_EXP_SEQ_POLL_DONE) || > > + ULONG_CMP_LT(rnp->exp_seq_poll_rq, s)) { > > WRITE_ONCE(rnp->exp_seq_poll_rq, s); > > if (rcu_init_invoked()) > > queue_work(rcu_gp_wq, &rnp->exp_poll_wq); > > -- > > 2.25.1 > >