From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 5A65CC433F5 for ; Sun, 2 Oct 2022 21:47:25 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229595AbiJBVrW (ORCPT ); Sun, 2 Oct 2022 17:47:22 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47332 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229449AbiJBVrR (ORCPT ); Sun, 2 Oct 2022 17:47:17 -0400 Received: from ams.source.kernel.org (ams.source.kernel.org [145.40.68.75]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 304D32229F; Sun, 2 Oct 2022 14:47:16 -0700 (PDT) Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ams.source.kernel.org (Postfix) with ESMTPS id D366CB80DD3; Sun, 2 Oct 2022 21:47:14 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 0A1B7C433C1; Sun, 2 Oct 2022 21:47:12 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1664747233; bh=ECXgpkNzKbEgqV2Aoq3cqkZc/QKqAF6GEAlq+k8sqQo=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=oGjQqxyhA/DTPJUQI0xkolxX4juf3nzxvfnS1d6ymOdOKWI/7k8WTmObYg2MJOQcg rd7rhx97BfEK6VUk2ntfuRmEa/VKhxKjj4LvIG6k0/Vi/BCM53G823y3kAFP1/Ig8d kwvZN61utHiQAw1q4Ds6Z4QL6mH+GJ8WMgb+4UvkyojVLRUG7lDlogw9gi3Gd8hrYK TxmfDEcYETxhk4PndcxX8WPmVU1fZR/n8iultuqMvnAjx6qPLw1Pju65cD0aXbph3D A6CFFEiJrO7/gvz7KTb7nXw6aOPMSCegBm8kiW0ts6M9ZUMd/WA39Yz8rLgOrZ9c/Q 7zfNycX70Uhtg== Date: Sun, 2 Oct 2022 23:47:10 +0200 From: Frederic Weisbecker To: "Paul E. McKenney" Cc: rcu@vger.kernel.org, linux-kernel@vger.kernel.org, kernel-team@fb.com, rostedt@goodmis.org, Randy Dunlap , Thomas Gleixner , John Ogness , Petr Mladek Subject: Re: [PATCH RFC v2 rcu 2/8] srcu: Create an srcu_read_lock_nmisafe() and srcu_read_unlock_nmisafe() Message-ID: <20221002214710.GA297965@lothringen> References: <20220929180714.GA2874192@paulmck-ThinkPad-P17-Gen-1> <20220929180731.2875722-2-paulmck@kernel.org> <20221002155516.GB292620@lothringen> <20221002160957.GP4196@paulmck-ThinkPad-P17-Gen-1> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20221002160957.GP4196@paulmck-ThinkPad-P17-Gen-1> Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Sun, Oct 02, 2022 at 09:09:57AM -0700, Paul E. McKenney wrote: > On Sun, Oct 02, 2022 at 05:55:16PM +0200, Frederic Weisbecker wrote: > > On Thu, Sep 29, 2022 at 11:07:25AM -0700, Paul E. McKenney wrote: > > > @@ -1090,7 +1121,7 @@ static unsigned long srcu_gp_start_if_needed(struct srcu_struct *ssp, > > > int ss_state; > > > > > > check_init_srcu_struct(ssp); > > > - idx = srcu_read_lock(ssp); > > > + idx = __srcu_read_lock_nmisafe(ssp); > > > > Why do we need to force the atomic based version here (even if CONFIG_NEED_SRCU_NMI_SAFE=y)? > > In kernels built with CONFIG_NEED_SRCU_NMI_SAFE=n, we of course need it. > As you say, in kernels built with CONFIG_NEED_SRCU_NMI_SAFE=y, we don't. > But it doesn't hurt to always use __srcu_read_lock_nmisafe() here, and > this is nowhere near a fastpath, so there is little benefit to using > __srcu_read_lock() when it is safe to do so. > > In addition, note that it is possible that a given srcu_struct structure's > first grace period is executed before its first reader. In that > case, we have no way of knowing which of __srcu_read_lock_nmisafe() > or __srcu_read_lock() to choose. > > So this code always does it the slow(ish) safe way. But then srcu_read_lock_nmisafe() would work as well, right? > > > > ss_state = smp_load_acquire(&ssp->srcu_size_state); > > > if (ss_state < SRCU_SIZE_WAIT_CALL) > > > sdp = per_cpu_ptr(ssp->sda, 0); > > > @@ -1123,7 +1154,7 @@ static unsigned long srcu_gp_start_if_needed(struct srcu_struct *ssp, > > > srcu_funnel_gp_start(ssp, sdp, s, do_norm); > > > else if (needexp) > > > srcu_funnel_exp_start(ssp, sdp_mynode, s); > > > - srcu_read_unlock(ssp, idx); > > > + __srcu_read_unlock_nmisafe(ssp, idx); > > > return s; > > > } > > > > > > @@ -1427,13 +1458,13 @@ void srcu_barrier(struct srcu_struct *ssp) > > > /* Initial count prevents reaching zero until all CBs are posted. */ > > > atomic_set(&ssp->srcu_barrier_cpu_cnt, 1); > > > > > > - idx = srcu_read_lock(ssp); > > > + idx = __srcu_read_lock_nmisafe(ssp); > > > > And same here? > > Yes, same here. ;-) Now bonus question: why do SRCU grace period starting/tracking need to be in an SRCU read side critical section? :o) Thanks.