From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753055AbbJMNVf (ORCPT ); Tue, 13 Oct 2015 09:21:35 -0400 Received: from foss.arm.com ([217.140.101.70]:52849 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751742AbbJMNVe (ORCPT ); Tue, 13 Oct 2015 09:21:34 -0400 Date: Tue, 13 Oct 2015 14:21:32 +0100 From: Will Deacon To: Boqun Feng Cc: linux-kernel@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, Peter Zijlstra , Ingo Molnar , Benjamin Herrenschmidt , Paul Mackerras , Michael Ellerman , Thomas Gleixner , "Paul E. McKenney" , Waiman Long , Davidlohr Bueso Subject: Re: [PATCH v3 4/6] powerpc: atomic: Implement atomic{,64}_*_return_* variants Message-ID: <20151013132132.GH21550@arm.com> References: <1444659246-24769-1-git-send-email-boqun.feng@gmail.com> <1444659246-24769-5-git-send-email-boqun.feng@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1444659246-24769-5-git-send-email-boqun.feng@gmail.com> User-Agent: Mutt/1.5.23 (2014-03-12) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Oct 12, 2015 at 10:14:04PM +0800, Boqun Feng wrote: > On powerpc, acquire and release semantics can be achieved with > lightweight barriers("lwsync" and "ctrl+isync"), which can be used to > implement __atomic_op_{acquire,release}. > > For release semantics, since we only need to ensure all memory accesses > that issue before must take effects before the -store- part of the > atomics, "lwsync" is what we only need. On the platform without > "lwsync", "sync" should be used. Therefore, smp_lwsync() is used here. > > For acquire semantics, "lwsync" is what we only need for the similar > reason. However on the platform without "lwsync", we can use "isync" > rather than "sync" as an acquire barrier. So a new kind of barrier > smp_acquire_barrier__after_atomic() is introduced, which is barrier() on > UP, "lwsync" if available and "isync" otherwise. > > __atomic_op_fence is defined as smp_lwsync() + _relaxed + > smp_mb__after_atomic() to guarantee a full barrier. > > Implement atomic{,64}_{add,sub,inc,dec}_return_relaxed, and build other > variants with these helpers. > > Signed-off-by: Boqun Feng > --- > arch/powerpc/include/asm/atomic.h | 122 +++++++++++++++++++++++++------------- > 1 file changed, 80 insertions(+), 42 deletions(-) > > diff --git a/arch/powerpc/include/asm/atomic.h b/arch/powerpc/include/asm/atomic.h > index 55f106e..3143af9 100644 > --- a/arch/powerpc/include/asm/atomic.h > +++ b/arch/powerpc/include/asm/atomic.h > @@ -12,6 +12,39 @@ > > #define ATOMIC_INIT(i) { (i) } > > +/* > + * Since {add,sub}_return_relaxed and xchg_relaxed are implemented with > + * a "bne-" instruction at the end, so an isync is enough as a acquire barrier > + * on the platform without lwsync. > + */ > +#ifdef CONFIG_SMP > +#define smp_acquire_barrier__after_atomic() \ > + __asm__ __volatile__(PPC_ACQUIRE_BARRIER : : : "memory") I'm not keen on this barrier, as it sounds like it's part of the kernel memory model, as opposed to an implementation detail on PowerPC (and we've already got enough of that in the generic code ;). Can you name it something different please (and maybe #undef it when you're done)? Will