From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pb0-f45.google.com (mail-pb0-f45.google.com [209.85.160.45]) by kanga.kvack.org (Postfix) with ESMTP id 1841B6B00D2 for ; Wed, 6 Nov 2013 07:22:06 -0500 (EST) Received: by mail-pb0-f45.google.com with SMTP id ma3so8885786pbc.4 for ; Wed, 06 Nov 2013 04:22:05 -0800 (PST) Received: from psmtp.com ([74.125.245.138]) by mx.google.com with SMTP id je1si16897559pbb.60.2013.11.06.04.22.03 for ; Wed, 06 Nov 2013 04:22:04 -0800 (PST) Date: Wed, 6 Nov 2013 12:20:20 +0000 From: Will Deacon Subject: Re: [PATCH v2 3/4] MCS Lock: Barrier corrections Message-ID: <20131106122019.GG21074@mudshark.cambridge.arm.com> References: <1383673356.11046.279.camel@schen9-DESK> <20131105183744.GJ26895@mudshark.cambridge.arm.com> <1383679317.11046.293.camel@schen9-DESK> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Sender: owner-linux-mm@kvack.org List-ID: To: "Figo.zhang" Cc: Tim Chen , Ingo Molnar , Andrew Morton , Thomas Gleixner , "linux-kernel@vger.kernel.org" , linux-mm , "linux-arch@vger.kernel.org" , Linus Torvalds , Waiman Long , Andrea Arcangeli , Alex Shi , Andi Kleen , Michel Lespinasse , Davidlohr Bueso , Matthew R Wilcox , Dave Hansen , Peter Zijlstra , Rik van Riel , Peter Hurley , "Paul E.McKenney" , Raghavendra K T , George Spelvin , "H. Peter Anvin" , Arnd Bergmann , Aswin Chandramouleeswaran , Scott J Norton On Wed, Nov 06, 2013 at 05:44:42AM +0000, Figo.zhang wrote: > 2013/11/6 Tim Chen > > On Tue, 2013-11-05 at 18:37 +0000, Will Deacon wrote: > > On Tue, Nov 05, 2013 at 05:42:36PM +0000, Tim Chen wrote: > > > diff --git a/include/linux/mcs_spinlock.h b/include/linux/mcs_spinlock.h > > > index 96f14299..93d445d 100644 > > > --- a/include/linux/mcs_spinlock.h > > > +++ b/include/linux/mcs_spinlock.h > > > @@ -36,16 +36,19 @@ void mcs_spin_lock(struct mcs_spinlock **lock, struct mcs_spinlock *node) > > > node->locked = 0; > > > node->next = NULL; > > > > > > + /* xchg() provides a memory barrier */ > > > prev = xchg(lock, node); > > > if (likely(prev == NULL)) { > > > /* Lock acquired */ > > > return; > > > } > > > ACCESS_ONCE(prev->next) = node; > > > - smp_wmb(); > > > /* Wait until the lock holder passes the lock down */ > > > while (!ACCESS_ONCE(node->locked)) > > > arch_mutex_cpu_relax(); > > > + > > > + /* Make sure subsequent operations happen after the lock is acquired */ > > > + smp_rmb(); > > > > Ok, so this is an smp_rmb() because we assume that stores aren't speculated, > > right? (i.e. the control dependency above is enough for stores to be ordered > > with respect to taking the lock)... > > > > > } > > > > > > /* > > > @@ -58,6 +61,7 @@ static void mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *nod > > > > > > if (likely(!next)) { > > > /* > > > + * cmpxchg() provides a memory barrier. > > > * Release the lock by setting it to NULL > > > */ > > > if (likely(cmpxchg(lock, node, NULL) == node)) > > > @@ -65,9 +69,14 @@ static void mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *nod > > > /* Wait until the next pointer is set */ > > > while (!(next = ACCESS_ONCE(node->next))) > > > arch_mutex_cpu_relax(); > > > + } else { > > > + /* > > > + * Make sure all operations within the critical section > > > + * happen before the lock is released. > > > + */ > > > + smp_wmb(); > > > > ...but I don't see what prevents reads inside the critical section from > > moving across the smp_wmb() here. > > This is to prevent any read in next critical section from > creeping up before write in the previous critical section > has completed Understood, but an smp_wmb() doesn't provide any ordering guarantees with respect to reads, hence why I think you need an smp_mb() here. > e.g. > CPU 1 execute > mcs_lock > x = 1; > ... > x = 2; > mcs_unlock > > and CPU 2 execute > > mcs_lock > y = x; > ... > mcs_unlock > > We expect y to be 2 after the "y = x" assignment. Without the proper > rmb in lock and wmb in unlock, y could be 1 for CPU 2 with > speculative read (i.e. before the x=2 assignment is completed). > > is it not a good example ? I think you need reads and writes by both CPUs to show the problem: // x, y are zero-initialised memory locations // a, b are registers CPU 1: mcs_lock a = x y = 1 mcs_unlock CPU 2: mcs_lock b = y x = 1 mcs_unlock In this case, you would hope that you can't observe a = b = 1. However, given the current barriers, I think you could end up with something equivalent to: CPU 1: y = 1 // Moved over read-barrier mcs_lock // smp_rmb mcs_unlock // smp_wmb a = x // Moved over write-barrier CPU 2: x = 1 // Moved over read-barrier mcs_lock // smp_rmb mcs_unlock // smp_wmb b = y // Moved over write-barrier which would permit a = b = 1, as well as other orderings. Will -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org