From: Peter Zijlstra <peterz@infradead.org>
To: Manfred Spraul <manfred@colorfullife.com>
Cc: benh@kernel.crashing.org, paulmck@linux.vnet.ibm.com,
Ingo Molnar <mingo@elte.hu>, Boqun Feng <boqun.feng@gmail.com>,
Andrew Morton <akpm@linux-foundation.org>,
LKML <linux-kernel@vger.kernel.org>,
1vier1@web.de, Davidlohr Bueso <dave@stgolabs.net>,
Will Deacon <will.deacon@arm.com>
Subject: Re: [PATCH 1/4] spinlock: Document memory barrier rules
Date: Wed, 31 Aug 2016 17:40:49 +0200 [thread overview]
Message-ID: <20160831154049.GY10121@twins.programming.kicks-ass.net> (raw)
In-Reply-To: <4859166f-ff39-e998-638b-6bf6912422a3@colorfullife.com>
On Wed, Aug 31, 2016 at 06:59:07AM +0200, Manfred Spraul wrote:
> The barrier must ensure that taking the spinlock (as observed by another cpu
> with spin_unlock_wait()) and a following read are ordered.
>
> start condition: sma->complex_mode = false;
>
> CPU 1:
> spin_lock(&sem->lock); /* sem_nsems instances */
> smp_mb__after_spin_lock();
> if (!smp_load_acquire(&sma->complex_mode)) {
> /* fast path successful! */
> return sops->sem_num;
> }
> /* slow path, not relevant */
>
> CPU 2: (holding sma->sem_perm.lock)
>
> smp_store_mb(sma->complex_mode, true);
>
> for (i = 0; i < sma->sem_nsems; i++) {
> spin_unlock_wait(&sma->sem_base[i].lock);
> }
>
> It must not happen that both CPUs proceed:
> Either CPU1 proceeds, then CPU2 must spin in spin_unlock_wait()
> or CPU2 proceeds, then CPU1 must enter the slow path.
>
> What about this?
> /*
> * spin_lock() provides ACQUIRE semantics regarding reading the lock.
> * There are no guarantees that the store of the lock is visible before
> * any read or write operation within the protected area is performed.
> * If the store of the lock must happen first, this function is required.
> */
> #define spin_lock_store_acquire()
So I think the fundamental problem is with our atomic_*_acquire()
primitives, where we've specified that the ACQUIRE only pertains to the
LOAD of the RmW.
The spinlock implementations suffer this problem mostly because of
that (not 100% accurate but close enough).
One solution would be to simply use smp_mb__after_atomic(). The
'problem' with that is __atomic_op_acquire() defaults to using that, so
the archs that use __atomic_op_acquire() will get a double smp_mb()
(arm64 and powerpc do not use __atomic_op_acquire()).
I'm not sure we want to introduce a new primitive for this specific to
spinlocks.
Will, any opinions?
next prev parent reply other threads:[~2016-08-31 15:41 UTC|newest]
Thread overview: 25+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-08-28 11:56 [PATCH 0/4] Clarify/standardize memory barriers for lock/unlock Manfred Spraul
2016-08-28 11:56 ` [PATCH 1/4] spinlock: Document memory barrier rules Manfred Spraul
2016-08-28 11:56 ` [PATCH 2/4] barrier.h: Move smp_mb__after_unlock_lock to barrier.h Manfred Spraul
2016-08-28 11:56 ` [PATCH 3/4] net/netfilter/nf_conntrack_core: update memory barriers Manfred Spraul
2016-08-28 11:56 ` [PATCH 4/4] qspinlock for x86: smp_mb__after_spin_lock() is free Manfred Spraul
2016-08-29 10:52 ` Peter Zijlstra
2016-08-29 10:51 ` [PATCH 3/4] net/netfilter/nf_conntrack_core: update memory barriers Peter Zijlstra
2016-08-28 13:43 ` [PATCH 2/4] barrier.h: Move smp_mb__after_unlock_lock to barrier.h Paul E. McKenney
2016-08-28 16:31 ` Manfred Spraul
2016-08-28 18:00 ` Manfred Spraul
2016-08-28 14:41 ` kbuild test robot
2016-08-28 17:43 ` [PATCH 2/4 v3] spinlock.h: Move smp_mb__after_unlock_lock to spinlock.h Manfred Spraul
2016-08-29 10:48 ` [PATCH 1/4] spinlock: Document memory barrier rules Peter Zijlstra
2016-08-29 12:54 ` Manfred Spraul
2016-08-29 13:44 ` Peter Zijlstra
2016-08-31 4:59 ` Manfred Spraul
2016-08-31 15:40 ` Peter Zijlstra [this message]
2016-08-31 16:40 ` Will Deacon
2016-08-31 18:32 ` Manfred Spraul
2016-09-01 8:44 ` Peter Zijlstra
2016-09-01 11:04 ` Manfred Spraul
2016-09-01 11:19 ` Will Deacon
2016-09-01 11:51 ` Peter Zijlstra
2016-09-01 14:05 ` Boqun Feng
2016-08-29 10:53 ` [PATCH 0/4] Clarify/standardize memory barriers for lock/unlock Peter Zijlstra
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20160831154049.GY10121@twins.programming.kicks-ass.net \
--to=peterz@infradead.org \
--cc=1vier1@web.de \
--cc=akpm@linux-foundation.org \
--cc=benh@kernel.crashing.org \
--cc=boqun.feng@gmail.com \
--cc=dave@stgolabs.net \
--cc=linux-kernel@vger.kernel.org \
--cc=manfred@colorfullife.com \
--cc=mingo@elte.hu \
--cc=paulmck@linux.vnet.ibm.com \
--cc=will.deacon@arm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox