linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Waiman Long <waiman.long@hp.com>
To: Will Deacon <will.deacon@arm.com>
Cc: "Figo.zhang" <figo1802@gmail.com>,
	Tim Chen <tim.c.chen@linux.intel.com>,
	Ingo Molnar <mingo@elte.hu>,
	Andrew Morton <akpm@linux-foundation.org>,
	Thomas Gleixner <tglx@linutronix.de>,
	"linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
	linux-mm <linux-mm@kvack.org>,
	"linux-arch@vger.kernel.org" <linux-arch@vger.kernel.org>,
	Linus Torvalds <torvalds@linux-foundation.org>,
	Andrea Arcangeli <aarcange@redhat.com>,
	Alex Shi <alex.shi@linaro.org>, Andi Kleen <andi@firstfloor.org>,
	Michel Lespinasse <walken@google.com>,
	Davidlohr Bueso <davidlohr.bueso@hp.com>,
	Matthew R Wilcox <matthew.r.wilcox@intel.com>,
	Dave Hansen <dave.hansen@intel.com>,
	Peter Zijlstra <a.p.zijlstra@chello.nl>,
	Rik van Riel <riel@redhat.com>,
	Peter Hurley <peter@hurleysoftware.com>,
	"Paul E.McKenney" <paulmck@linux.vnet.ibm.com>,
	Raghavendra K T <raghavendra.kt@linux.vnet.ibm.com>,
	George Spelvin <linux@horizon.com>,
	"H. Peter Anvin" <hpa@zytor.com>, Arnd Bergmann <arnd@arndb.de>,
	Aswin Chandramouleeswaran <aswin@hp.com>,
	Scott J Norton <scott.norton@hp.com>
Subject: Re: [PATCH v2 3/4] MCS Lock: Barrier corrections
Date: Wed, 06 Nov 2013 12:05:13 -0500	[thread overview]
Message-ID: <527A76C9.1030208@hp.com> (raw)
In-Reply-To: <20131106122019.GG21074@mudshark.cambridge.arm.com>

On 11/06/2013 07:20 AM, Will Deacon wrote:
> On Wed, Nov 06, 2013 at 05:44:42AM +0000, Figo.zhang wrote:
>> 2013/11/6 Tim Chen<tim.c.chen@linux.intel.com<mailto:tim.c.chen@linux.intel.com>>
>> On Tue, 2013-11-05 at 18:37 +0000, Will Deacon wrote:
>>> On Tue, Nov 05, 2013 at 05:42:36PM +0000, Tim Chen wrote:
>>>> diff --git a/include/linux/mcs_spinlock.h b/include/linux/mcs_spinlock.h
>>>> index 96f14299..93d445d 100644
>>>> --- a/include/linux/mcs_spinlock.h
>>>> +++ b/include/linux/mcs_spinlock.h
>>>> @@ -36,16 +36,19 @@ void mcs_spin_lock(struct mcs_spinlock **lock, struct mcs_spinlock *node)
>>>>      node->locked = 0;
>>>>      node->next   = NULL;
>>>>
>>>> +   /* xchg() provides a memory barrier */
>>>>      prev = xchg(lock, node);
>>>>      if (likely(prev == NULL)) {
>>>>              /* Lock acquired */
>>>>              return;
>>>>      }
>>>>      ACCESS_ONCE(prev->next) = node;
>>>> -   smp_wmb();
>>>>      /* Wait until the lock holder passes the lock down */
>>>>      while (!ACCESS_ONCE(node->locked))
>>>>              arch_mutex_cpu_relax();
>>>> +
>>>> +   /* Make sure subsequent operations happen after the lock is acquired */
>>>> +   smp_rmb();
>>> Ok, so this is an smp_rmb() because we assume that stores aren't speculated,
>>> right? (i.e. the control dependency above is enough for stores to be ordered
>>> with respect to taking the lock)...
>>>
>>>>   }
>>>>
>>>>   /*
>>>> @@ -58,6 +61,7 @@ static void mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *nod
>>>>
>>>>      if (likely(!next)) {
>>>>              /*
>>>> +            * cmpxchg() provides a memory barrier.
>>>>               * Release the lock by setting it to NULL
>>>>               */
>>>>              if (likely(cmpxchg(lock, node, NULL) == node))
>>>> @@ -65,9 +69,14 @@ static void mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *nod
>>>>              /* Wait until the next pointer is set */
>>>>              while (!(next = ACCESS_ONCE(node->next)))
>>>>                      arch_mutex_cpu_relax();
>>>> +   } else {
>>>> +           /*
>>>> +            * Make sure all operations within the critical section
>>>> +            * happen before the lock is released.
>>>> +            */
>>>> +           smp_wmb();
>>> ...but I don't see what prevents reads inside the critical section from
>>> moving across the smp_wmb() here.
>> This is to prevent any read in next critical section from
>> creeping up before write in the previous critical section
>> has completed
> Understood, but an smp_wmb() doesn't provide any ordering guarantees with
> respect to reads, hence why I think you need an smp_mb() here.

A major reason for the current design is to avoid overhead of a full 
memory barrier in x86 which doesn't need that. I do agree that the 
current code may not be enough for other architectures. I would like to 
propose that the following changes:

1) Move the lock/unlock functions to mcs_spinlock.c.
2) Define a set of primitives - smp_mb__before_critical_section(), 
smp_mb_after_critical_section() that will fall back to smp_mb() if they 
are not defined in asm/processor.h, for example.
3) Use the new primitives instead of the current smp_rmb() and smp_wmb() 
memory barrier.

That will allow each architecture to tailor what sort of memory barrier 
do they want to use.

Regards,
Longman

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  reply	other threads:[~2013-11-06 17:05 UTC|newest]

Thread overview: 22+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <cover.1383670202.git.tim.c.chen@linux.intel.com>
2013-11-05 17:42 ` [PATCH v2 0/4] MCS Lock: MCS lock code cleanup and optimizations Tim Chen
2013-11-05 21:14   ` Michel Lespinasse
2013-11-05 21:27     ` Tim Chen
2013-11-05 17:42 ` [PATCH v2 1/4] MCS Lock: Restructure the MCS lock defines and locking code into its own file Tim Chen
2013-11-05 17:42 ` [PATCH v2 2/4] MCS Lock: optimizations and extra comments Tim Chen
2013-11-05 17:42 ` [PATCH v2 3/4] MCS Lock: Barrier corrections Tim Chen
2013-11-05 18:37   ` Will Deacon
2013-11-05 19:21     ` Tim Chen
2013-11-05 21:18       ` Peter Zijlstra
2013-11-06  1:25         ` Tim Chen
2013-11-06 11:30           ` Will Deacon
2013-11-06 14:45         ` Paul E. McKenney
2013-11-06 18:22           ` Tim Chen
2013-11-06 19:13             ` Waiman Long
2013-11-06  5:44       ` Figo.zhang
2013-11-06 12:20         ` Will Deacon
2013-11-06 17:05           ` Waiman Long [this message]
2013-11-05 17:42 ` [PATCH v2 4/4] MCS Lock: Make mcs_spinlock.h includable in other files Tim Chen
2013-11-05 18:57   ` Peter Zijlstra
2013-11-05 19:30     ` Tim Chen
2013-11-06 15:31       ` Waiman Long
2013-11-06 16:08         ` Peter Zijlstra

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=527A76C9.1030208@hp.com \
    --to=waiman.long@hp.com \
    --cc=a.p.zijlstra@chello.nl \
    --cc=aarcange@redhat.com \
    --cc=akpm@linux-foundation.org \
    --cc=alex.shi@linaro.org \
    --cc=andi@firstfloor.org \
    --cc=arnd@arndb.de \
    --cc=aswin@hp.com \
    --cc=dave.hansen@intel.com \
    --cc=davidlohr.bueso@hp.com \
    --cc=figo1802@gmail.com \
    --cc=hpa@zytor.com \
    --cc=linux-arch@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux@horizon.com \
    --cc=matthew.r.wilcox@intel.com \
    --cc=mingo@elte.hu \
    --cc=paulmck@linux.vnet.ibm.com \
    --cc=peter@hurleysoftware.com \
    --cc=raghavendra.kt@linux.vnet.ibm.com \
    --cc=riel@redhat.com \
    --cc=scott.norton@hp.com \
    --cc=tglx@linutronix.de \
    --cc=tim.c.chen@linux.intel.com \
    --cc=torvalds@linux-foundation.org \
    --cc=walken@google.com \
    --cc=will.deacon@arm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).