From: Peter Zijlstra <peterz@infradead.org> To: Waiman.Long@hp.com Cc: raghavendra.kt@linux.vnet.ibm.com, kvm@vger.kernel.org, peterz@infradead.org, linux-kernel@vger.kernel.org, hpa@zytor.com, boris.ostrovsky@oracle.com, linux-arch@vger.kernel.org, x86@kernel.org, mingo@redhat.com, doug.hatch@hp.com, xen-devel@lists.xenproject.org, paulmck@linux.vnet.ibm.com, riel@redhat.com, scott.norton@hp.com, paolo.bonzini@gmail.com, tglx@linutronix.de, virtualization@lists.linux-foundation.org, oleg@redhat.com, luto@amacapital.net, david.vrabel@citrix.com, torvalds@linux-foundation.org Subject: [PATCH 4/9] qspinlock: Extract out code snippets for the next patch Date: Mon, 16 Mar 2015 14:16:17 +0100 [thread overview] Message-ID: <20150316133112.027716523@infradead.org> (raw) In-Reply-To: 20150316131613.720617163@infradead.org [-- Attachment #1: waiman_long-qspinlock-extract_out_code_snippets_for_the_next_patch.patch --] [-- Type: text/plain, Size: 4917 bytes --] From: Waiman Long <Waiman.Long@hp.com> This is a preparatory patch that extracts out the following 2 code snippets to prepare for the next performance optimization patch. 1) the logic for the exchange of new and previous tail code words into a new xchg_tail() function. 2) the logic for clearing the pending bit and setting the locked bit into a new clear_pending_set_locked() function. This patch also simplifies the trylock operation before queuing by calling queue_spin_trylock() directly. Cc: Ingo Molnar <mingo@redhat.com> Cc: David Vrabel <david.vrabel@citrix.com> Cc: Oleg Nesterov <oleg@redhat.com> Cc: Scott J Norton <scott.norton@hp.com> Cc: Paolo Bonzini <paolo.bonzini@gmail.com> Cc: Douglas Hatch <doug.hatch@hp.com> Cc: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com> Cc: Boris Ostrovsky <boris.ostrovsky@oracle.com> Cc: "Paul E. McKenney" <paulmck@linux.vnet.ibm.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Rik van Riel <riel@redhat.com> Cc: Raghavendra K T <raghavendra.kt@linux.vnet.ibm.com> Signed-off-by: Waiman Long <Waiman.Long@hp.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: http://lkml.kernel.org/r/1421784755-21945-5-git-send-email-Waiman.Long@hp.com --- include/asm-generic/qspinlock_types.h | 2 kernel/locking/qspinlock.c | 91 ++++++++++++++++++++++------------ 2 files changed, 62 insertions(+), 31 deletions(-) --- a/include/asm-generic/qspinlock_types.h +++ b/include/asm-generic/qspinlock_types.h @@ -58,6 +58,8 @@ typedef struct qspinlock { #define _Q_TAIL_CPU_BITS (32 - _Q_TAIL_CPU_OFFSET) #define _Q_TAIL_CPU_MASK _Q_SET_MASK(TAIL_CPU) +#define _Q_TAIL_MASK (_Q_TAIL_IDX_MASK | _Q_TAIL_CPU_MASK) + #define _Q_LOCKED_VAL (1U << _Q_LOCKED_OFFSET) #define _Q_PENDING_VAL (1U << _Q_PENDING_OFFSET) --- a/kernel/locking/qspinlock.c +++ b/kernel/locking/qspinlock.c @@ -97,6 +97,54 @@ static inline struct mcs_spinlock *decod #define _Q_LOCKED_PENDING_MASK (_Q_LOCKED_MASK | _Q_PENDING_MASK) /** + * clear_pending_set_locked - take ownership and clear the pending bit. + * @lock: Pointer to queue spinlock structure + * @val : Current value of the queue spinlock 32-bit word + * + * *,1,0 -> *,0,1 + */ +static __always_inline void +clear_pending_set_locked(struct qspinlock *lock, u32 val) +{ + u32 new, old; + + for (;;) { + new = (val & ~_Q_PENDING_MASK) | _Q_LOCKED_VAL; + + old = atomic_cmpxchg(&lock->val, val, new); + if (old == val) + break; + + val = old; + } +} + +/** + * xchg_tail - Put in the new queue tail code word & retrieve previous one + * @lock : Pointer to queue spinlock structure + * @tail : The new queue tail code word + * Return: The previous queue tail code word + * + * xchg(lock, tail) + * + * p,*,* -> n,*,* ; prev = xchg(lock, node) + */ +static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) +{ + u32 old, new, val = atomic_read(&lock->val); + + for (;;) { + new = (val & _Q_LOCKED_PENDING_MASK) | tail; + old = atomic_cmpxchg(&lock->val, val, new); + if (old == val) + break; + + val = old; + } + return old; +} + +/** * queue_spin_lock_slowpath - acquire the queue spinlock * @lock: Pointer to queue spinlock structure * @val: Current value of the queue spinlock 32-bit word @@ -178,15 +226,7 @@ void queue_spin_lock_slowpath(struct qsp * * *,1,0 -> *,0,1 */ - for (;;) { - new = (val & ~_Q_PENDING_MASK) | _Q_LOCKED_VAL; - - old = atomic_cmpxchg(&lock->val, val, new); - if (old == val) - break; - - val = old; - } + clear_pending_set_locked(lock, val); return; /* @@ -203,37 +243,26 @@ void queue_spin_lock_slowpath(struct qsp node->next = NULL; /* - * We have already touched the queueing cacheline; don't bother with - * pending stuff. - * - * trylock || xchg(lock, node) - * - * 0,0,0 -> 0,0,1 ; no tail, not locked -> no tail, locked. - * p,y,x -> n,y,x ; tail was p -> tail is n; preserving locked. + * We touched a (possibly) cold cacheline in the per-cpu queue node; + * attempt the trylock once more in the hope someone let go while we + * weren't watching. */ - for (;;) { - new = _Q_LOCKED_VAL; - if (val) - new = tail | (val & _Q_LOCKED_PENDING_MASK); - - old = atomic_cmpxchg(&lock->val, val, new); - if (old == val) - break; - - val = old; - } + if (queue_spin_trylock(lock)) + goto release; /* - * we won the trylock; forget about queueing. + * We have already touched the queueing cacheline; don't bother with + * pending stuff. + * + * p,*,* -> n,*,* */ - if (new == _Q_LOCKED_VAL) - goto release; + old = xchg_tail(lock, tail); /* * if there was a previous node; link it and wait until reaching the * head of the waitqueue. */ - if (old & ~_Q_LOCKED_PENDING_MASK) { + if (old & _Q_TAIL_MASK) { prev = decode_tail(old); WRITE_ONCE(prev->next, node);
WARNING: multiple messages have this Message-ID (diff)
From: Peter Zijlstra <peterz@infradead.org> To: Waiman.Long@hp.com Cc: tglx@linutronix.de, mingo@redhat.com, hpa@zytor.com, peterz@infradead.org, paolo.bonzini@gmail.com, konrad.wilk@oracle.com, boris.ostrovsky@oracle.com, paulmck@linux.vnet.ibm.com, riel@redhat.com, torvalds@linux-foundation.org, raghavendra.kt@linux.vnet.ibm.com, david.vrabel@citrix.com, oleg@redhat.com, scott.norton@hp.com, doug.hatch@hp.com, linux-arch@vger.kernel.org, x86@kernel.org, linux-kernel@vger.kernel.org, virtualization@lists.linux-foundation.org, xen-devel@lists.xenproject.org, kvm@vger.kernel.org, luto@amacapital.net Subject: [PATCH 4/9] qspinlock: Extract out code snippets for the next patch Date: Mon, 16 Mar 2015 14:16:17 +0100 [thread overview] Message-ID: <20150316133112.027716523@infradead.org> (raw) Message-ID: <20150316131617.8YFWs2QIUflboNUyfUNG51u0WEOKpmn_q4aTEasfPew@z> (raw) In-Reply-To: 20150316131613.720617163@infradead.org [-- Attachment #1: waiman_long-qspinlock-extract_out_code_snippets_for_the_next_patch.patch --] [-- Type: text/plain, Size: 4921 bytes --] From: Waiman Long <Waiman.Long@hp.com> This is a preparatory patch that extracts out the following 2 code snippets to prepare for the next performance optimization patch. 1) the logic for the exchange of new and previous tail code words into a new xchg_tail() function. 2) the logic for clearing the pending bit and setting the locked bit into a new clear_pending_set_locked() function. This patch also simplifies the trylock operation before queuing by calling queue_spin_trylock() directly. Cc: Ingo Molnar <mingo@redhat.com> Cc: David Vrabel <david.vrabel@citrix.com> Cc: Oleg Nesterov <oleg@redhat.com> Cc: Scott J Norton <scott.norton@hp.com> Cc: Paolo Bonzini <paolo.bonzini@gmail.com> Cc: Douglas Hatch <doug.hatch@hp.com> Cc: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com> Cc: Boris Ostrovsky <boris.ostrovsky@oracle.com> Cc: "Paul E. McKenney" <paulmck@linux.vnet.ibm.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Rik van Riel <riel@redhat.com> Cc: Raghavendra K T <raghavendra.kt@linux.vnet.ibm.com> Signed-off-by: Waiman Long <Waiman.Long@hp.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Link: http://lkml.kernel.org/r/1421784755-21945-5-git-send-email-Waiman.Long@hp.com --- include/asm-generic/qspinlock_types.h | 2 kernel/locking/qspinlock.c | 91 ++++++++++++++++++++++------------ 2 files changed, 62 insertions(+), 31 deletions(-) --- a/include/asm-generic/qspinlock_types.h +++ b/include/asm-generic/qspinlock_types.h @@ -58,6 +58,8 @@ typedef struct qspinlock { #define _Q_TAIL_CPU_BITS (32 - _Q_TAIL_CPU_OFFSET) #define _Q_TAIL_CPU_MASK _Q_SET_MASK(TAIL_CPU) +#define _Q_TAIL_MASK (_Q_TAIL_IDX_MASK | _Q_TAIL_CPU_MASK) + #define _Q_LOCKED_VAL (1U << _Q_LOCKED_OFFSET) #define _Q_PENDING_VAL (1U << _Q_PENDING_OFFSET) --- a/kernel/locking/qspinlock.c +++ b/kernel/locking/qspinlock.c @@ -97,6 +97,54 @@ static inline struct mcs_spinlock *decod #define _Q_LOCKED_PENDING_MASK (_Q_LOCKED_MASK | _Q_PENDING_MASK) /** + * clear_pending_set_locked - take ownership and clear the pending bit. + * @lock: Pointer to queue spinlock structure + * @val : Current value of the queue spinlock 32-bit word + * + * *,1,0 -> *,0,1 + */ +static __always_inline void +clear_pending_set_locked(struct qspinlock *lock, u32 val) +{ + u32 new, old; + + for (;;) { + new = (val & ~_Q_PENDING_MASK) | _Q_LOCKED_VAL; + + old = atomic_cmpxchg(&lock->val, val, new); + if (old == val) + break; + + val = old; + } +} + +/** + * xchg_tail - Put in the new queue tail code word & retrieve previous one + * @lock : Pointer to queue spinlock structure + * @tail : The new queue tail code word + * Return: The previous queue tail code word + * + * xchg(lock, tail) + * + * p,*,* -> n,*,* ; prev = xchg(lock, node) + */ +static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) +{ + u32 old, new, val = atomic_read(&lock->val); + + for (;;) { + new = (val & _Q_LOCKED_PENDING_MASK) | tail; + old = atomic_cmpxchg(&lock->val, val, new); + if (old == val) + break; + + val = old; + } + return old; +} + +/** * queue_spin_lock_slowpath - acquire the queue spinlock * @lock: Pointer to queue spinlock structure * @val: Current value of the queue spinlock 32-bit word @@ -178,15 +226,7 @@ void queue_spin_lock_slowpath(struct qsp * * *,1,0 -> *,0,1 */ - for (;;) { - new = (val & ~_Q_PENDING_MASK) | _Q_LOCKED_VAL; - - old = atomic_cmpxchg(&lock->val, val, new); - if (old == val) - break; - - val = old; - } + clear_pending_set_locked(lock, val); return; /* @@ -203,37 +243,26 @@ void queue_spin_lock_slowpath(struct qsp node->next = NULL; /* - * We have already touched the queueing cacheline; don't bother with - * pending stuff. - * - * trylock || xchg(lock, node) - * - * 0,0,0 -> 0,0,1 ; no tail, not locked -> no tail, locked. - * p,y,x -> n,y,x ; tail was p -> tail is n; preserving locked. + * We touched a (possibly) cold cacheline in the per-cpu queue node; + * attempt the trylock once more in the hope someone let go while we + * weren't watching. */ - for (;;) { - new = _Q_LOCKED_VAL; - if (val) - new = tail | (val & _Q_LOCKED_PENDING_MASK); - - old = atomic_cmpxchg(&lock->val, val, new); - if (old == val) - break; - - val = old; - } + if (queue_spin_trylock(lock)) + goto release; /* - * we won the trylock; forget about queueing. + * We have already touched the queueing cacheline; don't bother with + * pending stuff. + * + * p,*,* -> n,*,* */ - if (new == _Q_LOCKED_VAL) - goto release; + old = xchg_tail(lock, tail); /* * if there was a previous node; link it and wait until reaching the * head of the waitqueue. */ - if (old & ~_Q_LOCKED_PENDING_MASK) { + if (old & _Q_TAIL_MASK) { prev = decode_tail(old); WRITE_ONCE(prev->next, node);
next prev parent reply other threads:[~2015-03-16 13:16 UTC|newest] Thread overview: 78+ messages / expand[flat|nested] mbox.gz Atom feed top 2015-03-16 13:16 [PATCH 0/9] qspinlock stuff -v15 Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-16 13:16 ` [PATCH 1/9] qspinlock: A simple generic 4-byte queue spinlock Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-16 13:16 ` [PATCH 2/9] qspinlock, x86: Enable x86-64 to use " Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-16 13:16 ` [PATCH 3/9] qspinlock: Add pending bit Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra [this message] 2015-03-16 13:16 ` [PATCH 4/9] qspinlock: Extract out code snippets for the next patch Peter Zijlstra 2015-03-16 13:16 ` [PATCH 5/9] qspinlock: Optimize for smaller NR_CPUS Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-16 13:16 ` [PATCH 6/9] qspinlock: Use a simple write to grab the lock Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-16 13:16 ` [PATCH 7/9] qspinlock: Revert to test-and-set on hypervisors Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-16 13:16 ` [PATCH 8/9] qspinlock: Generic paravirt support Peter Zijlstra 2015-03-16 13:16 ` Peter Zijlstra 2015-03-18 20:50 ` Waiman Long 2015-03-19 10:12 ` Peter Zijlstra 2015-03-19 12:25 ` Peter Zijlstra 2015-03-19 12:25 ` Peter Zijlstra 2015-03-19 13:43 ` Peter Zijlstra 2015-03-19 23:25 ` Waiman Long 2015-04-01 16:20 ` Waiman Long 2015-04-01 16:20 ` Waiman Long 2015-04-01 17:12 ` Peter Zijlstra 2015-04-01 17:12 ` Peter Zijlstra 2015-04-01 17:42 ` Peter Zijlstra 2015-04-01 17:42 ` Peter Zijlstra 2015-04-01 18:17 ` Peter Zijlstra 2015-04-01 18:17 ` Peter Zijlstra 2015-04-01 18:54 ` Waiman Long 2015-04-01 18:48 ` Peter Zijlstra 2015-04-01 19:58 ` Waiman Long 2015-04-01 19:58 ` Waiman Long 2015-04-01 21:03 ` Peter Zijlstra 2015-04-01 21:03 ` Peter Zijlstra 2015-04-02 16:28 ` Waiman Long 2015-04-02 17:20 ` Peter Zijlstra 2015-04-02 17:20 ` Peter Zijlstra 2015-04-02 19:48 ` Peter Zijlstra 2015-04-03 3:39 ` Waiman Long 2015-04-03 3:39 ` Waiman Long 2015-04-03 13:43 ` Peter Zijlstra 2015-04-03 13:43 ` Peter Zijlstra 2015-04-01 18:54 ` Waiman Long 2015-04-01 20:10 ` Waiman Long 2015-04-01 20:10 ` Waiman Long 2015-03-16 13:16 ` [PATCH 9/9] qspinlock, x86, kvm: Implement KVM support for paravirt qspinlock Peter Zijlstra 2015-03-16 13:16 ` [PATCH 9/9] qspinlock,x86,kvm: " Peter Zijlstra 2015-03-19 2:45 ` [PATCH 9/9] qspinlock, x86, kvm: " Waiman Long 2015-03-19 10:01 ` [PATCH 9/9] qspinlock,x86,kvm: " Peter Zijlstra 2015-03-19 10:01 ` Peter Zijlstra 2015-03-19 21:08 ` [PATCH 9/9] qspinlock, x86, kvm: " Waiman Long 2015-03-19 21:08 ` [PATCH 9/9] qspinlock,x86,kvm: " Waiman Long 2015-03-20 7:43 ` [PATCH 9/9] qspinlock, x86, kvm: " Raghavendra K T 2015-03-20 7:43 ` [PATCH 9/9] qspinlock,x86,kvm: " Raghavendra K T 2015-03-16 14:08 ` [Xen-devel] [PATCH 0/9] qspinlock stuff -v15 David Vrabel 2015-03-16 14:08 ` David Vrabel 2015-03-18 20:36 ` Waiman Long 2015-03-19 18:01 ` [Xen-devel] " David Vrabel 2015-03-19 18:01 ` David Vrabel 2015-03-19 18:32 ` Peter Zijlstra 2015-03-19 18:32 ` Peter Zijlstra 2015-03-25 19:47 ` Konrad Rzeszutek Wilk 2015-03-26 20:21 ` Peter Zijlstra 2015-03-26 20:21 ` Peter Zijlstra 2015-03-27 14:07 ` Konrad Rzeszutek Wilk 2015-03-27 14:07 ` Konrad Rzeszutek Wilk 2015-03-30 16:41 ` Waiman Long 2015-03-30 16:25 ` Waiman Long 2015-03-30 16:29 ` Peter Zijlstra 2015-03-30 16:29 ` Peter Zijlstra 2015-03-30 16:43 ` Waiman Long 2015-03-30 16:43 ` Waiman Long 2015-03-27 6:40 ` Raghavendra K T 2015-03-27 6:40 ` Raghavendra K T
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20150316133112.027716523@infradead.org \ --to=peterz@infradead.org \ --cc=Waiman.Long@hp.com \ --cc=boris.ostrovsky@oracle.com \ --cc=david.vrabel@citrix.com \ --cc=doug.hatch@hp.com \ --cc=hpa@zytor.com \ --cc=kvm@vger.kernel.org \ --cc=linux-arch@vger.kernel.org \ --cc=linux-kernel@vger.kernel.org \ --cc=luto@amacapital.net \ --cc=mingo@redhat.com \ --cc=oleg@redhat.com \ --cc=paolo.bonzini@gmail.com \ --cc=paulmck@linux.vnet.ibm.com \ --cc=raghavendra.kt@linux.vnet.ibm.com \ --cc=riel@redhat.com \ --cc=scott.norton@hp.com \ --cc=tglx@linutronix.de \ --cc=torvalds@linux-foundation.org \ --cc=virtualization@lists.linux-foundation.org \ --cc=x86@kernel.org \ --cc=xen-devel@lists.xenproject.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).