From: Peter Zijlstra <peterz@infradead.org>
To: Alex Kogan <alex.kogan@oracle.com>
Cc: linux-arch@vger.kernel.org, arnd@arndb.de, dave.dice@oracle.com,
will.deacon@arm.com, linux@armlinux.org.uk,
linux-kernel@vger.kernel.org, rahul.x.yadav@oracle.com,
mingo@redhat.com, steven.sistare@oracle.com, longman@redhat.com,
daniel.m.jordan@oracle.com, linux-arm-kernel@lists.infradead.org
Subject: Re: [PATCH 0/3] Add NUMA-awareness to qspinlock
Date: Thu, 31 Jan 2019 10:56:38 +0100 [thread overview]
Message-ID: <20190131095638.GA31534@hirez.programming.kicks-ass.net> (raw)
In-Reply-To: <20190131030136.56999-1-alex.kogan@oracle.com>
On Wed, Jan 30, 2019 at 10:01:32PM -0500, Alex Kogan wrote:
> Lock throughput can be increased by handing a lock to a waiter on the
> same NUMA socket as the lock holder, provided care is taken to avoid
> starvation of waiters on other NUMA sockets. This patch introduces CNA
> (compact NUMA-aware lock) as the slow path for qspinlock.
Since you use NUMA, use the term node, not socket. The two are not
strictly related.
> CNA is a NUMA-aware version of the MCS spin-lock. Spinning threads are
> organized in two queues, a main queue for threads running on the same
> socket as the current lock holder, and a secondary queue for threads
> running on other sockets. Threads record the ID of the socket on which
> they are running in their queue nodes. At the unlock time, the lock
> holder scans the main queue looking for a thread running on the same
> socket. If found (call it thread T), all threads in the main queue
> between the current lock holder and T are moved to the end of the
> secondary queue, and the lock is passed to T. If such T is not found, the
> lock is passed to the first node in the secondary queue. Finally, if the
> secondary queue is empty, the lock is passed to the next thread in the
> main queue.
>
> Full details are available at https://arxiv.org/abs/1810.05600.
Full details really should also be in the Changelog. You can skip much
of the academic bla-bla, but the Changelog should be self contained.
> We have done some performance evaluation with the locktorture module
> as well as with several benchmarks from the will-it-scale repo.
> The following locktorture results are from an Oracle X5-4 server
> (four Intel Xeon E7-8895 v3 @ 2.60GHz sockets with 18 hyperthreaded
> cores each). Each number represents an average (over 5 runs) of the
> total number of ops (x10^7) reported at the end of each run. The stock
> kernel is v4.20.0-rc4+ compiled in the default configuration.
>
> #thr stock patched speedup (patched/stock)
> 1 2.710 2.715 1.002
> 2 3.108 3.001 0.966
> 4 4.194 3.919 0.934
So low contention is actually worse. Funnily low contention is the
majority of our locks and is _really_ important.
> 8 5.309 6.894 1.299
> 16 6.722 9.094 1.353
> 32 7.314 9.885 1.352
> 36 7.562 9.855 1.303
> 72 6.696 10.358 1.547
> 108 6.364 10.181 1.600
> 142 6.179 10.178 1.647
>
> When the kernel is compiled with lockstat enabled, CNA
I'll ignore that, lockstat/lockdep enabled runs are not what one would
call performance relevant.
WARNING: multiple messages have this Message-ID (diff)
From: Peter Zijlstra <peterz@infradead.org>
To: Alex Kogan <alex.kogan@oracle.com>
Cc: linux@armlinux.org.uk, mingo@redhat.com, will.deacon@arm.com,
arnd@arndb.de, longman@redhat.com, linux-arch@vger.kernel.org,
linux-arm-kernel@lists.infradead.org,
linux-kernel@vger.kernel.org, steven.sistare@oracle.com,
daniel.m.jordan@oracle.com, dave.dice@oracle.com,
rahul.x.yadav@oracle.com
Subject: Re: [PATCH 0/3] Add NUMA-awareness to qspinlock
Date: Thu, 31 Jan 2019 10:56:38 +0100 [thread overview]
Message-ID: <20190131095638.GA31534@hirez.programming.kicks-ass.net> (raw)
Message-ID: <20190131095638.BgeI5Pts9D_GeL_NN6FBaCa7vcV6RB-I75wapgB77KQ@z> (raw)
In-Reply-To: <20190131030136.56999-1-alex.kogan@oracle.com>
On Wed, Jan 30, 2019 at 10:01:32PM -0500, Alex Kogan wrote:
> Lock throughput can be increased by handing a lock to a waiter on the
> same NUMA socket as the lock holder, provided care is taken to avoid
> starvation of waiters on other NUMA sockets. This patch introduces CNA
> (compact NUMA-aware lock) as the slow path for qspinlock.
Since you use NUMA, use the term node, not socket. The two are not
strictly related.
> CNA is a NUMA-aware version of the MCS spin-lock. Spinning threads are
> organized in two queues, a main queue for threads running on the same
> socket as the current lock holder, and a secondary queue for threads
> running on other sockets. Threads record the ID of the socket on which
> they are running in their queue nodes. At the unlock time, the lock
> holder scans the main queue looking for a thread running on the same
> socket. If found (call it thread T), all threads in the main queue
> between the current lock holder and T are moved to the end of the
> secondary queue, and the lock is passed to T. If such T is not found, the
> lock is passed to the first node in the secondary queue. Finally, if the
> secondary queue is empty, the lock is passed to the next thread in the
> main queue.
>
> Full details are available at https://arxiv.org/abs/1810.05600.
Full details really should also be in the Changelog. You can skip much
of the academic bla-bla, but the Changelog should be self contained.
> We have done some performance evaluation with the locktorture module
> as well as with several benchmarks from the will-it-scale repo.
> The following locktorture results are from an Oracle X5-4 server
> (four Intel Xeon E7-8895 v3 @ 2.60GHz sockets with 18 hyperthreaded
> cores each). Each number represents an average (over 5 runs) of the
> total number of ops (x10^7) reported at the end of each run. The stock
> kernel is v4.20.0-rc4+ compiled in the default configuration.
>
> #thr stock patched speedup (patched/stock)
> 1 2.710 2.715 1.002
> 2 3.108 3.001 0.966
> 4 4.194 3.919 0.934
So low contention is actually worse. Funnily low contention is the
majority of our locks and is _really_ important.
> 8 5.309 6.894 1.299
> 16 6.722 9.094 1.353
> 32 7.314 9.885 1.352
> 36 7.562 9.855 1.303
> 72 6.696 10.358 1.547
> 108 6.364 10.181 1.600
> 142 6.179 10.178 1.647
>
> When the kernel is compiled with lockstat enabled, CNA
I'll ignore that, lockstat/lockdep enabled runs are not what one would
call performance relevant.
next prev parent reply other threads:[~2019-01-31 9:56 UTC|newest]
Thread overview: 28+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-01-31 3:01 [PATCH 0/3] Add NUMA-awareness to qspinlock Alex Kogan
2019-01-31 3:01 ` Alex Kogan
2019-01-31 3:01 ` [PATCH 1/3] locking/qspinlock: Make arch_mcs_spin_unlock_contended more generic Alex Kogan
2019-01-31 3:01 ` Alex Kogan
2019-01-31 3:01 ` [PATCH 2/3] locking/qspinlock: Introduce CNA into the slow path of qspinlock Alex Kogan
2019-01-31 3:01 ` Alex Kogan
2019-01-31 17:38 ` Waiman Long
2019-01-31 17:38 ` Waiman Long
2019-02-01 21:26 ` Alex Kogan
2019-02-01 21:26 ` Alex Kogan
2019-01-31 3:01 ` [PATCH 3/3] locking/qspinlock: Introduce starvation avoidance into CNA Alex Kogan
2019-01-31 3:01 ` Alex Kogan
2019-01-31 10:00 ` Peter Zijlstra
2019-01-31 10:00 ` Peter Zijlstra
2019-02-05 3:35 ` Alex Kogan
2019-02-05 3:35 ` Alex Kogan
2019-02-05 9:22 ` Peter Zijlstra
2019-02-05 9:22 ` Peter Zijlstra
2019-02-05 13:48 ` Waiman Long
2019-02-05 13:48 ` Waiman Long
2019-02-05 21:07 ` Alex Kogan
2019-02-05 21:07 ` Alex Kogan
2019-02-05 21:12 ` Waiman Long
2019-02-05 21:12 ` Waiman Long
2019-01-31 9:56 ` Peter Zijlstra [this message]
2019-01-31 9:56 ` [PATCH 0/3] Add NUMA-awareness to qspinlock Peter Zijlstra
2019-02-01 21:20 ` Alex Kogan
2019-02-01 21:20 ` Alex Kogan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190131095638.GA31534@hirez.programming.kicks-ass.net \
--to=peterz@infradead.org \
--cc=alex.kogan@oracle.com \
--cc=arnd@arndb.de \
--cc=daniel.m.jordan@oracle.com \
--cc=dave.dice@oracle.com \
--cc=linux-arch@vger.kernel.org \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux@armlinux.org.uk \
--cc=longman@redhat.com \
--cc=mingo@redhat.com \
--cc=rahul.x.yadav@oracle.com \
--cc=steven.sistare@oracle.com \
--cc=will.deacon@arm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox