From: Thomas Gleixner <tglx@linutronix.de>
To: K Prateek Nayak <kprateek.nayak@amd.com>,
LKML <linux-kernel@vger.kernel.org>
Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>,
"Paul E. McKenney" <paulmck@kernel.org>,
Boqun Feng <boqun.feng@gmail.com>,
Jonathan Corbet <corbet@lwn.net>,
Prakash Sangappa <prakash.sangappa@oracle.com>,
Madadi Vineeth Reddy <vineethr@linux.ibm.com>,
Steven Rostedt <rostedt@goodmis.org>,
Sebastian Andrzej Siewior <bigeasy@linutronix.de>,
Arnd Bergmann <arnd@arndb.de>,
linux-arch@vger.kernel.org, Randy Dunlap <rdunlap@infradead.org>,
Peter Zijlstra <peterz@infradead.org>
Subject: Re: [patch V4 00/12] rseq: Implement time slice extension mechanism
Date: Mon, 24 Nov 2025 11:33:56 +0100 [thread overview]
Message-ID: <87h5ujiusr.ffs@tglx> (raw)
In-Reply-To: <8cefc4be-b11a-414c-b3f4-280c900be67b@amd.com>
On Mon, Nov 24 2025 at 12:40, K. Prateek Nayak wrote:
> On 11/17/2025 2:21 AM, Thomas Gleixner wrote:
> I got a chance to test the series with Netflix's userspace locking
> benchmark [1] which ended up looking very similar to the test
> Steven had written initially when discussing the PoC except this
> can scale the number of threads.
>
> [1] https://www.github.com/Netflix/global-lock-bench/
>
> Critical section is just a single increment operation. Metric is
> average time taken to run a fixed amount of critical sections across
> #Threads over 3 runs.
>
> Here are the results of running the test with the default config on
> my 256CPU machine in a root cpuset containing 32CPUs to actually hit
> the contention:
>
> o rseq/slice with no benchmark modifications and "rseq_slice_ext=0"
>
> | Threads | Threaded (s) | Threaded/s |
> +---------+-----------------+--------------+
> | 1 | .026103 | 383493128.79 |
> | 2 | .086320 | 116134267.40 |
> | 4 | .669743 | 14937390.67 |
> | 8 | 1.105109 | 9053764.30 |
> | 16 | 1.863516 | 5366809.94 |
> | 32 | 7.249873 | 1379590.12 |
> | 64 | 14.360199 | 696486.76 |
> | 96 | 21.909887 | 456458.03 |
> | 128 | 29.126423 | 343358.95 |
> | 192 | 43.112188 | 231980.16 |
> | 256 | 57.628748 | 173554.39 |
> | 384 | 86.274354 | 115909.73 |
> | 512 | 114.564142 | 87289.97 |
>
>
> o rseq/slice with modified benchmark and "rseq_slice_ext=1"
>
> | Threads | Threaded (s) | Threaded/s | %diff (s) |
> +---------+-----------------+--------------+-----------+
> | 1 | .036438 | 274437690.71 | 40% |
> | 2 | .147520 | 68851845.82 | 71% |
> | 4 | .829240 | 12176948.03 | 24% |
> | 8 | 1.259632 | 7993476.42 | 14% |
> | 16 | 1.988396 | 5029209.62 | 7% |
> | 32 | 9.844307 | 1015837.43 | 36% |
> | 64 | 14.590723 | 685979.41 | 2% |
> | 96 | 18.898278 | 529171.84 | -14% |
> | 128 | 23.921747 | 418033.09 | -18% |
> | 192 | 33.284228 | 300673.66 | -23% |
> | 256 | 42.934755 | 232934.87 | -25% |
> | 384 | 61.794499 | 161924.64 | -28% |
> | 512 | 82.005069 | 121951.34 | -28% |
>
> ( Lower %diff is better )
>
>
> Until the contention begins (> 32 threads), there is a consistent
> regression which I believe can be attributed to the additional
> overhead in the critical section from setting the slice_ext
> request, however, once heavy contention begins, there is a clear
> win with slice extension.
Yes, that's about right.
I just played with that benchmark on a random machine and a cpuset of 28
CPUs (primary threads of second socket). First naive attempt doing:
request_slice(); // sets the request bit
lock();
unlock();
check_slice(); // clears the request bit, checks grant, yield
Threads noslice slice t ops
1 0.006 s 155101817 ops/s 0.008 s 128339525 ops/s 20% -17%
2 0.060 s 16709158 ops/s 0.071 s 14207646 ops/s 18% -14%
4 0.150 s 6681970 ops/s 0.150 s 6681428 ops/s 0% 0%
8 0.311 s 3217183 ops/s 0.290 s 3448088 ops/s -6% 7%
16 0.624 s 1602136 ops/s 0.581 s 1723563 ops/s -7% 7%
32 1.170 s 855494 ops/s 1.018 s 983453 ops/s -13% 14%
64 2.287 s 437461 ops/s 1.414 s 708871 ops/s -38% 62%
128 4.353 s 229926 ops/s 2.010 s 499227 ops/s -53% 117%
256 8.321 s 120311 ops/s 3.210 s 314349 ops/s -61% 161%
Then I modified lock():
request_slice();
while (!trylock) {
check_slice();
pause();
request_slice();
And it becomes this:
Threads noslice slice t ops
1 0.006 s 167464787 ops/s 0.008 s 126824764 ops/s 32% -24%
2 0.038 s 26868534 ops/s 0.044 s 23020570 ops/s 16% -14%
4 0.119 s 8411120 ops/s 0.126 s 7930586 ops/s 6% -5%
8 0.241 s 4157058 ops/s 0.257 s 3896290 ops/s 6% -6%
16 0.463 s 2161893 ops/s 0.468 s 2149853 ops/s 0% 0%
32 1.037 s 965391 ops/s 0.888 s 1127000 ops/s -14% 16%
64 2.018 s 495862 ops/s 0.880 s 1136807 ops/s -56% 129%
128 3.834 s 262185 ops/s 0.874 s 1144878 ops/s -77% 336%
256 7.704 s 129885 ops/s 0.875 s 1143316 ops/s -88% 780%
It's interesting that both runtime _and_ ops/s stay stable once contention is hit.
Thanks,
tglx
prev parent reply other threads:[~2025-11-24 10:33 UTC|newest]
Thread overview: 18+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-16 20:51 [patch V4 00/12] rseq: Implement time slice extension mechanism Thomas Gleixner
2025-11-16 20:51 ` [patch V4 01/12] sched: Provide and use set_need_resched_current() Thomas Gleixner
2025-11-20 21:29 ` [tip: sched/core] " tip-bot2 for Peter Zijlstra
2025-11-16 20:51 ` [patch V4 02/12] rseq: Add fields and constants for time slice extension Thomas Gleixner
2025-11-16 20:51 ` [patch V4 03/12] rseq: Provide static branch for time slice extensions Thomas Gleixner
2025-11-16 20:51 ` [patch V4 04/12] rseq: Add statistics " Thomas Gleixner
2025-11-16 20:51 ` [patch V4 05/12] rseq: Add prctl() to enable " Thomas Gleixner
2025-11-16 20:51 ` [patch V4 06/12] rseq: Implement sys_rseq_slice_yield() Thomas Gleixner
2025-11-16 21:51 ` Arnd Bergmann
2025-11-16 20:51 ` [patch V4 07/12] rseq: Implement syscall entry work for time slice extensions Thomas Gleixner
2025-11-16 20:51 ` [patch V4 08/12] rseq: Implement time slice extension enforcement timer Thomas Gleixner
2025-11-16 23:23 ` Hillf Danton
2025-11-16 20:51 ` [patch V4 09/12] rseq: Reset slice extension when scheduled Thomas Gleixner
2025-11-16 20:51 ` [patch V4 10/12] rseq: Implement rseq_grant_slice_extension() Thomas Gleixner
2025-11-16 20:51 ` [patch V4 11/12] entry: Hook up rseq time slice extension Thomas Gleixner
2025-11-16 20:51 ` [patch V4 12/12] selftests/rseq: Implement time slice extension test Thomas Gleixner
2025-11-24 7:10 ` [patch V4 00/12] rseq: Implement time slice extension mechanism K Prateek Nayak
2025-11-24 10:33 ` Thomas Gleixner [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=87h5ujiusr.ffs@tglx \
--to=tglx@linutronix.de \
--cc=arnd@arndb.de \
--cc=bigeasy@linutronix.de \
--cc=boqun.feng@gmail.com \
--cc=corbet@lwn.net \
--cc=kprateek.nayak@amd.com \
--cc=linux-arch@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mathieu.desnoyers@efficios.com \
--cc=paulmck@kernel.org \
--cc=peterz@infradead.org \
--cc=prakash.sangappa@oracle.com \
--cc=rdunlap@infradead.org \
--cc=rostedt@goodmis.org \
--cc=vineethr@linux.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox