From: "Paul E. McKenney" <paulmck@kernel.org>
To: Boqun Feng <boqun@kernel.org>
Cc: Joel Fernandes <joelagnelf@nvidia.com>,
Kumar Kartikeya Dwivedi <memxor@gmail.com>,
Sebastian Andrzej Siewior <bigeasy@linutronix.de>,
frederic@kernel.org, neeraj.iitr10@gmail.com, urezki@gmail.com,
boqun.feng@gmail.com, rcu@vger.kernel.org,
Tejun Heo <tj@kernel.org>,
bpf@vger.kernel.org, Alexei Starovoitov <ast@kernel.org>,
Daniel Borkmann <daniel@iogearbox.net>,
John Fastabend <john.fastabend@gmail.com>,
Andrea Righi <arighi@nvidia.com>, Zqiang <qiang.zhang@linux.dev>
Subject: Re: [PATCH] rcu: Use an intermediate irq_work to start process_srcu()
Date: Sat, 21 Mar 2026 03:10:05 -0700 [thread overview]
Message-ID: <b4857e4f-5a8d-4ccc-9e8b-df57718e4aee@paulmck-laptop> (raw)
In-Reply-To: <20260320181400.15909-1-boqun@kernel.org>
On Fri, Mar 20, 2026 at 11:14:00AM -0700, Boqun Feng wrote:
> Since commit c27cea4416a3 ("rcu: Re-implement RCU Tasks Trace in terms
> of SRCU-fast") we switched to SRCU in BPF. However as BPF instrument can
> happen basically everywhere (including where a scheduler lock is held),
> call_srcu() now needs to avoid acquiring scheduler lock because
> otherwise it could cause deadlock [1]. Fix this by following what the
> previous RCU Tasks Trace did: using an irq_work to delay the queuing of
> the work to start process_srcu().
>
> [boqun: Apply Joel's feedback]
>
> Reported-by: Andrea Righi <arighi@nvidia.com>
> Closes: https://lore.kernel.org/all/abjzvz_tL_siV17s@gpd4/
> Fixes: commit c27cea4416a3 ("rcu: Re-implement RCU Tasks Trace in terms of SRCU-fast")
> Link: https://lore.kernel.org/rcu/3c4c5a29-24ea-492d-aeee-e0d9605b4183@nvidia.com/ [1]
> Suggested-by: Zqiang <qiang.zhang@linux.dev>
> Signed-off-by: Boqun Feng <boqun@kernel.org>
First, thank you all for putting this together!
If I enable both early boot RCU testing and lockdep, for example,
by running the RUDE01 rcutorture scenario, I get the following splat,
which suggests that the raw_spin_unlock_irq_rcu_node() in srcu_irq_work()
might need help (see inline below):
[ 0.872594] ------------[ cut here ]------------
[ 0.873550] DEBUG_LOCKS_WARN_ON(lockdep_hardirq_context())
[ 0.873550] WARNING: kernel/locking/lockdep.c:4404 at lockdep_hardirqs_on_prepare+0x150/0x190, CPU#0: swapper/0/1
[ 0.873550] Modules linked in:
[ 0.873550] CPU: 0 UID: 0 PID: 1 Comm: swapper/0 Not tainted 7.0.0-rc3-00039-g35d354b6cd0f-dirty #8217 PREEMPT(full)
[ 0.873550] Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS rel-1.16.0-0-gd239552ce722-prebuilt.qemu.org 04/01/2014
[ 0.873550] RIP: 0010:lockdep_hardirqs_on_prepare+0x157/0x190
[ 0.873550] Code: 01 90 e8 ec c7 54 00 85 c0 74 0a 8b 35 12 c4 e0 01 85 f6 74 31 90 5d c3 cc cc cc cc 48 8d 3d 20 cf e1 01 48 c7 c6 ec c1 87 9f <67> 48 0f b9 3a eb ac 48 8d 3d 1b cf e1 01 48 c7 c6 87 be 87 9f 67
[ 0.873550] RSP: 0000:ffff9ff3c0003f50 EFLAGS: 00010046
[ 0.873550] RAX: 0000000000000001 RBX: ffffffff9fb608f8 RCX: 0000000000000001
[ 0.873550] RDX: 0000000000000000 RSI: ffffffff9f87c1ec RDI: ffffffff9fd44120
[ 0.873550] RBP: ffffffff9f00bae3 R08: 0000000000000001 R09: 0000000000000000
[ 0.873550] R10: 0000000000000001 R11: 0000000000000000 R12: 0000000000000000
[ 0.873550] R13: ffff9e1bc11e4bc0 R14: 0000000000000000 R15: 0000000000000000
[ 0.873550] FS: 0000000000000000(0000) GS:ffff9e1c3ed9b000(0000) knlGS:0000000000000000
[ 0.873550] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 0.873550] CR2: ffff9e1bcf3d8000 CR3: 000000000dc4e000 CR4: 00000000000006f0
[ 0.873550] Call Trace:
[ 0.873550] <IRQ>
[ 0.873550] ? _raw_spin_unlock_irq+0x23/0x40
[ 0.873550] trace_hardirqs_on+0x16/0xe0
[ 0.873550] _raw_spin_unlock_irq+0x23/0x40
[ 0.873550] srcu_irq_work+0x5e/0x90
[ 0.873550] irq_work_single+0x42/0x90
[ 0.873550] irq_work_run_list+0x26/0x40
[ 0.873550] irq_work_run+0x18/0x30
[ 0.873550] __sysvec_irq_work+0x30/0x180
[ 0.873550] sysvec_irq_work+0x6a/0x80
[ 0.873550] </IRQ>
[ 0.873550] <TASK>
[ 0.873550] asm_sysvec_irq_work+0x1a/0x20
[ 0.873550] RIP: 0010:_raw_spin_unlock_irqrestore+0x34/0x50
[ 0.873550] Code: c7 18 53 48 89 f3 48 8b 74 24 10 e8 06 d4 f1 fe 48 89 ef e8 2e 0c f2 fe 80 e7 02 74 06 e8 74 86 01 ff fb 65 ff 0d ec d4 66 01 <74> 07 5b 5d e9 53 1b 00 00 e8 5e 94 df fe 5b 5d e9 47 1b 00 00 0f
[ 0.873550] RSP: 0000:ffff9ff3c0013d50 EFLAGS: 00000286
[ 0.873550] RAX: 0000000000001417 RBX: 0000000000000297 RCX: 0000000000000000
[ 0.873550] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffffffff9f00bb3c
[ 0.873550] RBP: ffffffff9fb60630 R08: 0000000000000001 R09: 0000000000000000
[ 0.873550] R10: 0000000000000001 R11: 0000000000000001 R12: fffffffffffffe74
[ 0.873550] R13: 0000000000000001 R14: 0000000000000000 R15: 0000000000000001
[ 0.873550] ? _raw_spin_unlock_irqrestore+0x2c/0x50
[ 0.873550] srcu_gp_start_if_needed+0x354/0x530
[ 0.873550] __synchronize_srcu+0xd1/0x180
[ 0.873550] ? __pfx_wakeme_after_rcu+0x10/0x10
[ 0.873550] ? synchronize_srcu+0x3f/0x170
[ 0.873550] ? __pfx_rcu_init_tasks_generic+0x10/0x10
[ 0.873550] rcu_init_tasks_generic+0x10c/0x130
[ 0.873550] do_one_initcall+0x59/0x2e0
[ 0.873550] ? _printk+0x56/0x70
[ 0.873550] kernel_init_freeable+0x227/0x440
[ 0.873550] ? __pfx_kernel_init+0x10/0x10
[ 0.873550] kernel_init+0x15/0x1c0
[ 0.873550] ret_from_fork+0x2ac/0x330
[ 0.873550] ? __pfx_kernel_init+0x10/0x10
[ 0.873550] ret_from_fork_asm+0x1a/0x30
[ 0.873550] </TASK>
[ 0.873550] irq event stamp: 5144
[ 0.873550] hardirqs last enabled at (5143): [<ffffffff9f00bb3c>] _raw_spin_unlock_irqrestore+0x2c/0x50
[ 0.873550] hardirqs last disabled at (5144): [<ffffffff9eff7c9f>] sysvec_irq_work+0xf/0x80
[ 0.873550] softirqs last enabled at (5132): [<ffffffff9dea2501>] __irq_exit_rcu+0xa1/0xc0
[ 0.873550] softirqs last disabled at (5127): [<ffffffff9dea2501>] __irq_exit_rcu+0xa1/0xc0
[ 0.873550] ---[ end trace 0000000000000000 ]---
[ 0.873574] ------------[ cut here ]------------
> ---
> @Zqiang, I put your name as Suggested-by because you proposed the same
> idea, let me know if you rather not have it.
>
> @Joel, I did two updates (including your test feedback, other one is
> call irq_work_sync() when we clean the srcu_struct), please give it a
> try.
>
> include/linux/srcutree.h | 1 +
> kernel/rcu/srcutree.c | 29 +++++++++++++++++++++++++++--
> 2 files changed, 28 insertions(+), 2 deletions(-)
>
> diff --git a/include/linux/srcutree.h b/include/linux/srcutree.h
> index dfb31d11ff05..be76fa4fc170 100644
> --- a/include/linux/srcutree.h
> +++ b/include/linux/srcutree.h
> @@ -95,6 +95,7 @@ struct srcu_usage {
> unsigned long reschedule_jiffies;
> unsigned long reschedule_count;
> struct delayed_work work;
> + struct irq_work irq_work;
> struct srcu_struct *srcu_ssp;
> };
>
> diff --git a/kernel/rcu/srcutree.c b/kernel/rcu/srcutree.c
> index 2328827f8775..73aef361a524 100644
> --- a/kernel/rcu/srcutree.c
> +++ b/kernel/rcu/srcutree.c
> @@ -19,6 +19,7 @@
> #include <linux/mutex.h>
> #include <linux/percpu.h>
> #include <linux/preempt.h>
> +#include <linux/irq_work.h>
> #include <linux/rcupdate_wait.h>
> #include <linux/sched.h>
> #include <linux/smp.h>
> @@ -75,6 +76,7 @@ static bool __read_mostly srcu_init_done;
> static void srcu_invoke_callbacks(struct work_struct *work);
> static void srcu_reschedule(struct srcu_struct *ssp, unsigned long delay);
> static void process_srcu(struct work_struct *work);
> +static void srcu_irq_work(struct irq_work *work);
> static void srcu_delay_timer(struct timer_list *t);
>
> /*
> @@ -216,6 +218,7 @@ static int init_srcu_struct_fields(struct srcu_struct *ssp, bool is_static)
> mutex_init(&ssp->srcu_sup->srcu_barrier_mutex);
> atomic_set(&ssp->srcu_sup->srcu_barrier_cpu_cnt, 0);
> INIT_DELAYED_WORK(&ssp->srcu_sup->work, process_srcu);
> + init_irq_work(&ssp->srcu_sup->irq_work, srcu_irq_work);
> ssp->srcu_sup->sda_is_static = is_static;
> if (!is_static) {
> ssp->sda = alloc_percpu(struct srcu_data);
> @@ -713,6 +716,8 @@ void cleanup_srcu_struct(struct srcu_struct *ssp)
> return; /* Just leak it! */
> if (WARN_ON(srcu_readers_active(ssp)))
> return; /* Just leak it! */
> + /* Wait for irq_work to finish first as it may queue a new work. */
> + irq_work_sync(&sup->irq_work);
> flush_delayed_work(&sup->work);
> for_each_possible_cpu(cpu) {
> struct srcu_data *sdp = per_cpu_ptr(ssp->sda, cpu);
> @@ -1118,9 +1123,13 @@ static void srcu_funnel_gp_start(struct srcu_struct *ssp, struct srcu_data *sdp,
> // it isn't. And it does not have to be. After all, it
> // can only be executed during early boot when there is only
> // the one boot CPU running with interrupts still disabled.
> + //
> + // Use an irq_work here to avoid acquiring runqueue lock with
> + // srcu rcu_node::lock held. BPF instrument could introduce the
> + // opposite dependency, hence we need to break the possible
> + // locking dependency here.
> if (likely(srcu_init_done))
> - queue_delayed_work(rcu_gp_wq, &sup->work,
> - !!srcu_get_delay(ssp));
> + irq_work_queue(&sup->irq_work);
> else if (list_empty(&sup->work.work.entry))
> list_add(&sup->work.work.entry, &srcu_boot_list);
> }
> @@ -1979,6 +1988,22 @@ static void process_srcu(struct work_struct *work)
> srcu_reschedule(ssp, curdelay);
> }
>
> +static void srcu_irq_work(struct irq_work *work)
> +{
> + struct srcu_struct *ssp;
> + struct srcu_usage *sup;
> + unsigned long delay;
> +
> + sup = container_of(work, struct srcu_usage, irq_work);
> + ssp = sup->srcu_ssp;
> +
> + raw_spin_lock_irq_rcu_node(ssp->srcu_sup);
> + delay = srcu_get_delay(ssp);
> + raw_spin_unlock_irq_rcu_node(ssp->srcu_sup);
Removing the "_irq" from both avoids the lockdep splat in my test setup,
which makes sense given that interrupts are disabled in irq_work handlers.
Or at least it looks to me that they are. ;-)
Like this:
+ raw_spin_lock_rcu_node(ssp->srcu_sup);
+ delay = srcu_get_delay(ssp);
+ raw_spin_unlock_rcu_node(ssp->srcu_sup);
Thanx, Paul
> +
> + queue_delayed_work(rcu_gp_wq, &sup->work, !!delay);
> +}
> +
> void srcutorture_get_gp_data(struct srcu_struct *ssp, int *flags,
> unsigned long *gp_seq)
> {
> --
> 2.50.1 (Apple Git-155)
>
next prev parent reply other threads:[~2026-03-21 10:10 UTC|newest]
Thread overview: 100+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-03-17 13:34 Next-level bug in SRCU implementation of RCU Tasks Trace + PREEMPT_RT Paul E. McKenney
2026-03-18 10:50 ` Sebastian Andrzej Siewior
2026-03-18 11:49 ` Paul E. McKenney
2026-03-18 14:43 ` Sebastian Andrzej Siewior
2026-03-18 15:43 ` Paul E. McKenney
2026-03-18 16:04 ` Sebastian Andrzej Siewior
2026-03-18 16:32 ` Paul E. McKenney
2026-03-18 16:42 ` Boqun Feng
2026-03-18 18:45 ` Paul E. McKenney
2026-03-18 16:47 ` Sebastian Andrzej Siewior
2026-03-18 18:48 ` Paul E. McKenney
2026-03-19 8:55 ` Sebastian Andrzej Siewior
2026-03-19 10:05 ` Paul E. McKenney
2026-03-19 10:43 ` Paul E. McKenney
2026-03-19 10:51 ` Sebastian Andrzej Siewior
2026-03-18 15:51 ` Boqun Feng
2026-03-18 18:42 ` Paul E. McKenney
2026-03-18 20:04 ` Joel Fernandes
2026-03-18 20:11 ` Kumar Kartikeya Dwivedi
2026-03-18 20:25 ` Joel Fernandes
2026-03-18 21:52 ` Boqun Feng
2026-03-18 21:55 ` Boqun Feng
2026-03-18 22:15 ` Boqun Feng
2026-03-18 22:52 ` Joel Fernandes
2026-03-18 23:27 ` Boqun Feng
2026-03-19 1:08 ` Boqun Feng
2026-03-19 9:03 ` Sebastian Andrzej Siewior
2026-03-19 16:27 ` Boqun Feng
2026-03-19 16:33 ` Sebastian Andrzej Siewior
2026-03-19 16:48 ` Boqun Feng
2026-03-19 16:59 ` Kumar Kartikeya Dwivedi
2026-03-19 17:27 ` Boqun Feng
2026-03-19 18:41 ` Kumar Kartikeya Dwivedi
2026-03-19 20:14 ` Boqun Feng
2026-03-19 20:21 ` Joel Fernandes
2026-03-19 20:39 ` Boqun Feng
2026-03-20 15:34 ` Paul E. McKenney
2026-03-20 15:59 ` Boqun Feng
2026-03-20 16:24 ` Paul E. McKenney
2026-03-20 16:57 ` Boqun Feng
2026-03-20 17:54 ` Joel Fernandes
2026-03-20 18:14 ` [PATCH] rcu: Use an intermediate irq_work to start process_srcu() Boqun Feng
2026-03-20 19:18 ` Joel Fernandes
2026-03-20 20:47 ` Andrea Righi
2026-03-20 20:54 ` Boqun Feng
2026-03-20 21:00 ` Andrea Righi
2026-03-20 21:02 ` Andrea Righi
2026-03-20 21:06 ` Boqun Feng
2026-03-20 22:29 ` [PATCH v2] " Boqun Feng
2026-03-23 21:09 ` Joel Fernandes
2026-03-23 22:18 ` Boqun Feng
2026-03-23 22:50 ` Joel Fernandes
2026-03-24 11:27 ` Frederic Weisbecker
2026-03-24 14:56 ` Joel Fernandes
2026-03-24 14:56 ` Alexei Starovoitov
2026-03-24 17:36 ` Boqun Feng
2026-03-24 18:40 ` Joel Fernandes
2026-03-24 19:23 ` Paul E. McKenney
2026-03-21 4:27 ` [PATCH] " Zqiang
2026-03-21 18:15 ` Boqun Feng
2026-03-21 10:10 ` Paul E. McKenney [this message]
2026-03-21 17:15 ` Boqun Feng
2026-03-21 17:41 ` Paul E. McKenney
2026-03-21 18:06 ` Boqun Feng
2026-03-21 19:31 ` Paul E. McKenney
2026-03-21 19:45 ` Boqun Feng
2026-03-21 20:07 ` Paul E. McKenney
2026-03-21 20:08 ` Boqun Feng
2026-03-22 10:09 ` Paul E. McKenney
2026-03-22 16:16 ` Boqun Feng
2026-03-22 17:09 ` Paul E. McKenney
2026-03-22 17:31 ` Boqun Feng
2026-03-22 17:44 ` Paul E. McKenney
2026-03-22 18:17 ` Boqun Feng
2026-03-22 19:47 ` Paul E. McKenney
2026-03-22 20:26 ` Boqun Feng
2026-03-23 7:50 ` Paul E. McKenney
2026-03-20 18:20 ` Next-level bug in SRCU implementation of RCU Tasks Trace + PREEMPT_RT Boqun Feng
2026-03-20 23:11 ` Paul E. McKenney
2026-03-21 3:29 ` Paul E. McKenney
2026-03-21 17:03 ` [RFC PATCH] rcu-tasks: Avoid using mod_timer() in call_rcu_tasks_generic() Boqun Feng
2026-03-23 15:17 ` Boqun Feng
2026-03-23 20:37 ` Joel Fernandes
2026-03-23 21:50 ` Kumar Kartikeya Dwivedi
2026-03-23 22:13 ` Boqun Feng
2026-03-20 16:15 ` Next-level bug in SRCU implementation of RCU Tasks Trace + PREEMPT_RT Boqun Feng
2026-03-20 16:24 ` Paul E. McKenney
2026-03-19 17:02 ` Sebastian Andrzej Siewior
2026-03-19 17:44 ` Boqun Feng
2026-03-19 18:42 ` Joel Fernandes
2026-03-19 20:20 ` Boqun Feng
2026-03-19 20:26 ` Joel Fernandes
2026-03-19 20:45 ` Joel Fernandes
2026-03-19 10:02 ` Paul E. McKenney
2026-03-19 14:34 ` Boqun Feng
2026-03-19 16:10 ` Paul E. McKenney
2026-03-18 23:56 ` Kumar Kartikeya Dwivedi
2026-03-19 0:26 ` Zqiang
2026-03-19 1:13 ` Boqun Feng
2026-03-19 2:47 ` Joel Fernandes
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=b4857e4f-5a8d-4ccc-9e8b-df57718e4aee@paulmck-laptop \
--to=paulmck@kernel.org \
--cc=arighi@nvidia.com \
--cc=ast@kernel.org \
--cc=bigeasy@linutronix.de \
--cc=boqun.feng@gmail.com \
--cc=boqun@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=daniel@iogearbox.net \
--cc=frederic@kernel.org \
--cc=joelagnelf@nvidia.com \
--cc=john.fastabend@gmail.com \
--cc=memxor@gmail.com \
--cc=neeraj.iitr10@gmail.com \
--cc=qiang.zhang@linux.dev \
--cc=rcu@vger.kernel.org \
--cc=tj@kernel.org \
--cc=urezki@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox