From: Steven Rostedt <rostedt@goodmis.org>
To: Tejun Heo <tj@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>,
Ingo Molnar <mingo@redhat.com>,
Juri Lelli <juri.lelli@redhat.com>,
Vincent Guittot <vincent.guittot@linaro.org>,
Dietmar Eggemann <dietmar.eggemann@arm.com>,
Ben Segall <bsegall@google.com>, Mel Gorman <mgorman@suse.de>,
Valentin Schneider <vschneid@redhat.com>,
K Prateek Nayak <kprateek.nayak@amd.com>,
Kyle McMartin <jkkm@meta.com>,
linux-kernel@vger.kernel.org, stable@vger.kernel.org,
Linux RT Development <linux-rt-devel@lists.linux.dev>,
Clark Williams <williams@redhat.com>,
Sebastian Andrzej Siewior <bigeasy@linutronix.de>,
John Kacur <jkacur@redhat.com>
Subject: Re: [PATCH sched/core] sched/rt: Fix RT_PUSH_IPI soft lockup loop
Date: Wed, 13 May 2026 20:24:32 -0400 [thread overview]
Message-ID: <20260513202432.18dd7b9f@gandalf.local.home> (raw)
In-Reply-To: <20260513193914.1593369-1-tj@kernel.org>
On Wed, 13 May 2026 09:39:14 -1000
Tejun Heo <tj@kernel.org> wrote:
> So, here's a capture from a synthetic reproducer that I think
> models the dynamic and reaches the same end state.
Synthetic capture is fine.
>
> Test box, 192 CPUs, kernel without the fix:
>
> - Per-target hrtimer (HRTIMER_MODE_REL_PINNED_HARD) fires every
> 750us. Each fire schedules one tasklet round-robin from a pool
> of 20k distinct tasklets. Each tasklet body is a 500us cpu_relax
> loop, standing in for "process one item of softirq work".
So you are running a softirq for 500us every 750us?
This basically prevents any task from running on these CPUs while the
softirq is executing.
>
> - Storm driver: 190 SCHED_FIFO-50 nanosleep loops on non-target
> CPUs drive tell_cpu_to_push from balance_rt. Two synthetic
> psimon-shaped kthreads (FIFO 1) bound to the targets to pin
> them into rto_mask.
What exactly are these synthetic kthreads doing. Have code to share?
>
> Baseline (no storm helpers): ~85% softirq util, no lockup, runs
> indefinitely. The reproducer's baseline is higher than production -
> my guess is we need to scrape up against capacity to grow a backlog
> with the fixed-shape workload here, while production gets the same
> effect from bursty arrivals during brief slowdowns.
>
> With the storm: walker IPI overhead stretches each tasklet body
> from 500us to ~1.1ms. Service rate drops below arrival, backlog
> grows ~430/s. After ~46s, one tasklet_action_common snapshot has
> ~20k tasklets which it processes serially in BH-disabled softirq
> context. That's ~22s uninterruptible, watchdog fires.
The IPI walker should only go to the CPUs with overloaded RT tasks. Are you
making all the CPUS have overloaded RT tasks?
>
> Six soft-lockups in a 120s run:
>
> [61125.38] BUG: soft lockup - CPU#95 stuck for 22s! [kworker/95:0]
> [61145.38] BUG: soft lockup - CPU#47 stuck for 45s! [migration/47]
> [61173.38] BUG: soft lockup - CPU#47 stuck for 71s! [migration/47]
> [61197.38] BUG: soft lockup - CPU#95 stuck for 22s! [migration/95]
> [61209.38] BUG: soft lockup - CPU#47 stuck for 21s! [kworker/47:1]
> [61225.38] BUG: soft lockup - CPU#95 stuck for 48s! [migration/95]
>
> Stack at fire:
>
> rt_storm_wedge_fn+0x22/0xe0
> tasklet_action_common+0x100/0x2b0
> handle_softirqs+0xbe/0x280
> __irq_exit_rcu+0x47/0x100
> sysvec_apic_timer_interrupt+0x3a/0x80 <- watchdog hrtimer
> asm_sysvec_apic_timer_interrupt+0x16/0x20
> RIP: 0033:0x... <- user task (rt_storm_hog)
>
> Trace captured with your event list plus IPI:
>
> -e sched_switch -e sched_waking -e irq -e workqueue -e ipi
> -e irq_vectors:call_function_single_entry/exit
> -e irq_vectors:irq_work_entry/exit
> -e irq_vectors:reschedule_entry/exit
> -e irq_vectors:local_timer_entry/exit
>
> Sliced to a 17s window around the first RCU stall + first
> soft-lockup, filtered to CPUs 47 and 95, gzipped text (~11MB):
>
> https://drive.google.com/file/d/11AN6dyvOWiZLVNEEuVtQieRyAxJYbCbt/view?usp=sharing
>
So this is showing that the IPI logic is just extending the softirq work
load to something greater than the period of execution and causing a live
lock of softirqs.
This still doesn't explain to me why the current process is of a lower
priority than a waiting RT task.
I'm really starting to think you are fixing a symptom and not the cause.
-- Steve
next prev parent reply other threads:[~2026-05-14 0:24 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-05-06 23:57 [PATCH sched/core] sched/rt: Fix RT_PUSH_IPI soft lockup loop Tejun Heo
2026-05-07 14:14 ` Peter Zijlstra
2026-05-11 19:33 ` Tejun Heo
2026-05-12 15:37 ` Steven Rostedt
2026-05-12 18:07 ` Tejun Heo
2026-05-12 21:28 ` Steven Rostedt
2026-05-13 19:39 ` Tejun Heo
2026-05-14 0:24 ` Steven Rostedt [this message]
2026-05-14 0:53 ` Tejun Heo
2026-05-14 1:31 ` Steven Rostedt
2026-05-14 1:42 ` Tejun Heo
2026-05-14 2:01 ` Steven Rostedt
2026-05-14 4:48 ` Tejun Heo
2026-05-14 14:03 ` Steven Rostedt
2026-05-12 20:10 ` Valentin Schneider
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260513202432.18dd7b9f@gandalf.local.home \
--to=rostedt@goodmis.org \
--cc=bigeasy@linutronix.de \
--cc=bsegall@google.com \
--cc=dietmar.eggemann@arm.com \
--cc=jkacur@redhat.com \
--cc=jkkm@meta.com \
--cc=juri.lelli@redhat.com \
--cc=kprateek.nayak@amd.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-rt-devel@lists.linux.dev \
--cc=mgorman@suse.de \
--cc=mingo@redhat.com \
--cc=peterz@infradead.org \
--cc=stable@vger.kernel.org \
--cc=tj@kernel.org \
--cc=vincent.guittot@linaro.org \
--cc=vschneid@redhat.com \
--cc=williams@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox