From: Peter Zijlstra <peterz@infradead.org>
To: John Stultz <jstultz@google.com>
Cc: LKML <linux-kernel@vger.kernel.org>,
Joel Fernandes <joelagnelf@nvidia.com>,
Qais Yousef <qyousef@layalina.io>, Ingo Molnar <mingo@redhat.com>,
Juri Lelli <juri.lelli@redhat.com>,
Vincent Guittot <vincent.guittot@linaro.org>,
Dietmar Eggemann <dietmar.eggemann@arm.com>,
Valentin Schneider <vschneid@redhat.com>,
Steven Rostedt <rostedt@goodmis.org>,
Ben Segall <bsegall@google.com>,
Zimuzo Ezeozue <zezeozue@google.com>,
Mel Gorman <mgorman@suse.de>, Will Deacon <will@kernel.org>,
Waiman Long <longman@redhat.com>,
Boqun Feng <boqun.feng@gmail.com>,
"Paul E. McKenney" <paulmck@kernel.org>,
Metin Kaya <Metin.Kaya@arm.com>,
Xuewen Yan <xuewen.yan94@gmail.com>,
K Prateek Nayak <kprateek.nayak@amd.com>,
Thomas Gleixner <tglx@linutronix.de>,
Daniel Lezcano <daniel.lezcano@linaro.org>,
Suleiman Souhlal <suleiman@google.com>,
kernel-team@android.com,
Valentin Schneider <valentin.schneider@arm.com>,
Connor O'Brien <connoro@google.com>
Subject: Re: [RFC PATCH v15 7/7] sched: Start blocked_on chain processing in find_proxy_task()
Date: Mon, 17 Mar 2025 17:43:56 +0100 [thread overview]
Message-ID: <20250317164356.GB6888@noisy.programming.kicks-ass.net> (raw)
In-Reply-To: <20250312221147.1865364-8-jstultz@google.com>
On Wed, Mar 12, 2025 at 03:11:37PM -0700, John Stultz wrote:
> @@ -6668,47 +6676,138 @@ static bool proxy_deactivate(struct rq *rq, struct task_struct *donor)
> }
>
> /*
> + * Find runnable lock owner to proxy for mutex blocked donor
> + *
> + * Follow the blocked-on relation:
> + * task->blocked_on -> mutex->owner -> task...
> + *
> + * Lock order:
> + *
> + * p->pi_lock
> + * rq->lock
> + * mutex->wait_lock
> + *
> + * Returns the task that is going to be used as execution context (the one
> + * that is actually going to be run on cpu_of(rq)).
> */
> static struct task_struct *
> find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
> {
> + struct task_struct *owner = NULL;
> + struct task_struct *ret = NULL;
> + int this_cpu = cpu_of(rq);
> + struct task_struct *p;
> struct mutex *mutex;
>
> + /* Follow blocked_on chain. */
> + for (p = donor; task_is_blocked(p); p = owner) {
> + mutex = p->blocked_on;
> + /* Something changed in the chain, so pick again */
> + if (!mutex)
> + return NULL;
> /*
> + * By taking mutex->wait_lock we hold off concurrent mutex_unlock()
> + * and ensure @owner sticks around.
> */
> + raw_spin_lock(&mutex->wait_lock);
This comment -- that is only true if you kill __mutex_unlock_fast(),
which I don't think you did in the previous patches.
> +
> + /* Check again that p is blocked with wait_lock held */
> + if (mutex != __get_task_blocked_on(p)) {
> + /*
> + * Something changed in the blocked_on chain and
> + * we don't know if only at this level. So, let's
> + * just bail out completely and let __schedule
> + * figure things out (pick_again loop).
> + */
> + goto out;
> + }
> +
> + owner = __mutex_owner(mutex);
> + if (!owner) {
> + __clear_task_blocked_on(p, mutex);
> + ret = p;
> + goto out;
> + }
> +
> + if (task_cpu(owner) != this_cpu) {
> + /* XXX Don't handle migrations yet */
> + if (!proxy_deactivate(rq, donor))
> + goto deactivate_failed;
> + goto out;
> + }
> +
> + if (task_on_rq_migrating(owner)) {
> + /*
> + * One of the chain of mutex owners is currently migrating to this
> + * CPU, but has not yet been enqueued because we are holding the
> + * rq lock. As a simple solution, just schedule rq->idle to give
> + * the migration a chance to complete. Much like the migrate_task
> + * case we should end up back in find_proxy_task(), this time
> + * hopefully with all relevant tasks already enqueued.
> + */
> + raw_spin_unlock(&mutex->wait_lock);
> + return proxy_resched_idle(rq);
> + }
> +
> + if (!owner->on_rq) {
> + /* XXX Don't handle blocked owners yet */
> + if (!proxy_deactivate(rq, donor))
> + goto deactivate_failed;
> + goto out;
> + }
> +
> + if (owner->se.sched_delayed) {
> + /* XXX Don't handle delayed dequeue yet */
> + if (!proxy_deactivate(rq, donor))
> + goto deactivate_failed;
> + goto out;
> + }
> +
> + if (owner == p) {
> + /*
> + * It's possible we interleave with mutex_unlock like:
> + *
> + * lock(&rq->lock);
> + * find_proxy_task()
> + * mutex_unlock()
> + * lock(&wait_lock);
> + * donor(owner) = current->blocked_donor;
> + * unlock(&wait_lock);
> + *
> + * wake_up_q();
> + * ...
> + * ttwu_runnable()
> + * __task_rq_lock()
> + * lock(&wait_lock);
> + * owner == p
> + *
> + * Which leaves us to finish the ttwu_runnable() and make it go.
> + *
> + * So schedule rq->idle so that ttwu_runnable can get the rq lock
> + * and mark owner as running.
> + */
> + raw_spin_unlock(&mutex->wait_lock);
> + return proxy_resched_idle(rq);
> + }
>
> /*
> + * OK, now we're absolutely sure @owner is on this
> + * rq, therefore holding @rq->lock is sufficient to
> + * guarantee its existence, as per ttwu_remote().
> */
> raw_spin_unlock(&mutex->wait_lock);
> }
> +
> + WARN_ON_ONCE(owner && !owner->on_rq);
> + return owner;
> +
> +deactivate_failed:
> + /*
> + * XXX: For now, if deactivation failed, set donor
> + * as unblocked, as we aren't doing proxy-migrations
> + * yet (more logic will be needed then).
> + */
> + donor->blocked_on = NULL; /* XXX not following locking rules :( */
> out:
> raw_spin_unlock(&mutex->wait_lock);
> return NULL; /* do pick_next_task again */
Also, something like the below might be a cleanup -- I didn't check your
full series.
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -6653,7 +6653,7 @@ proxy_resched_idle(struct rq *rq)
return rq->idle;
}
-static bool proxy_deactivate(struct rq *rq, struct task_struct *donor)
+static bool __proxy_deactivate(struct rq *rq, struct task_struct *donor)
{
unsigned long state = READ_ONCE(donor->__state);
@@ -6673,6 +6673,19 @@ static bool proxy_deactivate(struct rq *
return try_to_block_task(rq, donor, state, true);
}
+static struct task_struct *proxy_deactivate(struct rq *rq, struct task_struct *donor)
+{
+ if (!__proxy_deactivate(rq, donor)) {
+ /*
+ * XXX: For now, if deactivation failed, set donor
+ * as unblocked, as we aren't doing proxy-migrations
+ * yet (more logic will be needed then).
+ */
+ donor->blocked_on = NULL;
+ }
+ return NULL;
+}
+
/*
* Find runnable lock owner to proxy for mutex blocked donor
*
@@ -6691,23 +6704,17 @@ static bool proxy_deactivate(struct rq *
static struct task_struct *
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
{
- struct task_struct *owner = NULL;
- struct task_struct *ret = NULL;
+ struct task_struct *owner, *p;
int this_cpu = cpu_of(rq);
- struct task_struct *p;
struct mutex *mutex;
/* Follow blocked_on chain. */
- for (p = donor; task_is_blocked(p); p = owner) {
- mutex = p->blocked_on;
- /* Something changed in the chain, so pick again */
- if (!mutex)
- return NULL;
+ for (p = donor; (mutex = p->blocked_on); p = owner) {
/*
* By taking mutex->wait_lock we hold off concurrent mutex_unlock()
* and ensure @owner sticks around.
*/
- raw_spin_lock(&mutex->wait_lock);
+ guard(raw_spinlock)(&mutex->wait_lock);
/* Check again that p is blocked with wait_lock held */
if (mutex != __get_task_blocked_on(p)) {
@@ -6717,22 +6724,17 @@ find_proxy_task(struct rq *rq, struct ta
* just bail out completely and let __schedule
* figure things out (pick_again loop).
*/
- goto out;
+ return NULL;
}
owner = __mutex_owner(mutex);
if (!owner) {
__clear_task_blocked_on(p, mutex);
- ret = p;
- goto out;
+ return p;
}
- if (task_cpu(owner) != this_cpu) {
- /* XXX Don't handle migrations yet */
- if (!proxy_deactivate(rq, donor))
- goto deactivate_failed;
- goto out;
- }
+ if (task_cpu(owner) != this_cpu)
+ return proxy_deactivate(rq, donor);
if (task_on_rq_migrating(owner)) {
/*
@@ -6743,22 +6745,17 @@ find_proxy_task(struct rq *rq, struct ta
* case we should end up back in find_proxy_task(), this time
* hopefully with all relevant tasks already enqueued.
*/
- raw_spin_unlock(&mutex->wait_lock);
- return proxy_resched_idle(rq);
+ goto ret_idle;
}
if (!owner->on_rq) {
/* XXX Don't handle blocked owners yet */
- if (!proxy_deactivate(rq, donor))
- goto deactivate_failed;
- goto out;
+ return proxy_deactivate(rq, donor);
}
if (owner->se.sched_delayed) {
/* XXX Don't handle delayed dequeue yet */
- if (!proxy_deactivate(rq, donor))
- goto deactivate_failed;
- goto out;
+ return proxy_deactivate(rq, donor);
}
if (owner == p) {
@@ -6784,8 +6781,7 @@ find_proxy_task(struct rq *rq, struct ta
* So schedule rq->idle so that ttwu_runnable can get the rq lock
* and mark owner as running.
*/
- raw_spin_unlock(&mutex->wait_lock);
- return proxy_resched_idle(rq);
+ goto ret_idle;
}
/*
@@ -6793,22 +6789,13 @@ find_proxy_task(struct rq *rq, struct ta
* rq, therefore holding @rq->lock is sufficient to
* guarantee its existence, as per ttwu_remote().
*/
- raw_spin_unlock(&mutex->wait_lock);
}
WARN_ON_ONCE(owner && !owner->on_rq);
return owner;
-deactivate_failed:
- /*
- * XXX: For now, if deactivation failed, set donor
- * as unblocked, as we aren't doing proxy-migrations
- * yet (more logic will be needed then).
- */
- donor->blocked_on = NULL; /* XXX not following locking rules :( */
-out:
- raw_spin_unlock(&mutex->wait_lock);
- return NULL; /* do pick_next_task again */
+ret_idle:
+ return proxy_resched_idle(rq);
}
#else /* SCHED_PROXY_EXEC */
static struct task_struct *
next prev parent reply other threads:[~2025-03-17 16:44 UTC|newest]
Thread overview: 36+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-03-12 22:11 [RFC PATCH v15 0/7] Single RunQueue Proxy Execution (v15) John Stultz
2025-03-12 22:11 ` [RFC PATCH v15 1/7] sched: Add CONFIG_SCHED_PROXY_EXEC & boot argument to enable/disable John Stultz
2025-03-13 10:09 ` Steven Rostedt
2025-03-14 0:48 ` John Stultz
2025-03-17 14:33 ` Peter Zijlstra
2025-03-17 14:44 ` John Stultz
2025-03-17 14:50 ` Peter Zijlstra
2025-03-12 22:11 ` [RFC PATCH v15 2/7] locking/mutex: Rework task_struct::blocked_on John Stultz
2025-03-13 10:13 ` Steven Rostedt
2025-03-14 6:12 ` John Stultz
2025-03-16 16:33 ` Steven Rostedt
2025-03-18 14:11 ` Masami Hiramatsu
2025-03-18 15:33 ` Lance Yang
2025-03-19 9:49 ` John Stultz
2025-03-19 12:05 ` Lance Yang
2025-03-19 8:54 ` John Stultz
2025-03-17 11:44 ` Peter Zijlstra
2025-03-12 22:11 ` [RFC PATCH v15 3/7] locking/mutex: Add p->blocked_on wrappers for correctness checks John Stultz
2025-03-12 22:11 ` [RFC PATCH v15 4/7] sched: Fix runtime accounting w/ split exec & sched contexts John Stultz
2025-03-13 10:26 ` Steven Rostedt
2025-03-15 6:05 ` John Stultz
2025-03-13 17:24 ` K Prateek Nayak
2025-03-12 22:11 ` [RFC PATCH v15 5/7] sched: Add an initial sketch of the find_proxy_task() function John Stultz
2025-03-15 16:35 ` K Prateek Nayak
2025-03-17 13:48 ` Peter Zijlstra
2025-03-12 22:11 ` [RFC PATCH v15 6/7] sched: Fix proxy/current (push,pull)ability John Stultz
2025-03-14 8:40 ` K Prateek Nayak
2025-03-15 5:10 ` John Stultz
2025-03-15 16:06 ` K Prateek Nayak
2025-03-17 14:07 ` Peter Zijlstra
2025-03-28 4:45 ` K Prateek Nayak
2025-03-12 22:11 ` [RFC PATCH v15 7/7] sched: Start blocked_on chain processing in find_proxy_task() John Stultz
2025-03-17 16:43 ` Peter Zijlstra [this message]
2025-03-18 6:09 ` Peter Zijlstra
2025-03-17 16:47 ` Peter Zijlstra
2025-03-17 16:49 ` Peter Zijlstra
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250317164356.GB6888@noisy.programming.kicks-ass.net \
--to=peterz@infradead.org \
--cc=Metin.Kaya@arm.com \
--cc=boqun.feng@gmail.com \
--cc=bsegall@google.com \
--cc=connoro@google.com \
--cc=daniel.lezcano@linaro.org \
--cc=dietmar.eggemann@arm.com \
--cc=joelagnelf@nvidia.com \
--cc=jstultz@google.com \
--cc=juri.lelli@redhat.com \
--cc=kernel-team@android.com \
--cc=kprateek.nayak@amd.com \
--cc=linux-kernel@vger.kernel.org \
--cc=longman@redhat.com \
--cc=mgorman@suse.de \
--cc=mingo@redhat.com \
--cc=paulmck@kernel.org \
--cc=qyousef@layalina.io \
--cc=rostedt@goodmis.org \
--cc=suleiman@google.com \
--cc=tglx@linutronix.de \
--cc=valentin.schneider@arm.com \
--cc=vincent.guittot@linaro.org \
--cc=vschneid@redhat.com \
--cc=will@kernel.org \
--cc=xuewen.yan94@gmail.com \
--cc=zezeozue@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox