Linux Power Management development
 help / color / mirror / Atom feed
From: Qais Yousef <qyousef@layalina.io>
To: Ingo Molnar <mingo@kernel.org>
Cc: "Rafael J. Wysocki" <rafael@kernel.org>,
	Viresh Kumar <viresh.kumar@linaro.org>,
	Peter Zijlstra <peterz@infradead.org>,
	Vincent Guittot <vincent.guittot@linaro.org>,
	Juri Lelli <juri.lelli@redhat.com>,
	Steven Rostedt <rostedt@goodmis.org>,
	Dietmar Eggemann <dietmar.eggemann@arm.com>,
	Ben Segall <bsegall@google.com>, Mel Gorman <mgorman@suse.de>,
	Valentin Schneider <vschneid@redhat.com>,
	Christian Loehle <christian.loehle@arm.com>,
	Hongyan Xia <hongyan.xia2@arm.com>,
	John Stultz <jstultz@google.com>,
	Anjali K <anjalik@linux.ibm.com>,
	linux-pm@vger.kernel.org, linux-kernel@vger.kernel.org
Subject: Re: [PATCH v8] sched: Consolidate cpufreq updates
Date: Sun, 23 Feb 2025 23:17:55 +0000	[thread overview]
Message-ID: <20250223231755.cvjcsl3jedkgpohc@airbuntu> (raw)
In-Reply-To: <Z7rm2XRqhCM8m9IU@gmail.com>

On 02/23/25 10:14, Ingo Molnar wrote:
> 
> * Qais Yousef <qyousef@layalina.io> wrote:
> 
> > On 02/21/25 16:47, Ingo Molnar wrote:
> > > 
> > > * Qais Yousef <qyousef@layalina.io> wrote:
> > > 
> > > > ---
> > > >  include/linux/sched/cpufreq.h    |   4 +-
> > > >  kernel/sched/core.c              | 116 +++++++++++++++++++++++++++--
> > > >  kernel/sched/cpufreq_schedutil.c | 122 +++++++++++++++++++------------
> > > >  kernel/sched/deadline.c          |  10 ++-
> > > >  kernel/sched/fair.c              |  84 +++++++++------------
> > > >  kernel/sched/rt.c                |   8 +-
> > > >  kernel/sched/sched.h             |   9 ++-
> > > >  kernel/sched/syscalls.c          |  30 ++++++--
> > > >  8 files changed, 266 insertions(+), 117 deletions(-)
> > > 
> > > The changelog is rather long, and the diffstat is non-trivial.
> > > 
> > > Could you please split this up into multiple patches?
> > 
> > Sure. I did consider that but what stopped me is that I couldn't see 
> > how I could break them into independent patches. A lot of corner 
> > cases needed to be addressed and if I moved them to their own patches 
> > I'd potentially break bisectability of this code. If this is not a 
> > problem then I can see how I can do a better split. If it is a 
> > problem, I'll still try to think it over but it might require a bit 
> > of stretching. But I admit I didn't try to think it over that hard.
> 
> Yeah, so bisectability should definitely be preserved.
> 
> I had a quick look, and these changes look fairly easy to split up to 
> reduce size/complexity of individual patches. The following split looks 
> pretty natural:
> 
>  # ============{ Preparatory changes with no change in functionality: }=========>
> 
>  [PATCH 1/9] Extend check_class_changed() with the 'class_changed' return bool
>              # But don't use it at call sites yet
> 
>  [PATCH 2/9] Introduce and maintain the sugov_cpu::last_iowait_update metric
>              # But don't use it yet
> 
>  [PATCH 3/9] Extend sugov_iowait_apply() with a 'flags' parameter
>              # But don't use it yet internally
> 
>  [PATCH 4/9] Extend sugov_next_freq_shared() with the 'flags' parameter
>              # But don't use it yet internally
> 
>  [PATCH 5/9] Clean up the enqueue_task_fair() control flow to make it easier to extend
>              # This adds the goto restructuring but doesn't change functionality
> 
>  [PATCH 6/9] Introduce and maintain the cfs_rq::decayed flag
>              # But don't use it yet
> 
>  [PATCH 7/8] Extend __setscheduler_uclamp() with the 'update_cpufreq' return bool
>              # But don't use it yet
> 
>  # ============{ Behavioral changes: }==========>
> 
>  [PATCH 8/9] Change sugov_iowait_apply() behavior
>  [PATCH 9/9] Change sugov_next_freq_shared() bahavior
> 
>  ... etc.
> 
> This is only a quick stab at the most trivial split-ups, it's not a 
> complete list, and I saw other opportunities for split-up too. Please 
> make these changes as finegrained as possible, as it changes behavior 
> and there is a fair chance of behavioral regressions down the road - 
> especially as the patch itself notes that even the new logic isn't 
> perfect yet.
> 
> If the behavioral changes can be split into further steps, that would 
> be preferable too.

Okay. I was focused on the functionality changes. Let me re-order and do more
preparatory work without introducing problems that I know will affect
performance bisectability. Thanks for helping to shift focus, I admit I had
a mental block on breaking this down to smaller pieces.

> 
> Also:
> 
>  - Please make the rq->cfs.decayed logic unconditional on UP too, even 
>    if it's not used. This reduces some of the ugly #ifdeffery AFAICS.
> 
>  - Please don't add prototypes for internal static functions like 
>    __update_cpufreq_ctx_switch(), define the functions in the right 
>    order instead.
> 
>  - Also, please read your comments and fix typos:
> 
> +                * This logic relied on PELT signal decays happening once every
> +                * 1ms. But due to changes to how updates are done now, we can
> +                * end up with more request coming up leading to iowait boost
> +                * to be prematurely reduced. Make the assumption explicit
> +                * until we improve the iowait boost logic to be better in
> +                * general as it is due for an overhaul.
> 
>   s/request
>    /requests
> 
> +        * We want to update cpufreq at context switch, but on systems with
> +        * long TICK values, this can happen after a long time while more tasks
> +        * would have been added meanwhile leaving us potentially running at
> +        * inadequate frequency for extended period of time.
> 
>   Either 'an inadequate frequency' or 'inadequate frequencies'.
> 
> +        * This logic should only apply when new fair task was added to the
> +        * CPU, we'd want to defer to context switch as much as possible, but
> +        * to avoid the potential delays mentioned above, let's check if this
> +        * additional tasks warrants sending an update sooner.
> 
>   s/when new fair task
>    /when a new fair task
> 
>   s/this additional tasks
>    /this additional task
> 
> (I haven't checked the spelling exhaustively, there might be more.)

Will do.

Thanks!

--
Qais Yousef

      reply	other threads:[~2025-02-23 23:18 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-02-09 23:52 [PATCH v8] sched: Consolidate cpufreq updates Qais Yousef
2025-02-21 15:47 ` Ingo Molnar
2025-02-23  0:03   ` Qais Yousef
2025-02-23  9:14     ` Ingo Molnar
2025-02-23 23:17       ` Qais Yousef [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20250223231755.cvjcsl3jedkgpohc@airbuntu \
    --to=qyousef@layalina.io \
    --cc=anjalik@linux.ibm.com \
    --cc=bsegall@google.com \
    --cc=christian.loehle@arm.com \
    --cc=dietmar.eggemann@arm.com \
    --cc=hongyan.xia2@arm.com \
    --cc=jstultz@google.com \
    --cc=juri.lelli@redhat.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-pm@vger.kernel.org \
    --cc=mgorman@suse.de \
    --cc=mingo@kernel.org \
    --cc=peterz@infradead.org \
    --cc=rafael@kernel.org \
    --cc=rostedt@goodmis.org \
    --cc=vincent.guittot@linaro.org \
    --cc=viresh.kumar@linaro.org \
    --cc=vschneid@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox