From: William Lee Irwin III <wli@holomorphy.com>
To: Srivatsa Vaddagiri <vatsa@in.ibm.com>
Cc: Ingo Molnar <mingo@elte.hu>,
efault@gmx.de, tingy@cs.umass.edu, linux-kernel@vger.kernel.org
Subject: Re: fair clock use in CFS
Date: Mon, 14 May 2007 03:29:29 -0700 [thread overview]
Message-ID: <20070514102929.GL31925@holomorphy.com> (raw)
In-Reply-To: <20070514083358.GA29775@in.ibm.com>
On Mon, May 14, 2007 at 02:03:58PM +0530, Srivatsa Vaddagiri wrote:
> I have been brooding over how fair clock is computed/used in
> CFS and thought I would ask the experts to avoid wrong guesses!
> As I understand, fair_clock is a monotonously increasing clock which
> advances at a pace inversely proportional to the load on the runqueue.
> If load = 1 (task), it will advance at same pace as wall clock, as
> load increases it advances slower than wall clock.
> In addition, following calculations depend on fair clock: task's wait
> time on runqueue and sleep time outside the runqueue (both reflected in
> p->wait_run_time).
It's not hard to see that that's a mistake. The great thing about virtual
deadline schedulers, though, is that all it takes to completely rewrite
the policy is changing the numbers calculated for these things.
On Mon, May 14, 2007 at 02:03:58PM +0530, Srivatsa Vaddagiri wrote:
> Few questions that come up are:
> 1. Why can't fair clock be same as wall clock at all times? i.e fair
> clock progresses at same pace as wall clock independent of the load on
> the runqueue.
> It would still give the ability to measure time spent waiting on runqueue
> or sleeping and use that calculated time to give latency/bandwidth
> credit?
> In case of EEVDF, the use of virtual clock seems more
> understandable, if we consider the fact that each client gets 'wi' real
> time units in 1 virtual time unit. That doesnt seem to be the case in
> CFS as Ting Yang explained +/- lags here
> http://lkml.org/lkml/2007/5/2/612 ..
It's not just more understandable, it doesn't break down with
increasing numbers of tasks.
On Mon, May 14, 2007 at 02:03:58PM +0530, Srivatsa Vaddagiri wrote:
> 2. Preemption granularity - sysctl_sched_granularity
> This seems to be measured in the fair clock scale rather than
> wall clock scale. As a consequence of this, the time taken
> for a task to relinquish to competetion is dependent on number N
> of tasks? For ex: if there a million cpu hungry tasks, then the
> time taken to switch between two tasks is more compared to the
> case where just two cpu hungry tasks are running. Is there
> any advantage of using fair clock scale to detect preemtion points?
I'm not convinced this is a great way to mitigate context switch
overhead. I'd recommend heuristically adjusting the latency parameters
(l_i) to try to mitigate context switch overhead or otherwise express
the tradeoff between latency and throughput instead of introducing an
arbitrary delay like sched_granularity_ns. I suspect it'll have to
restart from a point much closer to EEVDF for that to be effective.
-- wli
next prev parent reply other threads:[~2007-05-14 10:29 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2007-05-14 8:33 fair clock use in CFS Srivatsa Vaddagiri
2007-05-14 10:29 ` William Lee Irwin III [this message]
2007-05-14 10:31 ` Ingo Molnar
2007-05-14 11:05 ` William Lee Irwin III
2007-05-14 11:22 ` Srivatsa Vaddagiri
2007-05-14 11:20 ` William Lee Irwin III
2007-05-14 12:04 ` Ingo Molnar
2007-05-14 23:57 ` William Lee Irwin III
2007-05-14 20:20 ` Ting Yang
2007-05-14 11:50 ` Ingo Molnar
2007-05-14 14:31 ` Daniel Hazelton
2007-05-14 15:02 ` Srivatsa Vaddagiri
2007-05-14 15:08 ` Ingo Molnar
2007-05-15 2:59 ` David Schwartz
2007-05-14 21:24 ` Ting Yang
2007-05-15 0:57 ` Ting Yang
2007-05-14 23:23 ` William Lee Irwin III
2007-05-14 11:10 ` Ingo Molnar
2007-05-14 13:04 ` Srivatsa Vaddagiri
2007-05-14 13:15 ` Ingo Molnar
-- strict thread matches above, loose matches on Subject: below --
2007-05-14 15:02 Al Boldi
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20070514102929.GL31925@holomorphy.com \
--to=wli@holomorphy.com \
--cc=efault@gmx.de \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=tingy@cs.umass.edu \
--cc=vatsa@in.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox