public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Avi Kivity <avi@redhat.com>
To: Mike Galbraith <efault@gmx.de>
Cc: Rik van Riel <riel@redhat.com>,
	kvm@vger.kernel.org, linux-kernel@vger.kernel.org,
	Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>,
	Peter Zijlstra <a.p.zijlstra@chello.nl>,
	Chris Wright <chrisw@sous-sol.org>
Subject: Re: [RFC -v2 PATCH 2/3] sched: add yield_to function
Date: Sun, 19 Dec 2010 08:21:14 +0200	[thread overview]
Message-ID: <4D0DA45A.9070600@redhat.com> (raw)
In-Reply-To: <1292699204.1181.51.camel@marge.simson.net>

On 12/18/2010 09:06 PM, Mike Galbraith wrote:
> >  >  I can see the problem, and I'm not trying to be Mr. Negative here, I'm
> >  >  only trying to point out problems I see with what's been proposed.
> >  >
> >  >  If the yielding task had a concrete fee he could pay, that would be
> >  >  fine, but he does not.
> >
> >  It does.  The yielding task is entitled to its fair share of the cpu, as
> >  modified by priority and group scheduling.  The yielding task is willing
> >  to give up some of this cpu, in return for increasing another task's
> >  share.  Other tasks would not be negatively affected by this.
>
> The donor does absolutely NOT have any claim to any specific quantum of
> any cpu at any given time.  There is no reservation, only a running
> clock.

Of course not, but we aren't interested in a specific quantum of time.  
All we want is to boost task A, and to unboost the donor in order to 
maintain fairness.  We don't want A to run now (though it would be 
nice); all we want is to run it before B.

>   If you let one task move another task's clock backward, you open
> a can of starvation worms.  This is exactly why vruntimes are monotonic.
>
> >  >  If he did have something, how often do you think it should be possible
> >  >  for task X to bribe the scheduler into selecting task Y?
> >
> >  In extreme cases, very often.  Say 100KHz.
>
> Hm, so it needs to be very cheap, and highly repeatable.
>
> What if: so you're trying to get spinners out of the way right?  You
> somehow know they're spinning, so instead of trying to boost some task,
> can you do a directed yield in terms of directing a spinner that you
> have the right to diddle to yield.  Drop his lag, and resched him.  He's
> not accomplishing anything anyway.

There are a couple of problems with this approach:

- current yield() is a no-op
- even if it weren't, the process (containing the spinner and the 
lock-holder) would yield as a whole.  If it yielded for exactly the time 
needed (until the lock holder releases the lock), it wouldn't matter, 
since the spinner isn't accomplishing anything, but we don't know what 
the exact time is.  So we want to preserve our entitlement.

With a pure yield implementation the process would get less than its 
fair share, even discounting spin time, which we'd be happy to donate to 
the rest of the system.

> If the only thing running is virtualization, and nobody else can use the
> interface being invented, all is fair, but this passing of vruntime
> around is problematic when innocent bystanders may want to play too.

We definitely want to maintain fairness.  Both with a dedicated virt 
host and with a mixed workload.

> Forcing a spinning task to parity doesn't have the same problems.

Sorry, parse failure.

> >  >  Will his
> >  >  pockets be deep enough to actually solve the problem?  Once he's
> >  >  yielded, he's out of the picture for a while if he really gave anything
> >  >  up.
> >
> >  Unless the other task donates some cpu share back.  This is exactly what
> >  will happen in those extreme cases.
>
> So vruntime donation won't work.
>
> >  >  What happens to donated entitlement when the recipient goes to
> >  >  sleep?
> >
> >  Nothing.
>
> It's vaporized by the sleep.
>
> >  >  If you try to give it back, what happens if the donor exited?
> >
> >  It's lost, too bad.
>
> Yep, so much for accounting.

What's the problem exactly?  What's the difference, system-wide, with 
the donor continuing to run for that same entitlement?  Other tasks see 
the same thing.

> >  >  Where did the entitlement come from if task A running alone on cpu A
> >  >  tosses some entitlement over the fence to his pal task B on cpu B.. and
> >  >  keeps on trucking on cpu A?  Where does that leave task C, B's
> >  >  competition?
> >
> >  Eventually C would replace A, since its share will be exhausted.  If C
> >  is pinned... good question.  How does fairness work with pinned tasks?
>
> In the case I described, C had it's pocket picked by A.

Would that happen if global fairness was maintained?

> >  >  >   Do I correctly read between the lines that CFS maintains complete
> >  >  >   fairness only on a cpu, but not globally?
> >  >
> >  >  Nothing between the lines about it.  There are N individual engines,
> >  >  coupled via load balancing.
> >
> >  Is this not seen as a major deficiency?
>
> Doesn't seem to be.  That's what SMP-nice was all about.  It's not
> perfect, but seems to work well.

I guess random perturbations cause task migrations periodically and 
things balance out.  But it seems wierd to have this devotion to 
fairness on a single cpu and completely ignore fairness on a macro level.

> >  I can understand intra-cpu scheduling decisions at 300 Hz and inter-cpu
> >  decisions at 10 Hz (or even lower, with some intermediate rate for
> >  intra-socket scheduling).  But this looks like a major deviation from
> >  fairness - instead of 33%/33%/33% you get 50%/25%/25% depending on
> >  random placement.
>
> Yeah, you have to bounce tasks around regularly to make that work out.

That's fine as long as the bounce period is significantly larger than 
the cost of a migration.

-- 
I have a truly marvellous patch that fixes the bug which this
signature is too narrow to contain.


  reply	other threads:[~2010-12-19  6:21 UTC|newest]

Thread overview: 44+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2010-12-14  3:44 [RFC -v2 PATCH 0/3] directed yield for Pause Loop Exiting Rik van Riel
2010-12-14  3:45 ` [RFC -v2 PATCH 1/3] kvm: keep track of which task is running a KVM vcpu Rik van Riel
2010-12-14  3:46 ` [RFC -v2 PATCH 2/3] sched: add yield_to function Rik van Riel
2010-12-14  6:08   ` Mike Galbraith
2010-12-14 10:24     ` Srivatsa Vaddagiri
2010-12-14 11:03       ` Mike Galbraith
2010-12-14 11:26         ` Srivatsa Vaddagiri
2010-12-14 12:47           ` Mike Galbraith
2010-12-16 19:49     ` Rik van Riel
2010-12-17  6:56       ` Mike Galbraith
2010-12-17  7:15         ` Mike Galbraith
2010-12-18 17:08           ` Avi Kivity
2010-12-18 19:13             ` Mike Galbraith
2010-12-19  6:08               ` Avi Kivity
2010-12-20 15:40           ` Rik van Riel
2010-12-20 16:04             ` Mike Galbraith
2010-12-28  5:54               ` Mike Galbraith
2010-12-28  6:08                 ` Gene Heskett
2010-12-28  6:16                   ` Mike Galbraith
2010-12-28 16:18                     ` Gene Heskett
2010-12-28 22:34                 ` Rik van Riel
2010-12-17 15:09         ` Avi Kivity
2010-12-17 19:51           ` Mike Galbraith
2010-12-18 17:02             ` Avi Kivity
2010-12-18 19:06               ` Mike Galbraith
2010-12-19  6:21                 ` Avi Kivity [this message]
2010-12-19 10:05                   ` Mike Galbraith
2010-12-19  9:19                     ` Avi Kivity
2010-12-19 11:18                       ` Mike Galbraith
2010-12-20  8:39                       ` Mike Galbraith
2010-12-20  8:45                         ` Avi Kivity
2010-12-20  8:55                           ` Mike Galbraith
2010-12-20  9:03                             ` Avi Kivity
2010-12-20  9:30                               ` Mike Galbraith
2010-12-20  9:46                                 ` Avi Kivity
2010-12-20 10:33                                   ` Mike Galbraith
2010-12-20 10:39                                     ` Avi Kivity
2010-12-20 10:46                                       ` Mike Galbraith
2010-12-20 10:49                                         ` Avi Kivity
2010-12-20 10:50                                           ` Mike Galbraith
2010-12-20 11:06                                             ` Avi Kivity
2010-12-14 12:22   ` Peter Zijlstra
2010-12-18 14:50     ` Rik van Riel
2010-12-14  3:48 ` [RFC -v2 PATCH 3/3] kvm: use yield_to instead of sleep in kvm_vcpu_on_spin Rik van Riel

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4D0DA45A.9070600@redhat.com \
    --to=avi@redhat.com \
    --cc=a.p.zijlstra@chello.nl \
    --cc=chrisw@sous-sol.org \
    --cc=efault@gmx.de \
    --cc=kvm@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=riel@redhat.com \
    --cc=vatsa@linux.vnet.ibm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox