From: Peter Zijlstra <peterz@infradead.org>
To: Mike Galbraith <umgwanakikbuti@gmail.com>
Cc: Libo Chen <libo.chen@huawei.com>,
tglx@linutronix.de, mingo@elte.hu,
LKML <linux-kernel@vger.kernel.org>,
Greg KH <gregkh@linuxfoundation.org>,
Li Zefan <lizefan@huawei.com>
Subject: Re: balance storm
Date: Tue, 27 May 2014 12:43:49 +0200 [thread overview]
Message-ID: <20140527104349.GP30445@twins.programming.kicks-ass.net> (raw)
In-Reply-To: <1401185133.5134.119.camel@marge.simpson.net>
[-- Attachment #1: Type: text/plain, Size: 1199 bytes --]
On Tue, May 27, 2014 at 12:05:33PM +0200, Mike Galbraith wrote:
> On Tue, 2014-05-27 at 11:48 +0200, Peter Zijlstra wrote:
>
> > So I suppose this is due to the select_idle_sibling() nonsense again,
> > where we assumes L3 is a fair compromise between cheap enough and
> > effective enough.
>
> Nodz.
>
> > Of course, Intel keeps growing the cpu count covered by L3 to ridiculous
> > sizes, 8 cores isn't nowhere near their top silly, which shifts the
> > balance, and there's always going to be pathological cases (like the
> > proposed workload) where its just always going to suck eggs.
>
> Test is as pathological as it gets. 15 core + SMT wouldn't be pretty.
So one thing we could maybe do is measure the cost of
select_idle_sibling(), just like we do for idle_balance() and compare
this against the tasks avg runtime.
We can go all crazy and do reduced searches; like test every n-th cpu in
the mask, or make it statistical and do a full search ever n wakeups.
Not sure what's a good approach. But L3 spanning more and more CPUs is
not something that's going to get cured anytime soon I'm afraid.
Not to mention bloody SMT which makes the whole mess worse.
[-- Attachment #2: Type: application/pgp-signature, Size: 836 bytes --]
next prev parent reply other threads:[~2014-05-27 10:44 UTC|newest]
Thread overview: 33+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-05-26 3:04 balance storm Libo Chen
2014-05-26 5:11 ` Mike Galbraith
2014-05-26 12:16 ` Libo Chen
2014-05-26 14:19 ` Mike Galbraith
2014-05-27 7:56 ` Libo Chen
2014-05-27 9:55 ` Mike Galbraith
2014-05-27 12:50 ` Libo Chen
2014-05-27 13:20 ` Mike Galbraith
2014-05-28 1:04 ` Libo Chen
2014-05-28 1:53 ` Mike Galbraith
2014-05-28 6:54 ` Libo Chen
2014-05-28 8:16 ` Mike Galbraith
2014-05-28 9:08 ` Thomas Gleixner
2014-05-28 10:30 ` Peter Zijlstra
2014-05-28 10:52 ` Borislav Petkov
2014-05-28 11:43 ` Libo Chen
2014-05-28 11:55 ` Mike Galbraith
2014-05-29 7:58 ` Libo Chen
2014-05-29 7:57 ` Libo Chen
2014-05-27 20:53 ` Thomas Gleixner
2014-05-28 1:06 ` Libo Chen
2014-05-26 7:56 ` Mike Galbraith
2014-05-26 11:49 ` Libo Chen
2014-05-26 14:03 ` Mike Galbraith
2014-05-27 7:44 ` Libo Chen
2014-05-27 8:12 ` Mike Galbraith
2014-05-27 9:48 ` Peter Zijlstra
2014-05-27 10:05 ` Mike Galbraith
2014-05-27 10:43 ` Peter Zijlstra [this message]
2014-05-27 10:55 ` Mike Galbraith
2014-05-27 12:56 ` Libo Chen
2014-05-27 12:55 ` Libo Chen
2014-05-27 13:13 ` Peter Zijlstra
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20140527104349.GP30445@twins.programming.kicks-ass.net \
--to=peterz@infradead.org \
--cc=gregkh@linuxfoundation.org \
--cc=libo.chen@huawei.com \
--cc=linux-kernel@vger.kernel.org \
--cc=lizefan@huawei.com \
--cc=mingo@elte.hu \
--cc=tglx@linutronix.de \
--cc=umgwanakikbuti@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox