From: "Martin J. Bligh" <mbligh@aracnet.com>
To: Rick Lindsley <ricklind@us.ibm.com>, Anton Blanchard <anton@samba.org>
Cc: piggin@cyberone.com.au, akpm@osdl.org,
linux-kernel@vger.kernel.org, dvhltc@us.ibm.com
Subject: Re: [PATCH] Load balancing problem in 2.6.2-mm1
Date: Fri, 06 Feb 2004 10:39:31 -0800 [thread overview]
Message-ID: <207100000.1076092771@flay> (raw)
In-Reply-To: <200402061815.i16IFhY07073@owlet.beaverton.ibm.com>
> Good stuff, I just gave the patch a spin and things seem a little
> calmer. However Im still seeing a lot of balancing going on within a
> node.
>
> This is a clearly recognizable edge case, so I'll try drawing this up on
> some paper and see if I can suggest another patch. There's no good reason
> to move one lone process from a particular processor to another idle one.
>
> But it also approaches a question that's come up before: if you have 2
> tasks on processor A and 1 on processor B, do you move one from A to B?
> One argument is that the two tasks on A will take twice as long as
> the one on B if you do nothing. But another says that bouncing a task
> around can't correct the overall imbalance and so is wasteful. I know
> of benchmarks where both behaviors are considered important. Thoughts?
It's the classic fairness vs throughput thing we've argued about before.
Most workloads don't have that static a number of processes, but it
probably does need to do it if the imbalance is persistent ... but much
more reluctantly than normal balancing. See the patch I sent out a bit
earlier to test it - that may be *too* extreme in the other direction,
but it should confirm what's going on, at least.
M.
next prev parent reply other threads:[~2004-02-06 18:39 UTC|newest]
Thread overview: 37+ messages / expand[flat|nested] mbox.gz Atom feed top
2004-02-06 9:24 [PATCH] Load balancing problem in 2.6.2-mm1 Rick Lindsley
2004-02-06 9:38 ` Nick Piggin
2004-02-06 18:13 ` Rick Lindsley
2004-02-06 21:57 ` Nick Piggin
2004-02-06 22:30 ` Rick Lindsley
2004-02-06 22:40 ` Nick Piggin
2004-02-06 22:49 ` Rick Lindsley
2004-02-06 23:08 ` Nick Piggin
2004-02-06 10:30 ` Anton Blanchard
2004-02-06 18:15 ` Rick Lindsley
2004-02-06 18:39 ` Martin J. Bligh [this message]
2004-02-06 22:02 ` Nick Piggin
2004-02-06 22:34 ` Rick Lindsley
2004-02-06 22:48 ` Nick Piggin
2004-02-06 22:42 ` Martin J. Bligh
2004-02-06 22:53 ` Nick Piggin
2004-02-06 23:11 ` Rick Lindsley
2004-02-06 23:20 ` Nick Piggin
2004-02-06 23:33 ` Martin J. Bligh
2004-02-06 23:41 ` Nick Piggin
2004-02-06 23:47 ` Martin J. Bligh
2004-02-07 0:11 ` Nick Piggin
2004-02-07 0:25 ` Martin J. Bligh
2004-02-07 0:31 ` Nick Piggin
2004-02-07 9:50 ` Anton Blanchard
2004-02-08 0:40 ` Rick Lindsley
2004-02-08 1:12 ` Anton Blanchard
2004-02-08 1:21 ` Nick Piggin
2004-02-08 1:41 ` Anton Blanchard
2004-02-08 3:20 ` Nick Piggin
2004-02-08 3:57 ` Anton Blanchard
2004-02-08 4:05 ` Nick Piggin
2004-02-08 12:14 ` Anton Blanchard
2004-02-08 1:22 ` Anton Blanchard
2004-02-09 16:37 ` Timothy Miller
2004-02-09 16:43 ` Martin J. Bligh
2004-02-06 18:33 ` Martin J. Bligh
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=207100000.1076092771@flay \
--to=mbligh@aracnet.com \
--cc=akpm@osdl.org \
--cc=anton@samba.org \
--cc=dvhltc@us.ibm.com \
--cc=linux-kernel@vger.kernel.org \
--cc=piggin@cyberone.com.au \
--cc=ricklind@us.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox