From: Nick Piggin <nickpiggin@yahoo.com.au>
To: "Siddha, Suresh B" <suresh.b.siddha@intel.com>
Cc: mingo@elte.hu, akpm@osdl.org, linux-kernel@vger.kernel.org,
steiner@sgi.com
Subject: Re: [Patch] don't kick ALB in the presence of pinned task
Date: Tue, 02 Aug 2005 16:09:17 +1000 [thread overview]
Message-ID: <42EF0E0D.8000906@yahoo.com.au> (raw)
In-Reply-To: <20050801174221.B11610@unix-os.sc.intel.com>
[-- Attachment #1: Type: text/plain, Size: 745 bytes --]
Siddha, Suresh B wrote:
> Jack Steiner brought this issue at my OLS talk.
>
> Take a scenario where two tasks are pinned to two HT threads in a physical
> package. Idle packages in the system will keep kicking migration_thread
> on the busy package with out any success.
>
> We will run into similar scenarios in the presence of CMP/NUMA.
>
> Patch appended.
>
Hmm, I would have hoped the new "all_pinned" logic should have
handled this case properly. Are you actually seeing this happen?
I have a patch here which I still need to do more testing with,
which might help performance on HT systems.
I found that idle siblings could cause SMP and NUMA balancing to
be too aggressive in some cases.
Thanks,
Nick
--
SUSE Labs, Novell Inc.
[-- Attachment #2: sched-opt-ht.patch --]
[-- Type: text/plain, Size: 2941 bytes --]
If an idle sibling of an HT queue encounters a busy sibling, then
make higher level load balancing of the non-idle variety.
Index: linux-2.6/kernel/sched.c
===================================================================
--- linux-2.6.orig/kernel/sched.c 2005-07-29 19:30:39.000000000 +1000
+++ linux-2.6/kernel/sched.c 2005-07-29 19:35:01.000000000 +1000
@@ -1889,7 +1889,7 @@ out:
*/
static struct sched_group *
find_busiest_group(struct sched_domain *sd, int this_cpu,
- unsigned long *imbalance, enum idle_type idle)
+ unsigned long *imbalance, enum idle_type idle, int *sd_idle)
{
struct sched_group *busiest = NULL, *this = NULL, *group = sd->groups;
unsigned long max_load, avg_load, total_load, this_load, total_pwr;
@@ -1914,6 +1914,9 @@ find_busiest_group(struct sched_domain *
avg_load = 0;
for_each_cpu_mask(i, group->cpumask) {
+ if (*sd_idle && !idle_cpu(i))
+ *sd_idle = 0;
+
/* Bias balancing toward cpus of our domain */
if (local_group)
load = target_load(i, load_idx);
@@ -2057,11 +2060,15 @@ static int load_balance(int this_cpu, ru
unsigned long imbalance;
int nr_moved, all_pinned = 0;
int active_balance = 0;
+ int sd_idle = 0;
+
+ if (idle != NOT_IDLE && sd->flags & SD_SHARE_CPUPOWER)
+ sd_idle = 1;
spin_lock(&this_rq->lock);
schedstat_inc(sd, lb_cnt[idle]);
- group = find_busiest_group(sd, this_cpu, &imbalance, idle);
+ group = find_busiest_group(sd, this_cpu, &imbalance, idle, &sd_idle);
if (!group) {
schedstat_inc(sd, lb_nobusyg[idle]);
goto out_balanced;
@@ -2136,6 +2143,8 @@ static int load_balance(int this_cpu, ru
sd->balance_interval *= 2;
}
+ if (!nr_moved && !sd_idle && sd->flags & SD_SHARE_CPUPOWER)
+ return -1;
return nr_moved;
out_balanced:
@@ -2149,6 +2158,8 @@ out_balanced:
(sd->balance_interval < sd->max_interval))
sd->balance_interval *= 2;
+ if (!sd_idle && sd->flags & SD_SHARE_CPUPOWER)
+ return -1;
return 0;
}
@@ -2166,9 +2177,13 @@ static int load_balance_newidle(int this
runqueue_t *busiest = NULL;
unsigned long imbalance;
int nr_moved = 0;
+ int sd_idle = 0;
+ if (sd->flags & SD_SHARE_CPUPOWER)
+ sd_idle = 1;
+
schedstat_inc(sd, lb_cnt[NEWLY_IDLE]);
- group = find_busiest_group(sd, this_cpu, &imbalance, NEWLY_IDLE);
+ group = find_busiest_group(sd, this_cpu, &imbalance, NEWLY_IDLE, &sd_idle);
if (!group) {
schedstat_inc(sd, lb_nobusyg[NEWLY_IDLE]);
goto out_balanced;
@@ -2193,15 +2208,19 @@ static int load_balance_newidle(int this
spin_unlock(&busiest->lock);
}
- if (!nr_moved)
+ if (!nr_moved) {
schedstat_inc(sd, lb_failed[NEWLY_IDLE]);
- else
+ if (!sd_idle && sd->flags & SD_SHARE_CPUPOWER)
+ return -1;
+ } else
sd->nr_balance_failed = 0;
return nr_moved;
out_balanced:
schedstat_inc(sd, lb_balanced[NEWLY_IDLE]);
+ if (!sd_idle && sd->flags & SD_SHARE_CPUPOWER)
+ return -1;
sd->nr_balance_failed = 0;
return 0;
}
next prev parent reply other threads:[~2005-08-02 6:09 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2005-08-02 0:42 [Patch] don't kick ALB in the presence of pinned task Siddha, Suresh B
2005-08-02 6:09 ` Nick Piggin [this message]
2005-08-02 9:43 ` Ingo Molnar
2005-08-02 10:06 ` Nick Piggin
2005-08-02 21:12 ` Siddha, Suresh B
2005-08-02 9:27 ` Ingo Molnar
2005-08-09 23:08 ` allow the load to grow upto its cpu_power (was Re: [Patch] don't kick ALB in the presence of pinned task) Siddha, Suresh B
2005-08-10 0:27 ` Nick Piggin
2005-08-10 2:03 ` Siddha, Suresh B
2005-08-11 3:09 ` Nick Piggin
2005-08-11 18:14 ` Siddha, Suresh B
2005-08-11 23:49 ` Nick Piggin
2005-08-12 0:39 ` Siddha, Suresh B
2005-08-12 1:24 ` Nick Piggin
2005-08-12 1:44 ` Siddha, Suresh B
2005-08-10 7:16 ` Ingo Molnar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=42EF0E0D.8000906@yahoo.com.au \
--to=nickpiggin@yahoo.com.au \
--cc=akpm@osdl.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=steiner@sgi.com \
--cc=suresh.b.siddha@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox