From: John Hawkes <hawkes@google.engr.sgi.com>
To: John Hawkes <hawkes@oss.sgi.com>, Nick Piggin <nickpiggin@yahoo.com.au>
Cc: Ingo Molnar <mingo@elte.hu>,
hawkes@sgi.com, jbarnes@sgi.com, linux-kernel@vger.kernel.org,
akpm@osdl.org
Subject: Re: [PATCH, 2.6.9] improved load_balance() tolerance for pinned tasks
Date: Fri, 22 Oct 2004 12:38:50 -0700 (PDT) [thread overview]
Message-ID: <200410221938.MAA52152@google.engr.sgi.com> (raw)
In-Reply-To: 200410201936.i9KJa4FF026174@oss.sgi.com
> John and I also discussed a slightly more sophisticated algorithm
> for this that would try to be a bit smarter about the reason for
> balance failures (described in the changelog).
>
> John, this works quite well here, would you have any objections
> using it instead? Does it cure your problems there?
No, your variation doesn't solve the problem. This variation of your
patch does, however, solve the problem. The difference is in
move_tasks():
Signed-off-by: John Hawkes <hawkes@sgi.com>
linux-2.6-npiggin/kernel/sched.c | 34 +++++++++++++++++++++++-----------
1 files changed, 23 insertions(+), 11 deletions(-)
Index: linux/kernel/sched.c
===================================================================
--- linux.orig/kernel/sched.c 2004-10-22 09:11:12.000000000 -0700
+++ linux/kernel/sched.c 2004-10-22 11:45:10.000000000 -0700
@@ -1770,7 +1770,7 @@
*/
static inline
int can_migrate_task(task_t *p, runqueue_t *rq, int this_cpu,
- struct sched_domain *sd, enum idle_type idle)
+ struct sched_domain *sd, enum idle_type idle, int *pinned)
{
/*
* We do not migrate tasks that are:
@@ -1780,8 +1780,10 @@
*/
if (task_running(rq, p))
return 0;
- if (!cpu_isset(this_cpu, p->cpus_allowed))
+ if (!cpu_isset(this_cpu, p->cpus_allowed)) {
+ *pinned++;
return 0;
+ }
/* Aggressive migration if we've failed balancing */
if (idle == NEWLY_IDLE ||
@@ -1802,11 +1804,11 @@
*/
static int move_tasks(runqueue_t *this_rq, int this_cpu, runqueue_t *busiest,
unsigned long max_nr_move, struct sched_domain *sd,
- enum idle_type idle)
+ enum idle_type idle, int *all_pinned)
{
prio_array_t *array, *dst_array;
struct list_head *head, *curr;
- int idx, pulled = 0;
+ int idx, examined = 0, pulled = 0, pinned = 0;
task_t *tmp;
if (max_nr_move <= 0 || busiest->nr_running <= 1)
@@ -1850,7 +1852,8 @@
curr = curr->prev;
- if (!can_migrate_task(tmp, busiest, this_cpu, sd, idle)) {
+ examined++;
+ if (!can_migrate_task(tmp, busiest, this_cpu, sd, idle, &pinned)) {
if (curr != head)
goto skip_queue;
idx++;
@@ -1876,6 +1879,8 @@
goto skip_bitmap;
}
out:
+ if (unlikely(examined && examined == pinned))
+ *all_pinned = 1;
return pulled;
}
@@ -2056,7 +2061,7 @@
struct sched_group *group;
runqueue_t *busiest;
unsigned long imbalance;
- int nr_moved;
+ int nr_moved, all_pinned;
spin_lock(&this_rq->lock);
schedstat_inc(sd, lb_cnt[idle]);
@@ -2095,11 +2100,16 @@
*/
double_lock_balance(this_rq, busiest);
nr_moved = move_tasks(this_rq, this_cpu, busiest,
- imbalance, sd, idle);
+ imbalance, sd, idle,
+ &all_pinned);
spin_unlock(&busiest->lock);
}
- spin_unlock(&this_rq->lock);
+ /* All tasks on this runqueue were pinned by CPU affinity */
+ if (unlikely(all_pinned))
+ goto out_balanced;
+ spin_unlock(&this_rq->lock);
+
if (!nr_moved) {
schedstat_inc(sd, lb_failed[idle]);
sd->nr_balance_failed++;
@@ -2154,7 +2164,7 @@
struct sched_group *group;
runqueue_t *busiest = NULL;
unsigned long imbalance;
- int nr_moved = 0;
+ int nr_moved = 0, all_pinned;
schedstat_inc(sd, lb_cnt[NEWLY_IDLE]);
group = find_busiest_group(sd, this_cpu, &imbalance, NEWLY_IDLE);
@@ -2174,7 +2184,7 @@
schedstat_add(sd, lb_imbalance[NEWLY_IDLE], imbalance);
nr_moved = move_tasks(this_rq, this_cpu, busiest,
- imbalance, sd, NEWLY_IDLE);
+ imbalance, sd, NEWLY_IDLE, &all_pinned);
if (!nr_moved)
schedstat_inc(sd, lb_failed[NEWLY_IDLE]);
@@ -2236,6 +2246,7 @@
cpumask_t tmp;
runqueue_t *rq;
int push_cpu = 0;
+ int all_pinned;
if (group == busy_group)
goto next_group;
@@ -2261,7 +2272,8 @@
if (unlikely(busiest == rq))
goto next_group;
double_lock_balance(busiest, rq);
- if (move_tasks(rq, push_cpu, busiest, 1, sd, IDLE)) {
+ if (move_tasks(rq, push_cpu, busiest, 1,
+ sd, IDLE, &all_pinned)) {
schedstat_inc(busiest, alb_lost);
schedstat_inc(rq, alb_gained);
} else {
next prev parent reply other threads:[~2004-10-22 19:44 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2004-10-20 19:36 [PATCH, 2.6.9] improved load_balance() tolerance for pinned tasks John Hawkes
2004-10-20 19:59 ` Ingo Molnar
2004-10-22 13:08 ` Nick Piggin
2004-10-22 19:38 ` John Hawkes [this message]
[not found] ` <00ee01c4b870$030b80f0$6700a8c0@comcast.net>
2004-10-23 4:27 ` Nick Piggin
2004-10-25 16:02 ` John Hawkes
2004-10-25 23:59 ` Nick Piggin
2004-10-30 0:21 ` Matthew Dobson
-- strict thread matches above, loose matches on Subject: below --
2004-10-22 19:20 John Hawkes
2004-10-23 4:22 ` Nick Piggin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=200410221938.MAA52152@google.engr.sgi.com \
--to=hawkes@google.engr.sgi.com \
--cc=akpm@osdl.org \
--cc=hawkes@oss.sgi.com \
--cc=hawkes@sgi.com \
--cc=jbarnes@sgi.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=nickpiggin@yahoo.com.au \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox