From: Steven Rostedt <rostedt@goodmis.org>
To: LKML <linux-kernel@vger.kernel.org>
Cc: Ingo Molnar <mingo@elte.hu>,
Gregory Haskins <ghaskins@novell.com>,
Peter Zijlstra <a.p.zijlstra@chello.nl>,
Christoph Lameter <clameter@sgi.com>,
Steven Rostedt <srostedt@redhat.com>
Subject: [PATCH v4 08/20] Cache cpus_allowed weight for optimizing migration
Date: Tue, 20 Nov 2007 20:01:02 -0500 [thread overview]
Message-ID: <20071121011250.144586697@goodmis.org> (raw)
In-Reply-To: 20071121010054.663842380@goodmis.org
[-- Attachment #1: rt-balance-cpu-weight.patch --]
[-- Type: text/plain, Size: 7965 bytes --]
From: Gregory Haskins <ghaskins@novell.com>
Some RT tasks (particularly kthreads) are bound to one specific CPU.
It is fairly common for two or more bound tasks to get queued up at the
same time. Consider, for instance, softirq_timer and softirq_sched. A
timer goes off in an ISR which schedules softirq_thread to run at RT50.
Then the timer handler determines that it's time to smp-rebalance the
system so it schedules softirq_sched to run. So we are in a situation
where we have two RT50 tasks queued, and the system will go into
rt-overload condition to request other CPUs for help.
This causes two problems in the current code:
1) If a high-priority bound task and a low-priority unbounded task queue
up behind the running task, we will fail to ever relocate the unbounded
task because we terminate the search on the first unmovable task.
2) We spend precious futile cycles in the fast-path trying to pull
overloaded tasks over. It is therefore optimial to strive to avoid the
overhead all together if we can cheaply detect the condition before
overload even occurs.
This patch tries to achieve this optimization by utilizing the hamming
weight of the task->cpus_allowed mask. A weight of 1 indicates that
the task cannot be migrated. We will then utilize this information to
skip non-migratable tasks and to eliminate uncessary rebalance attempts.
We introduce a per-rq variable to count the number of migratable tasks
that are currently running. We only go into overload if we have more
than one rt task, AND at least one of them is migratable.
In addition, we introduce a per-task variable to cache the cpus_allowed
weight, since the hamming calculation is probably relatively expensive.
We only update the cached value when the mask is updated which should be
relatively infrequent, especially compared to scheduling frequency
in the fast path.
Signed-off-by: Gregory Haskins <ghaskins@novell.com>
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
---
include/linux/init_task.h | 1
include/linux/sched.h | 2 +
kernel/fork.c | 1
kernel/sched.c | 9 +++++++-
kernel/sched_rt.c | 50 +++++++++++++++++++++++++++++++++++++++++-----
5 files changed, 57 insertions(+), 6 deletions(-)
Index: linux-compile.git/include/linux/init_task.h
===================================================================
--- linux-compile.git.orig/include/linux/init_task.h 2007-11-20 19:52:44.000000000 -0500
+++ linux-compile.git/include/linux/init_task.h 2007-11-20 19:53:02.000000000 -0500
@@ -130,6 +130,7 @@ extern struct group_info init_groups;
.normal_prio = MAX_PRIO-20, \
.policy = SCHED_NORMAL, \
.cpus_allowed = CPU_MASK_ALL, \
+ .nr_cpus_allowed = NR_CPUS, \
.mm = NULL, \
.active_mm = &init_mm, \
.run_list = LIST_HEAD_INIT(tsk.run_list), \
Index: linux-compile.git/include/linux/sched.h
===================================================================
--- linux-compile.git.orig/include/linux/sched.h 2007-11-20 19:52:44.000000000 -0500
+++ linux-compile.git/include/linux/sched.h 2007-11-20 19:53:02.000000000 -0500
@@ -843,6 +843,7 @@ struct sched_class {
void (*set_curr_task) (struct rq *rq);
void (*task_tick) (struct rq *rq, struct task_struct *p);
void (*task_new) (struct rq *rq, struct task_struct *p);
+ void (*set_cpus_allowed)(struct task_struct *p, cpumask_t *newmask);
};
struct load_weight {
@@ -952,6 +953,7 @@ struct task_struct {
unsigned int policy;
cpumask_t cpus_allowed;
+ int nr_cpus_allowed;
unsigned int time_slice;
#if defined(CONFIG_SCHEDSTATS) || defined(CONFIG_TASK_DELAY_ACCT)
Index: linux-compile.git/kernel/fork.c
===================================================================
--- linux-compile.git.orig/kernel/fork.c 2007-11-20 19:52:44.000000000 -0500
+++ linux-compile.git/kernel/fork.c 2007-11-20 19:53:02.000000000 -0500
@@ -1237,6 +1237,7 @@ static struct task_struct *copy_process(
* parent's CPU). This avoids alot of nasty races.
*/
p->cpus_allowed = current->cpus_allowed;
+ p->nr_cpus_allowed = current->nr_cpus_allowed;
if (unlikely(!cpu_isset(task_cpu(p), p->cpus_allowed) ||
!cpu_online(task_cpu(p))))
set_task_cpu(p, smp_processor_id());
Index: linux-compile.git/kernel/sched.c
===================================================================
--- linux-compile.git.orig/kernel/sched.c 2007-11-20 19:53:00.000000000 -0500
+++ linux-compile.git/kernel/sched.c 2007-11-20 19:53:02.000000000 -0500
@@ -269,6 +269,7 @@ struct rt_rq {
int rt_load_balance_idx;
struct list_head *rt_load_balance_head, *rt_load_balance_curr;
unsigned long rt_nr_running;
+ unsigned long rt_nr_migratory;
/* highest queued rt task prio */
int highest_prio;
};
@@ -5070,7 +5071,13 @@ int set_cpus_allowed(struct task_struct
goto out;
}
- p->cpus_allowed = new_mask;
+ if (p->sched_class->set_cpus_allowed)
+ p->sched_class->set_cpus_allowed(p, &new_mask);
+ else {
+ p->cpus_allowed = new_mask;
+ p->nr_cpus_allowed = cpus_weight(new_mask);
+ }
+
/* Can the task run on the task's current CPU? If so, we're done */
if (cpu_isset(task_cpu(p), new_mask))
goto out;
Index: linux-compile.git/kernel/sched_rt.c
===================================================================
--- linux-compile.git.orig/kernel/sched_rt.c 2007-11-20 19:53:01.000000000 -0500
+++ linux-compile.git/kernel/sched_rt.c 2007-11-20 19:53:02.000000000 -0500
@@ -33,6 +33,14 @@ static inline void rt_clear_overload(str
atomic_dec(&rto_count);
cpu_clear(rq->cpu, rt_overload_mask);
}
+
+static void update_rt_migration(struct rq *rq)
+{
+ if (rq->rt.rt_nr_migratory && (rq->rt.rt_nr_running > 1))
+ rt_set_overload(rq);
+ else
+ rt_clear_overload(rq);
+}
#endif /* CONFIG_SMP */
/*
@@ -64,8 +72,10 @@ static inline void inc_rt_tasks(struct t
#ifdef CONFIG_SMP
if (p->prio < rq->rt.highest_prio)
rq->rt.highest_prio = p->prio;
- if (rq->rt.rt_nr_running > 1)
- rt_set_overload(rq);
+ if (p->nr_cpus_allowed > 1)
+ rq->rt.rt_nr_migratory++;
+
+ update_rt_migration(rq);
#endif /* CONFIG_SMP */
}
@@ -87,8 +97,10 @@ static inline void dec_rt_tasks(struct t
} /* otherwise leave rq->highest prio alone */
} else
rq->rt.highest_prio = MAX_RT_PRIO;
- if (rq->rt.rt_nr_running < 2)
- rt_clear_overload(rq);
+ if (p->nr_cpus_allowed > 1)
+ rq->rt.rt_nr_migratory--;
+
+ update_rt_migration(rq);
#endif /* CONFIG_SMP */
}
@@ -179,7 +191,8 @@ static void deactivate_task(struct rq *r
static int pick_rt_task(struct rq *rq, struct task_struct *p, int cpu)
{
if (!task_running(rq, p) &&
- (cpu < 0 || cpu_isset(cpu, p->cpus_allowed)))
+ (cpu < 0 || cpu_isset(cpu, p->cpus_allowed)) &&
+ (p->nr_cpus_allowed > 1))
return 1;
return 0;
}
@@ -581,6 +594,32 @@ move_one_task_rt(struct rq *this_rq, int
/* don't touch RT tasks */
return 0;
}
+static void set_cpus_allowed_rt(struct task_struct *p, cpumask_t *new_mask)
+{
+ int weight = cpus_weight(*new_mask);
+
+ BUG_ON(!rt_task(p));
+
+ /*
+ * Update the migration status of the RQ if we have an RT task
+ * which is running AND changing its weight value.
+ */
+ if (p->se.on_rq && (weight != p->nr_cpus_allowed)) {
+ struct rq *rq = task_rq(p);
+
+ if ((p->nr_cpus_allowed <= 1) && (weight > 1))
+ rq->rt.rt_nr_migratory++;
+ else if((p->nr_cpus_allowed > 1) && (weight <= 1)) {
+ BUG_ON(!rq->rt.rt_nr_migratory);
+ rq->rt.rt_nr_migratory--;
+ }
+
+ update_rt_migration(rq);
+ }
+
+ p->cpus_allowed = *new_mask;
+ p->nr_cpus_allowed = weight;
+}
#else /* CONFIG_SMP */
# define schedule_tail_balance_rt(rq) do { } while (0)
# define schedule_balance_rt(rq, prev) do { } while (0)
@@ -632,6 +671,7 @@ const struct sched_class rt_sched_class
#ifdef CONFIG_SMP
.load_balance = load_balance_rt,
.move_one_task = move_one_task_rt,
+ .set_cpus_allowed = set_cpus_allowed_rt,
#endif
.set_curr_task = set_curr_task_rt,
--
next prev parent reply other threads:[~2007-11-21 1:19 UTC|newest]
Thread overview: 36+ messages / expand[flat|nested] mbox.gz Atom feed top
2007-11-21 1:00 [PATCH v4 00/20] New RT Balancing version 4 Steven Rostedt
2007-11-21 1:00 ` [PATCH v4 01/20] Add rt_nr_running accounting Steven Rostedt
2007-11-21 1:00 ` [PATCH v4 02/20] track highest prio queued on runqueue Steven Rostedt
2007-11-21 1:00 ` [PATCH v4 03/20] push RT tasks Steven Rostedt
2007-11-21 1:00 ` [PATCH v4 04/20] RT overloaded runqueues accounting Steven Rostedt
2007-11-21 1:00 ` [PATCH v4 05/20] pull RT tasks Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 06/20] wake up balance RT Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 07/20] disable CFS RT load balancing Steven Rostedt
2007-11-21 1:01 ` Steven Rostedt [this message]
2007-11-21 1:01 ` [PATCH v4 09/20] RT: Consistency cleanup for this_rq usage Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 10/20] RT: Remove some CFS specific code from the wakeup path of RT tasks Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 11/20] RT: Break out the search function Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 12/20] RT: Allow current_cpu to be included in search Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 13/20] RT: Pre-route RT tasks on wakeup Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 14/20] RT: Optimize our cpu selection based on topology Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 15/20] RT: Optimize rebalancing Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 16/20] Avoid overload Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 17/20] RT: restore the migratable conditional Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 18/20] Optimize cpu search with hamming weight Steven Rostedt
2007-11-21 1:01 ` [PATCH v4 19/20] Optimize out cpu_clears Steven Rostedt
2007-11-21 2:10 ` Steven Rostedt
2007-11-21 3:10 ` [PATCH] Fix optimized search Gregory Haskins
2007-11-21 4:15 ` Steven Rostedt
2007-11-21 4:26 ` Steven Rostedt
2007-11-21 5:14 ` Gregory Haskins
2007-11-21 1:01 ` [PATCH v4 20/20] balance RT tasks no new wake up Steven Rostedt
2007-11-21 4:44 ` [PATCH 0/4] more RT balancing enhancements Gregory Haskins
2007-11-21 4:44 ` [PATCH 1/4] Fix optimized search Gregory Haskins
2007-11-21 4:44 ` [PATCH 2/4] RT: Add sched-domain roots Gregory Haskins
2007-11-21 4:44 ` [PATCH 3/4] RT: Only balance our RT tasks within our root-domain Gregory Haskins
2007-11-21 4:44 ` [PATCH 4/4] RT: Use a 2-d bitmap for searching lowest-pri CPU Gregory Haskins
2007-11-21 19:51 ` [PATCH 0/4] more RT balancing enhancements v6a Gregory Haskins
2007-11-21 19:52 ` [PATCH 1/4] SCHED: Add sched-domain roots Gregory Haskins
2007-11-21 19:52 ` [PATCH 2/4] SCHED: Track online cpus in the root-domain Gregory Haskins
2007-11-21 19:52 ` [PATCH 3/4] SCHED: Only balance our RT tasks within our root-domain Gregory Haskins
2007-11-21 19:52 ` [PATCH 4/4] SCHED: Use a 2-d bitmap for searching lowest-pri CPU Gregory Haskins
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20071121011250.144586697@goodmis.org \
--to=rostedt@goodmis.org \
--cc=a.p.zijlstra@chello.nl \
--cc=clameter@sgi.com \
--cc=ghaskins@novell.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=srostedt@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox