From: Srikar Dronamraju <srikar@linux.vnet.ibm.com>
To: Mel Gorman <mgorman@suse.de>,
Peter Zijlstra <a.p.zijlstra@chello.nl>,
Ingo Molnar <mingo@kernel.org>
Cc: Andrea Arcangeli <aarcange@redhat.com>,
Johannes Weiner <hannes@cmpxchg.org>,
Linux-MM <linux-mm@kvack.org>,
LKML <linux-kernel@vger.kernel.org>,
Preeti U Murthy <preeti@linux.vnet.ibm.com>,
Linus Torvalds <torvalds@linux-foundation.org>,
Srikar Dronamraju <srikar@linux.vnet.ibm.com>
Subject: [RFC PATCH 05/10] sched: Extend idle balancing to look for consolidation of tasks
Date: Tue, 30 Jul 2013 13:18:20 +0530 [thread overview]
Message-ID: <1375170505-5967-6-git-send-email-srikar@linux.vnet.ibm.com> (raw)
In-Reply-To: <1375170505-5967-1-git-send-email-srikar@linux.vnet.ibm.com>
If the cpu is idle even after a regular load balance, then
try to move a task from another node to this node, such that
node locality improves.
While choosing a task to pull, choose a task/address-space from the
currently running set of tasks on this node. Make sure that the chosen
address-space has a numa affinity to the current node. Choose another
node that has the least number of tasks that belong to this address
space.
This change might induce a slight imbalance but there are enough checks
to make sure that the imbalance is within limits. This slight imbalance
that is created can act as a catalyst/opportunity for the other node to
pull its node affine tasks.
TODO: current checks that look at nr_running should be modified to
look at task loads instead.
Signed-off-by: Srikar Dronamraju <srikar@linux.vnet.ibm.com>
---
kernel/sched/fair.c | 156 +++++++++++++++++++++++++++++++++++++++++++++++++++
1 files changed, 156 insertions(+), 0 deletions(-)
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index debb75a..43af8d9 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -5609,6 +5609,77 @@ void update_max_interval(void)
max_load_balance_interval = HZ*num_online_cpus()/10;
}
+#ifdef CONFIG_NUMA_BALANCING
+static struct task_struct *
+select_task_to_pull(struct mm_struct *this_mm, int this_cpu, int nid)
+{
+ struct mm_struct *mm;
+ struct rq *rq;
+ int cpu;
+
+ for_each_cpu(cpu, cpumask_of_node(nid)) {
+ rq = cpu_rq(cpu);
+ mm = rq->curr->mm;
+
+ if (mm == this_mm) {
+ if (cpumask_test_cpu(this_cpu, tsk_cpus_allowed(rq->curr)))
+ return rq->curr;
+ }
+ }
+ return NULL;
+}
+
+static int
+select_node_to_pull(struct mm_struct *mm, unsigned int nr_running, int nid)
+{
+ atomic_t *weights = mm->numa_weights;
+ int least_running, running, other_nr_running, other_running;
+ int least_node = -1;
+ int other_node, cpu;
+
+ least_running = atomic_read(&weights[nid]);
+ running = least_running;
+ for_each_online_node(other_node) {
+ /* our own node? skip */
+ if (other_node == nid)
+ continue;
+
+ other_running = atomic_read(&weights[other_node]);
+ /* no interesting thread in this node */
+ if (other_running == 0)
+ continue;
+
+ /* other_node has more numa affinity? Dont move. */
+ if (other_running > least_running)
+ continue;
+
+ other_nr_running = 0;
+ for_each_cpu(cpu, cpumask_of_node(other_node))
+ other_nr_running += cpu_rq(cpu)->nr_running;
+
+ /* other_node is already lightly loaded? */
+ if (nr_running > other_nr_running)
+ continue;
+
+ /*
+ * If the other node has significant proportion of load of
+ * the process in question. Or relatively has more affinity
+ * to this address space than the current node, then dont
+ * move
+ */
+ if (other_nr_running < 2 * other_running)
+ continue;
+
+ if (nr_running * other_running >= other_nr_running * running)
+ continue;
+
+ least_running = other_running;
+ least_node = other_node;
+ }
+ return least_node;
+}
+#endif
+
/*
* It checks each scheduling domain to see if it is due to be balanced,
* and initiates a balancing operation if so.
@@ -5674,6 +5745,91 @@ static void rebalance_domains(int cpu, enum cpu_idle_type idle)
if (!balance)
break;
}
+#ifdef CONFIG_NUMA_BALANCING
+ if (!rq->nr_running) {
+ struct mm_struct *prev_mm, *mm;
+ struct task_struct *p = NULL;
+ unsigned int nr_running = 0;
+ int curr_running, total_running;
+ int other_node, nid, dcpu;
+
+ nid = cpu_to_node(cpu);
+ prev_mm = NULL;
+
+ for_each_cpu(dcpu, cpumask_of_node(nid))
+ nr_running += cpu_rq(dcpu)->nr_running;
+
+ for_each_cpu(dcpu, cpumask_of_node(nid)) {
+ struct rq *this_rq;
+
+ this_rq = cpu_rq(dcpu);
+ mm = this_rq->curr->mm;
+ if (!mm || !mm->numa_weights)
+ continue;
+
+ /*
+ * Dont retry if the previous and the current
+ * requests share the same address space
+ */
+ if (mm == prev_mm)
+ continue;
+
+ curr_running = atomic_read(&mm->numa_weights[nid]);
+ total_running = atomic_read(&mm->numa_weights[nr_node_ids]);
+
+ if (curr_running < 2 || total_running < 2)
+ continue;
+
+ prev_mm = mm;
+
+ /* all threads have consolidated */
+ if (curr_running == total_running)
+ continue;
+
+ /*
+ * in-significant proportion of load running on
+ * this node?
+ */
+ if (total_running > curr_running * (nr_node_ids + 1)) {
+ if (nr_running > 2 * curr_running)
+ continue;
+ }
+
+ other_node = select_node_to_pull(mm, nr_running, nid);
+ if (other_node == -1)
+ continue;
+ p = select_task_to_pull(mm, cpu, other_node);
+ if (p)
+ break;
+ }
+ if (p) {
+ struct rq *this_rq;
+ unsigned long flags;
+ int active_balance;
+
+ this_rq = task_rq(p);
+ active_balance = 0;
+
+ /*
+ * ->active_balance synchronizes accesses to
+ * ->active_balance_work. Once set, it's cleared
+ * only after active load balance is finished.
+ */
+ raw_spin_lock_irqsave(&this_rq->lock, flags);
+ if (task_rq(p) == this_rq) {
+ if (!this_rq->active_balance) {
+ this_rq->active_balance = 1;
+ this_rq->push_cpu = cpu;
+ active_balance = 1;
+ }
+ }
+ raw_spin_unlock_irqrestore(&this_rq->lock, flags);
+
+ if (active_balance)
+ active_load_balance(this_rq);
+ }
+ }
+#endif
rcu_read_unlock();
/*
--
1.7.1
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2013-07-30 7:49 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-07-30 7:48 [RFC PATCH 00/10] Improve numa scheduling by consolidating tasks Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 01/10] sched: Introduce per node numa weights Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 02/10] sched: Use numa weights while migrating tasks Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 03/10] sched: Select a better task to pull across node using iterations Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 04/10] sched: Move active_load_balance_cpu_stop to a new helper function Srikar Dronamraju
2013-07-30 7:48 ` Srikar Dronamraju [this message]
2013-07-30 7:48 ` [RFC PATCH 06/10] sched: Limit migrations from a node Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 07/10] sched: Pass hint to active balancer about the task to be chosen Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 08/10] sched: Prevent a task from migrating immediately after an active balance Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 09/10] sched: Choose a runqueue that has lesser local affinity tasks Srikar Dronamraju
2013-07-30 7:48 ` [RFC PATCH 10/10] x86, mm: Prevent gcc to re-read the pagetables Srikar Dronamraju
2013-07-30 8:17 ` [RFC PATCH 00/10] Improve numa scheduling by consolidating tasks Peter Zijlstra
2013-07-30 8:20 ` Peter Zijlstra
2013-07-30 9:03 ` Srikar Dronamraju
2013-07-30 9:10 ` Peter Zijlstra
2013-07-30 9:26 ` Peter Zijlstra
2013-07-30 9:46 ` Srikar Dronamraju
2013-07-31 15:09 ` Peter Zijlstra
2013-07-31 18:06 ` Srikar Dronamraju
2013-07-30 9:15 ` Srikar Dronamraju
2013-07-30 9:33 ` Peter Zijlstra
2013-07-31 17:35 ` Srikar Dronamraju
2013-07-31 13:33 ` Andrew Theurer
2013-07-31 15:43 ` Srikar Dronamraju
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1375170505-5967-6-git-send-email-srikar@linux.vnet.ibm.com \
--to=srikar@linux.vnet.ibm.com \
--cc=a.p.zijlstra@chello.nl \
--cc=aarcange@redhat.com \
--cc=hannes@cmpxchg.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mgorman@suse.de \
--cc=mingo@kernel.org \
--cc=preeti@linux.vnet.ibm.com \
--cc=torvalds@linux-foundation.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).