From: Alex Shi <alex.shi@intel.com>
To: pjt@google.com
Cc: Alex Shi <alex.shi@intel.com>,
mingo@redhat.com, peterz@infradead.org, tglx@linutronix.de,
akpm@linux-foundation.org, arjan@linux.intel.com, bp@alien8.de,
namhyung@kernel.org, efault@gmx.de, vincent.guittot@linaro.org,
gregkh@linuxfoundation.org, preeti@linux.vnet.ibm.com,
linux-kernel@vger.kernel.org
Subject: Re: [PATCH v3 09/22] sched: compute runnable load avg in cpu_load and cpu_avg_load_per_task
Date: Sat, 05 Jan 2013 16:56:17 +0800 [thread overview]
Message-ID: <50E7EAB1.6020302@intel.com> (raw)
In-Reply-To: <1357375071-11793-10-git-send-email-alex.shi@intel.com>
On 01/05/2013 04:37 PM, Alex Shi wrote:
> They are the base values in load balance, update them with rq runnable
> load average, then the load balance will consider runnable load avg
> naturally.
>
> Signed-off-by: Alex Shi <alex.shi@intel.com>
> ---
> kernel/sched/core.c | 8 ++++++++
> kernel/sched/fair.c | 4 ++--
> 2 files changed, 10 insertions(+), 2 deletions(-)
>
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index 06d27af..5feed5e 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -2544,7 +2544,11 @@ static void __update_cpu_load(struct rq *this_rq, unsigned long this_load,
> void update_idle_cpu_load(struct rq *this_rq)
> {
> unsigned long curr_jiffies = ACCESS_ONCE(jiffies);
> +#if defined(CONFIG_SMP) && defined(CONFIG_FAIR_GROUP_SCHED)
> + unsigned long load = (unsigned long)this_rq->cfs.runnable_load_avg;
> +#else
> unsigned long load = this_rq->load.weight;
> +#endif
> unsigned long pending_updates;
>
> /*
> @@ -2594,7 +2598,11 @@ static void update_cpu_load_active(struct rq *this_rq)
> * See the mess around update_idle_cpu_load() / update_cpu_load_nohz().
> */
> this_rq->last_load_update_tick = jiffies;
> +#if defined(CONFIG_SMP) && defined(CONFIG_FAIR_GROUP_SCHED)
> + __update_cpu_load(this_rq, this_rq->cfs.runnable_load_avg, 1);
> +#else
> __update_cpu_load(this_rq, this_rq->load.weight, 1);
> +#endif
>
> calc_load_account_active(this_rq);
> }
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index 5c545e4..84a6517 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -2906,7 +2906,7 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags)
> /* Used instead of source_load when we know the type == 0 */
> static unsigned long weighted_cpuload(const int cpu)
> {
> - return cpu_rq(cpu)->load.weight;
> + return (unsigned long)cpu_rq(cpu)->cfs.runnable_load_avg;
Above line change cause aim9 multitask benchmark drop about 10%
performance on many x86 machines. Profile just show there are more
cpuidle enter called.
The testing command:
#( echo $hostname ; echo test ; echo 1 ; echo 2000 ; echo 2 ; echo 2000
; echo 100 ) | ./multitask -nl
The oprofile output here:
with this patch set
101978 total 0.0134
54406 cpuidle_wrap_enter 499.1376
2098 __do_page_fault 2.0349
1976 rwsem_wake 29.0588
1824 finish_task_switch 12.4932
1560 copy_user_generic_string 24.3750
1346 clear_page_c 84.1250
1249 unmap_single_vma 0.6885
1141 copy_page_rep 71.3125
1093 anon_vma_interval_tree_insert 8.1567
3.8-rc2
68982 total 0.0090
22166 cpuidle_wrap_enter 203.3578
2188 rwsem_wake 32.1765
2136 __do_page_fault 2.0718
1920 finish_task_switch 13.1507
1724 poll_idle 15.2566
1433 copy_user_generic_string 22.3906
1237 clear_page_c 77.3125
1222 unmap_single_vma 0.6736
1053 anon_vma_interval_tree_insert 7.8582
Without load avg in periodic balancing, each cpu will weighted with all
tasks load.
with new load tracking, we just update the cfs_rq load avg with each
task at enqueue/dequeue moment, and with just update current task in
scheduler_tick. I am wondering if it's the sample is a bit rare.
What's your opinion of this, Paul?
> }
>
> /*
> @@ -2953,7 +2953,7 @@ static unsigned long cpu_avg_load_per_task(int cpu)
> unsigned long nr_running = ACCESS_ONCE(rq->nr_running);
>
> if (nr_running)
> - return rq->load.weight / nr_running;
> + return (unsigned long)rq->cfs.runnable_load_avg / nr_running;
>
> return 0;
> }
>
--
Thanks Alex
next prev parent reply other threads:[~2013-01-05 8:55 UTC|newest]
Thread overview: 91+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-01-05 8:37 [PATCH V3 0/22] sched: simplified fork, enable load average into LB and power awareness scheduling Alex Shi
2013-01-05 8:37 ` [PATCH v3 01/22] sched: set SD_PREFER_SIBLING on MC domain to reduce a domain level Alex Shi
2013-01-05 8:37 ` [PATCH v3 02/22] sched: select_task_rq_fair clean up Alex Shi
2013-01-11 4:57 ` Preeti U Murthy
2013-01-05 8:37 ` [PATCH v3 03/22] sched: fix find_idlest_group mess logical Alex Shi
2013-01-11 4:59 ` Preeti U Murthy
2013-01-05 8:37 ` [PATCH v3 04/22] sched: don't need go to smaller sched domain Alex Shi
2013-01-09 17:38 ` Morten Rasmussen
2013-01-10 3:16 ` Mike Galbraith
2013-01-11 5:02 ` Preeti U Murthy
2013-01-05 8:37 ` [PATCH v3 05/22] sched: remove domain iterations in fork/exec/wake Alex Shi
2013-01-09 18:21 ` Morten Rasmussen
2013-01-11 2:46 ` Alex Shi
2013-01-11 10:07 ` Morten Rasmussen
2013-01-11 14:50 ` Alex Shi
2013-01-14 8:55 ` li guang
2013-01-14 9:18 ` Alex Shi
2013-01-11 4:56 ` Preeti U Murthy
2013-01-11 8:01 ` li guang
2013-01-11 14:56 ` Alex Shi
2013-01-14 9:03 ` li guang
2013-01-15 2:34 ` Alex Shi
2013-01-16 1:54 ` li guang
2013-01-11 10:54 ` Morten Rasmussen
2013-01-16 5:43 ` Alex Shi
2013-01-16 7:41 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 06/22] sched: load tracking bug fix Alex Shi
2013-01-05 8:37 ` [PATCH v3 07/22] sched: set initial load avg of new forked task Alex Shi
2013-01-11 5:10 ` Preeti U Murthy
2013-01-11 5:44 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 08/22] sched: update cpu load after task_tick Alex Shi
2013-01-05 8:37 ` [PATCH v3 09/22] sched: compute runnable load avg in cpu_load and cpu_avg_load_per_task Alex Shi
2013-01-05 8:56 ` Alex Shi [this message]
2013-01-06 7:54 ` Alex Shi
2013-01-06 18:31 ` Linus Torvalds
2013-01-07 7:00 ` Preeti U Murthy
2013-01-08 14:27 ` Alex Shi
2013-01-11 6:31 ` Alex Shi
2013-01-21 14:47 ` Alex Shi
2013-01-22 3:20 ` Alex Shi
2013-01-22 6:55 ` Mike Galbraith
2013-01-22 7:50 ` Alex Shi
2013-01-22 9:52 ` Mike Galbraith
2013-01-23 0:36 ` Alex Shi
2013-01-23 1:47 ` Mike Galbraith
2013-01-23 2:01 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 10/22] sched: consider runnable load average in move_tasks Alex Shi
2013-01-05 8:37 ` [PATCH v3 11/22] sched: consider runnable load average in effective_load Alex Shi
2013-01-10 11:28 ` Morten Rasmussen
2013-01-11 3:26 ` Alex Shi
2013-01-14 12:01 ` Morten Rasmussen
2013-01-16 5:30 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 12/22] Revert "sched: Introduce temporary FAIR_GROUP_SCHED dependency for load-tracking" Alex Shi
2013-01-05 8:37 ` [PATCH v3 13/22] sched: add sched_policy in kernel Alex Shi
2013-01-05 8:37 ` [PATCH v3 14/22] sched: add sched_policy and it's sysfs interface Alex Shi
2013-01-14 6:53 ` Namhyung Kim
2013-01-14 8:11 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 15/22] sched: log the cpu utilization at rq Alex Shi
2013-01-10 11:40 ` Morten Rasmussen
2013-01-11 3:30 ` Alex Shi
2013-01-14 13:59 ` Morten Rasmussen
2013-01-16 5:53 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 16/22] sched: add power aware scheduling in fork/exec/wake Alex Shi
2013-01-10 15:01 ` Morten Rasmussen
2013-01-11 7:08 ` Alex Shi
2013-01-14 16:09 ` Morten Rasmussen
2013-01-16 6:02 ` Alex Shi
2013-01-16 14:27 ` Morten Rasmussen
2013-01-17 5:47 ` Namhyung Kim
2013-01-18 13:41 ` Alex Shi
2013-01-14 7:03 ` Namhyung Kim
2013-01-14 8:30 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 17/22] sched: packing small tasks in wake/exec balancing Alex Shi
2013-01-10 17:17 ` Morten Rasmussen
2013-01-11 3:47 ` Alex Shi
2013-01-14 7:13 ` Namhyung Kim
2013-01-16 6:11 ` Alex Shi
2013-01-16 12:52 ` Namhyung Kim
2013-01-14 17:00 ` Morten Rasmussen
2013-01-16 7:32 ` Alex Shi
2013-01-16 15:08 ` Morten Rasmussen
2013-01-18 14:06 ` Alex Shi
2013-01-05 8:37 ` [PATCH v3 18/22] sched: add power/performance balance allowed flag Alex Shi
2013-01-05 8:37 ` [PATCH v3 19/22] sched: pull all tasks from source group Alex Shi
2013-01-05 8:37 ` [PATCH v3 20/22] sched: don't care if the local group has capacity Alex Shi
2013-01-05 8:37 ` [PATCH v3 21/22] sched: power aware load balance, Alex Shi
2013-01-05 8:37 ` [PATCH v3 22/22] sched: lazy powersaving balance Alex Shi
2013-01-14 8:39 ` Namhyung Kim
2013-01-14 8:45 ` Alex Shi
2013-01-09 17:16 ` [PATCH V3 0/22] sched: simplified fork, enable load average into LB and power awareness scheduling Morten Rasmussen
2013-01-10 3:49 ` Alex Shi
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=50E7EAB1.6020302@intel.com \
--to=alex.shi@intel.com \
--cc=akpm@linux-foundation.org \
--cc=arjan@linux.intel.com \
--cc=bp@alien8.de \
--cc=efault@gmx.de \
--cc=gregkh@linuxfoundation.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@redhat.com \
--cc=namhyung@kernel.org \
--cc=peterz@infradead.org \
--cc=pjt@google.com \
--cc=preeti@linux.vnet.ibm.com \
--cc=tglx@linutronix.de \
--cc=vincent.guittot@linaro.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).