From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753332AbcFTMXu (ORCPT ); Mon, 20 Jun 2016 08:23:50 -0400 Received: from mx1.redhat.com ([209.132.183.28]:47685 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1753252AbcFTMXc (ORCPT ); Mon, 20 Jun 2016 08:23:32 -0400 From: Jiri Olsa To: Ingo Molnar , Peter Zijlstra Cc: lkml , James Hartsock , Rik van Riel , Srivatsa Vaddagiri , Kirill Tkhai Subject: [PATCH 4/4] sched/fair: Add schedstat debug values for REBALANCE_AFFINITY Date: Mon, 20 Jun 2016 14:15:14 +0200 Message-Id: <1466424914-8981-5-git-send-email-jolsa@kernel.org> In-Reply-To: <1466424914-8981-1-git-send-email-jolsa@kernel.org> References: <1466424914-8981-1-git-send-email-jolsa@kernel.org> X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.26]); Mon, 20 Jun 2016 12:15:29 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org It was helpful to watch several stats when testing this feature. Adding the most useful. Signed-off-by: Jiri Olsa --- include/linux/sched.h | 2 ++ kernel/sched/debug.c | 4 ++++ kernel/sched/fair.c | 15 ++++++++++++++- kernel/sched/sched.h | 5 +++++ 4 files changed, 25 insertions(+), 1 deletion(-) diff --git a/include/linux/sched.h b/include/linux/sched.h index 0e6ac882283b..4b772820436b 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -1305,6 +1305,8 @@ struct sched_statistics { u64 nr_failed_migrations_running; u64 nr_failed_migrations_hot; u64 nr_forced_migrations; + u64 nr_dont_balance; + u64 nr_balanced_affinity; u64 nr_wakeups; u64 nr_wakeups_sync; diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c index 2a0a9995256d..5947558dab65 100644 --- a/kernel/sched/debug.c +++ b/kernel/sched/debug.c @@ -635,6 +635,9 @@ do { \ P(sched_goidle); P(ttwu_count); P(ttwu_local); + P(nr_dont_balance); + P(nr_affinity_out); + P(nr_affinity_in); } #undef P @@ -912,6 +915,7 @@ void proc_sched_show_task(struct task_struct *p, struct seq_file *m) P(se.statistics.nr_wakeups_affine_attempts); P(se.statistics.nr_wakeups_passive); P(se.statistics.nr_wakeups_idle); + P(se.statistics.nr_dont_balance); avg_atom = p->se.sum_exec_runtime; if (nr_switches) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 736e525e189c..a4f1ed403f1e 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -902,7 +902,15 @@ update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se) static bool dont_balance(struct task_struct *p) { - return sched_feat(REBALANCE_AFFINITY) && p->se.dont_balance; + bool dont_balance = false; + + if (sched_feat(REBALANCE_AFFINITY) && p->se.dont_balance) { + dont_balance = true; + schedstat_inc(task_rq(p), nr_dont_balance); + schedstat_inc(p, se.statistics.nr_dont_balance); + } + + return dont_balance; } /************************************************** @@ -7903,6 +7911,7 @@ static void rebalance_affinity(struct rq *rq) if (cpu >= nr_cpu_ids) continue; + schedstat_inc(rq, nr_affinity_out); __detach_task(p, rq, cpu); raw_spin_unlock(&rq->lock); @@ -7911,6 +7920,10 @@ static void rebalance_affinity(struct rq *rq) raw_spin_lock(&dst_rq->lock); attach_task(dst_rq, p); p->se.dont_balance = true; + + schedstat_inc(p, se.statistics.nr_balanced_affinity); + schedstat_inc(dst_rq, nr_affinity_in); + raw_spin_unlock(&dst_rq->lock); local_irq_restore(flags); diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index d1a6224cd140..f086e233f7e6 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h @@ -701,6 +701,11 @@ struct rq { /* try_to_wake_up() stats */ unsigned int ttwu_count; unsigned int ttwu_local; + + /* rebalance_affinity() stats */ + unsigned int nr_dont_balance; + unsigned int nr_affinity_out; + unsigned int nr_affinity_in; #endif #ifdef CONFIG_SMP -- 2.4.11