From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754901AbZHYJvi (ORCPT ); Tue, 25 Aug 2009 05:51:38 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1752220AbZHYJvh (ORCPT ); Tue, 25 Aug 2009 05:51:37 -0400 Received: from e7.ny.us.ibm.com ([32.97.182.137]:50859 "EHLO e7.ny.us.ibm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751238AbZHYJvg (ORCPT ); Tue, 25 Aug 2009 05:51:36 -0400 Date: Tue, 25 Aug 2009 15:21:50 +0530 From: Bharata B Rao To: linux-kernel@vger.kernel.org Cc: Dhaval Giani , Balbir Singh , Vaidyanathan Srinivasan , Gautham R Shenoy , Srivatsa Vaddagiri , Ingo Molnar , Peter Zijlstra , Pavel Emelyanov , Herbert Poetzl , Avi Kivity , Chris Friesen , Paul Menage , Mike Waychison Subject: [RFC v1 PATCH 6/7] sched: Add throttle time statistics to /proc/sched_debug Message-ID: <20090825095150.GV3663@in.ibm.com> Reply-To: bharata@linux.vnet.ibm.com References: <20090825094729.GP3663@in.ibm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20090825094729.GP3663@in.ibm.com> User-Agent: Mutt/1.5.18 (2008-05-17) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org sched: Add throttle time statistics to /proc/sched_debug From: Bharata B Rao With hard limits, provide stats about throttle time, throttle count and max throttle time for group sched entities in /proc/sched_debug Throttle stats are collected only for group entities. Signed-off-by: Bharata B Rao --- include/linux/sched.h | 6 ++++++ kernel/sched_debug.c | 17 ++++++++++++++++- kernel/sched_fair.c | 20 ++++++++++++++++++++ 3 files changed, 42 insertions(+), 1 deletion(-) --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -1137,6 +1137,12 @@ struct sched_entity { u64 nr_wakeups_affine_attempts; u64 nr_wakeups_passive; u64 nr_wakeups_idle; +#ifdef CONFIG_CFS_HARD_LIMITS + u64 throttle_start; + u64 throttle_max; + u64 throttle_count; + u64 throttle_sum; +#endif #endif #ifdef CONFIG_FAIR_GROUP_SCHED --- a/kernel/sched_debug.c +++ b/kernel/sched_debug.c @@ -80,6 +80,11 @@ static void print_cfs_group_stats(struct PN(se->wait_max); PN(se->wait_sum); P(se->wait_count); +#ifdef CONFIG_CFS_HARD_LIMITS + PN(se->throttle_max); + PN(se->throttle_sum); + P(se->throttle_count); +#endif #endif P(se->load.weight); #undef PN @@ -216,6 +221,16 @@ void print_cfs_rq(struct seq_file *m, in #endif SEQ_printf(m, " .%-30s: %ld\n", "nr_tasks_running", cfs_rq->nr_tasks_running); +#ifdef CONFIG_CFS_HARD_LIMITS + spin_lock_irqsave(&rq->lock, flags); + SEQ_printf(m, " .%-30s: %d\n", "cfs_throttled", + cfs_rq->cfs_throttled); + SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "cfs_time", + SPLIT_NS(cfs_rq->cfs_time)); + SEQ_printf(m, " .%-30s: %Ld.%06ld\n", "cfs_runtime", + SPLIT_NS(cfs_rq->cfs_runtime)); + spin_unlock_irqrestore(&rq->lock, flags); +#endif print_cfs_group_stats(m, cpu, cfs_rq->tg); #endif } @@ -312,7 +327,7 @@ static int sched_debug_show(struct seq_f u64 now = ktime_to_ns(ktime_get()); int cpu; - SEQ_printf(m, "Sched Debug Version: v0.09, %s %.*s\n", + SEQ_printf(m, "Sched Debug Version: v0.10, %s %.*s\n", init_utsname()->release, (int)strcspn(init_utsname()->version, " "), init_utsname()->version); --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c @@ -188,6 +188,23 @@ find_matching_se(struct sched_entity **s #ifdef CONFIG_CFS_HARD_LIMITS +static inline void update_stats_throttle_start(struct cfs_rq *cfs_rq, + struct sched_entity *se) +{ + schedstat_set(se->throttle_start, rq_of(cfs_rq)->clock); +} + +static inline void update_stats_throttle_end(struct cfs_rq *cfs_rq, + struct sched_entity *se) +{ + schedstat_set(se->throttle_max, max(se->throttle_max, + rq_of(cfs_rq)->clock - se->throttle_start)); + schedstat_set(se->throttle_count, se->throttle_count + 1); + schedstat_set(se->throttle_sum, se->throttle_sum + + rq_of(cfs_rq)->clock - se->throttle_start); + schedstat_set(se->throttle_start, 0); +} + static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) { return cfs_rq->cfs_throttled; @@ -217,6 +234,7 @@ static void sched_cfs_runtime_exceeded(s if (cfs_rq->cfs_time > cfs_rq->cfs_runtime) { cfs_rq->cfs_throttled = 1; + update_stats_throttle_start(cfs_rq, se); resched_task(tsk_curr); } } @@ -314,6 +332,8 @@ void do_sched_cfs_period_timer(struct cf spin_lock_irqsave(&rq->lock, flags); cfs_rq->cfs_time = 0; if (cfs_rq_throttled(cfs_rq)) { + update_rq_clock(rq); + update_stats_throttle_end(cfs_rq, se); cfs_rq->cfs_throttled = 0; enqueue_throttled_entity(rq, se); }