From: Peter Zijlstra <a.p.zijlstra@chello.nl>
To: mingo@elte.hu, paulus@samba.org, tglx@linutronix.de
Cc: linux-kernel@vger.kernel.org, Peter Zijlstra <a.p.zijlstra@chello.nl>
Subject: [PATCH 10/11] perf_counter: hrtimer based sampling for software time events
Date: Fri, 13 Mar 2009 12:21:35 +0100 [thread overview]
Message-ID: <20090313112301.862384083@chello.nl> (raw)
In-Reply-To: 20090313112125.886730125@chello.nl
[-- Attachment #1: perf_swcounter_hrtimer.patch --]
[-- Type: text/plain, Size: 8245 bytes --]
Use hrtimers to profile timer based sampling for the software time counters.
This allows platforms without hardware counter support to still perform sample
based profiling.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
---
include/linux/perf_counter.h | 20 ++++--
kernel/perf_counter.c | 123 ++++++++++++++++++++++++++++++-------------
2 files changed, 100 insertions(+), 43 deletions(-)
Index: linux-2.6/kernel/perf_counter.c
===================================================================
--- linux-2.6.orig/kernel/perf_counter.c
+++ linux-2.6/kernel/perf_counter.c
@@ -1395,7 +1395,7 @@ static void perf_swcounter_handle_group(
struct perf_counter *counter, *group_leader = sibling->group_leader;
list_for_each_entry(counter, &group_leader->sibling_list, list_entry) {
- perf_swcounter_update(counter);
+ counter->hw_ops->read(counter);
perf_swcounter_store_irq(sibling, counter->hw_event.type);
perf_swcounter_store_irq(sibling, atomic64_read(&counter->count));
}
@@ -1404,8 +1404,6 @@ static void perf_swcounter_handle_group(
static void perf_swcounter_interrupt(struct perf_counter *counter,
int nmi, struct pt_regs *regs)
{
- perf_swcounter_save_and_restart(counter);
-
switch (counter->hw_event.record_type) {
case PERF_RECORD_SIMPLE:
break;
@@ -1426,6 +1424,38 @@ static void perf_swcounter_interrupt(str
wake_up(&counter->waitq);
}
+static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer)
+{
+ struct perf_counter *counter;
+ struct pt_regs *regs;
+
+ counter = container_of(hrtimer, struct perf_counter, hw.hrtimer);
+ counter->hw_ops->read(counter);
+
+ regs = get_irq_regs();
+ /*
+ * In case we exclude kernel IPs or are somehow not in interrupt
+ * context, provide the next best thing, the user IP.
+ */
+ if ((counter->hw_event.exclude_kernel || !regs) &&
+ !counter->hw_event.exclude_user)
+ regs = task_pt_regs(current);
+
+ if (regs)
+ perf_swcounter_interrupt(counter, 0, regs);
+
+ hrtimer_forward_now(hrtimer, ns_to_ktime(counter->hw.irq_period));
+
+ return HRTIMER_RESTART;
+}
+
+static void perf_swcounter_overflow(struct perf_counter *counter,
+ int nmi, struct pt_regs *regs)
+{
+ perf_swcounter_save_and_restart(counter);
+ perf_swcounter_interrupt(counter, nmi, regs);
+}
+
static int perf_swcounter_match(struct perf_counter *counter,
enum hw_event_types event,
struct pt_regs *regs)
@@ -1448,13 +1478,20 @@ static int perf_swcounter_match(struct p
return 1;
}
+static void perf_swcounter_add(struct perf_counter *counter, u64 nr,
+ int nmi, struct pt_regs *regs)
+{
+ int neg = atomic64_add_negative(nr, &counter->hw.count);
+ if (counter->hw.irq_period && !neg)
+ perf_swcounter_overflow(counter, nmi, regs);
+}
+
static void perf_swcounter_ctx_event(struct perf_counter_context *ctx,
enum hw_event_types event, u64 nr,
int nmi, struct pt_regs *regs)
{
struct perf_counter *counter;
unsigned long flags;
- int neg;
if (list_empty(&ctx->counter_list))
return;
@@ -1465,11 +1502,8 @@ static void perf_swcounter_ctx_event(str
* XXX: make counter_list RCU safe
*/
list_for_each_entry(counter, &ctx->counter_list, list_entry) {
- if (perf_swcounter_match(counter, event, regs)) {
- neg = atomic64_add_negative(nr, &counter->hw.count);
- if (counter->hw.irq_period && !neg)
- perf_swcounter_interrupt(counter, nmi, regs);
- }
+ if (perf_swcounter_match(counter, event, regs))
+ perf_swcounter_add(counter, nr, nmi, regs);
}
spin_unlock_irqrestore(&ctx->lock, flags);
@@ -1513,14 +1547,6 @@ static const struct hw_perf_counter_ops
* Software counter: cpu wall time clock
*/
-static int cpu_clock_perf_counter_enable(struct perf_counter *counter)
-{
- int cpu = raw_smp_processor_id();
-
- atomic64_set(&counter->hw.prev_count, cpu_clock(cpu));
- return 0;
-}
-
static void cpu_clock_perf_counter_update(struct perf_counter *counter)
{
int cpu = raw_smp_processor_id();
@@ -1533,8 +1559,26 @@ static void cpu_clock_perf_counter_updat
atomic64_add(now - prev, &counter->count);
}
+static int cpu_clock_perf_counter_enable(struct perf_counter *counter)
+{
+ struct hw_perf_counter *hwc = &counter->hw;
+ int cpu = raw_smp_processor_id();
+
+ atomic64_set(&hwc->prev_count, cpu_clock(cpu));
+ if (hwc->irq_period) {
+ hrtimer_init(&hwc->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
+ hwc->hrtimer.function = perf_swcounter_hrtimer;
+ __hrtimer_start_range_ns(&hwc->hrtimer,
+ ns_to_ktime(hwc->irq_period), 0,
+ HRTIMER_MODE_REL, 0);
+ }
+
+ return 0;
+}
+
static void cpu_clock_perf_counter_disable(struct perf_counter *counter)
{
+ hrtimer_cancel(&counter->hw.hrtimer);
cpu_clock_perf_counter_update(counter);
}
@@ -1580,27 +1624,33 @@ static void task_clock_perf_counter_upda
atomic64_add(delta, &counter->count);
}
-static void task_clock_perf_counter_read(struct perf_counter *counter)
-{
- u64 now = task_clock_perf_counter_val(counter, 1);
-
- task_clock_perf_counter_update(counter, now);
-}
-
static int task_clock_perf_counter_enable(struct perf_counter *counter)
{
- if (counter->prev_state <= PERF_COUNTER_STATE_OFF)
- atomic64_set(&counter->hw.prev_count,
- task_clock_perf_counter_val(counter, 0));
+ struct hw_perf_counter *hwc = &counter->hw;
+
+ atomic64_set(&hwc->prev_count, task_clock_perf_counter_val(counter, 0));
+ if (hwc->irq_period) {
+ hrtimer_init(&hwc->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
+ hwc->hrtimer.function = perf_swcounter_hrtimer;
+ __hrtimer_start_range_ns(&hwc->hrtimer,
+ ns_to_ktime(hwc->irq_period), 0,
+ HRTIMER_MODE_REL, 0);
+ }
return 0;
}
static void task_clock_perf_counter_disable(struct perf_counter *counter)
{
- u64 now = task_clock_perf_counter_val(counter, 0);
+ hrtimer_cancel(&counter->hw.hrtimer);
+ task_clock_perf_counter_update(counter,
+ task_clock_perf_counter_val(counter, 0));
+}
- task_clock_perf_counter_update(counter, now);
+static void task_clock_perf_counter_read(struct perf_counter *counter)
+{
+ task_clock_perf_counter_update(counter,
+ task_clock_perf_counter_val(counter, 1));
}
static const struct hw_perf_counter_ops perf_ops_task_clock = {
@@ -1729,16 +1779,12 @@ sw_perf_counter_init(struct perf_counter
*/
switch (counter->hw_event.type) {
case PERF_COUNT_CPU_CLOCK:
- if (!(counter->hw_event.exclude_user ||
- counter->hw_event.exclude_kernel ||
- counter->hw_event.exclude_hv))
- hw_ops = &perf_ops_cpu_clock;
+ hw_ops = &perf_ops_cpu_clock;
+
+ if (hw_event->irq_period && hw_event->irq_period < 10000)
+ hw_event->irq_period = 10000;
break;
case PERF_COUNT_TASK_CLOCK:
- if (counter->hw_event.exclude_user ||
- counter->hw_event.exclude_kernel ||
- counter->hw_event.exclude_hv)
- break;
/*
* If the user instantiates this as a per-cpu counter,
* use the cpu_clock counter instead.
@@ -1747,6 +1793,9 @@ sw_perf_counter_init(struct perf_counter
hw_ops = &perf_ops_task_clock;
else
hw_ops = &perf_ops_cpu_clock;
+
+ if (hw_event->irq_period && hw_event->irq_period < 10000)
+ hw_event->irq_period = 10000;
break;
case PERF_COUNT_PAGE_FAULTS:
case PERF_COUNT_PAGE_FAULTS_MIN:
Index: linux-2.6/include/linux/perf_counter.h
===================================================================
--- linux-2.6.orig/include/linux/perf_counter.h
+++ linux-2.6/include/linux/perf_counter.h
@@ -114,6 +114,7 @@ struct perf_counter_hw_event {
#include <linux/rculist.h>
#include <linux/rcupdate.h>
#include <linux/spinlock.h>
+#include <linux/hrtimer.h>
#include <asm/atomic.h>
struct task_struct;
@@ -123,12 +124,19 @@ struct task_struct;
*/
struct hw_perf_counter {
#ifdef CONFIG_PERF_COUNTERS
- u64 config;
- unsigned long config_base;
- unsigned long counter_base;
- int nmi;
- unsigned int idx;
- atomic64_t count; /* software */
+ union {
+ struct { /* hardware */
+ u64 config;
+ unsigned long config_base;
+ unsigned long counter_base;
+ int nmi;
+ unsigned int idx;
+ };
+ union { /* software */
+ atomic64_t count;
+ struct hrtimer hrtimer;
+ };
+ };
atomic64_t prev_count;
u64 irq_period;
atomic64_t period_left;
--
next prev parent reply other threads:[~2009-03-13 11:27 UTC|newest]
Thread overview: 40+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-03-13 11:21 [PATCH 00/11] generic software counters -v2 Peter Zijlstra
2009-03-13 11:21 ` [PATCH 01/11] sched: remove extra call overhead for schedule() Peter Zijlstra
2009-03-13 13:00 ` [tip:core/locking] " Peter Zijlstra
2009-04-20 19:00 ` [tip:sched/core] " tip-bot for Peter Zijlstra
2009-03-13 11:21 ` [PATCH 02/11] hrtimer: fix rq->lock inversion (again) Peter Zijlstra
2009-03-13 13:00 ` [tip:core/locking] " Peter Zijlstra
2009-03-13 13:27 ` Peter Zijlstra
2009-03-13 14:57 ` Peter Zijlstra
2009-03-31 12:57 ` Peter Zijlstra
2009-04-02 19:45 ` [tip:timers/urgent] " Peter Zijlstra
2009-03-13 11:21 ` [PATCH 03/11] perf_counter: x86: fix 32bit irq_period assumption Peter Zijlstra
2009-03-13 13:00 ` [tip:perfcounters/core] perf_counter: x86: fix 32-bit " Peter Zijlstra
2009-03-13 13:06 ` Peter Zijlstra
2009-03-13 11:21 ` [PATCH 04/11] perf_counter: use list_move_tail Peter Zijlstra
2009-03-13 13:00 ` [tip:perfcounters/core] perf_counter: use list_move_tail() Peter Zijlstra
2009-03-13 13:06 ` Peter Zijlstra
2009-03-13 11:21 ` [PATCH 05/11] perf_counter: add comment to barrier Peter Zijlstra
2009-03-13 13:01 ` [tip:perfcounters/core] " Peter Zijlstra
2009-03-13 13:06 ` Peter Zijlstra
2009-03-13 11:21 ` [PATCH 06/11] perf_counter: x86: use ULL postfix for 64bit constants Peter Zijlstra
2009-03-13 13:01 ` [tip:perfcounters/core] " Peter Zijlstra
2009-03-13 13:06 ` Peter Zijlstra
2009-03-13 11:21 ` [PATCH 07/11] perf_counter: software counter event infrastructure Peter Zijlstra
2009-03-13 13:01 ` [tip:perfcounters/core] " Peter Zijlstra
2009-03-13 13:07 ` Peter Zijlstra
2009-03-13 11:21 ` [PATCH 08/11] perf_counter: provide pagefault software events Peter Zijlstra
2009-03-13 13:01 ` [tip:perfcounters/core] " Peter Zijlstra
2009-03-13 13:07 ` Peter Zijlstra
2009-03-13 11:21 ` [PATCH 09/11] perf_counter: provide major/minor page fault " Peter Zijlstra
2009-03-13 13:01 ` [tip:perfcounters/core] " Peter Zijlstra
2009-03-13 13:07 ` Peter Zijlstra
2009-03-13 11:21 ` Peter Zijlstra [this message]
2009-03-13 13:01 ` [tip:perfcounters/core] perf_counter: hrtimer based sampling for software time events Peter Zijlstra
2009-03-13 13:07 ` Peter Zijlstra
2009-03-13 15:43 ` [PATCH 10.5/11] perf_counter: fix hrtimer sampling Peter Zijlstra
2009-03-13 16:09 ` [tip:perfcounters/core] " Peter Zijlstra
2009-03-13 11:21 ` [PATCH 11/11] perf_counter: add an event_list Peter Zijlstra
2009-03-13 13:02 ` [tip:perfcounters/core] " Peter Zijlstra
2009-03-13 13:07 ` Peter Zijlstra
2009-03-13 13:07 ` [PATCH 00/11] generic software counters -v2 Ingo Molnar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20090313112301.862384083@chello.nl \
--to=a.p.zijlstra@chello.nl \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=paulus@samba.org \
--cc=tglx@linutronix.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox