* [PATCH] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
@ 2015-04-21 8:54 Jiri Olsa
2015-04-21 10:52 ` Ingo Molnar
2015-04-21 15:12 ` [PATCH] " Peter Zijlstra
0 siblings, 2 replies; 8+ messages in thread
From: Jiri Olsa @ 2015-04-21 8:54 UTC (permalink / raw)
To: lkml
Cc: Andi Kleen, Arnaldo Carvalho de Melo, Ingo Molnar, Paul Mackerras,
Peter Zijlstra, Stephane Eranian
The core_pmu does not define cpu_* callbacks, which handles
allocation of 'struct cpu_hw_events::shared_regs' data,
initialization of debug store and PMU_FL_EXCL_CNTRS counters.
While this probably won't happen on bare metal, virtual CPU can
define x86_pmu.extra_regs together with PMU version 1 and thus
be using core_pmu -> using shared_regs data without it being
allocated. That could could leave to following panic:
BUG: unable to handle kernel NULL pointer dereference at (null)
IP: [<ffffffff8152cd4f>] _spin_lock_irqsave+0x1f/0x40
SNIP
[<ffffffff81024bd9>] __intel_shared_reg_get_constraints+0x69/0x1e0
[<ffffffff81024deb>] intel_get_event_constraints+0x9b/0x180
[<ffffffff8101e815>] x86_schedule_events+0x75/0x1d0
[<ffffffff810586dc>] ? check_preempt_curr+0x7c/0x90
[<ffffffff810649fe>] ? try_to_wake_up+0x24e/0x3e0
[<ffffffff81064ba2>] ? default_wake_function+0x12/0x20
[<ffffffff8109eb16>] ? autoremove_wake_function+0x16/0x40
[<ffffffff810577e9>] ? __wake_up_common+0x59/0x90
[<ffffffff811a9517>] ? __d_lookup+0xa7/0x150
[<ffffffff8119db5f>] ? do_lookup+0x9f/0x230
[<ffffffff811a993a>] ? dput+0x9a/0x150
[<ffffffff8119c8f5>] ? path_to_nameidata+0x25/0x60
[<ffffffff8119e90a>] ? __link_path_walk+0x7da/0x1000
[<ffffffff8101d8f9>] ? x86_pmu_add+0xb9/0x170
[<ffffffff8101d7a7>] x86_pmu_commit_txn+0x67/0xc0
[<ffffffff811b07b0>] ? mntput_no_expire+0x30/0x110
[<ffffffff8119c731>] ? path_put+0x31/0x40
[<ffffffff8107c297>] ? current_fs_time+0x27/0x30
[<ffffffff8117d170>] ? mem_cgroup_get_reclaim_stat_from_page+0x20/0x70
[<ffffffff8111b7aa>] group_sched_in+0x13a/0x170
[<ffffffff81014a29>] ? sched_clock+0x9/0x10
[<ffffffff8111bac8>] ctx_sched_in+0x2e8/0x330
[<ffffffff8111bb7b>] perf_event_sched_in+0x6b/0xb0
[<ffffffff8111bc36>] perf_event_context_sched_in+0x76/0xc0
[<ffffffff8111eb3b>] perf_event_comm+0x1bb/0x2e0
[<ffffffff81195ee9>] set_task_comm+0x69/0x80
[<ffffffff81195fe1>] setup_new_exec+0xe1/0x2e0
[<ffffffff811ea68e>] load_elf_binary+0x3ce/0x1ab0
Adding cpu_(prepare|starting|dying) for core_pmu to have shared_regs
data allocated for core_pmu. AFAICS there's no harm to initialize
debug store and PMU_FL_EXCL_CNTRS either for core_pmu.
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
arch/x86/kernel/cpu/perf_event_intel.c | 7 +++++++
1 file changed, 7 insertions(+)
diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c
index 9da2400c2ec3..0a61a9a021de 100644
--- a/arch/x86/kernel/cpu/perf_event_intel.c
+++ b/arch/x86/kernel/cpu/perf_event_intel.c
@@ -2533,6 +2533,10 @@ ssize_t intel_event_sysfs_show(char *page, u64 config)
return x86_event_sysfs_show(page, config, event);
}
+static int intel_pmu_cpu_prepare(int cpu);
+static void intel_pmu_cpu_starting(int cpu);
+static void intel_pmu_cpu_dying(int cpu);
+
static __initconst const struct x86_pmu core_pmu = {
.name = "core",
.handle_irq = x86_pmu_handle_irq,
@@ -2559,6 +2563,9 @@ static __initconst const struct x86_pmu core_pmu = {
.guest_get_msrs = core_guest_get_msrs,
.format_attrs = intel_arch_formats_attr,
.events_sysfs_show = intel_event_sysfs_show,
+ .cpu_prepare = intel_pmu_cpu_prepare,
+ .cpu_starting = intel_pmu_cpu_starting,
+ .cpu_dying = intel_pmu_cpu_dying,
};
struct intel_shared_regs *allocate_shared_regs(int cpu)
--
1.9.3
^ permalink raw reply related [flat|nested] 8+ messages in thread
* Re: [PATCH] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
2015-04-21 8:54 [PATCH] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu Jiri Olsa
@ 2015-04-21 10:52 ` Ingo Molnar
2015-04-21 15:14 ` [PATCHv2] " Jiri Olsa
2015-04-21 15:12 ` [PATCH] " Peter Zijlstra
1 sibling, 1 reply; 8+ messages in thread
From: Ingo Molnar @ 2015-04-21 10:52 UTC (permalink / raw)
To: Jiri Olsa
Cc: lkml, Andi Kleen, Arnaldo Carvalho de Melo, Paul Mackerras,
Peter Zijlstra, Stephane Eranian
* Jiri Olsa <jolsa@kernel.org> wrote:
> The core_pmu does not define cpu_* callbacks, which handles
> allocation of 'struct cpu_hw_events::shared_regs' data,
> initialization of debug store and PMU_FL_EXCL_CNTRS counters.
>
> While this probably won't happen on bare metal, virtual CPU can
> define x86_pmu.extra_regs together with PMU version 1 and thus
> be using core_pmu -> using shared_regs data without it being
> allocated. That could could leave to following panic:
>
> BUG: unable to handle kernel NULL pointer dereference at (null)
> IP: [<ffffffff8152cd4f>] _spin_lock_irqsave+0x1f/0x40
ok.
> diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c
> index 9da2400c2ec3..0a61a9a021de 100644
> --- a/arch/x86/kernel/cpu/perf_event_intel.c
> +++ b/arch/x86/kernel/cpu/perf_event_intel.c
> @@ -2533,6 +2533,10 @@ ssize_t intel_event_sysfs_show(char *page, u64 config)
> return x86_event_sysfs_show(page, config, event);
> }
>
> +static int intel_pmu_cpu_prepare(int cpu);
> +static void intel_pmu_cpu_starting(int cpu);
> +static void intel_pmu_cpu_dying(int cpu);
> +
> static __initconst const struct x86_pmu core_pmu = {
> .name = "core",
> .handle_irq = x86_pmu_handle_irq,
> @@ -2559,6 +2563,9 @@ static __initconst const struct x86_pmu core_pmu = {
> .guest_get_msrs = core_guest_get_msrs,
> .format_attrs = intel_arch_formats_attr,
> .events_sysfs_show = intel_event_sysfs_show,
> + .cpu_prepare = intel_pmu_cpu_prepare,
> + .cpu_starting = intel_pmu_cpu_starting,
> + .cpu_dying = intel_pmu_cpu_dying,
> };
Instead of adding prototype declarations, please arrange the x86_pmu
definition's position so that it comes after the required functions -
so that no prototypes are needed.
Thanks,
Ingo
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
2015-04-21 8:54 [PATCH] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu Jiri Olsa
2015-04-21 10:52 ` Ingo Molnar
@ 2015-04-21 15:12 ` Peter Zijlstra
2015-04-21 15:18 ` Jiri Olsa
1 sibling, 1 reply; 8+ messages in thread
From: Peter Zijlstra @ 2015-04-21 15:12 UTC (permalink / raw)
To: Jiri Olsa
Cc: lkml, Andi Kleen, Arnaldo Carvalho de Melo, Ingo Molnar,
Paul Mackerras, Stephane Eranian
On Tue, Apr 21, 2015 at 10:54:25AM +0200, Jiri Olsa wrote:
> @@ -2559,6 +2563,9 @@ static __initconst const struct x86_pmu core_pmu = {
> .guest_get_msrs = core_guest_get_msrs,
> .format_attrs = intel_arch_formats_attr,
> .events_sysfs_show = intel_event_sysfs_show,
Could you also insert a little comment here how these methods are for
'funny' 'hardware' ?
> + .cpu_prepare = intel_pmu_cpu_prepare,
> + .cpu_starting = intel_pmu_cpu_starting,
> + .cpu_dying = intel_pmu_cpu_dying,
> };
>
> struct intel_shared_regs *allocate_shared_regs(int cpu)
> --
> 1.9.3
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [PATCHv2] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
2015-04-21 10:52 ` Ingo Molnar
@ 2015-04-21 15:14 ` Jiri Olsa
0 siblings, 0 replies; 8+ messages in thread
From: Jiri Olsa @ 2015-04-21 15:14 UTC (permalink / raw)
To: Ingo Molnar
Cc: Jiri Olsa, lkml, Andi Kleen, Arnaldo Carvalho de Melo,
Paul Mackerras, Peter Zijlstra, Stephane Eranian
On Tue, Apr 21, 2015 at 12:52:40PM +0200, Ingo Molnar wrote:
>
> * Jiri Olsa <jolsa@kernel.org> wrote:
SNIP
> > .guest_get_msrs = core_guest_get_msrs,
> > .format_attrs = intel_arch_formats_attr,
> > .events_sysfs_show = intel_event_sysfs_show,
> > + .cpu_prepare = intel_pmu_cpu_prepare,
> > + .cpu_starting = intel_pmu_cpu_starting,
> > + .cpu_dying = intel_pmu_cpu_dying,
> > };
>
> Instead of adding prototype declarations, please arrange the x86_pmu
> definition's position so that it comes after the required functions -
> so that no prototypes are needed.
ok, v2 is attached
thanks,
jirka
---
The core_pmu does not define cpu_* callbacks, which handles
allocation of 'struct cpu_hw_events::shared_regs' data,
initialization of debug store and PMU_FL_EXCL_CNTRS counters.
While this probably won't happen on bare metal, virtual CPU can
define x86_pmu.extra_regs together with PMU version 1 and thus
be using core_pmu -> using shared_regs data without it being
allocated. That could could leave to following panic:
BUG: unable to handle kernel NULL pointer dereference at (null)
IP: [<ffffffff8152cd4f>] _spin_lock_irqsave+0x1f/0x40
SNIP
[<ffffffff81024bd9>] __intel_shared_reg_get_constraints+0x69/0x1e0
[<ffffffff81024deb>] intel_get_event_constraints+0x9b/0x180
[<ffffffff8101e815>] x86_schedule_events+0x75/0x1d0
[<ffffffff810586dc>] ? check_preempt_curr+0x7c/0x90
[<ffffffff810649fe>] ? try_to_wake_up+0x24e/0x3e0
[<ffffffff81064ba2>] ? default_wake_function+0x12/0x20
[<ffffffff8109eb16>] ? autoremove_wake_function+0x16/0x40
[<ffffffff810577e9>] ? __wake_up_common+0x59/0x90
[<ffffffff811a9517>] ? __d_lookup+0xa7/0x150
[<ffffffff8119db5f>] ? do_lookup+0x9f/0x230
[<ffffffff811a993a>] ? dput+0x9a/0x150
[<ffffffff8119c8f5>] ? path_to_nameidata+0x25/0x60
[<ffffffff8119e90a>] ? __link_path_walk+0x7da/0x1000
[<ffffffff8101d8f9>] ? x86_pmu_add+0xb9/0x170
[<ffffffff8101d7a7>] x86_pmu_commit_txn+0x67/0xc0
[<ffffffff811b07b0>] ? mntput_no_expire+0x30/0x110
[<ffffffff8119c731>] ? path_put+0x31/0x40
[<ffffffff8107c297>] ? current_fs_time+0x27/0x30
[<ffffffff8117d170>] ? mem_cgroup_get_reclaim_stat_from_page+0x20/0x70
[<ffffffff8111b7aa>] group_sched_in+0x13a/0x170
[<ffffffff81014a29>] ? sched_clock+0x9/0x10
[<ffffffff8111bac8>] ctx_sched_in+0x2e8/0x330
[<ffffffff8111bb7b>] perf_event_sched_in+0x6b/0xb0
[<ffffffff8111bc36>] perf_event_context_sched_in+0x76/0xc0
[<ffffffff8111eb3b>] perf_event_comm+0x1bb/0x2e0
[<ffffffff81195ee9>] set_task_comm+0x69/0x80
[<ffffffff81195fe1>] setup_new_exec+0xe1/0x2e0
[<ffffffff811ea68e>] load_elf_binary+0x3ce/0x1ab0
Adding cpu_(prepare|starting|dying) for core_pmu to have shared_regs
data allocated for core_pmu. AFAICS there's no harm to initialize
debug store and PMU_FL_EXCL_CNTRS either for core_pmu.
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
arch/x86/kernel/cpu/perf_event_intel.c | 59 ++++++++++++++++++----------------
1 file changed, 31 insertions(+), 28 deletions(-)
diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c
index 9da2400c2ec3..973abdc796fb 100644
--- a/arch/x86/kernel/cpu/perf_event_intel.c
+++ b/arch/x86/kernel/cpu/perf_event_intel.c
@@ -2533,34 +2533,6 @@ ssize_t intel_event_sysfs_show(char *page, u64 config)
return x86_event_sysfs_show(page, config, event);
}
-static __initconst const struct x86_pmu core_pmu = {
- .name = "core",
- .handle_irq = x86_pmu_handle_irq,
- .disable_all = x86_pmu_disable_all,
- .enable_all = core_pmu_enable_all,
- .enable = core_pmu_enable_event,
- .disable = x86_pmu_disable_event,
- .hw_config = x86_pmu_hw_config,
- .schedule_events = x86_schedule_events,
- .eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
- .perfctr = MSR_ARCH_PERFMON_PERFCTR0,
- .event_map = intel_pmu_event_map,
- .max_events = ARRAY_SIZE(intel_perfmon_event_map),
- .apic = 1,
- /*
- * Intel PMCs cannot be accessed sanely above 32 bit width,
- * so we install an artificial 1<<31 period regardless of
- * the generic event period:
- */
- .max_period = (1ULL << 31) - 1,
- .get_event_constraints = intel_get_event_constraints,
- .put_event_constraints = intel_put_event_constraints,
- .event_constraints = intel_core_event_constraints,
- .guest_get_msrs = core_guest_get_msrs,
- .format_attrs = intel_arch_formats_attr,
- .events_sysfs_show = intel_event_sysfs_show,
-};
-
struct intel_shared_regs *allocate_shared_regs(int cpu)
{
struct intel_shared_regs *regs;
@@ -2743,6 +2715,37 @@ static struct attribute *intel_arch3_formats_attr[] = {
NULL,
};
+static __initconst const struct x86_pmu core_pmu = {
+ .name = "core",
+ .handle_irq = x86_pmu_handle_irq,
+ .disable_all = x86_pmu_disable_all,
+ .enable_all = core_pmu_enable_all,
+ .enable = core_pmu_enable_event,
+ .disable = x86_pmu_disable_event,
+ .hw_config = x86_pmu_hw_config,
+ .schedule_events = x86_schedule_events,
+ .eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
+ .perfctr = MSR_ARCH_PERFMON_PERFCTR0,
+ .event_map = intel_pmu_event_map,
+ .max_events = ARRAY_SIZE(intel_perfmon_event_map),
+ .apic = 1,
+ /*
+ * Intel PMCs cannot be accessed sanely above 32 bit width,
+ * so we install an artificial 1<<31 period regardless of
+ * the generic event period:
+ */
+ .max_period = (1ULL << 31) - 1,
+ .get_event_constraints = intel_get_event_constraints,
+ .put_event_constraints = intel_put_event_constraints,
+ .event_constraints = intel_core_event_constraints,
+ .guest_get_msrs = core_guest_get_msrs,
+ .format_attrs = intel_arch_formats_attr,
+ .events_sysfs_show = intel_event_sysfs_show,
+ .cpu_prepare = intel_pmu_cpu_prepare,
+ .cpu_starting = intel_pmu_cpu_starting,
+ .cpu_dying = intel_pmu_cpu_dying,
+};
+
static __initconst const struct x86_pmu intel_pmu = {
.name = "Intel",
.handle_irq = intel_pmu_handle_irq,
--
1.9.3
^ permalink raw reply related [flat|nested] 8+ messages in thread
* Re: [PATCH] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
2015-04-21 15:12 ` [PATCH] " Peter Zijlstra
@ 2015-04-21 15:18 ` Jiri Olsa
2015-04-21 15:26 ` [PATCHv3] " Jiri Olsa
0 siblings, 1 reply; 8+ messages in thread
From: Jiri Olsa @ 2015-04-21 15:18 UTC (permalink / raw)
To: Peter Zijlstra
Cc: Jiri Olsa, lkml, Andi Kleen, Arnaldo Carvalho de Melo,
Ingo Molnar, Paul Mackerras, Stephane Eranian
On Tue, Apr 21, 2015 at 05:12:16PM +0200, Peter Zijlstra wrote:
> On Tue, Apr 21, 2015 at 10:54:25AM +0200, Jiri Olsa wrote:
> > @@ -2559,6 +2563,9 @@ static __initconst const struct x86_pmu core_pmu = {
> > .guest_get_msrs = core_guest_get_msrs,
> > .format_attrs = intel_arch_formats_attr,
> > .events_sysfs_show = intel_event_sysfs_show,
>
> Could you also insert a little comment here how these methods are for
> 'funny' 'hardware' ?
ook, will do v3
jirka
>
> > + .cpu_prepare = intel_pmu_cpu_prepare,
> > + .cpu_starting = intel_pmu_cpu_starting,
> > + .cpu_dying = intel_pmu_cpu_dying,
> > };
> >
> > struct intel_shared_regs *allocate_shared_regs(int cpu)
> > --
> > 1.9.3
> >
^ permalink raw reply [flat|nested] 8+ messages in thread
* [PATCHv3] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
2015-04-21 15:18 ` Jiri Olsa
@ 2015-04-21 15:26 ` Jiri Olsa
2015-04-21 21:00 ` Peter Zijlstra
2015-04-22 14:10 ` [tip:perf/urgent] perf/x86/intel: Add cpu_(prepare|starting|dying ) " tip-bot for Jiri Olsa
0 siblings, 2 replies; 8+ messages in thread
From: Jiri Olsa @ 2015-04-21 15:26 UTC (permalink / raw)
To: Peter Zijlstra
Cc: Jiri Olsa, lkml, Andi Kleen, Arnaldo Carvalho de Melo,
Ingo Molnar, Paul Mackerras, Stephane Eranian
On Tue, Apr 21, 2015 at 05:18:27PM +0200, Jiri Olsa wrote:
> On Tue, Apr 21, 2015 at 05:12:16PM +0200, Peter Zijlstra wrote:
> > On Tue, Apr 21, 2015 at 10:54:25AM +0200, Jiri Olsa wrote:
> > > @@ -2559,6 +2563,9 @@ static __initconst const struct x86_pmu core_pmu = {
> > > .guest_get_msrs = core_guest_get_msrs,
> > > .format_attrs = intel_arch_formats_attr,
> > > .events_sysfs_show = intel_event_sysfs_show,
> >
> > Could you also insert a little comment here how these methods are for
> > 'funny' 'hardware' ?
>
> ook, will do v3
attached,
jirka
---
The core_pmu does not define cpu_* callbacks, which handles
allocation of 'struct cpu_hw_events::shared_regs' data,
initialization of debug store and PMU_FL_EXCL_CNTRS counters.
While this probably won't happen on bare metal, virtual CPU can
define x86_pmu.extra_regs together with PMU version 1 and thus
be using core_pmu -> using shared_regs data without it being
allocated. That could could leave to following panic:
BUG: unable to handle kernel NULL pointer dereference at (null)
IP: [<ffffffff8152cd4f>] _spin_lock_irqsave+0x1f/0x40
SNIP
[<ffffffff81024bd9>] __intel_shared_reg_get_constraints+0x69/0x1e0
[<ffffffff81024deb>] intel_get_event_constraints+0x9b/0x180
[<ffffffff8101e815>] x86_schedule_events+0x75/0x1d0
[<ffffffff810586dc>] ? check_preempt_curr+0x7c/0x90
[<ffffffff810649fe>] ? try_to_wake_up+0x24e/0x3e0
[<ffffffff81064ba2>] ? default_wake_function+0x12/0x20
[<ffffffff8109eb16>] ? autoremove_wake_function+0x16/0x40
[<ffffffff810577e9>] ? __wake_up_common+0x59/0x90
[<ffffffff811a9517>] ? __d_lookup+0xa7/0x150
[<ffffffff8119db5f>] ? do_lookup+0x9f/0x230
[<ffffffff811a993a>] ? dput+0x9a/0x150
[<ffffffff8119c8f5>] ? path_to_nameidata+0x25/0x60
[<ffffffff8119e90a>] ? __link_path_walk+0x7da/0x1000
[<ffffffff8101d8f9>] ? x86_pmu_add+0xb9/0x170
[<ffffffff8101d7a7>] x86_pmu_commit_txn+0x67/0xc0
[<ffffffff811b07b0>] ? mntput_no_expire+0x30/0x110
[<ffffffff8119c731>] ? path_put+0x31/0x40
[<ffffffff8107c297>] ? current_fs_time+0x27/0x30
[<ffffffff8117d170>] ? mem_cgroup_get_reclaim_stat_from_page+0x20/0x70
[<ffffffff8111b7aa>] group_sched_in+0x13a/0x170
[<ffffffff81014a29>] ? sched_clock+0x9/0x10
[<ffffffff8111bac8>] ctx_sched_in+0x2e8/0x330
[<ffffffff8111bb7b>] perf_event_sched_in+0x6b/0xb0
[<ffffffff8111bc36>] perf_event_context_sched_in+0x76/0xc0
[<ffffffff8111eb3b>] perf_event_comm+0x1bb/0x2e0
[<ffffffff81195ee9>] set_task_comm+0x69/0x80
[<ffffffff81195fe1>] setup_new_exec+0xe1/0x2e0
[<ffffffff811ea68e>] load_elf_binary+0x3ce/0x1ab0
Adding cpu_(prepare|starting|dying) for core_pmu to have shared_regs
data allocated for core_pmu. AFAICS there's no harm to initialize
debug store and PMU_FL_EXCL_CNTRS either for core_pmu.
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
arch/x86/kernel/cpu/perf_event_intel.c | 66 +++++++++++++++++++---------------
1 file changed, 38 insertions(+), 28 deletions(-)
diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c
index 9da2400c2ec3..8a4e21eb2087 100644
--- a/arch/x86/kernel/cpu/perf_event_intel.c
+++ b/arch/x86/kernel/cpu/perf_event_intel.c
@@ -2533,34 +2533,6 @@ ssize_t intel_event_sysfs_show(char *page, u64 config)
return x86_event_sysfs_show(page, config, event);
}
-static __initconst const struct x86_pmu core_pmu = {
- .name = "core",
- .handle_irq = x86_pmu_handle_irq,
- .disable_all = x86_pmu_disable_all,
- .enable_all = core_pmu_enable_all,
- .enable = core_pmu_enable_event,
- .disable = x86_pmu_disable_event,
- .hw_config = x86_pmu_hw_config,
- .schedule_events = x86_schedule_events,
- .eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
- .perfctr = MSR_ARCH_PERFMON_PERFCTR0,
- .event_map = intel_pmu_event_map,
- .max_events = ARRAY_SIZE(intel_perfmon_event_map),
- .apic = 1,
- /*
- * Intel PMCs cannot be accessed sanely above 32 bit width,
- * so we install an artificial 1<<31 period regardless of
- * the generic event period:
- */
- .max_period = (1ULL << 31) - 1,
- .get_event_constraints = intel_get_event_constraints,
- .put_event_constraints = intel_put_event_constraints,
- .event_constraints = intel_core_event_constraints,
- .guest_get_msrs = core_guest_get_msrs,
- .format_attrs = intel_arch_formats_attr,
- .events_sysfs_show = intel_event_sysfs_show,
-};
-
struct intel_shared_regs *allocate_shared_regs(int cpu)
{
struct intel_shared_regs *regs;
@@ -2743,6 +2715,44 @@ static struct attribute *intel_arch3_formats_attr[] = {
NULL,
};
+static __initconst const struct x86_pmu core_pmu = {
+ .name = "core",
+ .handle_irq = x86_pmu_handle_irq,
+ .disable_all = x86_pmu_disable_all,
+ .enable_all = core_pmu_enable_all,
+ .enable = core_pmu_enable_event,
+ .disable = x86_pmu_disable_event,
+ .hw_config = x86_pmu_hw_config,
+ .schedule_events = x86_schedule_events,
+ .eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
+ .perfctr = MSR_ARCH_PERFMON_PERFCTR0,
+ .event_map = intel_pmu_event_map,
+ .max_events = ARRAY_SIZE(intel_perfmon_event_map),
+ .apic = 1,
+ /*
+ * Intel PMCs cannot be accessed sanely above 32 bit width,
+ * so we install an artificial 1<<31 period regardless of
+ * the generic event period:
+ */
+ .max_period = (1ULL << 31) - 1,
+ .get_event_constraints = intel_get_event_constraints,
+ .put_event_constraints = intel_put_event_constraints,
+ .event_constraints = intel_core_event_constraints,
+ .guest_get_msrs = core_guest_get_msrs,
+ .format_attrs = intel_arch_formats_attr,
+ .events_sysfs_show = intel_event_sysfs_show,
+
+ /*
+ * Virtual (or funny metal) CPU can define x86_pmu.extra_regs
+ * together with PMU version 1 and thus be using core_pmu with
+ * shared_regs. We need following callbacks here to allocate
+ * it properly.
+ */
+ .cpu_prepare = intel_pmu_cpu_prepare,
+ .cpu_starting = intel_pmu_cpu_starting,
+ .cpu_dying = intel_pmu_cpu_dying,
+};
+
static __initconst const struct x86_pmu intel_pmu = {
.name = "Intel",
.handle_irq = intel_pmu_handle_irq,
--
1.9.3
^ permalink raw reply related [flat|nested] 8+ messages in thread
* Re: [PATCHv3] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
2015-04-21 15:26 ` [PATCHv3] " Jiri Olsa
@ 2015-04-21 21:00 ` Peter Zijlstra
2015-04-22 14:10 ` [tip:perf/urgent] perf/x86/intel: Add cpu_(prepare|starting|dying ) " tip-bot for Jiri Olsa
1 sibling, 0 replies; 8+ messages in thread
From: Peter Zijlstra @ 2015-04-21 21:00 UTC (permalink / raw)
To: Jiri Olsa
Cc: Jiri Olsa, lkml, Andi Kleen, Arnaldo Carvalho de Melo,
Ingo Molnar, Paul Mackerras, Stephane Eranian
On Tue, Apr 21, 2015 at 05:26:23PM +0200, Jiri Olsa wrote:
> The core_pmu does not define cpu_* callbacks, which handles
> allocation of 'struct cpu_hw_events::shared_regs' data,
> initialization of debug store and PMU_FL_EXCL_CNTRS counters.
>
> While this probably won't happen on bare metal, virtual CPU can
> define x86_pmu.extra_regs together with PMU version 1 and thus
> be using core_pmu -> using shared_regs data without it being
> allocated. That could could leave to following panic:
>
> BUG: unable to handle kernel NULL pointer dereference at (null)
> IP: [<ffffffff8152cd4f>] _spin_lock_irqsave+0x1f/0x40
>
> SNIP
>
> [<ffffffff81024bd9>] __intel_shared_reg_get_constraints+0x69/0x1e0
> [<ffffffff81024deb>] intel_get_event_constraints+0x9b/0x180
> [<ffffffff8101e815>] x86_schedule_events+0x75/0x1d0
> [<ffffffff810586dc>] ? check_preempt_curr+0x7c/0x90
> [<ffffffff810649fe>] ? try_to_wake_up+0x24e/0x3e0
> [<ffffffff81064ba2>] ? default_wake_function+0x12/0x20
> [<ffffffff8109eb16>] ? autoremove_wake_function+0x16/0x40
> [<ffffffff810577e9>] ? __wake_up_common+0x59/0x90
> [<ffffffff811a9517>] ? __d_lookup+0xa7/0x150
> [<ffffffff8119db5f>] ? do_lookup+0x9f/0x230
> [<ffffffff811a993a>] ? dput+0x9a/0x150
> [<ffffffff8119c8f5>] ? path_to_nameidata+0x25/0x60
> [<ffffffff8119e90a>] ? __link_path_walk+0x7da/0x1000
> [<ffffffff8101d8f9>] ? x86_pmu_add+0xb9/0x170
> [<ffffffff8101d7a7>] x86_pmu_commit_txn+0x67/0xc0
> [<ffffffff811b07b0>] ? mntput_no_expire+0x30/0x110
> [<ffffffff8119c731>] ? path_put+0x31/0x40
> [<ffffffff8107c297>] ? current_fs_time+0x27/0x30
> [<ffffffff8117d170>] ? mem_cgroup_get_reclaim_stat_from_page+0x20/0x70
> [<ffffffff8111b7aa>] group_sched_in+0x13a/0x170
> [<ffffffff81014a29>] ? sched_clock+0x9/0x10
> [<ffffffff8111bac8>] ctx_sched_in+0x2e8/0x330
> [<ffffffff8111bb7b>] perf_event_sched_in+0x6b/0xb0
> [<ffffffff8111bc36>] perf_event_context_sched_in+0x76/0xc0
> [<ffffffff8111eb3b>] perf_event_comm+0x1bb/0x2e0
> [<ffffffff81195ee9>] set_task_comm+0x69/0x80
> [<ffffffff81195fe1>] setup_new_exec+0xe1/0x2e0
> [<ffffffff811ea68e>] load_elf_binary+0x3ce/0x1ab0
>
> Adding cpu_(prepare|starting|dying) for core_pmu to have shared_regs
> data allocated for core_pmu. AFAICS there's no harm to initialize
> debug store and PMU_FL_EXCL_CNTRS either for core_pmu.
>
> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
^ permalink raw reply [flat|nested] 8+ messages in thread
* [tip:perf/urgent] perf/x86/intel: Add cpu_(prepare|starting|dying ) for core_pmu
2015-04-21 15:26 ` [PATCHv3] " Jiri Olsa
2015-04-21 21:00 ` Peter Zijlstra
@ 2015-04-22 14:10 ` tip-bot for Jiri Olsa
1 sibling, 0 replies; 8+ messages in thread
From: tip-bot for Jiri Olsa @ 2015-04-22 14:10 UTC (permalink / raw)
To: linux-tip-commits
Cc: peterz, jolsa, jolsa, acme, mingo, linux-kernel, hpa, eranian,
tglx, paulus
Commit-ID: 3b6e042188994466ec257b71296b5f85b894dcd9
Gitweb: http://git.kernel.org/tip/3b6e042188994466ec257b71296b5f85b894dcd9
Author: Jiri Olsa <jolsa@redhat.com>
AuthorDate: Tue, 21 Apr 2015 17:26:23 +0200
Committer: Ingo Molnar <mingo@kernel.org>
CommitDate: Wed, 22 Apr 2015 08:24:33 +0200
perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu
The core_pmu does not define cpu_* callbacks, which handles
allocation of 'struct cpu_hw_events::shared_regs' data,
initialization of debug store and PMU_FL_EXCL_CNTRS counters.
While this probably won't happen on bare metal, virtual CPU can
define x86_pmu.extra_regs together with PMU version 1 and thus
be using core_pmu -> using shared_regs data without it being
allocated. That could could leave to following panic:
BUG: unable to handle kernel NULL pointer dereference at (null)
IP: [<ffffffff8152cd4f>] _spin_lock_irqsave+0x1f/0x40
SNIP
[<ffffffff81024bd9>] __intel_shared_reg_get_constraints+0x69/0x1e0
[<ffffffff81024deb>] intel_get_event_constraints+0x9b/0x180
[<ffffffff8101e815>] x86_schedule_events+0x75/0x1d0
[<ffffffff810586dc>] ? check_preempt_curr+0x7c/0x90
[<ffffffff810649fe>] ? try_to_wake_up+0x24e/0x3e0
[<ffffffff81064ba2>] ? default_wake_function+0x12/0x20
[<ffffffff8109eb16>] ? autoremove_wake_function+0x16/0x40
[<ffffffff810577e9>] ? __wake_up_common+0x59/0x90
[<ffffffff811a9517>] ? __d_lookup+0xa7/0x150
[<ffffffff8119db5f>] ? do_lookup+0x9f/0x230
[<ffffffff811a993a>] ? dput+0x9a/0x150
[<ffffffff8119c8f5>] ? path_to_nameidata+0x25/0x60
[<ffffffff8119e90a>] ? __link_path_walk+0x7da/0x1000
[<ffffffff8101d8f9>] ? x86_pmu_add+0xb9/0x170
[<ffffffff8101d7a7>] x86_pmu_commit_txn+0x67/0xc0
[<ffffffff811b07b0>] ? mntput_no_expire+0x30/0x110
[<ffffffff8119c731>] ? path_put+0x31/0x40
[<ffffffff8107c297>] ? current_fs_time+0x27/0x30
[<ffffffff8117d170>] ? mem_cgroup_get_reclaim_stat_from_page+0x20/0x70
[<ffffffff8111b7aa>] group_sched_in+0x13a/0x170
[<ffffffff81014a29>] ? sched_clock+0x9/0x10
[<ffffffff8111bac8>] ctx_sched_in+0x2e8/0x330
[<ffffffff8111bb7b>] perf_event_sched_in+0x6b/0xb0
[<ffffffff8111bc36>] perf_event_context_sched_in+0x76/0xc0
[<ffffffff8111eb3b>] perf_event_comm+0x1bb/0x2e0
[<ffffffff81195ee9>] set_task_comm+0x69/0x80
[<ffffffff81195fe1>] setup_new_exec+0xe1/0x2e0
[<ffffffff811ea68e>] load_elf_binary+0x3ce/0x1ab0
Adding cpu_(prepare|starting|dying) for core_pmu to have
shared_regs data allocated for core_pmu. AFAICS there's no harm
to initialize debug store and PMU_FL_EXCL_CNTRS either for
core_pmu.
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Arnaldo Carvalho de Melo <acme@kernel.org>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Stephane Eranian <eranian@google.com>
Link: http://lkml.kernel.org/r/20150421152623.GC13169@krava.redhat.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
---
arch/x86/kernel/cpu/perf_event_intel.c | 66 +++++++++++++++++++---------------
1 file changed, 38 insertions(+), 28 deletions(-)
diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c
index 219d3fb..960e85d 100644
--- a/arch/x86/kernel/cpu/perf_event_intel.c
+++ b/arch/x86/kernel/cpu/perf_event_intel.c
@@ -2533,34 +2533,6 @@ ssize_t intel_event_sysfs_show(char *page, u64 config)
return x86_event_sysfs_show(page, config, event);
}
-static __initconst const struct x86_pmu core_pmu = {
- .name = "core",
- .handle_irq = x86_pmu_handle_irq,
- .disable_all = x86_pmu_disable_all,
- .enable_all = core_pmu_enable_all,
- .enable = core_pmu_enable_event,
- .disable = x86_pmu_disable_event,
- .hw_config = x86_pmu_hw_config,
- .schedule_events = x86_schedule_events,
- .eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
- .perfctr = MSR_ARCH_PERFMON_PERFCTR0,
- .event_map = intel_pmu_event_map,
- .max_events = ARRAY_SIZE(intel_perfmon_event_map),
- .apic = 1,
- /*
- * Intel PMCs cannot be accessed sanely above 32 bit width,
- * so we install an artificial 1<<31 period regardless of
- * the generic event period:
- */
- .max_period = (1ULL << 31) - 1,
- .get_event_constraints = intel_get_event_constraints,
- .put_event_constraints = intel_put_event_constraints,
- .event_constraints = intel_core_event_constraints,
- .guest_get_msrs = core_guest_get_msrs,
- .format_attrs = intel_arch_formats_attr,
- .events_sysfs_show = intel_event_sysfs_show,
-};
-
struct intel_shared_regs *allocate_shared_regs(int cpu)
{
struct intel_shared_regs *regs;
@@ -2743,6 +2715,44 @@ static struct attribute *intel_arch3_formats_attr[] = {
NULL,
};
+static __initconst const struct x86_pmu core_pmu = {
+ .name = "core",
+ .handle_irq = x86_pmu_handle_irq,
+ .disable_all = x86_pmu_disable_all,
+ .enable_all = core_pmu_enable_all,
+ .enable = core_pmu_enable_event,
+ .disable = x86_pmu_disable_event,
+ .hw_config = x86_pmu_hw_config,
+ .schedule_events = x86_schedule_events,
+ .eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
+ .perfctr = MSR_ARCH_PERFMON_PERFCTR0,
+ .event_map = intel_pmu_event_map,
+ .max_events = ARRAY_SIZE(intel_perfmon_event_map),
+ .apic = 1,
+ /*
+ * Intel PMCs cannot be accessed sanely above 32-bit width,
+ * so we install an artificial 1<<31 period regardless of
+ * the generic event period:
+ */
+ .max_period = (1ULL<<31) - 1,
+ .get_event_constraints = intel_get_event_constraints,
+ .put_event_constraints = intel_put_event_constraints,
+ .event_constraints = intel_core_event_constraints,
+ .guest_get_msrs = core_guest_get_msrs,
+ .format_attrs = intel_arch_formats_attr,
+ .events_sysfs_show = intel_event_sysfs_show,
+
+ /*
+ * Virtual (or funny metal) CPU can define x86_pmu.extra_regs
+ * together with PMU version 1 and thus be using core_pmu with
+ * shared_regs. We need following callbacks here to allocate
+ * it properly.
+ */
+ .cpu_prepare = intel_pmu_cpu_prepare,
+ .cpu_starting = intel_pmu_cpu_starting,
+ .cpu_dying = intel_pmu_cpu_dying,
+};
+
static __initconst const struct x86_pmu intel_pmu = {
.name = "Intel",
.handle_irq = intel_pmu_handle_irq,
^ permalink raw reply related [flat|nested] 8+ messages in thread
end of thread, other threads:[~2015-04-22 14:11 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-04-21 8:54 [PATCH] perf/x86/intel: Add cpu_(prepare|starting|dying) for core_pmu Jiri Olsa
2015-04-21 10:52 ` Ingo Molnar
2015-04-21 15:14 ` [PATCHv2] " Jiri Olsa
2015-04-21 15:12 ` [PATCH] " Peter Zijlstra
2015-04-21 15:18 ` Jiri Olsa
2015-04-21 15:26 ` [PATCHv3] " Jiri Olsa
2015-04-21 21:00 ` Peter Zijlstra
2015-04-22 14:10 ` [tip:perf/urgent] perf/x86/intel: Add cpu_(prepare|starting|dying ) " tip-bot for Jiri Olsa
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox