From: "Liang, Kan" <kan.liang@linux.intel.com>
To: weilin.wang@intel.com, Ian Rogers <irogers@google.com>,
Peter Zijlstra <peterz@infradead.org>,
Ingo Molnar <mingo@redhat.com>,
Arnaldo Carvalho de Melo <acme@kernel.org>,
Alexander Shishkin <alexander.shishkin@linux.intel.com>,
Jiri Olsa <jolsa@kernel.org>, Namhyung Kim <namhyung@kernel.org>,
Adrian Hunter <adrian.hunter@intel.com>
Cc: linux-perf-users@vger.kernel.org, linux-kernel@vger.kernel.org,
Perry Taylor <perry.taylor@intel.com>,
Samantha Alt <samantha.alt@intel.com>,
Caleb Biggers <caleb.biggers@intel.com>,
Mark Rutland <mark.rutland@arm.com>
Subject: Re: [RFC PATCH 01/25] perf stat: Add hardware-grouping cmd option to perf stat
Date: Tue, 26 Sep 2023 10:50:21 -0400 [thread overview]
Message-ID: <52b13a04-2d14-5011-01c9-0989aa47d40d@linux.intel.com> (raw)
In-Reply-To: <20230925061824.3818631-2-weilin.wang@intel.com>
On 2023-09-25 2:18 a.m., weilin.wang@intel.com wrote:
> From: Weilin Wang <weilin.wang@intel.com>
>
> Hardware counter and event information could be used to help creating event
> groups that better utilize hardware counters and improve multiplexing.
>
> Add a cmd option to allow user to choose this new metric grouping method.
>
It's better to mention this option is only available with -M.
> $ perf stat -M TopdownL1 -a --hardware-grouping
>
I think it should be counter aware grouping. Hardware grouping sounds
like the group is created by hardware.
Eventually we should make it default and let the default grouping code
take the counter information into account.
Thanks,
Kan
> Signed-off-by: Weilin Wang <weilin.wang@intel.com>
> ---
> tools/perf/builtin-stat.c | 7 +++++++
> tools/perf/util/metricgroup.c | 5 +++++
> tools/perf/util/metricgroup.h | 1 +
> tools/perf/util/stat.h | 1 +
> 4 files changed, 14 insertions(+)
>
> diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c
> index 07b48f6df..1a7dbfff5 100644
> --- a/tools/perf/builtin-stat.c
> +++ b/tools/perf/builtin-stat.c
> @@ -1256,6 +1256,8 @@ static struct option stat_options[] = {
> "don't try to share events between metrics in a group"),
> OPT_BOOLEAN(0, "metric-no-threshold", &stat_config.metric_no_threshold,
> "don't try to share events between metrics in a group "),
> + OPT_BOOLEAN(0, "hardware-grouping", &stat_config.hardware_aware_grouping,
> + "Use hardware aware metric grouping method"),
> OPT_BOOLEAN(0, "topdown", &topdown_run,
> "measure top-down statistics"),
> OPT_UINTEGER(0, "td-level", &stat_config.topdown_level,
> @@ -2061,6 +2063,7 @@ static int add_default_attributes(void)
> stat_config.metric_no_threshold,
> stat_config.user_requested_cpu_list,
> stat_config.system_wide,
> + stat_config.hardware_aware_grouping,
> &stat_config.metric_events);
> }
>
> @@ -2094,6 +2097,7 @@ static int add_default_attributes(void)
> stat_config.metric_no_threshold,
> stat_config.user_requested_cpu_list,
> stat_config.system_wide,
> + stat_config.hardware_aware_grouping,
> &stat_config.metric_events);
> }
>
> @@ -2128,6 +2132,7 @@ static int add_default_attributes(void)
> /*metric_no_threshold=*/true,
> stat_config.user_requested_cpu_list,
> stat_config.system_wide,
> + stat_config.hardware_aware_grouping,
> &stat_config.metric_events) < 0)
> return -1;
> }
> @@ -2169,6 +2174,7 @@ static int add_default_attributes(void)
> /*metric_no_threshold=*/true,
> stat_config.user_requested_cpu_list,
> stat_config.system_wide,
> + stat_config.hardware_aware_grouping,
> &stat_config.metric_events) < 0)
> return -1;
>
> @@ -2702,6 +2708,7 @@ int cmd_stat(int argc, const char **argv)
> stat_config.metric_no_threshold,
> stat_config.user_requested_cpu_list,
> stat_config.system_wide,
> + stat_config.hardware_aware_grouping,
> &stat_config.metric_events);
> zfree(&metrics);
> }
> diff --git a/tools/perf/util/metricgroup.c b/tools/perf/util/metricgroup.c
> index 6231044a4..b08af6860 100644
> --- a/tools/perf/util/metricgroup.c
> +++ b/tools/perf/util/metricgroup.c
> @@ -1690,12 +1690,17 @@ int metricgroup__parse_groups(struct evlist *perf_evlist,
> bool metric_no_threshold,
> const char *user_requested_cpu_list,
> bool system_wide,
> + bool hardware_aware_grouping,
> struct rblist *metric_events)
> {
> const struct pmu_metrics_table *table = pmu_metrics_table__find();
>
> if (!table)
> return -EINVAL;
> + if (hardware_aware_grouping) {
> + pr_debug("Use hardware aware grouping instead of traditional metric grouping method\n");
> + }
> +
>
> return parse_groups(perf_evlist, pmu, str, metric_no_group, metric_no_merge,
> metric_no_threshold, user_requested_cpu_list, system_wide,
> diff --git a/tools/perf/util/metricgroup.h b/tools/perf/util/metricgroup.h
> index d5325c6ec..779f6ede1 100644
> --- a/tools/perf/util/metricgroup.h
> +++ b/tools/perf/util/metricgroup.h
> @@ -77,6 +77,7 @@ int metricgroup__parse_groups(struct evlist *perf_evlist,
> bool metric_no_threshold,
> const char *user_requested_cpu_list,
> bool system_wide,
> + bool hardware_aware_grouping,
> struct rblist *metric_events);
> int metricgroup__parse_groups_test(struct evlist *evlist,
> const struct pmu_metrics_table *table,
> diff --git a/tools/perf/util/stat.h b/tools/perf/util/stat.h
> index 325d0fad1..f97d6282b 100644
> --- a/tools/perf/util/stat.h
> +++ b/tools/perf/util/stat.h
> @@ -87,6 +87,7 @@ struct perf_stat_config {
> bool metric_no_group;
> bool metric_no_merge;
> bool metric_no_threshold;
> + bool hardware_aware_grouping;
> bool stop_read_counter;
> bool iostat_run;
> char *user_requested_cpu_list;
next prev parent reply other threads:[~2023-09-26 14:50 UTC|newest]
Thread overview: 49+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-09-25 6:17 [RFC PATCH 00/25] Perf stat metric grouping with hardware information weilin.wang
2023-09-25 6:18 ` [RFC PATCH 01/25] perf stat: Add hardware-grouping cmd option to perf stat weilin.wang
2023-09-26 14:50 ` Liang, Kan [this message]
2023-09-25 6:18 ` [RFC PATCH 02/25] perf stat: Add basic functions for the hardware-grouping stat cmd option weilin.wang
2023-09-26 15:10 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 03/25] perf pmu-events: Add functions in jevent.py weilin.wang
2023-09-26 15:17 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 04/25] perf pmu-events: Add counter info into JSON files for SapphireRapids weilin.wang
2023-09-26 15:20 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 05/25] perf pmu-events: Add event counter data for Cascadelakex weilin.wang
2023-09-25 6:18 ` [RFC PATCH 06/25] perf pmu-events: Add event counter data for Icelakex weilin.wang
2023-09-25 6:18 ` [RFC PATCH 07/25] perf stat: Add helper functions for hardware-grouping method weilin.wang
2023-09-26 15:28 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 08/25] perf stat: Add functions to get counter info weilin.wang
2023-09-26 15:37 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 09/25] perf stat: Add helper functions for hardware-grouping method weilin.wang
2023-09-26 3:37 ` Yang Jihong
2023-09-26 20:51 ` Wang, Weilin
2023-09-25 6:18 ` [RFC PATCH 10/25] perf stat: Add helper functions to " weilin.wang
2023-09-26 3:44 ` Yang Jihong
2023-09-26 15:55 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 11/25] perf stat: Add utility " weilin.wang
2023-09-26 16:02 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 12/25] perf stat: Add more functions for " weilin.wang
2023-09-25 6:18 ` [RFC PATCH 13/25] perf stat: Add functions to " weilin.wang
2023-09-26 16:18 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 14/25] perf stat: Add build string function and topdown events handling in hardware-grouping weilin.wang
2023-09-26 16:21 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 15/25] perf stat: Add function to combine metrics for hardware-grouping weilin.wang
2023-09-25 6:18 ` [RFC PATCH 16/25] perf stat: Update keyword core to default_core to adjust to the changes for events with no unit weilin.wang
2023-09-26 16:25 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 17/25] perf stat: Handle taken alone in hardware-grouping weilin.wang
2023-09-25 6:18 ` [RFC PATCH 18/25] perf stat: Handle NMI " weilin.wang
2023-09-25 6:18 ` [RFC PATCH 19/25] perf stat: Handle grouping method fall back " weilin.wang
2023-09-25 6:18 ` [RFC PATCH 20/25] perf stat: Code refactoring " weilin.wang
2023-09-25 6:18 ` [RFC PATCH 21/25] perf stat: Add tool events support " weilin.wang
2023-09-25 6:18 ` [RFC PATCH 22/25] perf stat: Add TSC " weilin.wang
2023-09-26 16:35 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 23/25] perf stat: Fix a return error issue " weilin.wang
2023-09-26 16:36 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 24/25] perf stat: Add check to ensure correctness in platform that does not support hardware-grouping weilin.wang
2023-09-26 16:38 ` Liang, Kan
2023-09-25 6:18 ` [RFC PATCH 25/25] perf pmu-events: Add event counter data for Tigerlake weilin.wang
2023-09-26 16:41 ` Liang, Kan
2023-09-25 18:29 ` [RFC PATCH 00/25] Perf stat metric grouping with hardware information Ian Rogers
2023-09-26 20:40 ` Wang, Weilin
2023-09-26 14:43 ` Liang, Kan
2023-09-26 16:48 ` Liang, Kan
2023-09-26 20:40 ` Wang, Weilin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=52b13a04-2d14-5011-01c9-0989aa47d40d@linux.intel.com \
--to=kan.liang@linux.intel.com \
--cc=acme@kernel.org \
--cc=adrian.hunter@intel.com \
--cc=alexander.shishkin@linux.intel.com \
--cc=caleb.biggers@intel.com \
--cc=irogers@google.com \
--cc=jolsa@kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-perf-users@vger.kernel.org \
--cc=mark.rutland@arm.com \
--cc=mingo@redhat.com \
--cc=namhyung@kernel.org \
--cc=perry.taylor@intel.com \
--cc=peterz@infradead.org \
--cc=samantha.alt@intel.com \
--cc=weilin.wang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).