* [PATCH] perf stat: Fix wrong skipping for per-die aggregation
@ 2020-12-16 7:01 Jin Yao
2020-12-23 22:35 ` Jiri Olsa
0 siblings, 1 reply; 3+ messages in thread
From: Jin Yao @ 2020-12-16 7:01 UTC (permalink / raw)
To: acme, jolsa, peterz, mingo, alexander.shishkin
Cc: Linux-kernel, ak, kan.liang, yao.jin, Jin Yao
Uncore becomes die-scope on Xeon Cascade Lake-AP and perf has supported
--per-die aggregation yet.
One issue is found in check_per_pkg() for uncore events running on
AP system. On cascade Lake-AP, we have:
S0-D0
S0-D1
S1-D0
S1-D1
But in check_per_pkg(), S0-D1 and S1-D1 are skipped because the
mask bits for S0 and S1 have been set for S0-D0 and S1-D0. It doesn't
check die_id. So the counting for S0-D1 and S1-D1 are set to zero.
That's not correct.
root@lkp-csl-2ap4 ~# ./perf stat -a -I 1000 -e llc_misses.mem_read --per-die -- sleep 5
1.001460963 S0-D0 1 1317376 Bytes llc_misses.mem_read
1.001460963 S0-D1 1 998016 Bytes llc_misses.mem_read
1.001460963 S1-D0 1 970496 Bytes llc_misses.mem_read
1.001460963 S1-D1 1 1291264 Bytes llc_misses.mem_read
2.003488021 S0-D0 1 1082048 Bytes llc_misses.mem_read
2.003488021 S0-D1 1 1919040 Bytes llc_misses.mem_read
2.003488021 S1-D0 1 890752 Bytes llc_misses.mem_read
2.003488021 S1-D1 1 2380800 Bytes llc_misses.mem_read
3.005613270 S0-D0 1 1126080 Bytes llc_misses.mem_read
3.005613270 S0-D1 1 2898176 Bytes llc_misses.mem_read
3.005613270 S1-D0 1 870912 Bytes llc_misses.mem_read
3.005613270 S1-D1 1 3388608 Bytes llc_misses.mem_read
4.007627598 S0-D0 1 1124608 Bytes llc_misses.mem_read
4.007627598 S0-D1 1 3884416 Bytes llc_misses.mem_read
4.007627598 S1-D0 1 921088 Bytes llc_misses.mem_read
4.007627598 S1-D1 1 4451840 Bytes llc_misses.mem_read
5.001479927 S0-D0 1 963328 Bytes llc_misses.mem_read
5.001479927 S0-D1 1 4831936 Bytes llc_misses.mem_read
5.001479927 S1-D0 1 895104 Bytes llc_misses.mem_read
5.001479927 S1-D1 1 5496640 Bytes llc_misses.mem_read
From above output, we can see S0-D1 and S1-D1 don't report the interval
values, they are continued to grow. That's because check_per_pkg() wrongly
decides to use zero counts for S0-D1 and S1-D1.
So in check_per_pkg(), we should use a pair of die_id + socket_id to
decide if this cpu counts needs to skip. Only considering socket_id is
not enough.
Now with this patch,
root@lkp-csl-2ap4 ~# ./perf stat -a -I 1000 -e llc_misses.mem_read --per-die -- sleep 5
1.001622560 S0-D0 1 1210176 Bytes llc_misses.mem_read
1.001622560 S0-D1 1 950208 Bytes llc_misses.mem_read
1.001622560 S1-D0 1 968704 Bytes llc_misses.mem_read
1.001622560 S1-D1 1 1113280 Bytes llc_misses.mem_read
2.004399430 S0-D0 1 1258560 Bytes llc_misses.mem_read
2.004399430 S0-D1 1 1011008 Bytes llc_misses.mem_read
2.004399430 S1-D0 1 1036544 Bytes llc_misses.mem_read
2.004399430 S1-D1 1 1237376 Bytes llc_misses.mem_read
3.006535657 S0-D0 1 1036608 Bytes llc_misses.mem_read
3.006535657 S0-D1 1 906560 Bytes llc_misses.mem_read
3.006535657 S1-D0 1 994112 Bytes llc_misses.mem_read
3.006535657 S1-D1 1 1025472 Bytes llc_misses.mem_read
4.008934402 S0-D0 1 1005120 Bytes llc_misses.mem_read
4.008934402 S0-D1 1 882368 Bytes llc_misses.mem_read
4.008934402 S1-D0 1 961408 Bytes llc_misses.mem_read
4.008934402 S1-D1 1 988352 Bytes llc_misses.mem_read
5.001605406 S0-D0 1 1930240 Bytes llc_misses.mem_read
5.001605406 S0-D1 1 1134272 Bytes llc_misses.mem_read
5.001605406 S1-D0 1 1531136 Bytes llc_misses.mem_read
5.001605406 S1-D1 1 2230848 Bytes llc_misses.mem_read
On no-die system, die_id is 0, this patch keeps original behavior unchanged.
Reported-by: Huang Ying <ying.huang@intel.com>
Signed-off-by: Jin Yao <yao.jin@linux.intel.com>
---
tools/perf/util/cpumap.h | 2 ++
tools/perf/util/stat.c | 21 +++++++++++++++++----
2 files changed, 19 insertions(+), 4 deletions(-)
diff --git a/tools/perf/util/cpumap.h b/tools/perf/util/cpumap.h
index 3a442f021468..58604c047121 100644
--- a/tools/perf/util/cpumap.h
+++ b/tools/perf/util/cpumap.h
@@ -7,6 +7,8 @@
#include <internal/cpumap.h>
#include <perf/cpumap.h>
+#define CPU_MAP_MAX_DIE 256
+
struct perf_record_cpu_map_data;
struct perf_cpu_map *perf_cpu_map__empty_new(int nr);
diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c
index 1e125e39ff84..b3fc6e261fe0 100644
--- a/tools/perf/util/stat.c
+++ b/tools/perf/util/stat.c
@@ -282,9 +282,9 @@ static void zero_per_pkg(struct evsel *counter)
static int check_per_pkg(struct evsel *counter,
struct perf_counts_values *vals, int cpu, bool *skip)
{
- unsigned long *mask = counter->per_pkg_mask;
+ unsigned long *mask = counter->per_pkg_mask, *l;
struct perf_cpu_map *cpus = evsel__cpus(counter);
- int s;
+ int s, d;
*skip = false;
@@ -295,7 +295,7 @@ static int check_per_pkg(struct evsel *counter,
return 0;
if (!mask) {
- mask = zalloc(cpu__max_cpu());
+ mask = zalloc(cpu__max_cpu() * CPU_MAP_MAX_DIE);
if (!mask)
return -ENOMEM;
@@ -317,7 +317,20 @@ static int check_per_pkg(struct evsel *counter,
if (s < 0)
return -1;
- *skip = test_and_set_bit(s, mask) == 1;
+ d = cpu_map__get_die(cpus, cpu, NULL);
+ if (d < 0)
+ return -1;
+
+ /*
+ * On multi-die system, die_id < 256. We use a pair
+ * of socket + die to identify the used bit.
+ * On no-die system, die_id is 0, l = &mask[0]. It doesn't
+ * change original "test_and_set_bit(s, mask)" behavior.
+ */
+ d &= 0xff;
+ l = &mask[d];
+
+ *skip = test_and_set_bit(s, l) == 1;
return 0;
}
--
2.17.1
^ permalink raw reply related [flat|nested] 3+ messages in thread* Re: [PATCH] perf stat: Fix wrong skipping for per-die aggregation 2020-12-16 7:01 [PATCH] perf stat: Fix wrong skipping for per-die aggregation Jin Yao @ 2020-12-23 22:35 ` Jiri Olsa 2020-12-24 1:24 ` Jin, Yao 0 siblings, 1 reply; 3+ messages in thread From: Jiri Olsa @ 2020-12-23 22:35 UTC (permalink / raw) To: Jin Yao Cc: acme, jolsa, peterz, mingo, alexander.shishkin, Linux-kernel, ak, kan.liang, yao.jin, James Clark On Wed, Dec 16, 2020 at 03:01:46PM +0800, Jin Yao wrote: > Uncore becomes die-scope on Xeon Cascade Lake-AP and perf has supported > --per-die aggregation yet. > > One issue is found in check_per_pkg() for uncore events running on > AP system. On cascade Lake-AP, we have: > > S0-D0 > S0-D1 > S1-D0 > S1-D1 > > But in check_per_pkg(), S0-D1 and S1-D1 are skipped because the > mask bits for S0 and S1 have been set for S0-D0 and S1-D0. It doesn't > check die_id. So the counting for S0-D1 and S1-D1 are set to zero. > That's not correct. > > root@lkp-csl-2ap4 ~# ./perf stat -a -I 1000 -e llc_misses.mem_read --per-die -- sleep 5 > 1.001460963 S0-D0 1 1317376 Bytes llc_misses.mem_read > 1.001460963 S0-D1 1 998016 Bytes llc_misses.mem_read > 1.001460963 S1-D0 1 970496 Bytes llc_misses.mem_read > 1.001460963 S1-D1 1 1291264 Bytes llc_misses.mem_read > 2.003488021 S0-D0 1 1082048 Bytes llc_misses.mem_read > 2.003488021 S0-D1 1 1919040 Bytes llc_misses.mem_read > 2.003488021 S1-D0 1 890752 Bytes llc_misses.mem_read > 2.003488021 S1-D1 1 2380800 Bytes llc_misses.mem_read > 3.005613270 S0-D0 1 1126080 Bytes llc_misses.mem_read > 3.005613270 S0-D1 1 2898176 Bytes llc_misses.mem_read > 3.005613270 S1-D0 1 870912 Bytes llc_misses.mem_read > 3.005613270 S1-D1 1 3388608 Bytes llc_misses.mem_read > 4.007627598 S0-D0 1 1124608 Bytes llc_misses.mem_read > 4.007627598 S0-D1 1 3884416 Bytes llc_misses.mem_read > 4.007627598 S1-D0 1 921088 Bytes llc_misses.mem_read > 4.007627598 S1-D1 1 4451840 Bytes llc_misses.mem_read > 5.001479927 S0-D0 1 963328 Bytes llc_misses.mem_read > 5.001479927 S0-D1 1 4831936 Bytes llc_misses.mem_read > 5.001479927 S1-D0 1 895104 Bytes llc_misses.mem_read > 5.001479927 S1-D1 1 5496640 Bytes llc_misses.mem_read > > From above output, we can see S0-D1 and S1-D1 don't report the interval > values, they are continued to grow. That's because check_per_pkg() wrongly > decides to use zero counts for S0-D1 and S1-D1. > > So in check_per_pkg(), we should use a pair of die_id + socket_id to > decide if this cpu counts needs to skip. Only considering socket_id is > not enough. > > Now with this patch, > > root@lkp-csl-2ap4 ~# ./perf stat -a -I 1000 -e llc_misses.mem_read --per-die -- sleep 5 > 1.001622560 S0-D0 1 1210176 Bytes llc_misses.mem_read > 1.001622560 S0-D1 1 950208 Bytes llc_misses.mem_read > 1.001622560 S1-D0 1 968704 Bytes llc_misses.mem_read > 1.001622560 S1-D1 1 1113280 Bytes llc_misses.mem_read > 2.004399430 S0-D0 1 1258560 Bytes llc_misses.mem_read > 2.004399430 S0-D1 1 1011008 Bytes llc_misses.mem_read > 2.004399430 S1-D0 1 1036544 Bytes llc_misses.mem_read > 2.004399430 S1-D1 1 1237376 Bytes llc_misses.mem_read > 3.006535657 S0-D0 1 1036608 Bytes llc_misses.mem_read > 3.006535657 S0-D1 1 906560 Bytes llc_misses.mem_read > 3.006535657 S1-D0 1 994112 Bytes llc_misses.mem_read > 3.006535657 S1-D1 1 1025472 Bytes llc_misses.mem_read > 4.008934402 S0-D0 1 1005120 Bytes llc_misses.mem_read > 4.008934402 S0-D1 1 882368 Bytes llc_misses.mem_read > 4.008934402 S1-D0 1 961408 Bytes llc_misses.mem_read > 4.008934402 S1-D1 1 988352 Bytes llc_misses.mem_read > 5.001605406 S0-D0 1 1930240 Bytes llc_misses.mem_read > 5.001605406 S0-D1 1 1134272 Bytes llc_misses.mem_read > 5.001605406 S1-D0 1 1531136 Bytes llc_misses.mem_read > 5.001605406 S1-D1 1 2230848 Bytes llc_misses.mem_read > > On no-die system, die_id is 0, this patch keeps original behavior unchanged. > > Reported-by: Huang Ying <ying.huang@intel.com> > Signed-off-by: Jin Yao <yao.jin@linux.intel.com> > --- > tools/perf/util/cpumap.h | 2 ++ > tools/perf/util/stat.c | 21 +++++++++++++++++---- > 2 files changed, 19 insertions(+), 4 deletions(-) > > diff --git a/tools/perf/util/cpumap.h b/tools/perf/util/cpumap.h > index 3a442f021468..58604c047121 100644 > --- a/tools/perf/util/cpumap.h > +++ b/tools/perf/util/cpumap.h > @@ -7,6 +7,8 @@ > #include <internal/cpumap.h> > #include <perf/cpumap.h> > > +#define CPU_MAP_MAX_DIE 256 > + > struct perf_record_cpu_map_data; > > struct perf_cpu_map *perf_cpu_map__empty_new(int nr); > diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c > index 1e125e39ff84..b3fc6e261fe0 100644 > --- a/tools/perf/util/stat.c > +++ b/tools/perf/util/stat.c > @@ -282,9 +282,9 @@ static void zero_per_pkg(struct evsel *counter) > static int check_per_pkg(struct evsel *counter, > struct perf_counts_values *vals, int cpu, bool *skip) > { > - unsigned long *mask = counter->per_pkg_mask; > + unsigned long *mask = counter->per_pkg_mask, *l; > struct perf_cpu_map *cpus = evsel__cpus(counter); > - int s; > + int s, d; > > *skip = false; > > @@ -295,7 +295,7 @@ static int check_per_pkg(struct evsel *counter, > return 0; > > if (!mask) { > - mask = zalloc(cpu__max_cpu()); > + mask = zalloc(cpu__max_cpu() * CPU_MAP_MAX_DIE); > if (!mask) > return -ENOMEM; > > @@ -317,7 +317,20 @@ static int check_per_pkg(struct evsel *counter, > if (s < 0) > return -1; > > - *skip = test_and_set_bit(s, mask) == 1; > + d = cpu_map__get_die(cpus, cpu, NULL); > + if (d < 0) > + return -1; > + > + /* > + * On multi-die system, die_id < 256. We use a pair > + * of socket + die to identify the used bit. > + * On no-die system, die_id is 0, l = &mask[0]. It doesn't > + * change original "test_and_set_bit(s, mask)" behavior. > + */ > + d &= 0xff; > + l = &mask[d]; James is fixing the aggregation, because we could get large socket numbers: https://lore.kernel.org/lkml/20201126141328.6509-1-james.clark@arm.com/ I think we should fix that in here as well maybe just use hashmap (socket,die) -> bool ? jirka > + > + *skip = test_and_set_bit(s, l) == 1; > return 0; > } > > -- > 2.17.1 > ^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [PATCH] perf stat: Fix wrong skipping for per-die aggregation 2020-12-23 22:35 ` Jiri Olsa @ 2020-12-24 1:24 ` Jin, Yao 0 siblings, 0 replies; 3+ messages in thread From: Jin, Yao @ 2020-12-24 1:24 UTC (permalink / raw) To: Jiri Olsa Cc: acme, jolsa, peterz, mingo, alexander.shishkin, Linux-kernel, ak, kan.liang, yao.jin, James Clark On 12/24/2020 6:35 AM, Jiri Olsa wrote: > On Wed, Dec 16, 2020 at 03:01:46PM +0800, Jin Yao wrote: >> Uncore becomes die-scope on Xeon Cascade Lake-AP and perf has supported >> --per-die aggregation yet. >> >> One issue is found in check_per_pkg() for uncore events running on >> AP system. On cascade Lake-AP, we have: >> >> S0-D0 >> S0-D1 >> S1-D0 >> S1-D1 >> >> But in check_per_pkg(), S0-D1 and S1-D1 are skipped because the >> mask bits for S0 and S1 have been set for S0-D0 and S1-D0. It doesn't >> check die_id. So the counting for S0-D1 and S1-D1 are set to zero. >> That's not correct. >> >> root@lkp-csl-2ap4 ~# ./perf stat -a -I 1000 -e llc_misses.mem_read --per-die -- sleep 5 >> 1.001460963 S0-D0 1 1317376 Bytes llc_misses.mem_read >> 1.001460963 S0-D1 1 998016 Bytes llc_misses.mem_read >> 1.001460963 S1-D0 1 970496 Bytes llc_misses.mem_read >> 1.001460963 S1-D1 1 1291264 Bytes llc_misses.mem_read >> 2.003488021 S0-D0 1 1082048 Bytes llc_misses.mem_read >> 2.003488021 S0-D1 1 1919040 Bytes llc_misses.mem_read >> 2.003488021 S1-D0 1 890752 Bytes llc_misses.mem_read >> 2.003488021 S1-D1 1 2380800 Bytes llc_misses.mem_read >> 3.005613270 S0-D0 1 1126080 Bytes llc_misses.mem_read >> 3.005613270 S0-D1 1 2898176 Bytes llc_misses.mem_read >> 3.005613270 S1-D0 1 870912 Bytes llc_misses.mem_read >> 3.005613270 S1-D1 1 3388608 Bytes llc_misses.mem_read >> 4.007627598 S0-D0 1 1124608 Bytes llc_misses.mem_read >> 4.007627598 S0-D1 1 3884416 Bytes llc_misses.mem_read >> 4.007627598 S1-D0 1 921088 Bytes llc_misses.mem_read >> 4.007627598 S1-D1 1 4451840 Bytes llc_misses.mem_read >> 5.001479927 S0-D0 1 963328 Bytes llc_misses.mem_read >> 5.001479927 S0-D1 1 4831936 Bytes llc_misses.mem_read >> 5.001479927 S1-D0 1 895104 Bytes llc_misses.mem_read >> 5.001479927 S1-D1 1 5496640 Bytes llc_misses.mem_read >> >> From above output, we can see S0-D1 and S1-D1 don't report the interval >> values, they are continued to grow. That's because check_per_pkg() wrongly >> decides to use zero counts for S0-D1 and S1-D1. >> >> So in check_per_pkg(), we should use a pair of die_id + socket_id to >> decide if this cpu counts needs to skip. Only considering socket_id is >> not enough. >> >> Now with this patch, >> >> root@lkp-csl-2ap4 ~# ./perf stat -a -I 1000 -e llc_misses.mem_read --per-die -- sleep 5 >> 1.001622560 S0-D0 1 1210176 Bytes llc_misses.mem_read >> 1.001622560 S0-D1 1 950208 Bytes llc_misses.mem_read >> 1.001622560 S1-D0 1 968704 Bytes llc_misses.mem_read >> 1.001622560 S1-D1 1 1113280 Bytes llc_misses.mem_read >> 2.004399430 S0-D0 1 1258560 Bytes llc_misses.mem_read >> 2.004399430 S0-D1 1 1011008 Bytes llc_misses.mem_read >> 2.004399430 S1-D0 1 1036544 Bytes llc_misses.mem_read >> 2.004399430 S1-D1 1 1237376 Bytes llc_misses.mem_read >> 3.006535657 S0-D0 1 1036608 Bytes llc_misses.mem_read >> 3.006535657 S0-D1 1 906560 Bytes llc_misses.mem_read >> 3.006535657 S1-D0 1 994112 Bytes llc_misses.mem_read >> 3.006535657 S1-D1 1 1025472 Bytes llc_misses.mem_read >> 4.008934402 S0-D0 1 1005120 Bytes llc_misses.mem_read >> 4.008934402 S0-D1 1 882368 Bytes llc_misses.mem_read >> 4.008934402 S1-D0 1 961408 Bytes llc_misses.mem_read >> 4.008934402 S1-D1 1 988352 Bytes llc_misses.mem_read >> 5.001605406 S0-D0 1 1930240 Bytes llc_misses.mem_read >> 5.001605406 S0-D1 1 1134272 Bytes llc_misses.mem_read >> 5.001605406 S1-D0 1 1531136 Bytes llc_misses.mem_read >> 5.001605406 S1-D1 1 2230848 Bytes llc_misses.mem_read >> >> On no-die system, die_id is 0, this patch keeps original behavior unchanged. >> >> Reported-by: Huang Ying <ying.huang@intel.com> >> Signed-off-by: Jin Yao <yao.jin@linux.intel.com> >> --- >> tools/perf/util/cpumap.h | 2 ++ >> tools/perf/util/stat.c | 21 +++++++++++++++++---- >> 2 files changed, 19 insertions(+), 4 deletions(-) >> >> diff --git a/tools/perf/util/cpumap.h b/tools/perf/util/cpumap.h >> index 3a442f021468..58604c047121 100644 >> --- a/tools/perf/util/cpumap.h >> +++ b/tools/perf/util/cpumap.h >> @@ -7,6 +7,8 @@ >> #include <internal/cpumap.h> >> #include <perf/cpumap.h> >> >> +#define CPU_MAP_MAX_DIE 256 >> + >> struct perf_record_cpu_map_data; >> >> struct perf_cpu_map *perf_cpu_map__empty_new(int nr); >> diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c >> index 1e125e39ff84..b3fc6e261fe0 100644 >> --- a/tools/perf/util/stat.c >> +++ b/tools/perf/util/stat.c >> @@ -282,9 +282,9 @@ static void zero_per_pkg(struct evsel *counter) >> static int check_per_pkg(struct evsel *counter, >> struct perf_counts_values *vals, int cpu, bool *skip) >> { >> - unsigned long *mask = counter->per_pkg_mask; >> + unsigned long *mask = counter->per_pkg_mask, *l; >> struct perf_cpu_map *cpus = evsel__cpus(counter); >> - int s; >> + int s, d; >> >> *skip = false; >> >> @@ -295,7 +295,7 @@ static int check_per_pkg(struct evsel *counter, >> return 0; >> >> if (!mask) { >> - mask = zalloc(cpu__max_cpu()); >> + mask = zalloc(cpu__max_cpu() * CPU_MAP_MAX_DIE); >> if (!mask) >> return -ENOMEM; >> >> @@ -317,7 +317,20 @@ static int check_per_pkg(struct evsel *counter, >> if (s < 0) >> return -1; >> >> - *skip = test_and_set_bit(s, mask) == 1; >> + d = cpu_map__get_die(cpus, cpu, NULL); >> + if (d < 0) >> + return -1; >> + >> + /* >> + * On multi-die system, die_id < 256. We use a pair >> + * of socket + die to identify the used bit. >> + * On no-die system, die_id is 0, l = &mask[0]. It doesn't >> + * change original "test_and_set_bit(s, mask)" behavior. >> + */ >> + d &= 0xff; >> + l = &mask[d]; > > James is fixing the aggregation, because we could get large socket numbers: > https://lore.kernel.org/lkml/20201126141328.6509-1-james.clark@arm.com/ > James is using "struct cpu_aggr_map" to replace "struct perf_cpu_map", but do we also need to replace the "struct perf_cpu_map" in "struct perf_evsel"? Otherwise the evsel->cpus will be inconsistent, right? struct perf_evsel { ... struct perf_cpu_map *cpus; struct perf_cpu_map *own_cpus; ... }; > I think we should fix that in here as well Yes, we need to fix that in check_per_pkg() as well. As I said above, in check_per_pkg(), struct perf_cpu_map *cpus = evsel__cpus(counter); we should use "struct cpu_aggr_map" here. > maybe just use hashmap (socket,die) -> bool ? > Yes, hashmap should be better, saving more memory. Thanks Jin Yao > jirka > >> + >> + *skip = test_and_set_bit(s, l) == 1; >> return 0; >> } >> >> -- >> 2.17.1 >> > ^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2020-12-24 1:26 UTC | newest] Thread overview: 3+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2020-12-16 7:01 [PATCH] perf stat: Fix wrong skipping for per-die aggregation Jin Yao 2020-12-23 22:35 ` Jiri Olsa 2020-12-24 1:24 ` Jin, Yao
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox