From: Ravi Bangoria <ravi.bangoria@amd.com>
To: <acme@kernel.org>
Cc: <ravi.bangoria@amd.com>, <kan.liang@linux.intel.com>,
<namhyung@kernel.org>, <jolsa@kernel.org>, <irogers@google.com>,
<peterz@infradead.org>, <rrichter@amd.com>, <mingo@redhat.com>,
<mark.rutland@arm.com>, <tglx@linutronix.de>, <bp@alien8.de>,
<james.clark@arm.com>, <leo.yan@linaro.org>, <ak@linux.intel.com>,
<eranian@google.com>, <like.xu.linux@gmail.com>, <x86@kernel.org>,
<linux-perf-users@vger.kernel.org>,
<linux-kernel@vger.kernel.org>, <sandipan.das@amd.com>,
<ananth.narayan@amd.com>, <kim.phillips@amd.com>,
<santosh.shukla@amd.com>
Subject: [PATCH v6 4/8] perf headers: Store pmu caps in an array of strings
Date: Sat, 4 Jun 2022 10:15:15 +0530 [thread overview]
Message-ID: <20220604044519.594-5-ravi.bangoria@amd.com> (raw)
In-Reply-To: <20220604044519.594-1-ravi.bangoria@amd.com>
Currently all capabilities are stored in a single string separated
by NULL character. Instead, store them in an array which makes
searching of capability easier.
Signed-off-by: Ravi Bangoria <ravi.bangoria@amd.com>
Reviewed-by: Kan Liang <kan.liang@linux.intel.com>
---
tools/perf/util/env.c | 6 +++-
tools/perf/util/env.h | 4 +--
| 63 ++++++++++++++++++++++------------------
3 files changed, 41 insertions(+), 32 deletions(-)
diff --git a/tools/perf/util/env.c b/tools/perf/util/env.c
index 579e44c59914..7d3aeb2e4622 100644
--- a/tools/perf/util/env.c
+++ b/tools/perf/util/env.c
@@ -179,7 +179,7 @@ static void perf_env__purge_bpf(struct perf_env *env __maybe_unused)
void perf_env__exit(struct perf_env *env)
{
- int i;
+ int i, j;
perf_env__purge_bpf(env);
perf_env__purge_cgroups(env);
@@ -196,6 +196,8 @@ void perf_env__exit(struct perf_env *env)
zfree(&env->sibling_threads);
zfree(&env->pmu_mappings);
zfree(&env->cpu);
+ for (i = 0; i < env->nr_cpu_pmu_caps; i++)
+ zfree(&env->cpu_pmu_caps[i]);
zfree(&env->cpu_pmu_caps);
zfree(&env->numa_map);
@@ -218,6 +220,8 @@ void perf_env__exit(struct perf_env *env)
zfree(&env->hybrid_nodes);
for (i = 0; i < env->nr_hybrid_cpc_nodes; i++) {
+ for (j = 0; j < env->hybrid_cpc_nodes[i].nr_cpu_pmu_caps; j++)
+ zfree(&env->hybrid_cpc_nodes[i].cpu_pmu_caps[j]);
zfree(&env->hybrid_cpc_nodes[i].cpu_pmu_caps);
zfree(&env->hybrid_cpc_nodes[i].pmu_name);
}
diff --git a/tools/perf/util/env.h b/tools/perf/util/env.h
index a3541f98e1fc..43aab59f7322 100644
--- a/tools/perf/util/env.h
+++ b/tools/perf/util/env.h
@@ -46,7 +46,7 @@ struct hybrid_node {
struct hybrid_cpc_node {
int nr_cpu_pmu_caps;
unsigned int max_branches;
- char *cpu_pmu_caps;
+ char **cpu_pmu_caps;
char *pmu_name;
};
@@ -81,7 +81,7 @@ struct perf_env {
char *sibling_dies;
char *sibling_threads;
char *pmu_mappings;
- char *cpu_pmu_caps;
+ char **cpu_pmu_caps;
struct cpu_topology_map *cpu;
struct cpu_cache_level *caches;
int caches_cnt;
--git a/tools/perf/util/header.c b/tools/perf/util/header.c
index ee7ccd94e272..ca5b1f03810d 100644
--- a/tools/perf/util/header.c
+++ b/tools/perf/util/header.c
@@ -2051,26 +2051,21 @@ static void print_compressed(struct feat_fd *ff, FILE *fp)
ff->ph->env.comp_level, ff->ph->env.comp_ratio);
}
-static void print_per_cpu_pmu_caps(FILE *fp, int nr_caps, char *cpu_pmu_caps,
+static void print_per_cpu_pmu_caps(FILE *fp, int nr_caps, char **cpu_pmu_caps,
char *pmu_name)
{
- const char *delimiter;
- char *str, buf[128];
+ const char *delimiter = "";
+ int i;
if (!nr_caps) {
fprintf(fp, "# %s pmu capabilities: not available\n", pmu_name);
return;
}
- scnprintf(buf, sizeof(buf), "# %s pmu capabilities: ", pmu_name);
-
- delimiter = buf;
-
- str = cpu_pmu_caps;
- while (nr_caps--) {
- fprintf(fp, "%s%s", delimiter, str);
+ fprintf(fp, "# %s pmu capabilities: ", pmu_name);
+ for (i = 0; i < nr_caps; i++) {
+ fprintf(fp, "%s%s", delimiter, cpu_pmu_caps[i]);
delimiter = ", ";
- str += strlen(str) + 1;
}
fprintf(fp, "\n");
@@ -3202,27 +3197,26 @@ static int process_compressed(struct feat_fd *ff,
}
static int process_per_cpu_pmu_caps(struct feat_fd *ff, int *nr_cpu_pmu_caps,
- char **cpu_pmu_caps,
+ char ***cpu_pmu_caps,
unsigned int *max_branches)
{
- char *name, *value;
- struct strbuf sb;
- u32 nr_caps;
+ char *name, *value, *ptr;
+ u32 nr_caps, i;
+
+ *nr_cpu_pmu_caps = 0;
+ *cpu_pmu_caps = NULL;
if (do_read_u32(ff, &nr_caps))
return -1;
- if (!nr_caps) {
- pr_debug("cpu pmu capabilities not available\n");
+ if (!nr_caps)
return 0;
- }
-
- *nr_cpu_pmu_caps = nr_caps;
- if (strbuf_init(&sb, 128) < 0)
+ *cpu_pmu_caps = zalloc(sizeof(char *) * nr_caps);
+ if (!*cpu_pmu_caps)
return -1;
- while (nr_caps--) {
+ for (i = 0; i < nr_caps; i++) {
name = do_read_string(ff);
if (!name)
goto error;
@@ -3231,12 +3225,10 @@ static int process_per_cpu_pmu_caps(struct feat_fd *ff, int *nr_cpu_pmu_caps,
if (!value)
goto free_name;
- if (strbuf_addf(&sb, "%s=%s", name, value) < 0)
+ if (asprintf(&ptr, "%s=%s", name, value) < 0)
goto free_value;
- /* include a NULL character at the end */
- if (strbuf_add(&sb, "", 1) < 0)
- goto free_value;
+ (*cpu_pmu_caps)[i] = ptr;
if (!strcmp(name, "branches"))
*max_branches = atoi(value);
@@ -3244,7 +3236,7 @@ static int process_per_cpu_pmu_caps(struct feat_fd *ff, int *nr_cpu_pmu_caps,
free(value);
free(name);
}
- *cpu_pmu_caps = strbuf_detach(&sb, NULL);
+ *nr_cpu_pmu_caps = nr_caps;
return 0;
free_value:
@@ -3252,16 +3244,24 @@ static int process_per_cpu_pmu_caps(struct feat_fd *ff, int *nr_cpu_pmu_caps,
free_name:
free(name);
error:
- strbuf_release(&sb);
+ for (; i > 0; i--)
+ free((*cpu_pmu_caps)[i - 1]);
+ free(*cpu_pmu_caps);
+ *cpu_pmu_caps = NULL;
+ *nr_cpu_pmu_caps = 0;
return -1;
}
static int process_cpu_pmu_caps(struct feat_fd *ff,
void *data __maybe_unused)
{
- return process_per_cpu_pmu_caps(ff, &ff->ph->env.nr_cpu_pmu_caps,
+ int ret = process_per_cpu_pmu_caps(ff, &ff->ph->env.nr_cpu_pmu_caps,
&ff->ph->env.cpu_pmu_caps,
&ff->ph->env.max_branches);
+
+ if (!ret && !ff->ph->env.cpu_pmu_caps)
+ pr_debug("cpu pmu capabilities not available\n");
+ return ret;
}
static int process_hybrid_cpu_pmu_caps(struct feat_fd *ff,
@@ -3270,6 +3270,7 @@ static int process_hybrid_cpu_pmu_caps(struct feat_fd *ff,
struct hybrid_cpc_node *nodes;
u32 nr_pmu, i;
int ret;
+ int j;
if (do_read_u32(ff, &nr_pmu))
return -1;
@@ -3297,6 +3298,8 @@ static int process_hybrid_cpu_pmu_caps(struct feat_fd *ff,
ret = -1;
goto err;
}
+ if (!n->nr_cpu_pmu_caps)
+ pr_debug("%s pmu capabilities not available\n", n->pmu_name);
}
ff->ph->env.nr_hybrid_cpc_nodes = nr_pmu;
@@ -3305,6 +3308,8 @@ static int process_hybrid_cpu_pmu_caps(struct feat_fd *ff,
err:
for (i = 0; i < nr_pmu; i++) {
+ for (j = 0; j < nodes[i].nr_cpu_pmu_caps; j++)
+ free(nodes[i].cpu_pmu_caps[j]);
free(nodes[i].cpu_pmu_caps);
free(nodes[i].pmu_name);
}
--
2.31.1
next prev parent reply other threads:[~2022-06-04 4:47 UTC|newest]
Thread overview: 13+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-06-04 4:45 [PATCH v6 0/8] perf/amd: Zen4 IBS extensions support (tool changes) Ravi Bangoria
2022-06-04 4:45 ` [PATCH v6 1/8] perf record ibs: Warn about sampling period skew Ravi Bangoria
2022-06-04 4:45 ` [PATCH v6 2/8] perf tool: Parse pmu caps sysfs only once Ravi Bangoria
2022-06-04 4:45 ` [PATCH v6 3/8] perf headers: Pass "cpu" pmu name while printing caps Ravi Bangoria
2022-06-04 4:45 ` Ravi Bangoria [this message]
2022-06-04 4:45 ` [PATCH v6 5/8] perf headers: Record non-cpu pmu capabilities Ravi Bangoria
2022-06-04 4:45 ` [PATCH v6 6/8] perf/x86/ibs: Add new IBS register bits into header Ravi Bangoria
2022-07-27 12:09 ` [tip: perf/core] " tip-bot2 for Ravi Bangoria
2022-06-04 4:45 ` [PATCH v6 7/8] perf tool ibs: Sync amd ibs header file Ravi Bangoria
2022-06-04 4:45 ` [PATCH v6 8/8] perf script ibs: Support new IBS bits in raw trace dump Ravi Bangoria
2022-06-06 23:46 ` [PATCH v6 0/8] perf/amd: Zen4 IBS extensions support (tool changes) Namhyung Kim
2022-06-24 16:19 ` Arnaldo Carvalho de Melo
2022-06-27 4:20 ` Ravi Bangoria
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20220604044519.594-5-ravi.bangoria@amd.com \
--to=ravi.bangoria@amd.com \
--cc=acme@kernel.org \
--cc=ak@linux.intel.com \
--cc=ananth.narayan@amd.com \
--cc=bp@alien8.de \
--cc=eranian@google.com \
--cc=irogers@google.com \
--cc=james.clark@arm.com \
--cc=jolsa@kernel.org \
--cc=kan.liang@linux.intel.com \
--cc=kim.phillips@amd.com \
--cc=leo.yan@linaro.org \
--cc=like.xu.linux@gmail.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-perf-users@vger.kernel.org \
--cc=mark.rutland@arm.com \
--cc=mingo@redhat.com \
--cc=namhyung@kernel.org \
--cc=peterz@infradead.org \
--cc=rrichter@amd.com \
--cc=sandipan.das@amd.com \
--cc=santosh.shukla@amd.com \
--cc=tglx@linutronix.de \
--cc=x86@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox