From: Arnaldo Carvalho de Melo <acme@kernel.org>
To: Namhyung Kim <namhyung@kernel.org>
Cc: Ingo Molnar <mingo@kernel.org>,
Thomas Gleixner <tglx@linutronix.de>,
James Clark <james.clark@linaro.org>,
Jiri Olsa <jolsa@kernel.org>, Ian Rogers <irogers@google.com>,
Adrian Hunter <adrian.hunter@intel.com>,
Kan Liang <kan.liang@linux.intel.com>,
Clark Williams <williams@redhat.com>,
linux-kernel@vger.kernel.org, linux-perf-users@vger.kernel.org,
Arnaldo Carvalho de Melo <acme@redhat.com>,
sashiko-bot@kernel.org,
"Claude Opus 4.6 (1M context)" <noreply@anthropic.com>
Subject: [PATCH 26/28] perf timechart: Bounds check cpu_id and fix topology_map allocation
Date: Sun, 10 May 2026 00:34:17 -0300 [thread overview]
Message-ID: <20260510033424.255812-27-acme@kernel.org> (raw)
In-Reply-To: <20260510033424.255812-1-acme@kernel.org>
From: Arnaldo Carvalho de Melo <acme@redhat.com>
The cpu_idle, cpu_frequency, power_start, and power_frequency
tracepoint handlers extract cpu_id from the event payload via
evsel__intval() and use it directly as an array index into
cpus_cstate_start_times[] and cpus_pstate_start_times[], which
are allocated with MAX_CPUS (4096) entries.
Unlike sample->cpu which is validated in perf_session__deliver_event(),
cpu_id comes from the tracepoint data and is never bounds checked.
A crafted perf.data with a malicious cpu_id in a tracepoint event
causes out-of-bounds array accesses.
Validate cpu_id against tchart->numcpus (nr_cpus_avail from the
file header) and reject the event with an error if it is out of
range, as this indicates a corrupted or crafted file.
The power_end handler uses sample->cpu (not a tracepoint cpu_id
field). Add a bounds check there too since a crafted file could
omit PERF_SAMPLE_CPU, leaving sample->cpu as the (u32)-1 sentinel
which would cause out-of-bounds access in c_state_end().
Also validate sample->cpu in sched_switch and sched_wakeup
handlers, which store it in cpu_sample structs later used as
array indices into topology_map[] during SVG generation.
Fix svg_build_topology_map() to allocate topology_map using
nr_cpus_avail instead of nr_cpus_online. When offline CPUs exist,
nr_cpus_online < nr_cpus_avail, and a valid cpu_id that passes
the numcpus check could still exceed the topology_map allocation,
causing a heap out-of-bounds read in cpu2y(). Reject negative CPU
values in str_to_bitmap() to prevent perf_cpu_map__new("") on an
empty topology string from passing -1 to __set_bit(), which would
write at offset ULONG_MAX/BITS_PER_LONG.
Fix the pre-existing backtrace memory leak: change the
tracepoint_handler typedef to pass const char **backtrace
(pointer-to-pointer). Handlers that consume the string
(sched_switch, sched_wakeup) set *backtrace = NULL to claim
ownership. The caller always calls free() after the handler
returns — if ownership was taken the pointer is NULL and
free(NULL) is a no-op. Skip cat_backtrace() entirely when
tchart->with_backtrace is not set.
Cap tchart->numcpus at MAX_CPUS in the HEADER_NRCPUS callback
so the bounds check cannot exceed the array allocation size.
Reported-by: sashiko-bot@kernel.org # Running on a local machine
Cc: Ian Rogers <irogers@google.com>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Assisted-by: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
---
tools/perf/builtin-timechart.c | 115 ++++++++++++++++++++++++++++-----
tools/perf/util/svghelper.c | 6 +-
2 files changed, 104 insertions(+), 17 deletions(-)
diff --git a/tools/perf/builtin-timechart.c b/tools/perf/builtin-timechart.c
index 40297f2dcd0353cc..bccc48cfb99a1d57 100644
--- a/tools/perf/builtin-timechart.c
+++ b/tools/perf/builtin-timechart.c
@@ -71,6 +71,7 @@ struct timechart {
bool io_only,
skip_eagain;
u64 io_events;
+ u32 nr_invalid_cpu;
u64 min_time,
merge_dist;
};
@@ -569,7 +570,7 @@ static const char *cat_backtrace(union perf_event *event,
typedef int (*tracepoint_handler)(struct timechart *tchart,
struct evsel *evsel,
struct perf_sample *sample,
- const char *backtrace);
+ const char **backtrace);
static int process_sample_event(const struct perf_tool *tool,
union perf_event *event,
@@ -588,22 +589,46 @@ static int process_sample_event(const struct perf_tool *tool,
if (evsel->handler != NULL) {
tracepoint_handler f = evsel->handler;
- return f(tchart, evsel, sample,
- cat_backtrace(event, sample, machine));
+ const char *bt = NULL;
+ int ret;
+
+ if (tchart->with_backtrace)
+ bt = cat_backtrace(event, sample, machine);
+ ret = f(tchart, evsel, sample, &bt);
+ /*
+ * Handlers that consume backtrace (sched_switch,
+ * sched_wakeup) store the pointer and set *bt = NULL
+ * to claim ownership. For all other handlers bt is
+ * still ours to free. free(NULL) is safe.
+ */
+ free((void *)bt);
+ return ret;
}
return 0;
}
static int
-process_sample_cpu_idle(struct timechart *tchart __maybe_unused,
+process_sample_cpu_idle(struct timechart *tchart,
struct evsel *evsel,
struct perf_sample *sample,
- const char *backtrace __maybe_unused)
+ const char **backtrace __maybe_unused)
{
u32 state = evsel__intval(evsel, sample, "state");
u32 cpu_id = evsel__intval(evsel, sample, "cpu_id");
+ /*
+ * cpu_id from tracepoint data indexes cpus_cstate_start_times[]
+ * and cpus_pstate_start_times[], both allocated as MAX_CPUS
+ * entries. Reject out-of-range values to prevent OOB writes;
+ * numcpus (from nr_cpus_avail) is the tighter, valid bound.
+ */
+ if (cpu_id >= tchart->numcpus) {
+ pr_err("cpu_idle event cpu_id %u >= nr_cpus_avail %u\n",
+ cpu_id, tchart->numcpus);
+ return -EINVAL;
+ }
+
if (state == (u32)PWR_EVENT_EXIT)
c_state_end(tchart, cpu_id, sample->time);
else
@@ -615,11 +640,18 @@ static int
process_sample_cpu_frequency(struct timechart *tchart,
struct evsel *evsel,
struct perf_sample *sample,
- const char *backtrace __maybe_unused)
+ const char **backtrace __maybe_unused)
{
u32 state = evsel__intval(evsel, sample, "state");
u32 cpu_id = evsel__intval(evsel, sample, "cpu_id");
+ /* Same bounds check as process_sample_cpu_idle — see comment there */
+ if (cpu_id >= tchart->numcpus) {
+ pr_err("cpu_frequency event cpu_id %u >= nr_cpus_avail %u\n",
+ cpu_id, tchart->numcpus);
+ return -EINVAL;
+ }
+
p_state_change(tchart, cpu_id, sample->time, state);
return 0;
}
@@ -628,13 +660,20 @@ static int
process_sample_sched_wakeup(struct timechart *tchart,
struct evsel *evsel,
struct perf_sample *sample,
- const char *backtrace)
+ const char **backtrace)
{
u8 flags = evsel__intval(evsel, sample, "common_flags");
int waker = evsel__intval(evsel, sample, "common_pid");
int wakee = evsel__intval(evsel, sample, "pid");
- sched_wakeup(tchart, sample->cpu, sample->time, waker, wakee, flags, backtrace);
+ /* sample->cpu used as index into topology_map[] during SVG generation */
+ if (sample->cpu >= tchart->numcpus) {
+ tchart->nr_invalid_cpu++;
+ return 0;
+ }
+
+ sched_wakeup(tchart, sample->cpu, sample->time, waker, wakee, flags, *backtrace);
+ *backtrace = NULL;
return 0;
}
@@ -642,27 +681,41 @@ static int
process_sample_sched_switch(struct timechart *tchart,
struct evsel *evsel,
struct perf_sample *sample,
- const char *backtrace)
+ const char **backtrace)
{
int prev_pid = evsel__intval(evsel, sample, "prev_pid");
int next_pid = evsel__intval(evsel, sample, "next_pid");
u64 prev_state = evsel__intval(evsel, sample, "prev_state");
+ /* sample->cpu used as index into topology_map[] during SVG generation */
+ if (sample->cpu >= tchart->numcpus) {
+ tchart->nr_invalid_cpu++;
+ return 0;
+ }
+
sched_switch(tchart, sample->cpu, sample->time, prev_pid, next_pid,
- prev_state, backtrace);
+ prev_state, *backtrace);
+ *backtrace = NULL;
return 0;
}
#ifdef SUPPORT_OLD_POWER_EVENTS
static int
-process_sample_power_start(struct timechart *tchart __maybe_unused,
+process_sample_power_start(struct timechart *tchart,
struct evsel *evsel,
struct perf_sample *sample,
- const char *backtrace __maybe_unused)
+ const char **backtrace __maybe_unused)
{
u64 cpu_id = evsel__intval(evsel, sample, "cpu_id");
u64 value = evsel__intval(evsel, sample, "value");
+ /* Same bounds check as process_sample_cpu_idle — see comment there */
+ if (cpu_id >= tchart->numcpus) {
+ pr_err("power_start event cpu_id %" PRIu64 " >= nr_cpus_avail %u\n",
+ cpu_id, tchart->numcpus);
+ return -EINVAL;
+ }
+
c_state_start(cpu_id, sample->time, value);
return 0;
}
@@ -671,8 +724,16 @@ static int
process_sample_power_end(struct timechart *tchart,
struct evsel *evsel __maybe_unused,
struct perf_sample *sample,
- const char *backtrace __maybe_unused)
+ const char **backtrace __maybe_unused)
{
+ /*
+ * sample->cpu is validated centrally when PERF_SAMPLE_CPU is
+ * set, but a crafted file could omit it from sample_type.
+ */
+ if (sample->cpu >= tchart->numcpus) {
+ tchart->nr_invalid_cpu++;
+ return 0;
+ }
c_state_end(tchart, sample->cpu, sample->time);
return 0;
}
@@ -681,11 +742,18 @@ static int
process_sample_power_frequency(struct timechart *tchart,
struct evsel *evsel,
struct perf_sample *sample,
- const char *backtrace __maybe_unused)
+ const char **backtrace __maybe_unused)
{
u64 cpu_id = evsel__intval(evsel, sample, "cpu_id");
u64 value = evsel__intval(evsel, sample, "value");
+ /* Same bounds check as process_sample_cpu_idle — see comment there */
+ if (cpu_id >= tchart->numcpus) {
+ pr_err("power_frequency event cpu_id %" PRIu64 " >= nr_cpus_avail %u\n",
+ cpu_id, tchart->numcpus);
+ return -EINVAL;
+ }
+
p_state_change(tchart, cpu_id, sample->time, value);
return 0;
}
@@ -1519,7 +1587,8 @@ static int process_header(struct perf_file_section *section __maybe_unused,
switch (feat) {
case HEADER_NRCPUS:
- tchart->numcpus = ph->env.nr_cpus_avail;
+ /* Cap at MAX_CPUS — the allocation size of cpus_cstate/pstate arrays */
+ tchart->numcpus = min((int)ph->env.nr_cpus_avail, MAX_CPUS);
break;
case HEADER_CPU_TOPOLOGY:
@@ -1625,6 +1694,16 @@ static int __cmd_timechart(struct timechart *tchart, const char *output_name)
tchart,
process_header);
+ /*
+ * Truncated files (interrupted recording) lose all feature
+ * sections so the HEADER_NRCPUS callback never fires, and
+ * pipe mode doesn't use perf_header__process_sections at all.
+ * Fall back to MAX_CPUS — the actual allocation size of the
+ * cpus_cstate/pstate arrays.
+ */
+ if (!tchart->numcpus)
+ tchart->numcpus = MAX_CPUS;
+
if (!perf_session__has_traces(session, "timechart record"))
goto out_delete;
@@ -1646,6 +1725,12 @@ static int __cmd_timechart(struct timechart *tchart, const char *output_name)
pr_info("Written %2.1f seconds of trace to %s.\n",
(tchart->last_time - tchart->first_time) / (double)NSEC_PER_SEC, output_name);
+
+ if (tchart->nr_invalid_cpu) {
+ pr_warning("WARNING: %u events had invalid CPU values and were skipped.\n"
+ " Scheduling and power state data may be incomplete.\n",
+ tchart->nr_invalid_cpu);
+ }
out_delete:
perf_session__delete(session);
return ret;
diff --git a/tools/perf/util/svghelper.c b/tools/perf/util/svghelper.c
index e360e7736c7ba65b..a3c7cfecc072f3e3 100644
--- a/tools/perf/util/svghelper.c
+++ b/tools/perf/util/svghelper.c
@@ -736,7 +736,8 @@ static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus)
return -1;
perf_cpu_map__for_each_cpu(cpu, idx, map) {
- if (cpu.cpu >= nr_cpus) {
+ /* perf_cpu_map__new("") yields cpu=-1; reject to prevent __set_bit OOB */
+ if (cpu.cpu < 0 || cpu.cpu >= nr_cpus) {
ret = -1;
break;
}
@@ -756,7 +757,8 @@ int svg_build_topology_map(struct perf_env *env)
char *sib_core, *sib_thr;
int ret = -1;
- nr_cpus = min(env->nr_cpus_online, MAX_NR_CPUS);
+ /* Use nr_cpus_avail: offline CPUs still need slots in the topology map */
+ nr_cpus = min(env->nr_cpus_avail, MAX_NR_CPUS);
t.sib_core_nr = env->nr_sibling_cores;
t.sib_thr_nr = env->nr_sibling_threads;
--
2.54.0
next prev parent reply other threads:[~2026-05-10 3:37 UTC|newest]
Thread overview: 44+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-05-10 3:33 [PATCH 00/28] perf: Harden perf.data parsing against crafted/corrupted files Arnaldo Carvalho de Melo
2026-05-10 3:33 ` [PATCH 01/28] perf session: Add minimum event size validation table Arnaldo Carvalho de Melo
2026-05-11 19:01 ` Ian Rogers
2026-05-10 3:33 ` [PATCH 02/28] perf tools: Fix event_contains() macro to verify full field extent Arnaldo Carvalho de Melo
2026-05-11 23:46 ` sashiko-bot
2026-05-10 3:33 ` [PATCH 03/28] perf zstd: Fix compression error path in zstd_compress_stream_to_records() Arnaldo Carvalho de Melo
2026-05-12 0:13 ` sashiko-bot
2026-05-10 3:33 ` [PATCH 04/28] perf zstd: Fix multi-iteration decompression and error handling Arnaldo Carvalho de Melo
2026-05-10 3:33 ` [PATCH 05/28] perf session: Fix PERF_RECORD_READ swap and dump for variable-length events Arnaldo Carvalho de Melo
2026-05-10 3:33 ` [PATCH 06/28] perf session: Align auxtrace_info priv size before byte-swapping Arnaldo Carvalho de Melo
2026-05-10 3:33 ` [PATCH 07/28] perf session: Add validated swap infrastructure with null-termination checks Arnaldo Carvalho de Melo
2026-05-12 4:08 ` sashiko-bot
2026-05-10 3:33 ` [PATCH 08/28] perf session: Use bounded copy for PERF_RECORD_TIME_CONV Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 09/28] perf session: Validate HEADER_ATTR alignment and attr.size before swapping Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 10/28] perf session: Validate nr fields against event size on both swap and common paths Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 11/28] perf header: Byte-swap build ID event pid and bounds check section entries Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 12/28] perf cpumap: Reject RANGE_CPUS with start_cpu > end_cpu Arnaldo Carvalho de Melo
2026-05-12 21:37 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 13/28] perf auxtrace: Harden auxtrace_error event handling Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 14/28] perf session: Add byte-swap and bounds check for PERF_RECORD_BPF_METADATA events Arnaldo Carvalho de Melo
2026-05-12 22:58 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 15/28] perf header: Validate null-termination in PERF_RECORD_EVENT_UPDATE string fields Arnaldo Carvalho de Melo
2026-05-12 23:45 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 16/28] perf tools: Bounds check perf_event_attr fields against attr.size before printing Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 17/28] perf header: Propagate feature section processing errors Arnaldo Carvalho de Melo
2026-05-13 3:21 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 18/28] perf header: Validate f_attr.ids section before use in perf_session__read_header() Arnaldo Carvalho de Melo
2026-05-13 4:36 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 19/28] perf header: Validate feature section size and add read path bounds checking Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 20/28] perf header: Sanity check HEADER_EVENT_DESC attr.size before swap Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 21/28] perf header: Validate bitmap size before allocating in do_read_bitmap() Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 22/28] perf session: Add byte-swap for PERF_RECORD_COMPRESSED2 events Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 23/28] perf tools: Harden compressed event processing Arnaldo Carvalho de Melo
2026-05-13 21:56 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 24/28] perf session: Check for decompression buffer size overflow Arnaldo Carvalho de Melo
2026-05-10 3:34 ` [PATCH 25/28] perf session: Bound nr_cpus_avail and validate sample CPU Arnaldo Carvalho de Melo
2026-05-10 3:34 ` Arnaldo Carvalho de Melo [this message]
2026-05-12 18:32 ` [PATCH 26/28] perf timechart: Bounds check cpu_id and fix topology_map allocation Ian Rogers
2026-05-12 19:48 ` Arnaldo Carvalho de Melo
2026-05-13 23:43 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 27/28] perf kwork: Bounds check work->cpu before indexing cpus_runtime[] Arnaldo Carvalho de Melo
2026-05-14 0:06 ` sashiko-bot
2026-05-10 3:34 ` [PATCH 28/28] perf test: Add truncated perf.data robustness test Arnaldo Carvalho de Melo
2026-05-14 0:18 ` sashiko-bot
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260510033424.255812-27-acme@kernel.org \
--to=acme@kernel.org \
--cc=acme@redhat.com \
--cc=adrian.hunter@intel.com \
--cc=irogers@google.com \
--cc=james.clark@linaro.org \
--cc=jolsa@kernel.org \
--cc=kan.liang@linux.intel.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-perf-users@vger.kernel.org \
--cc=mingo@kernel.org \
--cc=namhyung@kernel.org \
--cc=noreply@anthropic.com \
--cc=sashiko-bot@kernel.org \
--cc=tglx@linutronix.de \
--cc=williams@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.