The Linux Kernel Mailing List
 help / color / mirror / Atom feed
From: Arnaldo Carvalho de Melo <acme@kernel.org>
To: Namhyung Kim <namhyung@kernel.org>
Cc: Ingo Molnar <mingo@kernel.org>,
	Thomas Gleixner <tglx@linutronix.de>,
	James Clark <james.clark@linaro.org>,
	Jiri Olsa <jolsa@kernel.org>, Ian Rogers <irogers@google.com>,
	Adrian Hunter <adrian.hunter@intel.com>,
	Kan Liang <kan.liang@linux.intel.com>,
	Clark Williams <williams@redhat.com>,
	linux-kernel@vger.kernel.org, linux-perf-users@vger.kernel.org,
	Arnaldo Carvalho de Melo <acme@redhat.com>,
	sashiko-bot@kernel.org,
	"Claude Opus 4.6 (1M context)" <noreply@anthropic.com>
Subject: [PATCH 26/28] perf timechart: Bounds check cpu_id and fix topology_map allocation
Date: Sun, 10 May 2026 00:34:17 -0300	[thread overview]
Message-ID: <20260510033424.255812-27-acme@kernel.org> (raw)
In-Reply-To: <20260510033424.255812-1-acme@kernel.org>

From: Arnaldo Carvalho de Melo <acme@redhat.com>

The cpu_idle, cpu_frequency, power_start, and power_frequency
tracepoint handlers extract cpu_id from the event payload via
evsel__intval() and use it directly as an array index into
cpus_cstate_start_times[] and cpus_pstate_start_times[], which
are allocated with MAX_CPUS (4096) entries.

Unlike sample->cpu which is validated in perf_session__deliver_event(),
cpu_id comes from the tracepoint data and is never bounds checked.
A crafted perf.data with a malicious cpu_id in a tracepoint event
causes out-of-bounds array accesses.

Validate cpu_id against tchart->numcpus (nr_cpus_avail from the
file header) and reject the event with an error if it is out of
range, as this indicates a corrupted or crafted file.

The power_end handler uses sample->cpu (not a tracepoint cpu_id
field).  Add a bounds check there too since a crafted file could
omit PERF_SAMPLE_CPU, leaving sample->cpu as the (u32)-1 sentinel
which would cause out-of-bounds access in c_state_end().

Also validate sample->cpu in sched_switch and sched_wakeup
handlers, which store it in cpu_sample structs later used as
array indices into topology_map[] during SVG generation.

Fix svg_build_topology_map() to allocate topology_map using
nr_cpus_avail instead of nr_cpus_online.  When offline CPUs exist,
nr_cpus_online < nr_cpus_avail, and a valid cpu_id that passes
the numcpus check could still exceed the topology_map allocation,
causing a heap out-of-bounds read in cpu2y().  Reject negative CPU
values in str_to_bitmap() to prevent perf_cpu_map__new("") on an
empty topology string from passing -1 to __set_bit(), which would
write at offset ULONG_MAX/BITS_PER_LONG.

Fix the pre-existing backtrace memory leak: change the
tracepoint_handler typedef to pass const char **backtrace
(pointer-to-pointer).  Handlers that consume the string
(sched_switch, sched_wakeup) set *backtrace = NULL to claim
ownership.  The caller always calls free() after the handler
returns — if ownership was taken the pointer is NULL and
free(NULL) is a no-op.  Skip cat_backtrace() entirely when
tchart->with_backtrace is not set.

Cap tchart->numcpus at MAX_CPUS in the HEADER_NRCPUS callback
so the bounds check cannot exceed the array allocation size.

Reported-by: sashiko-bot@kernel.org # Running on a local machine
Cc: Ian Rogers <irogers@google.com>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Assisted-by: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
---
 tools/perf/builtin-timechart.c | 115 ++++++++++++++++++++++++++++-----
 tools/perf/util/svghelper.c    |   6 +-
 2 files changed, 104 insertions(+), 17 deletions(-)

diff --git a/tools/perf/builtin-timechart.c b/tools/perf/builtin-timechart.c
index 40297f2dcd0353cc..bccc48cfb99a1d57 100644
--- a/tools/perf/builtin-timechart.c
+++ b/tools/perf/builtin-timechart.c
@@ -71,6 +71,7 @@ struct timechart {
 	bool			io_only,
 				skip_eagain;
 	u64			io_events;
+	u32			nr_invalid_cpu;
 	u64			min_time,
 				merge_dist;
 };
@@ -569,7 +570,7 @@ static const char *cat_backtrace(union perf_event *event,
 typedef int (*tracepoint_handler)(struct timechart *tchart,
 				  struct evsel *evsel,
 				  struct perf_sample *sample,
-				  const char *backtrace);
+				  const char **backtrace);
 
 static int process_sample_event(const struct perf_tool *tool,
 				union perf_event *event,
@@ -588,22 +589,46 @@ static int process_sample_event(const struct perf_tool *tool,
 
 	if (evsel->handler != NULL) {
 		tracepoint_handler f = evsel->handler;
-		return f(tchart, evsel, sample,
-			 cat_backtrace(event, sample, machine));
+		const char *bt = NULL;
+		int ret;
+
+		if (tchart->with_backtrace)
+			bt = cat_backtrace(event, sample, machine);
+		ret = f(tchart, evsel, sample, &bt);
+		/*
+		 * Handlers that consume backtrace (sched_switch,
+		 * sched_wakeup) store the pointer and set *bt = NULL
+		 * to claim ownership.  For all other handlers bt is
+		 * still ours to free.  free(NULL) is safe.
+		 */
+		free((void *)bt);
+		return ret;
 	}
 
 	return 0;
 }
 
 static int
-process_sample_cpu_idle(struct timechart *tchart __maybe_unused,
+process_sample_cpu_idle(struct timechart *tchart,
 			struct evsel *evsel,
 			struct perf_sample *sample,
-			const char *backtrace __maybe_unused)
+			const char **backtrace __maybe_unused)
 {
 	u32 state  = evsel__intval(evsel, sample, "state");
 	u32 cpu_id = evsel__intval(evsel, sample, "cpu_id");
 
+	/*
+	 * cpu_id from tracepoint data indexes cpus_cstate_start_times[]
+	 * and cpus_pstate_start_times[], both allocated as MAX_CPUS
+	 * entries.  Reject out-of-range values to prevent OOB writes;
+	 * numcpus (from nr_cpus_avail) is the tighter, valid bound.
+	 */
+	if (cpu_id >= tchart->numcpus) {
+		pr_err("cpu_idle event cpu_id %u >= nr_cpus_avail %u\n",
+		       cpu_id, tchart->numcpus);
+		return -EINVAL;
+	}
+
 	if (state == (u32)PWR_EVENT_EXIT)
 		c_state_end(tchart, cpu_id, sample->time);
 	else
@@ -615,11 +640,18 @@ static int
 process_sample_cpu_frequency(struct timechart *tchart,
 			     struct evsel *evsel,
 			     struct perf_sample *sample,
-			     const char *backtrace __maybe_unused)
+			     const char **backtrace __maybe_unused)
 {
 	u32 state  = evsel__intval(evsel, sample, "state");
 	u32 cpu_id = evsel__intval(evsel, sample, "cpu_id");
 
+	/* Same bounds check as process_sample_cpu_idle — see comment there */
+	if (cpu_id >= tchart->numcpus) {
+		pr_err("cpu_frequency event cpu_id %u >= nr_cpus_avail %u\n",
+		       cpu_id, tchart->numcpus);
+		return -EINVAL;
+	}
+
 	p_state_change(tchart, cpu_id, sample->time, state);
 	return 0;
 }
@@ -628,13 +660,20 @@ static int
 process_sample_sched_wakeup(struct timechart *tchart,
 			    struct evsel *evsel,
 			    struct perf_sample *sample,
-			    const char *backtrace)
+			    const char **backtrace)
 {
 	u8 flags  = evsel__intval(evsel, sample, "common_flags");
 	int waker = evsel__intval(evsel, sample, "common_pid");
 	int wakee = evsel__intval(evsel, sample, "pid");
 
-	sched_wakeup(tchart, sample->cpu, sample->time, waker, wakee, flags, backtrace);
+	/* sample->cpu used as index into topology_map[] during SVG generation */
+	if (sample->cpu >= tchart->numcpus) {
+		tchart->nr_invalid_cpu++;
+		return 0;
+	}
+
+	sched_wakeup(tchart, sample->cpu, sample->time, waker, wakee, flags, *backtrace);
+	*backtrace = NULL;
 	return 0;
 }
 
@@ -642,27 +681,41 @@ static int
 process_sample_sched_switch(struct timechart *tchart,
 			    struct evsel *evsel,
 			    struct perf_sample *sample,
-			    const char *backtrace)
+			    const char **backtrace)
 {
 	int prev_pid   = evsel__intval(evsel, sample, "prev_pid");
 	int next_pid   = evsel__intval(evsel, sample, "next_pid");
 	u64 prev_state = evsel__intval(evsel, sample, "prev_state");
 
+	/* sample->cpu used as index into topology_map[] during SVG generation */
+	if (sample->cpu >= tchart->numcpus) {
+		tchart->nr_invalid_cpu++;
+		return 0;
+	}
+
 	sched_switch(tchart, sample->cpu, sample->time, prev_pid, next_pid,
-		     prev_state, backtrace);
+		     prev_state, *backtrace);
+	*backtrace = NULL;
 	return 0;
 }
 
 #ifdef SUPPORT_OLD_POWER_EVENTS
 static int
-process_sample_power_start(struct timechart *tchart __maybe_unused,
+process_sample_power_start(struct timechart *tchart,
 			   struct evsel *evsel,
 			   struct perf_sample *sample,
-			   const char *backtrace __maybe_unused)
+			   const char **backtrace __maybe_unused)
 {
 	u64 cpu_id = evsel__intval(evsel, sample, "cpu_id");
 	u64 value  = evsel__intval(evsel, sample, "value");
 
+	/* Same bounds check as process_sample_cpu_idle — see comment there */
+	if (cpu_id >= tchart->numcpus) {
+		pr_err("power_start event cpu_id %" PRIu64 " >= nr_cpus_avail %u\n",
+		       cpu_id, tchart->numcpus);
+		return -EINVAL;
+	}
+
 	c_state_start(cpu_id, sample->time, value);
 	return 0;
 }
@@ -671,8 +724,16 @@ static int
 process_sample_power_end(struct timechart *tchart,
 			 struct evsel *evsel __maybe_unused,
 			 struct perf_sample *sample,
-			 const char *backtrace __maybe_unused)
+			 const char **backtrace __maybe_unused)
 {
+	/*
+	 * sample->cpu is validated centrally when PERF_SAMPLE_CPU is
+	 * set, but a crafted file could omit it from sample_type.
+	 */
+	if (sample->cpu >= tchart->numcpus) {
+		tchart->nr_invalid_cpu++;
+		return 0;
+	}
 	c_state_end(tchart, sample->cpu, sample->time);
 	return 0;
 }
@@ -681,11 +742,18 @@ static int
 process_sample_power_frequency(struct timechart *tchart,
 			       struct evsel *evsel,
 			       struct perf_sample *sample,
-			       const char *backtrace __maybe_unused)
+			       const char **backtrace __maybe_unused)
 {
 	u64 cpu_id = evsel__intval(evsel, sample, "cpu_id");
 	u64 value  = evsel__intval(evsel, sample, "value");
 
+	/* Same bounds check as process_sample_cpu_idle — see comment there */
+	if (cpu_id >= tchart->numcpus) {
+		pr_err("power_frequency event cpu_id %" PRIu64 " >= nr_cpus_avail %u\n",
+		       cpu_id, tchart->numcpus);
+		return -EINVAL;
+	}
+
 	p_state_change(tchart, cpu_id, sample->time, value);
 	return 0;
 }
@@ -1519,7 +1587,8 @@ static int process_header(struct perf_file_section *section __maybe_unused,
 
 	switch (feat) {
 	case HEADER_NRCPUS:
-		tchart->numcpus = ph->env.nr_cpus_avail;
+		/* Cap at MAX_CPUS — the allocation size of cpus_cstate/pstate arrays */
+		tchart->numcpus = min((int)ph->env.nr_cpus_avail, MAX_CPUS);
 		break;
 
 	case HEADER_CPU_TOPOLOGY:
@@ -1625,6 +1694,16 @@ static int __cmd_timechart(struct timechart *tchart, const char *output_name)
 					    tchart,
 					    process_header);
 
+	/*
+	 * Truncated files (interrupted recording) lose all feature
+	 * sections so the HEADER_NRCPUS callback never fires, and
+	 * pipe mode doesn't use perf_header__process_sections at all.
+	 * Fall back to MAX_CPUS — the actual allocation size of the
+	 * cpus_cstate/pstate arrays.
+	 */
+	if (!tchart->numcpus)
+		tchart->numcpus = MAX_CPUS;
+
 	if (!perf_session__has_traces(session, "timechart record"))
 		goto out_delete;
 
@@ -1646,6 +1725,12 @@ static int __cmd_timechart(struct timechart *tchart, const char *output_name)
 
 	pr_info("Written %2.1f seconds of trace to %s.\n",
 		(tchart->last_time - tchart->first_time) / (double)NSEC_PER_SEC, output_name);
+
+	if (tchart->nr_invalid_cpu) {
+		pr_warning("WARNING: %u events had invalid CPU values and were skipped.\n"
+			   "  Scheduling and power state data may be incomplete.\n",
+			   tchart->nr_invalid_cpu);
+	}
 out_delete:
 	perf_session__delete(session);
 	return ret;
diff --git a/tools/perf/util/svghelper.c b/tools/perf/util/svghelper.c
index e360e7736c7ba65b..a3c7cfecc072f3e3 100644
--- a/tools/perf/util/svghelper.c
+++ b/tools/perf/util/svghelper.c
@@ -736,7 +736,8 @@ static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus)
 		return -1;
 
 	perf_cpu_map__for_each_cpu(cpu, idx, map) {
-		if (cpu.cpu >= nr_cpus) {
+		/* perf_cpu_map__new("") yields cpu=-1; reject to prevent __set_bit OOB */
+		if (cpu.cpu < 0 || cpu.cpu >= nr_cpus) {
 			ret = -1;
 			break;
 		}
@@ -756,7 +757,8 @@ int svg_build_topology_map(struct perf_env *env)
 	char *sib_core, *sib_thr;
 	int ret = -1;
 
-	nr_cpus = min(env->nr_cpus_online, MAX_NR_CPUS);
+	/* Use nr_cpus_avail: offline CPUs still need slots in the topology map */
+	nr_cpus = min(env->nr_cpus_avail, MAX_NR_CPUS);
 
 	t.sib_core_nr = env->nr_sibling_cores;
 	t.sib_thr_nr = env->nr_sibling_threads;
-- 
2.54.0


  parent reply	other threads:[~2026-05-10  3:37 UTC|newest]

Thread overview: 30+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-05-10  3:33 [PATCH 00/28] perf: Harden perf.data parsing against crafted/corrupted files Arnaldo Carvalho de Melo
2026-05-10  3:33 ` [PATCH 01/28] perf session: Add minimum event size validation table Arnaldo Carvalho de Melo
2026-05-11 19:01   ` Ian Rogers
2026-05-10  3:33 ` [PATCH 02/28] perf tools: Fix event_contains() macro to verify full field extent Arnaldo Carvalho de Melo
2026-05-10  3:33 ` [PATCH 03/28] perf zstd: Fix compression error path in zstd_compress_stream_to_records() Arnaldo Carvalho de Melo
2026-05-10  3:33 ` [PATCH 04/28] perf zstd: Fix multi-iteration decompression and error handling Arnaldo Carvalho de Melo
2026-05-10  3:33 ` [PATCH 05/28] perf session: Fix PERF_RECORD_READ swap and dump for variable-length events Arnaldo Carvalho de Melo
2026-05-10  3:33 ` [PATCH 06/28] perf session: Align auxtrace_info priv size before byte-swapping Arnaldo Carvalho de Melo
2026-05-10  3:33 ` [PATCH 07/28] perf session: Add validated swap infrastructure with null-termination checks Arnaldo Carvalho de Melo
2026-05-10  3:33 ` [PATCH 08/28] perf session: Use bounded copy for PERF_RECORD_TIME_CONV Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 09/28] perf session: Validate HEADER_ATTR alignment and attr.size before swapping Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 10/28] perf session: Validate nr fields against event size on both swap and common paths Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 11/28] perf header: Byte-swap build ID event pid and bounds check section entries Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 12/28] perf cpumap: Reject RANGE_CPUS with start_cpu > end_cpu Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 13/28] perf auxtrace: Harden auxtrace_error event handling Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 14/28] perf session: Add byte-swap and bounds check for PERF_RECORD_BPF_METADATA events Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 15/28] perf header: Validate null-termination in PERF_RECORD_EVENT_UPDATE string fields Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 16/28] perf tools: Bounds check perf_event_attr fields against attr.size before printing Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 17/28] perf header: Propagate feature section processing errors Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 18/28] perf header: Validate f_attr.ids section before use in perf_session__read_header() Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 19/28] perf header: Validate feature section size and add read path bounds checking Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 20/28] perf header: Sanity check HEADER_EVENT_DESC attr.size before swap Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 21/28] perf header: Validate bitmap size before allocating in do_read_bitmap() Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 22/28] perf session: Add byte-swap for PERF_RECORD_COMPRESSED2 events Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 23/28] perf tools: Harden compressed event processing Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 24/28] perf session: Check for decompression buffer size overflow Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 25/28] perf session: Bound nr_cpus_avail and validate sample CPU Arnaldo Carvalho de Melo
2026-05-10  3:34 ` Arnaldo Carvalho de Melo [this message]
2026-05-10  3:34 ` [PATCH 27/28] perf kwork: Bounds check work->cpu before indexing cpus_runtime[] Arnaldo Carvalho de Melo
2026-05-10  3:34 ` [PATCH 28/28] perf test: Add truncated perf.data robustness test Arnaldo Carvalho de Melo

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20260510033424.255812-27-acme@kernel.org \
    --to=acme@kernel.org \
    --cc=acme@redhat.com \
    --cc=adrian.hunter@intel.com \
    --cc=irogers@google.com \
    --cc=james.clark@linaro.org \
    --cc=jolsa@kernel.org \
    --cc=kan.liang@linux.intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-perf-users@vger.kernel.org \
    --cc=mingo@kernel.org \
    --cc=namhyung@kernel.org \
    --cc=noreply@anthropic.com \
    --cc=sashiko-bot@kernel.org \
    --cc=tglx@linutronix.de \
    --cc=williams@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox