From: Peter Zijlstra <peterz@infradead.org>
To: kan.liang@linux.intel.com
Cc: mingo@redhat.com, acme@kernel.org, namhyung@kernel.org,
linux-kernel@vger.kernel.org, ak@linux.intel.com,
eranian@google.com
Subject: Re: [PATCH] perf: Extend per event callchain limit to branch stack
Date: Tue, 11 Mar 2025 12:40:12 +0100 [thread overview]
Message-ID: <20250311114012.GE19424@noisy.programming.kicks-ass.net> (raw)
In-Reply-To: <20250310181536.3645382-1-kan.liang@linux.intel.com>
On Mon, Mar 10, 2025 at 11:15:36AM -0700, kan.liang@linux.intel.com wrote:
> From: Kan Liang <kan.liang@linux.intel.com>
>
> The commit 97c79a38cd45 ("perf core: Per event callchain limit")
> introduced a per-event term to allow finer tuning of the depth of
> callchains to save space.
>
> It should be applied to the branch stack as well. For example, autoFDO
> collections require maximum LBR entries. In the meantime, other
> system-wide LBR users may only be interested in the latest a few number
> of LBRs. A per-event LBR depth would save the perf output buffer.
>
> The patch simply drops the uninterested branches, but HW still collects
> the maximum branches. There may be a model-specific optimization that
> can reduce the HW depth for some cases to reduce the overhead further.
> But it isn't included in the patch set. Because it's not useful for all
> cases. For example, ARCH LBR can utilize the PEBS and XSAVE to collect
> LBRs. The depth should have less impact on the collecting overhead.
> The model-specific optimization may be implemented later separately.
>
> Signed-off-by: Kan Liang <kan.liang@linux.intel.com>
Thanks!
> ---
> include/linux/perf_event.h | 3 +++
> include/uapi/linux/perf_event.h | 2 ++
> 2 files changed, 5 insertions(+)
>
> diff --git a/include/linux/perf_event.h b/include/linux/perf_event.h
> index 24f2eba200ac..bca1dfd30276 100644
> --- a/include/linux/perf_event.h
> +++ b/include/linux/perf_event.h
> @@ -1347,6 +1347,9 @@ static inline void perf_sample_save_brstack(struct perf_sample_data *data,
>
> if (branch_sample_hw_index(event))
> size += sizeof(u64);
> +
> + brs->nr = min_t(u16, event->attr.sample_max_stack, brs->nr);
> +
> size += brs->nr * sizeof(struct perf_branch_entry);
>
> /*
> diff --git a/include/uapi/linux/perf_event.h b/include/uapi/linux/perf_event.h
> index 0524d541d4e3..5fc753c23734 100644
> --- a/include/uapi/linux/perf_event.h
> +++ b/include/uapi/linux/perf_event.h
> @@ -385,6 +385,8 @@ enum perf_event_read_format {
> *
> * @sample_max_stack: Max number of frame pointers in a callchain,
> * should be < /proc/sys/kernel/perf_event_max_stack
> + * Max number of entries of branch stack
> + * should be < hardware limit
> */
> struct perf_event_attr {
>
> --
> 2.38.1
>
next prev parent reply other threads:[~2025-03-11 11:40 UTC|newest]
Thread overview: 3+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-03-10 18:15 [PATCH] perf: Extend per event callchain limit to branch stack kan.liang
2025-03-11 11:40 ` Peter Zijlstra [this message]
2025-03-17 10:34 ` [tip: perf/core] " tip-bot2 for Kan Liang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250311114012.GE19424@noisy.programming.kicks-ass.net \
--to=peterz@infradead.org \
--cc=acme@kernel.org \
--cc=ak@linux.intel.com \
--cc=eranian@google.com \
--cc=kan.liang@linux.intel.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@redhat.com \
--cc=namhyung@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox