public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Jiri Olsa <jolsa@redhat.com>
To: kan.liang@linux.intel.com
Cc: acme@kernel.org, jolsa@kernel.org, mingo@redhat.com,
	linux-kernel@vger.kernel.org, peterz@infradead.org,
	ak@linux.intel.com
Subject: Re: [PATCH 2/3] perf stat: Support per-die aggregation
Date: Tue, 28 May 2019 11:00:20 +0200	[thread overview]
Message-ID: <20190528090020.GG27906@krava> (raw)
In-Reply-To: <1558644081-17738-2-git-send-email-kan.liang@linux.intel.com>

On Thu, May 23, 2019 at 01:41:20PM -0700, kan.liang@linux.intel.com wrote:

SNIP

>  		if (cpu_map__build_core_map(evsel_list->cpus, &stat_config.aggr_map)) {
>  			perror("cannot build core map");
> @@ -936,21 +957,41 @@ static int perf_env__get_socket(struct cpu_map *map, int idx, void *data)
>  	return cpu == -1 ? -1 : env->cpu[cpu].socket_id;
>  }
>  
> +static int perf_env__get_die(struct cpu_map *map, int idx, void *data)
> +{
> +	struct perf_env *env = data;
> +	int die = -1, cpu = perf_env__get_cpu(env, map, idx);
> +
> +	if (cpu != -1) {
> +		/*
> +		 * Encode socket in upper 8 bits
> +		 * die_id is relative to socket,
> +		 * we need a global id. So we combine
> +		 * socket + die id
> +		 */
> +		die = (env->cpu[cpu].socket_id << 8) |
> +		      (env->cpu[cpu].die_id & 0xff);
> +	}
> +
> +	return die;
> +}
> +
>  static int perf_env__get_core(struct cpu_map *map, int idx, void *data)
>  {
>  	struct perf_env *env = data;
>  	int core = -1, cpu = perf_env__get_cpu(env, map, idx);
>  
>  	if (cpu != -1) {
> -		int socket_id = env->cpu[cpu].socket_id;
> -
>  		/*
> -		 * Encode socket in upper 16 bits
> -		 * core_id is relative to socket, and
> +		 * Encode socket in upper 24 bits
> +		 * encode die id in upper 16 bits
> +		 * core_id is relative to socket and die,
>  		 * we need a global id. So we combine
> -		 * socket + core id.
> +		 * socket + die id + core id
>  		 */
> -		core = (socket_id << 16) | (env->cpu[cpu].core_id & 0xffff);
> +		core = (env->cpu[cpu].socket_id << 24) |
> +		       (env->cpu[cpu].die_id << 16) |
> +		       (env->cpu[cpu].core_id & 0xffff);

I guess we're still safe with 1 byte for socket and die id,
but could we still check the size fits, and warn and bail
out otherwise?

thanks,
jirka

  reply	other threads:[~2019-05-28  9:00 UTC|newest]

Thread overview: 13+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-05-23 20:41 [PATCH 1/3] perf header: Add die information in CPU topology kan.liang
2019-05-23 20:41 ` [PATCH 2/3] perf stat: Support per-die aggregation kan.liang
2019-05-28  9:00   ` Jiri Olsa [this message]
2019-05-23 20:41 ` [PATCH 3/3] perf header: Rename "sibling cores" to "sibling sockets" kan.liang
2019-05-28  8:59   ` Jiri Olsa
2019-05-28 19:05     ` Liang, Kan
2019-05-28  8:59 ` [PATCH 1/3] perf header: Add die information in CPU topology Jiri Olsa
2019-05-28  8:59 ` Jiri Olsa
2019-05-28  9:00 ` Jiri Olsa
2019-05-28 19:06   ` Liang, Kan
2019-05-28 19:29     ` Jiri Olsa
2019-05-28  9:00 ` Jiri Olsa
2019-05-28  9:00 ` Jiri Olsa

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20190528090020.GG27906@krava \
    --to=jolsa@redhat.com \
    --cc=acme@kernel.org \
    --cc=ak@linux.intel.com \
    --cc=jolsa@kernel.org \
    --cc=kan.liang@linux.intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=mingo@redhat.com \
    --cc=peterz@infradead.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox