From: Valentin Schneider <valentin.schneider@arm.com>
To: Sudeep Holla <sudeep.holla@arm.com>
Cc: linux-kernel@vger.kernel.org,
linux-arm-kernel@lists.infradead.org,
Catalin Marinas <catalin.marinas@arm.com>,
Will Deacon <will@kernel.org>,
Robin Murphy <robin.murphy@arm.com>,
Jeremy Linton <Jeremy.Linton@arm.com>,
Dietmar Eggemann <dietmar.eggemann@arm.com>,
Morten Rasmussen <morten.rasmussen@arm.com>,
"Zengtao \(B\)" <prime.zeng@hisilicon.com>
Subject: Re: [PATCH] arm64: topology: Stop using MPIDR for topology information
Date: Wed, 02 Sep 2020 11:52:34 +0100 [thread overview]
Message-ID: <jhj8sdso3gt.mognet@arm.com> (raw)
In-Reply-To: <20200902100422.GA25462@bogus>
On 02/09/20 11:04, Sudeep Holla wrote:
> On Sat, Aug 29, 2020 at 02:00:16PM +0100, Valentin Schneider wrote:
>> In the absence of ACPI or DT topology data, we fallback to haphazardly
>> decoding *something* out of MPIDR. Sadly, the contents of that register are
>> mostly unusable due to the implementation leniancy and things like Aff0
>> having to be capped to 15 (despite being encoded on 8 bits).
>>
>> Consider a simple system with a single package of 32 cores, all under the
>> same LLC. We ought to be shoving them in the same core_sibling mask, but
>> MPIDR is going to look like:
>>
>> | CPU | 0 | ... | 15 | 16 | ... | 31 |
>> |------+---+-----+----+----+-----+----+
>> | Aff0 | 0 | ... | 15 | 0 | ... | 15 |
>> | Aff1 | 0 | ... | 0 | 1 | ... | 1 |
>> | Aff2 | 0 | ... | 0 | 0 | ... | 0 |
>>
>> Which will eventually yield
>>
>> core_sibling(0-15) == 0-15
>> core_sibling(16-31) == 16-31
>>
>> NUMA woes
>> =========
>>
>> If we try to play games with this and set up NUMA boundaries within those
>> groups of 16 cores via e.g. QEMU:
>>
>> # Node0: 0-9; Node1: 10-19
>> $ qemu-system-aarch64 <blah> \
>> -smp 20 -numa node,cpus=0-9,nodeid=0 -numa node,cpus=10-19,nodeid=1
>>
>> The scheduler's MC domain (all CPUs with same LLC) is going to be built via
>>
>> arch_topology.c::cpu_coregroup_mask()
>>
>> In there we try to figure out a sensible mask out of the topology
>> information we have. In short, here we'll pick the smallest of NUMA or
>> core sibling mask.
>>
>> node_mask(CPU9) == 0-9
>> core_sibling(CPU9) == 0-15
>>
>> MC mask for CPU9 will thus be 0-9, not a problem.
>>
>> node_mask(CPU10) == 10-19
>> core_sibling(CPU10) == 0-15
>>
>> MC mask for CPU10 will thus be 10-19, not a problem.
>>
>> node_mask(CPU16) == 10-19
>> core_sibling(CPU16) == 16-19
>>
>> MC mask for CPU16 will thus be 16-19... Uh oh. CPUs 16-19 are in two
>> different unique MC spans, and the scheduler has no idea what to make of
>> that. That triggers the WARN_ON() added by commit
>>
>> ccf74128d66c ("sched/topology: Assert non-NUMA topology masks don't (partially) overlap")
>>
>> Fixing MPIDR-derived topology
>> =============================
>>
>> We could try to come up with some cleverer scheme to figure out which of
>> the available masks to pick, but really if one of those masks resulted from
>> MPIDR then it should be discarded because it's bound to be bogus.
>>
>> I was hoping to give MPIDR a chance for SMT, to figure out which threads are
>> in the same core using Aff1-3 as core ID, but Sudeep and Robin pointed out
>> to me that there are systems out there where *all* cores have non-zero
>> values in their higher affinity fields (e.g. RK3288 has "5" in all of its
>> cores' MPIDR.Aff1), which would expose a bogus core ID to userspace.
>>
>> Stop using MPIDR for topology information. When no other source of topology
>> information is available, mark each CPU as its own core and its NUMA node
>> as its LLC domain.
>>
>
> Looks good to me, so:
>
> Reviewed-by: Sudeep Holla <sudeep.holla@arm.com>
>
Thanks!
> However, we need to get it tested on some systems with *weird* MPIDR
> values and don't have topology described in DT cpu-maps and somehow
> (wrongly) relied on below logic. Also though these affect user ABI via
> sysfs topology, I expect systems w/o DT cpu-maps and weird MPIDR are
> broken either way.
>
Agreed, it's the one bit that may be regarded as a regression, but what is
already out there is quite broken :(
> Luckily found only one such mpidr in arm64 DTS files:
> arch/arm64/boot/dts/sprd/sc9860.dtsi
So those have 0x53 for Aff2 for all cores, which is going to end up in the
package_id. AFAICT that means that
/sys/devices/system/cpu/cpu*/topology/physical_package_id
is going to look pretty wild.
next prev parent reply other threads:[~2020-09-02 10:52 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-08-29 13:00 [PATCH] arm64: topology: Stop using MPIDR for topology information Valentin Schneider
2020-09-02 3:24 ` Zengtao (B)
2020-09-02 10:50 ` Sudeep Holla
2020-09-02 10:52 ` Valentin Schneider
2020-09-03 1:44 ` Zengtao (B)
2020-09-09 11:21 ` Valentin Schneider
2020-09-02 10:04 ` Sudeep Holla
2020-09-02 10:52 ` Valentin Schneider [this message]
2020-09-07 21:35 ` Will Deacon
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=jhj8sdso3gt.mognet@arm.com \
--to=valentin.schneider@arm.com \
--cc=Jeremy.Linton@arm.com \
--cc=catalin.marinas@arm.com \
--cc=dietmar.eggemann@arm.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-kernel@vger.kernel.org \
--cc=morten.rasmussen@arm.com \
--cc=prime.zeng@hisilicon.com \
--cc=robin.murphy@arm.com \
--cc=sudeep.holla@arm.com \
--cc=will@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox