From: "Daniel P. Berrangé" <berrange@redhat.com>
To: Gavin Shan <gshan@redhat.com>
Cc: qemu-arm@nongnu.org, qemu-devel@nongnu.org,
qemu-riscv@nongnu.org, rad@semihalf.com,
peter.maydell@linaro.org, quic_llindhol@quicinc.com,
eduardo@habkost.net, marcel.apfelbaum@gmail.com,
philmd@linaro.org, wangyanan55@huawei.com, palmer@dabbelt.com,
alistair.francis@wdc.com, bin.meng@windriver.com,
thuth@redhat.com, lvivier@redhat.com, pbonzini@redhat.com,
imammedo@redhat.com, yihyu@redhat.com, shan.gavin@gmail.com
Subject: Re: [PATCH v2 0/4] NUMA: Apply socket-NUMA-node boundary for aarch64 and RiscV machines
Date: Thu, 23 Feb 2023 12:57:22 +0000 [thread overview]
Message-ID: <Y/disinKmr6gLby1@redhat.com> (raw)
In-Reply-To: <20230223081401.248835-1-gshan@redhat.com>
On Thu, Feb 23, 2023 at 04:13:57PM +0800, Gavin Shan wrote:
> For arm64 and RiscV architecture, the driver (/base/arch_topology.c) is
> used to populate the CPU topology in the Linux guest. It's required that
> the CPUs in one socket can't span mutiple NUMA nodes. Otherwise, the Linux
> scheduling domain can't be sorted out, as the following warning message
> indicates. To avoid the unexpected confusion, this series attempts to
> rejects such kind of insane configurations.
>
> -smp 6,maxcpus=6,sockets=2,clusters=1,cores=3,threads=1 \
> -numa node,nodeid=0,cpus=0-1,memdev=ram0 \
> -numa node,nodeid=1,cpus=2-3,memdev=ram1 \
> -numa node,nodeid=2,cpus=4-5,memdev=ram2 \
This is somewhat odd as a config, because core 2 is in socket 0
and core 3 is in socket 1, so it wouldn't make much conceptual
sense to have them in the same NUMA node, while other cores within
the same socket are in different NUMA nodes. Basically the split
of NUMA nodes is not aligned with any level in the topology.
This series, however, also rejects configurations that I would
normally consider to be reasonable. I've not tested linux kernel
behaviour though, but as a user I would expect to be able to
ask for:
-smp 6,maxcpus=6,sockets=2,clusters=1,cores=3,threads=1 \
-numa node,nodeid=0,cpus=0,memdev=ram0 \
-numa node,nodeid=1,cpus=1,memdev=ram1 \
-numa node,nodeid=2,cpus=2,memdev=ram2 \
-numa node,nodeid=3,cpus=3,memdev=ram3 \
-numa node,nodeid=4,cpus=4,memdev=ram4 \
-numa node,nodeid=5,cpus=5,memdev=ram5 \
ie, every core gets its own NUMA node
Or to aask for every cluster as a numa node:
-smp 6,maxcpus=6,sockets=2,clusters=3,cores=1,threads=1 \
-numa node,nodeid=0,cpus=0,memdev=ram0 \
-numa node,nodeid=1,cpus=1,memdev=ram1 \
-numa node,nodeid=2,cpus=2,memdev=ram2 \
-numa node,nodeid=3,cpus=3,memdev=ram3 \
-numa node,nodeid=4,cpus=4,memdev=ram4 \
-numa node,nodeid=5,cpus=5,memdev=ram5 \
In both cases the NUMA split is aligned with a given level
in the topology, which was not the case with your example.
Rejecting these feels overly strict to me, and may risk breaking
existing valid deployments, unless we can demonstrate those
scenarios were unambiguously already broken ?
If there was something in the hardware specs that requires
this, then it is more valid to do, than if it is merely an
specific guest kernel limitation that might be fixed any day.
With regards,
Daniel
--
|: https://berrange.com -o- https://www.flickr.com/photos/dberrange :|
|: https://libvirt.org -o- https://fstop138.berrange.com :|
|: https://entangle-photo.org -o- https://www.instagram.com/dberrange :|
next prev parent reply other threads:[~2023-02-23 12:59 UTC|newest]
Thread overview: 18+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-02-23 8:13 [PATCH v2 0/4] NUMA: Apply socket-NUMA-node boundary for aarch64 and RiscV machines Gavin Shan
2023-02-23 8:13 ` [PATCH v2 1/4] qtest/numa-test: Follow socket-NUMA-node boundary for aarch64 Gavin Shan
2023-02-23 8:13 ` [PATCH v2 2/4] numa: Validate socket and NUMA node boundary if required Gavin Shan
2023-02-23 9:05 ` Philippe Mathieu-Daudé
2023-02-23 10:27 ` Gavin Shan
2023-02-23 8:14 ` [PATCH v2 3/4] hw/arm: Validate socket and NUMA node boundary Gavin Shan
2023-02-23 8:14 ` [PATCH v2 4/4] hw/riscv: " Gavin Shan
2023-02-23 12:25 ` [PATCH v2 0/4] NUMA: Apply socket-NUMA-node boundary for aarch64 and RiscV machines Andrew Jones
2023-02-24 7:20 ` Gavin Shan
2023-02-23 12:57 ` Daniel P. Berrangé [this message]
2023-02-24 5:47 ` Gavin Shan
2023-02-23 13:18 ` Daniel Henrique Barboza
2023-02-24 7:09 ` Gavin Shan
2023-02-24 9:26 ` Daniel Henrique Barboza
2023-02-24 10:16 ` Gavin Shan
2023-02-24 10:39 ` Andrew Jones
2023-02-24 14:20 ` Igor Mammedov
2023-02-25 0:05 ` Gavin Shan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=Y/disinKmr6gLby1@redhat.com \
--to=berrange@redhat.com \
--cc=alistair.francis@wdc.com \
--cc=bin.meng@windriver.com \
--cc=eduardo@habkost.net \
--cc=gshan@redhat.com \
--cc=imammedo@redhat.com \
--cc=lvivier@redhat.com \
--cc=marcel.apfelbaum@gmail.com \
--cc=palmer@dabbelt.com \
--cc=pbonzini@redhat.com \
--cc=peter.maydell@linaro.org \
--cc=philmd@linaro.org \
--cc=qemu-arm@nongnu.org \
--cc=qemu-devel@nongnu.org \
--cc=qemu-riscv@nongnu.org \
--cc=quic_llindhol@quicinc.com \
--cc=rad@semihalf.com \
--cc=shan.gavin@gmail.com \
--cc=thuth@redhat.com \
--cc=wangyanan55@huawei.com \
--cc=yihyu@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).