From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-9.8 required=3.0 tests=BAYES_00,DKIM_INVALID, DKIM_SIGNED,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI, SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E3985C433DF for ; Thu, 30 Jul 2020 01:01:16 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 934CC2075D for ; Thu, 30 Jul 2020 01:01:16 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=fail reason="signature verification failed" (1024-bit key) header.d=gibson.dropbear.id.au header.i=@gibson.dropbear.id.au header.b="pdD2CKWE" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 934CC2075D Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=gibson.dropbear.id.au Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Received: from localhost ([::1]:35142 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1k0wwd-0002DP-Pj for qemu-devel@archiver.kernel.org; Wed, 29 Jul 2020 21:01:15 -0400 Received: from eggs.gnu.org ([2001:470:142:3::10]:58696) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1k0wvY-00019E-GE; Wed, 29 Jul 2020 21:00:08 -0400 Received: from bilbo.ozlabs.org ([2401:3900:2:1::2]:37097 helo=ozlabs.org) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1k0wvV-0006Vg-LQ; Wed, 29 Jul 2020 21:00:08 -0400 Received: by ozlabs.org (Postfix, from userid 1007) id 4BHBsB66hkz9sRX; Thu, 30 Jul 2020 10:59:58 +1000 (AEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=gibson.dropbear.id.au; s=201602; t=1596070798; bh=x/RpU1BK0WaitmCkntUeF2XOttbRK+o5uxRb0gEnT+0=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=pdD2CKWEqfWd6VTetWLrWGs0tKfMYdg4UcP3LYny81XkF4GcfsZeEQ3l1kNwbH4Rn uuu0OFY/9Q0ABhKr44y469fDr7AZ7YDSnCU16olEPT/vkNpgvQSYoLtNnz6QR0d93W AyG7FgGN1tfLRF8OXH24xJWXegP/BEkmNnclnzYM= Date: Thu, 30 Jul 2020 10:58:52 +1000 From: David Gibson To: Daniel Henrique Barboza Subject: Re: [PATCH 1/1] docs: adding NUMA documentation for pseries Message-ID: <20200730005852.GN84173@umbus.fritz.box> References: <20200729125756.224846-1-danielhb413@gmail.com> MIME-Version: 1.0 Content-Type: multipart/signed; micalg=pgp-sha256; protocol="application/pgp-signature"; boundary="H7cT1SUwsqXggVRO" Content-Disposition: inline In-Reply-To: <20200729125756.224846-1-danielhb413@gmail.com> Received-SPF: pass client-ip=2401:3900:2:1::2; envelope-from=dgibson@ozlabs.org; helo=ozlabs.org X-detected-operating-system: by eggs.gnu.org: No matching host in p0f cache. That's all we know. X-Spam_score_int: -9 X-Spam_score: -1.0 X-Spam_bar: - X-Spam_report: (-1.0 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, HEADER_FROM_DIFFERENT_DOMAINS=1, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001, URIBL_BLOCKED=0.001 autolearn=no autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: qemu-ppc@nongnu.org, qemu-devel@nongnu.org Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" --H7cT1SUwsqXggVRO Content-Type: text/plain; charset=us-ascii Content-Disposition: inline Content-Transfer-Encoding: quoted-printable On Wed, Jul 29, 2020 at 09:57:56AM -0300, Daniel Henrique Barboza wrote: > This patch adds a new documentation file, ppc-spapr-numa.rst, > informing what developers and user can expect of the NUMA distance > support for the pseries machine, up to QEMU 5.1. >=20 > In the (hopefully soon) future, when we rework the NUMA mechanics > of the pseries machine to at least attempt to contemplate user > choice, this doc will be extended to inform about the new > support. >=20 > Signed-off-by: Daniel Henrique Barboza Applied to ppc-for-5.2, thanks. > --- > docs/specs/ppc-spapr-numa.rst | 191 ++++++++++++++++++++++++++++++++++ > 1 file changed, 191 insertions(+) > create mode 100644 docs/specs/ppc-spapr-numa.rst >=20 > diff --git a/docs/specs/ppc-spapr-numa.rst b/docs/specs/ppc-spapr-numa.rst > new file mode 100644 > index 0000000000..e762038022 > --- /dev/null > +++ b/docs/specs/ppc-spapr-numa.rst > @@ -0,0 +1,191 @@ > + > +NUMA mechanics for sPAPR (pseries machines) > +=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D > + > +NUMA in sPAPR works different than the System Locality Distance > +Information Table (SLIT) in ACPI. The logic is explained in the LOPAPR > +1.1 chapter 15, "Non Uniform Memory Access (NUMA) Option". This > +document aims to complement this specification, providing details > +of the elements that impacts how QEMU views NUMA in pseries. > + > +Associativity and ibm,associativity property > +-------------------------------------------- > + > +Associativity is defined as a group of platform resources that has > +similar mean performance (or in our context here, distance) relative to > +everyone else outside of the group. > + > +The format of the ibm,associativity property varies with the value of > +bit 0 of byte 5 of the ibm,architecture-vec-5 property. The format with > +bit 0 equal to zero is deprecated. The current format, with the bit 0 > +with the value of one, makes ibm,associativity property represent the > +physical hierarchy of the platform, as one or more lists that starts > +with the highest level grouping up to the smallest. Considering the > +following topology: > + > +:: > + > + Mem M1 ---- Proc P1 | > + ----------------- | Socket S1 ---| > + chip C1 | | > + | HW module 1 (MOD1) > + Mem M2 ---- Proc P2 | | > + ----------------- | Socket S2 ---| > + chip C2 | > + > +The ibm,associativity property for the processors would be: > + > +* P1: {MOD1, S1, C1, P1} > +* P2: {MOD1, S2, C2, P2} > + > +Each allocable resource has an ibm,associativity property. The LOPAPR > +specification allows multiple lists to be present in this property, > +considering that the same resource can have multiple connections to the > +platform. > + > +Relative Performance Distance and ibm,associativity-reference-points > +-------------------------------------------------------------------- > + > +The ibm,associativity-reference-points property is an array that is used > +to define the relevant performance/distance related boundaries, defining > +the NUMA levels for the platform. > + > +The definition of its elements also varies with the value of bit 0 of by= te 5 > +of the ibm,architecture-vec-5 property. The format with bit 0 equal to z= ero > +is also deprecated. With the current format, each integer of the > +ibm,associativity-reference-points represents an 1 based ordinal index (= i.e. > +the first element is 1) of the ibm,associativity array. The first > +boundary is the most significant to application performance, followed by > +less significant boundaries. Allocated resources that belongs to the > +same performance boundaries are expected to have relative NUMA distance > +that matches the relevancy of the boundary itself. Resources that belongs > +to the same first boundary will have the shortest distance from each > +other. Subsequent boundaries represents greater distances and degraded > +performance. > + > +Using the previous example, the following setting reference points defin= es > +three NUMA levels: > + > +* ibm,associativity-reference-points =3D {0x3, 0x2, 0x1} > + > +The first NUMA level (0x3) is interpreted as the third element of each > +ibm,associativity array, the second level is the second element and > +the third level is the first element. Let's also consider that elements > +belonging to the first NUMA level have distance equal to 10 from each > +other, and each NUMA level doubles the distance from the previous. This > +means that the second would be 20 and the third level 40. For the P1 and > +P2 processors, we would have the following NUMA levels: > + > +:: > + > + * ibm,associativity-reference-points =3D {0x3, 0x2, 0x1} > + > + * P1: associativity{MOD1, S1, C1, P1} > + > + First NUMA level (0x3) =3D> associativity[2] =3D C1 > + Second NUMA level (0x2) =3D> associativity[1] =3D S1 > + Third NUMA level (0x1) =3D> associativity[0] =3D MOD1 > + > + * P2: associativity{MOD1, S2, C2, P2} > + > + First NUMA level (0x3) =3D> associativity[2] =3D C2 > + Second NUMA level (0x2) =3D> associativity[1] =3D S2 > + Third NUMA level (0x1) =3D> associativity[0] =3D MOD1 > + > + P1 and P2 have the same third NUMA level, MOD1: Distance between them = =3D 40 > + > +Changing the ibm,associativity-reference-points array changes the perfor= mance > +distance attributes for the same associativity arrays, as the following > +example illustrates: > + > +:: > + > + * ibm,associativity-reference-points =3D {0x2} > + > + * P1: associativity{MOD1, S1, C1, P1} > + > + First NUMA level (0x2) =3D> associativity[1] =3D S1 > + > + * P2: associativity{MOD1, S2, C2, P2} > + > + First NUMA level (0x2) =3D> associativity[1] =3D S2 > + > + P1 and P2 does not have a common performance boundary. Since this is a= one level > + NUMA configuration, distance between them is one boundary above the fi= rst > + level, 20. > + > + > +In a hypothetical platform where all resources inside the same hardware = module > +is considered to be on the same performance boundary: > + > +:: > + > + * ibm,associativity-reference-points =3D {0x1} > + > + * P1: associativity{MOD1, S1, C1, P1} > + > + First NUMA level (0x1) =3D> associativity[0] =3D MOD0 > + > + * P2: associativity{MOD1, S2, C2, P2} > + > + First NUMA level (0x1) =3D> associativity[0] =3D MOD0 > + > + P1 and P2 belongs to the same first order boundary. The distance betwe= en then > + is 10. > + > + > +How the pseries Linux guest calculates NUMA distances > +=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D > + > +Another key difference between ACPI SLIT and the LOPAPR regarding NUMA is > +how the distances are expressed. The SLIT table provides the NUMA distan= ce > +value between the relevant resources. LOPAPR does not provide a standard > +way to calculate it. We have the ibm,associativity for each resource, wh= ich > +provides a common-performance hierarchy, and the ibm,associativity-refe= rence-points > +array that tells which level of associativity is considered to be releva= nt > +or not. > + > +The result is that each OS is free to implement and to interpret the dis= tance > +as it sees fit. For the pseries Linux guest, each level of NUMA duplicat= es > +the distance of the previous level, and the maximum amount of levels is > +limited to MAX_DISTANCE_REF_POINTS =3D 4 (from arch/powerpc/mm/numa.c in= the > +kernel tree). This results in the following distances: > + > +* both resources in the first NUMA level: 10 > +* resources one NUMA level apart: 20 > +* resources two NUMA levels apart: 40 > +* resources three NUMA levels apart: 80 > +* resources four NUMA levels apart: 160 > + > + > +Consequences for QEMU NUMA tuning > +--------------------------------- > + > +The way the pseries Linux guest calculates NUMA distances has a direct e= ffect > +on what QEMU users can expect when doing NUMA tuning. As of QEMU 5.1, th= is is > +the default ibm,associativity-reference-points being used in the pseries > +machine: > + > +ibm,associativity-reference-points =3D {0x4, 0x4, 0x2} > + > +The first and second level are equal, 0x4, and a third one was added in > +commit a6030d7e0b35 exclusively for NVLink GPUs support. This means that > +regardless of how the ibm,associativity properties are being created in > +the device tree, the pseries Linux guest will only recognize three scena= rios > +as far as NUMA distance goes: > + > +* if the resources belongs to the same first NUMA level =3D 10 > +* second level is skipped since it's equal to the first > +* all resources that aren't a NVLink GPU, it is guaranteed that they wil= l belong > + to the same third NUMA level, having distance =3D 40 > +* for NVLink GPUs, distance =3D 80 from everything else > + > +In short, we can summarize the NUMA distances seem in pseries Linux gues= ts, using > +QEMU up to 5.1, as follows: > + > +* local distance, i.e. the distance of the resource to its own NUMA node= : 10 > +* if it's a NVLink GPU device, distance: 80 > +* every other resource, distance: 40 > + > +This also means that user input in QEMU command line does not change the > +NUMA distancing inside the guest for the pseries machine. --=20 David Gibson | I'll have my music baroque, and my code david AT gibson.dropbear.id.au | minimalist, thank you. NOT _the_ _other_ | _way_ _around_! http://www.ozlabs.org/~dgibson --H7cT1SUwsqXggVRO Content-Type: application/pgp-signature; name="signature.asc" -----BEGIN PGP SIGNATURE----- iQIzBAEBCAAdFiEEdfRlhq5hpmzETofcbDjKyiDZs5IFAl8iG0YACgkQbDjKyiDZ s5Jxig/+JlMbwXOn6vg5rDR/FBMbGdLmYz2adXWeCp/M2SDO97sYgee5jCu3pUPX fSee6kxbR+RCqUAlw1lZVYBYf2PgJbveeRXhB7m1Hjt10d5nEMLn5HeEt+d/SAjb 0LNhkkELKfU2wbzoa1jE+vhaQTMszrhgfOPF3EduxuGrQbVLBU7bQorCTr5vqUhx 4xJKwuKLCl3WDS1E40Y73WkDfWqLruPG22+UwdUMA4jUJBPMaez8CzSYJL33gzug GkyQuxis6QUC7R5I+zw84gKaUJ/c/DESCrO0TKvgnKw57w93o8n0WJQ/jsPigWtd bLCNvhln2cO5T8KRothyI5syouuBiKE84kNqhKhqC0DIY/tcq/C7N32u/aPDsjQt vpap3p6io7I6eqe9xTeWuf6Ceq5JXCocBnwVL5RSNGLPFsrR6ipJvWr3+gQajoUQ G5Kv36nEmzE/TcCagSnH92HfHmGpqAfOtVPhQugGb7nIBc2EZ2s2cH6dBB1My8b7 weFa9f5wkK+yiUZEYy5+HDmyQrrvKL2doOsED3KjOr2HRjVTjUbmusmudRmW+kSv +Zq7f+iDoUwopEyICs3+ojHlbFbM5yjAE3CchK6mWVQ4xtld2vaMwwR20v5Ash5N 8j4QF61xlZ4DhVunEJmEkYXzYesGmCjYtemRlmA+DG78FzxcFGo= =b4Jw -----END PGP SIGNATURE----- --H7cT1SUwsqXggVRO--