From: Mark Rutland <mark.rutland@arm.com>
To: linux-arm-kernel@lists.infradead.org
Cc: ardb@kernel.org, bertrand.marquis@arm.com,
boris.ostrovsky@oracle.com, broonie@kernel.org,
catalin.marinas@arm.com, daniel.lezcano@linaro.org,
james.morse@arm.com, jgross@suse.com, mark.rutland@arm.com,
maz@kernel.org, oliver.upton@linux.dev, pcc@google.com,
sstabellini@kernel.org, suzuki.poulose@arm.com,
tglx@linutronix.de, vladimir.murzin@arm.com, will@kernel.org
Subject: [PATCH 28/37] arm64: Avoid cpus_have_const_cap() for ARM64_{SVE,SME,SME2,FA64}
Date: Tue, 19 Sep 2023 10:28:41 +0100 [thread overview]
Message-ID: <20230919092850.1940729-29-mark.rutland@arm.com> (raw)
In-Reply-To: <20230919092850.1940729-1-mark.rutland@arm.com>
In system_supports_{sve,sme,sme2,fa64}() we use cpus_have_const_cap() to
check for the relevant cpucaps, but this is only necessary so that
sve_setup() and sme_setup() can run prior to alternatives being patched,
and otherwise alternative_has_cap_*() would be preferable.
For historical reasons, cpus_have_const_cap() is more complicated than
it needs to be. Before cpucaps are finalized, it will perform a bitmap
test of the system_cpucaps bitmap, and once cpucaps are finalized it
will use an alternative branch. This used to be necessary to handle some
race conditions in the window between cpucap detection and the
subsequent patching of alternatives and static branches, where different
branches could be out-of-sync with one another (or w.r.t. alternative
sequences). Now that we use alternative branches instead of static
branches, these are all patched atomically w.r.t. one another, and there
are only a handful of cases that need special care in the window between
cpucap detection and alternative patching.
Due to the above, it would be nice to remove cpus_have_const_cap(), and
migrate callers over to alternative_has_cap_*(), cpus_have_final_cap(),
or cpus_have_cap() depending on when their requirements. This will
remove redundant instructions and improve code generation, and will make
it easier to determine how each callsite will behave before, during, and
after alternative patching.
All of system_supports_{sve,sme,sme2,fa64}() will return false prior to
system cpucaps being detected. In the window between system cpucaps being
detected and patching alternatives, we need system_supports_sve() and
system_supports_sme() to run to initialize SVE and SME properties, but
all other users of system_supports_{sve,sme,sme2,fa64}() don't depend on
the relevant cpucap becoming true until alternatives are patched:
* No KVM code runs until after alternatives are patched, and so this can
safely use cpus_have_final_cap() or alternative_has_cap_*().
* The cpuid_cpu_online() callback in arch/arm64/kernel/cpuinfo.c is
registered later from cpuinfo_regs_init() as a device_initcall, and so
this can safely use cpus_have_final_cap() or alternative_has_cap_*().
* The entry, signal, and ptrace code isn't reachable until userspace has
run, and so this can safely use cpus_have_final_cap() or
alternative_has_cap_*().
* Currently perf_reg_validate() will un-reserve the PERF_REG_ARM64_VG
pseudo-register before alternatives are patched, and before
sve_setup() has run. If a sampling event is created early enough, this
would allow perf_ext_reg_value() to sample (the as-yet uninitialized)
thread_struct::vl[] prior to alternatives being patched.
It would be preferable to defer this until alternatives are patched,
and this can safely use alternative_has_cap_*().
* The context-switch code will run during this window as part of
stop_machine() used during alternatives_patch_all(), and potentially
for other work if other kernel threads are created early. No threads
require the use of SVE/SME/SME2/FA64 prior to alternatives being
patched, and it would be preferable for the related context-switch
logic to take effect after alternatives are patched so that ths is
guaranteed to see a consistent system-wide state (e.g. anything
initialized by sve_setup() and sme_setup().
This can safely ues alternative_has_cap_*().
This patch replaces the use of cpus_have_const_cap() with
alternative_has_cap_unlikely(), which will avoid generating code to test
the system_cpucaps bitmap and should be better for all subsequent calls
at runtime. The sve_setup() and sme_setup() functions are modified to
use cpus_have_cap() directly so that they can observe the cpucaps being
set prior to alternatives being patched.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Mark Brown <broonie@kernel.org>
Cc: Suzuki K Poulose <suzuki.poulose@arm.com>
Cc: Will Deacon <will@kernel.org>
---
arch/arm64/include/asm/cpufeature.h | 8 ++++----
arch/arm64/kernel/fpsimd.c | 4 ++--
2 files changed, 6 insertions(+), 6 deletions(-)
diff --git a/arch/arm64/include/asm/cpufeature.h b/arch/arm64/include/asm/cpufeature.h
index 7c218828c06d5..4deaa94de36e8 100644
--- a/arch/arm64/include/asm/cpufeature.h
+++ b/arch/arm64/include/asm/cpufeature.h
@@ -776,22 +776,22 @@ static inline bool system_uses_ttbr0_pan(void)
static __always_inline bool system_supports_sve(void)
{
- return cpus_have_const_cap(ARM64_SVE);
+ return alternative_has_cap_unlikely(ARM64_SVE);
}
static __always_inline bool system_supports_sme(void)
{
- return cpus_have_const_cap(ARM64_SME);
+ return alternative_has_cap_unlikely(ARM64_SME);
}
static __always_inline bool system_supports_sme2(void)
{
- return cpus_have_const_cap(ARM64_SME2);
+ return alternative_has_cap_unlikely(ARM64_SME2);
}
static __always_inline bool system_supports_fa64(void)
{
- return cpus_have_const_cap(ARM64_SME_FA64);
+ return alternative_has_cap_unlikely(ARM64_SME_FA64);
}
static __always_inline bool system_supports_tpidr2(void)
diff --git a/arch/arm64/kernel/fpsimd.c b/arch/arm64/kernel/fpsimd.c
index 6b7960c445daf..85bba83dda996 100644
--- a/arch/arm64/kernel/fpsimd.c
+++ b/arch/arm64/kernel/fpsimd.c
@@ -1199,7 +1199,7 @@ void __init sve_setup(void)
DECLARE_BITMAP(tmp_map, SVE_VQ_MAX);
unsigned long b;
- if (!system_supports_sve())
+ if (!cpus_have_cap(ARM64_SVE))
return;
/*
@@ -1358,7 +1358,7 @@ void __init sme_setup(void)
u64 smcr;
int min_bit;
- if (!system_supports_sme())
+ if (!cpus_have_cap(ARM64_SME))
return;
/*
--
2.30.2
_______________________________________________
linux-arm-kernel mailing list
linux-arm-kernel@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
next prev parent reply other threads:[~2023-09-19 9:31 UTC|newest]
Thread overview: 56+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-09-19 9:28 [PATCH 00/37] arm64: Remove cpus_have_const_cap() Mark Rutland
2023-09-19 9:28 ` [PATCH 01/37] clocksource/drivers/arm_arch_timer: Initialize evtstrm after finalizing cpucaps Mark Rutland
2023-09-21 7:41 ` Marc Zyngier
2023-09-21 16:27 ` Mark Rutland
2023-09-19 9:28 ` [PATCH 02/37] arm64/arm: xen: enlighten: Fix KPTI checks Mark Rutland
2023-09-19 9:28 ` [PATCH 03/37] arm64: Factor out cpucap definitions Mark Rutland
2023-09-19 9:28 ` [PATCH 04/37] arm64: Add cpucap_is_possible() Mark Rutland
2023-09-19 9:28 ` [PATCH 05/37] arm64: Add cpus_have_final_boot_cap() Mark Rutland
2023-09-21 9:13 ` Suzuki K Poulose
2023-09-21 16:36 ` Mark Rutland
2023-09-22 10:26 ` Suzuki K Poulose
2023-10-02 10:25 ` Mark Rutland
2023-10-05 9:23 ` Mark Rutland
2023-10-05 9:39 ` Suzuki K Poulose
2023-09-19 9:28 ` [PATCH 06/37] arm64: Rework setup_cpu_features() Mark Rutland
2023-09-25 13:04 ` Suzuki K Poulose
2023-09-19 9:28 ` [PATCH 07/37] arm64: Fixup user features at boot time Mark Rutland
2023-09-19 9:28 ` [PATCH 08/37] arm64: Split kpti_install_ng_mappings() Mark Rutland
2023-09-19 9:28 ` [PATCH 09/37] arm64: kvm: Use cpus_have_final_cap() explicitly Mark Rutland
2023-09-21 7:49 ` Marc Zyngier
2023-09-19 9:28 ` [PATCH 10/37] arm64: Explicitly save/restore CPACR when probing SVE and SME Mark Rutland
2023-09-19 9:28 ` [PATCH 11/37] arm64: Rename SVE/SME cpu_enable functions Mark Rutland
2023-09-19 10:52 ` Mark Brown
2023-09-21 16:50 ` Mark Rutland
2023-09-19 9:28 ` [PATCH 12/37] arm64: Use a positive cpucap for FP/SIMD Mark Rutland
2023-09-19 11:21 ` Mark Brown
2023-09-19 9:28 ` [PATCH 13/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_{ADDRESS,GENERIC}_AUTH Mark Rutland
2023-09-19 9:28 ` [PATCH 14/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_ARMv8_4_TTL Mark Rutland
2023-09-19 9:28 ` [PATCH 15/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_BTI Mark Rutland
2023-09-19 11:23 ` Mark Brown
2023-09-19 9:28 ` [PATCH 16/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_CACHE_DIC Mark Rutland
2023-09-19 9:28 ` [PATCH 17/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_CNP Mark Rutland
2023-09-19 9:28 ` [PATCH 18/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_DIT Mark Rutland
2023-09-19 9:28 ` [PATCH 19/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_GIC_PRIO_MASKING Mark Rutland
2023-09-19 9:28 ` [PATCH 20/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_PAN Mark Rutland
2023-09-19 9:28 ` [PATCH 21/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_EPAN Mark Rutland
2023-09-19 9:28 ` [PATCH 22/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_RNG Mark Rutland
2023-09-19 11:24 ` Mark Brown
2023-09-19 9:28 ` [PATCH 23/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_WFXT Mark Rutland
2023-09-19 9:28 ` [PATCH 24/37] arm64: Avoid cpus_have_const_cap() for ARM64_HAS_TLB_RANGE Mark Rutland
2023-09-19 9:28 ` [PATCH 25/37] arm64: Avoid cpus_have_const_cap() for ARM64_MTE Mark Rutland
2023-09-19 9:28 ` [PATCH 26/37] arm64: Avoid cpus_have_const_cap() for ARM64_SSBS Mark Rutland
2023-09-19 9:28 ` [PATCH 27/37] arm64: Avoid cpus_have_const_cap() for ARM64_SPECTRE_V2 Mark Rutland
2023-09-19 9:28 ` Mark Rutland [this message]
2023-09-19 11:27 ` [PATCH 28/37] arm64: Avoid cpus_have_const_cap() for ARM64_{SVE,SME,SME2,FA64} Mark Brown
2023-09-19 9:28 ` [PATCH 29/37] arm64: Avoid cpus_have_const_cap() for ARM64_UNMAP_KERNEL_AT_EL0 Mark Rutland
2023-09-19 9:28 ` [PATCH 30/37] arm64: Avoid cpus_have_const_cap() for ARM64_WORKAROUND_843419 Mark Rutland
2023-09-19 9:28 ` [PATCH 31/37] arm64: Avoid cpus_have_const_cap() for ARM64_WORKAROUND_1542419 Mark Rutland
2023-09-19 9:28 ` [PATCH 32/37] arm64: Avoid cpus_have_const_cap() for ARM64_WORKAROUND_1742098 Mark Rutland
2023-09-19 9:28 ` [PATCH 33/37] arm64: Avoid cpus_have_const_cap() for ARM64_WORKAROUND_2645198 Mark Rutland
2023-09-19 9:28 ` [PATCH 34/37] arm64: Avoid cpus_have_const_cap() for ARM64_WORKAROUND_CAVIUM_23154 Mark Rutland
2023-09-19 9:28 ` [PATCH 35/37] arm64: Avoid cpus_have_const_cap() for ARM64_WORKAROUND_NVIDIA_CARMEL_CNP Mark Rutland
2023-09-19 9:28 ` [PATCH 36/37] arm64: Avoid cpus_have_const_cap() for ARM64_WORKAROUND_REPEAT_TLBI Mark Rutland
2023-09-19 9:28 ` [PATCH 37/37] arm64: Remove cpus_have_const_cap() Mark Rutland
2023-10-03 17:20 ` Kristina Martsenko
2023-10-05 9:35 ` Mark Rutland
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20230919092850.1940729-29-mark.rutland@arm.com \
--to=mark.rutland@arm.com \
--cc=ardb@kernel.org \
--cc=bertrand.marquis@arm.com \
--cc=boris.ostrovsky@oracle.com \
--cc=broonie@kernel.org \
--cc=catalin.marinas@arm.com \
--cc=daniel.lezcano@linaro.org \
--cc=james.morse@arm.com \
--cc=jgross@suse.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=maz@kernel.org \
--cc=oliver.upton@linux.dev \
--cc=pcc@google.com \
--cc=sstabellini@kernel.org \
--cc=suzuki.poulose@arm.com \
--cc=tglx@linutronix.de \
--cc=vladimir.murzin@arm.com \
--cc=will@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).