[PATCH v4 06/49] KVM: arm64: GICv3: Detect and work around the lack of ICV_DIR_EL1 trapping
Marc Zyngier
maz at kernel.org
Tue Nov 25 05:48:04 PST 2025
On Tue, 25 Nov 2025 11:26:10 +0000,
Suzuki K Poulose <suzuki.poulose at arm.com> wrote:
>
> On 20/11/2025 17:24, Marc Zyngier wrote:
> > A long time ago, an unsuspecting architect forgot to add a trap
> > bit for ICV_DIR_EL1 in ICH_HCR_EL2. Which was unfortunate, but
> > what's a bit of spec between friends? Thankfully, this was fixed
> > in a later revision, and ARM "deprecates" the lack of trapping
> > ability.
> >
> > Unfortuantely, a few (billion) CPUs went out with that defect,
> > anything ARMv8.0 from ARM, give or take. And on these CPUs,
> > you can't trap DIR on its own, full stop.
> >
> > As the next best thing, we can trap everything in the common group,
> > which is a tad expensive, but hey ho, that's what you get. You can
> > otherwise recycle the HW in the neaby bin.
> >
> > Tested-by: Fuad Tabba <tabba at google.com>
> > Signed-off-by: Marc Zyngier <maz at kernel.org>
> > ---
> > arch/arm64/include/asm/virt.h | 7 ++++-
> > arch/arm64/kernel/cpufeature.c | 52 ++++++++++++++++++++++++++++++++++
> > arch/arm64/kernel/hyp-stub.S | 5 ++++
> > arch/arm64/kvm/vgic/vgic-v3.c | 3 ++
> > arch/arm64/tools/cpucaps | 1 +
> > 5 files changed, 67 insertions(+), 1 deletion(-)
> >
> > diff --git a/arch/arm64/include/asm/virt.h b/arch/arm64/include/asm/virt.h
> > index aa280f356b96a..8eb63d3294974 100644
> > --- a/arch/arm64/include/asm/virt.h
> > +++ b/arch/arm64/include/asm/virt.h
> > @@ -40,8 +40,13 @@
> > */
> > #define HVC_FINALISE_EL2 3
> > +/*
> > + * HVC_GET_ICH_VTR_EL2 - Retrieve the ICH_VTR_EL2 value
> > + */
> > +#define HVC_GET_ICH_VTR_EL2 4
> > +
> > /* Max number of HYP stub hypercalls */
> > -#define HVC_STUB_HCALL_NR 4
> > +#define HVC_STUB_HCALL_NR 5
> > /* Error returned when an invalid stub number is passed into x0
> > */
> > #define HVC_STUB_ERR 0xbadca11
> > diff --git a/arch/arm64/kernel/cpufeature.c b/arch/arm64/kernel/cpufeature.c
> > index 5ed401ff79e3e..5de51cb1b8fe2 100644
> > --- a/arch/arm64/kernel/cpufeature.c
> > +++ b/arch/arm64/kernel/cpufeature.c
> > @@ -2303,6 +2303,49 @@ static bool has_gic_prio_relaxed_sync(const struct arm64_cpu_capabilities *entry
> > }
> > #endif
> > +static bool can_trap_icv_dir_el1(const struct
> > arm64_cpu_capabilities *entry,
> > + int scope)
> > +{
> > + static const struct midr_range has_vgic_v3[] = {
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M1_ICESTORM),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M1_FIRESTORM),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M1_ICESTORM_PRO),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M1_FIRESTORM_PRO),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M1_ICESTORM_MAX),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M1_FIRESTORM_MAX),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M2_BLIZZARD),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M2_AVALANCHE),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M2_BLIZZARD_PRO),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M2_AVALANCHE_PRO),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M2_BLIZZARD_MAX),
> > + MIDR_ALL_VERSIONS(MIDR_APPLE_M2_AVALANCHE_MAX),
> > + {},
> > + };
> > + struct arm_smccc_res res = {};
> > +
> > + BUILD_BUG_ON(ARM64_HAS_ICH_HCR_EL2_TDIR <= ARM64_HAS_GICV3_CPUIF);
> > + BUILD_BUG_ON(ARM64_HAS_ICH_HCR_EL2_TDIR <= ARM64_HAS_GICV5_LEGACY);
> > + if (!cpus_have_cap(ARM64_HAS_GICV3_CPUIF) &&
> > + !is_midr_in_range_list(has_vgic_v3))
> > + return false;
> > +
> > + if (!is_hyp_mode_available())
> > + return false;
> > +
> > + if (cpus_have_cap(ARM64_HAS_GICV5_LEGACY))
> > + return true;
> > +
> > + if (is_kernel_in_hyp_mode())
> > + res.a1 = read_sysreg_s(SYS_ICH_VTR_EL2);
> > + else
> > + arm_smccc_1_1_hvc(HVC_GET_ICH_VTR_EL2, &res);
>
> We are reading the register on the current CPU and this capability,
> being a SYSTEM_FEATURE, relies on the "probing CPU". If there CPUs
> with differing values (which I don't think is practical, but hey,
> never say..). This is would better be a
> ARM64_CPUCAP_EARLY_LOCAL_CPU_FEATURE, which would run through all
> boot CPUs and would set the capability when it matches.
While I agree that SYSTEM_FEATURE is most probably the wrong thing, I
can't help but notice that
- ARM64_HAS_GICV3_CPUIF,
- ARM64_HAS_GIC_PRIO_MASKING
- ARM64_HAS_GIC_PRIO_RELAXED_SYNC
are all ARM64_CPUCAP_STRICT_BOOT_CPU_FEATURE.
On the other ARM64_HAS_GICV5_LEGACY is ARM64_CPUCAP_EARLY_LOCAL_CPU_FEATURE.
Given that ARM64_HAS_ICH_HCR_EL2_TDIR is dependent on both
ARM64_HAS_GICV3_CPUIF and ARM64_HAS_GICV5_LEGACY, shouldn't these two
(and their dependencies) be aligned to have the same behaviour?
Thanks,
M.
--
Without deviation from the norm, progress is not possible.
More information about the linux-arm-kernel
mailing list