[PATCH v2 05/25] KVM: arm64: nv: Add sanitising to VNCR-backed FGT sysregs
Joey Gouly
joey.gouly at arm.com
Thu Feb 1 06:38:06 PST 2024
On Tue, Jan 30, 2024 at 08:45:12PM +0000, Marc Zyngier wrote:
> Fine Grained Traps are controlled by a whole bunch of features.
> Each one of them must be checked and the corresponding masks
> computed so that we don't let the guest apply traps it shouldn't
> be using.
>
> This takes care of HFGxTR_EL2, HDFGxTR_EL2, and HAFGRTR_EL2.
^^ and HFGITR_EL2
>
> Signed-off-by: Marc Zyngier <maz at kernel.org>
> ---
> arch/arm64/kvm/nested.c | 128 ++++++++++++++++++++++++++++++++++++++++
> 1 file changed, 128 insertions(+)
>
> diff --git a/arch/arm64/kvm/nested.c b/arch/arm64/kvm/nested.c
> index ee461e630527..cdeef3259193 100644
> --- a/arch/arm64/kvm/nested.c
> +++ b/arch/arm64/kvm/nested.c
> @@ -263,6 +263,134 @@ int kvm_init_nv_sysregs(struct kvm *kvm)
> res1 |= HCR_E2H;
> set_sysreg_masks(kvm, HCR_EL2, res0, res1);
>
> + /* HFG[RW]TR_EL2 */
> + res0 = res1 = 0;
> + if (!(__vcpu_has_feature(&kvm->arch, KVM_ARM_VCPU_PTRAUTH_ADDRESS) &&
> + __vcpu_has_feature(&kvm->arch, KVM_ARM_VCPU_PTRAUTH_ADDRESS)))
> + res0 |= (HFGxTR_EL2_APDAKey | HFGxTR_EL2_APDBKey |
> + HFGxTR_EL2_APGAKey | HFGxTR_EL2_APIAKey |
> + HFGxTR_EL2_APIBKey);
> + if (!kvm_has_feat(kvm, ID_AA64MMFR1_EL1, LO, IMP))
> + res0 |= (HFGxTR_EL2_LORC_EL1 | HFGxTR_EL2_LOREA_EL1 |
> + HFGxTR_EL2_LORID_EL1 | HFGxTR_EL2_LORN_EL1 |
> + HFGxTR_EL2_LORSA_EL1);
> + if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, CSV2, CSV2_2) &&
> + !kvm_has_feat(kvm, ID_AA64PFR1_EL1, CSV2_frac, CSV2_1p2))
> + res0 |= (HFGxTR_EL2_SCXTNUM_EL1 | HFGxTR_EL2_SCXTNUM_EL0);
> + if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, GIC, IMP))
> + res0 |= HFGxTR_EL2_ICC_IGRPENn_EL1;
> + if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, RAS, IMP))
> + res0 |= (HFGxTR_EL2_ERRIDR_EL1 | HFGxTR_EL2_ERRSELR_EL1 |
> + HFGxTR_EL2_ERXFR_EL1 | HFGxTR_EL2_ERXCTLR_EL1 |
> + HFGxTR_EL2_ERXSTATUS_EL1 | HFGxTR_EL2_ERXMISCn_EL1 |
> + HFGxTR_EL2_ERXPFGF_EL1 | HFGxTR_EL2_ERXPFGCTL_EL1 |
> + HFGxTR_EL2_ERXPFGCDN_EL1 | HFGxTR_EL2_ERXADDR_EL1);
> + if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, LS64, LS64_ACCDATA))
> + res0 |= HFGxTR_EL2_nACCDATA_EL1;
> + if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, GCS, IMP))
> + res0 |= (HFGxTR_EL2_nGCS_EL0 | HFGxTR_EL2_nGCS_EL1);
> + if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, SME, IMP))
> + res0 |= (HFGxTR_EL2_nSMPRI_EL1 | HFGxTR_EL2_nTPIDR2_EL0);
> + if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, THE, IMP))
> + res0 |= HFGxTR_EL2_nRCWMASK_EL1;
> + if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, S1PIE, IMP))
> + res0 |= (HFGxTR_EL2_nPIRE0_EL1 | HFGxTR_EL2_nPIR_EL1);
> + if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, S1POE, IMP))
> + res0 |= (HFGxTR_EL2_nPOR_EL0 | HFGxTR_EL2_nPOR_EL1);
> + if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, S2POE, IMP))
> + res0 |= HFGxTR_EL2_nS2POR_EL1;
> + if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, AIE, IMP))
> + res0 |= (HFGxTR_EL2_nMAIR2_EL1 | HFGxTR_EL2_nAMAIR2_EL1);
> + set_sysreg_masks(kvm, HFGRTR_EL2, res0 | __HFGRTR_EL2_RES0, res1);
> + set_sysreg_masks(kvm, HFGWTR_EL2, res0 | __HFGWTR_EL2_RES0, res1);
> +
> + /* HDFG[RW]TR_EL2 */
> + res0 = res1 = 0;
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, DoubleLock, IMP))
> + res0 |= HDFGRTR_EL2_OSDLR_EL1;
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMUVer, IMP))
> + res0 |= (HDFGRTR_EL2_PMEVCNTRn_EL0 | HDFGRTR_EL2_PMEVTYPERn_EL0 |
> + HDFGRTR_EL2_PMCCFILTR_EL0 | HDFGRTR_EL2_PMCCNTR_EL0 |
> + HDFGRTR_EL2_PMCNTEN | HDFGRTR_EL2_PMINTEN |
> + HDFGRTR_EL2_PMOVS | HDFGRTR_EL2_PMSELR_EL0 |
> + HDFGRTR_EL2_PMMIR_EL1 | HDFGRTR_EL2_PMUSERENR_EL0 |
> + HDFGRTR_EL2_PMCEIDn_EL0);
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMSVer, IMP))
> + res0 |= (HDFGRTR_EL2_PMBLIMITR_EL1 | HDFGRTR_EL2_PMBPTR_EL1 |
> + HDFGRTR_EL2_PMBSR_EL1 | HDFGRTR_EL2_PMSCR_EL1 |
> + HDFGRTR_EL2_PMSEVFR_EL1 | HDFGRTR_EL2_PMSFCR_EL1 |
> + HDFGRTR_EL2_PMSICR_EL1 | HDFGRTR_EL2_PMSIDR_EL1 |
> + HDFGRTR_EL2_PMSIRR_EL1 | HDFGRTR_EL2_PMSLATFR_EL1 |
> + HDFGRTR_EL2_PMBIDR_EL1);
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceVer, IMP))
> + res0 |= (HDFGRTR_EL2_TRC | HDFGRTR_EL2_TRCAUTHSTATUS |
> + HDFGRTR_EL2_TRCAUXCTLR | HDFGRTR_EL2_TRCCLAIM |
> + HDFGRTR_EL2_TRCCNTVRn | HDFGRTR_EL2_TRCID |
> + HDFGRTR_EL2_TRCIMSPECn | HDFGRTR_EL2_TRCOSLSR |
> + HDFGRTR_EL2_TRCPRGCTLR | HDFGRTR_EL2_TRCSEQSTR |
> + HDFGRTR_EL2_TRCSSCSRn | HDFGRTR_EL2_TRCSTATR |
> + HDFGRTR_EL2_TRCVICTLR);
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceBuffer, IMP))
> + res0 |= (HDFGRTR_EL2_TRBBASER_EL1 | HDFGRTR_EL2_TRBIDR_EL1 |
> + HDFGRTR_EL2_TRBLIMITR_EL1 | HDFGRTR_EL2_TRBMAR_EL1 |
> + HDFGRTR_EL2_TRBPTR_EL1 | HDFGRTR_EL2_TRBSR_EL1 |
> + HDFGRTR_EL2_TRBTRG_EL1);
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, BRBE, IMP))
> + res0 |= (HDFGRTR_EL2_nBRBIDR | HDFGRTR_EL2_nBRBCTL |
> + HDFGRTR_EL2_nBRBDATA);
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMSVer, V1P2))
> + res0 |= HDFGRTR_EL2_nPMSNEVFR_EL1;
> + set_sysreg_masks(kvm, HDFGRTR_EL2, res0 | HDFGRTR_EL2_RES0, res1);
> +
> + /* Reuse the bits from the read-side and add the write-specific stuff */
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMUVer, IMP))
> + res0 |= (HDFGWTR_EL2_PMCR_EL0 | HDFGWTR_EL2_PMSWINC_EL0);
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceVer, IMP))
> + res0 |= HDFGWTR_EL2_TRCOSLAR;
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceFilt, IMP))
> + res0 |= HDFGWTR_EL2_TRFCR_EL1;
> + set_sysreg_masks(kvm, HFGWTR_EL2, res0 | HDFGWTR_EL2_RES0, res1);
> +
> + /* HFGITR_EL2 */
> + res0 = HFGITR_EL2_RES0;
> + res1 = HFGITR_EL2_RES1;
> + if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, DPB, DPB2))
> + res0 |= HFGITR_EL2_DCCVADP;
> + if (!kvm_has_feat(kvm, ID_AA64MMFR1_EL1, PAN, PAN2))
> + res0 |= (HFGITR_EL2_ATS1E1RP | HFGITR_EL2_ATS1E1WP);
> + if (!kvm_has_feat(kvm, ID_AA64ISAR0_EL1, TLB, OS))
> + res0 |= (HFGITR_EL2_TLBIRVAALE1OS | HFGITR_EL2_TLBIRVALE1OS |
> + HFGITR_EL2_TLBIRVAAE1OS | HFGITR_EL2_TLBIRVAE1OS |
> + HFGITR_EL2_TLBIVAALE1OS | HFGITR_EL2_TLBIVALE1OS |
> + HFGITR_EL2_TLBIVAAE1OS | HFGITR_EL2_TLBIASIDE1OS |
> + HFGITR_EL2_TLBIVAE1OS | HFGITR_EL2_TLBIVMALLE1OS);
> + if (!kvm_has_feat(kvm, ID_AA64ISAR0_EL1, TLB, RANGE))
> + res0 |= (HFGITR_EL2_TLBIRVAALE1 | HFGITR_EL2_TLBIRVALE1 |
> + HFGITR_EL2_TLBIRVAAE1 | HFGITR_EL2_TLBIRVAE1 |
> + HFGITR_EL2_TLBIRVAALE1IS | HFGITR_EL2_TLBIRVALE1IS |
> + HFGITR_EL2_TLBIRVAAE1IS | HFGITR_EL2_TLBIRVAE1IS |
> + HFGITR_EL2_TLBIRVAALE1OS | HFGITR_EL2_TLBIRVALE1OS |
> + HFGITR_EL2_TLBIRVAAE1OS | HFGITR_EL2_TLBIRVAE1OS);
> + if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, SPECRES, IMP))
> + res0 |= (HFGITR_EL2_CFPRCTX | HFGITR_EL2_DVPRCTX |
> + HFGITR_EL2_CPPRCTX);
> + if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, BRBE, IMP))
> + res0 |= (HFGITR_EL2_nBRBINJ | HFGITR_EL2_nBRBIALL);
> + if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, GCS, IMP))
> + res0 |= (HFGITR_EL2_nGCSPUSHM_EL1 | HFGITR_EL2_nGCSSTR_EL1 |
> + HFGITR_EL2_nGCSEPP);
> + if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, SPECRES, COSP_RCTX))
> + res0 |= HFGITR_EL2_COSPRCTX;
> + if (!kvm_has_feat(kvm, ID_AA64ISAR2_EL1, ATS1A, IMP))
> + res0 |= HFGITR_EL2_ATS1E1A;
> + set_sysreg_masks(kvm, HFGITR_EL2, res0, res1);
> +
> + /* HAFGRTR_EL2 - not a lot to see here*/
*very* minor nitpick: add a space before the */
> + res0 = HAFGRTR_EL2_RES0;
> + res1 = HAFGRTR_EL2_RES1;
> + if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, AMU, V1P1))
> + res0 |= ~(res0 | res1);
> + set_sysreg_masks(kvm, HAFGRTR_EL2, res0, res1);
> out:
> mutex_unlock(&kvm->arch.config_lock);
>
Looked through every trap bit, that was tedious! and I just realised HCRX_EL2
is still left to do...
Reviewed-by: Joey Gouly <joey.gouly at arm.com>
Thanks,
Joey
More information about the linux-arm-kernel
mailing list