[PATCH v2 7/7] KVM: arm64: Consolidate initializing the host data's fpsimd_state/sve in pKVM
Fuad Tabba
tabba at google.com
Wed May 22 07:49:07 PDT 2024
Hi Marc,
On Tue, May 21, 2024 at 11:56 PM Marc Zyngier <maz at kernel.org> wrote:
>
> On Tue, 21 May 2024 17:37:20 +0100,
> Fuad Tabba <tabba at google.com> wrote:
> >
> > Now that we have introduced finalize_init_hyp_mode(), lets
> > consolidate the initializing of the host_data fpsimd_state and
> > sve state.
> >
> > Signed-off-by: Fuad Tabba <tabba at google.com>
> > ---
> > arch/arm64/include/asm/kvm_host.h | 10 ++++++++--
> > arch/arm64/kvm/arm.c | 18 ++++++++++++------
> > arch/arm64/kvm/hyp/include/nvhe/pkvm.h | 1 -
> > arch/arm64/kvm/hyp/nvhe/pkvm.c | 11 -----------
> > arch/arm64/kvm/hyp/nvhe/setup.c | 1 -
> > 5 files changed, 20 insertions(+), 21 deletions(-)
> >
> > diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h
> > index 7b3745ef1d73..8a170f314498 100644
> > --- a/arch/arm64/include/asm/kvm_host.h
> > +++ b/arch/arm64/include/asm/kvm_host.h
> > @@ -536,8 +536,14 @@ struct kvm_cpu_context {
> > struct kvm_host_data {
> > struct kvm_cpu_context host_ctxt;
> >
> > - struct user_fpsimd_state *fpsimd_state; /* hyp VA */
> > - struct user_sve_state *sve_state; /* hyp VA */
> > + /*
> > + * All pointers in this union are hyp VA.
> > + * sve_state is only used in pKVM and if system_supports_sve().
> > + */
> > + union {
> > + struct user_fpsimd_state *fpsimd_state;
> > + struct user_sve_state *sve_state;
> > + };
> >
> > /* Ownership of the FP regs */
> > enum {
> > diff --git a/arch/arm64/kvm/arm.c b/arch/arm64/kvm/arm.c
> > index a9b1b0e9c319..a1c7e0ad6951 100644
> > --- a/arch/arm64/kvm/arm.c
> > +++ b/arch/arm64/kvm/arm.c
> > @@ -2445,14 +2445,20 @@ static void finalize_init_hyp_mode(void)
> > {
> > int cpu;
> >
> > - if (!is_protected_kvm_enabled() || !system_supports_sve())
> > - return;
> > -
> > for_each_possible_cpu(cpu) {
> > - struct user_sve_state *sve_state;
> > + if (system_supports_sve() && is_protected_kvm_enabled()) {
> > + struct user_sve_state *sve_state;
> >
> > - sve_state = per_cpu_ptr_nvhe_sym(kvm_host_data, cpu)->sve_state;
> > - per_cpu_ptr_nvhe_sym(kvm_host_data, cpu)->sve_state = kern_hyp_va(sve_state);
> > + sve_state = per_cpu_ptr_nvhe_sym(kvm_host_data, cpu)->sve_state;
> > + per_cpu_ptr_nvhe_sym(kvm_host_data, cpu)->sve_state =
> > + kern_hyp_va(sve_state);
> > + } else {
> > + struct user_fpsimd_state *fpsimd_state;
> > +
> > + fpsimd_state = &per_cpu_ptr_nvhe_sym(kvm_host_data, cpu)->host_ctxt.fp_regs;
> > + per_cpu_ptr_nvhe_sym(kvm_host_data, cpu)->fpsimd_state =
> > + kern_hyp_va(fpsimd_state);
> > + }
>
> nit: SVE support and protected state do not change on a per CPU basis,
> so checking for these inside the loop is pretty counter intuitive.
I'll fix this.
Thanks for all the reviews!
/fuad
> Thanks,
>
> M.
>
> --
> Without deviation from the norm, progress is not possible.
More information about the linux-arm-kernel
mailing list