[PATCH v4 13/40] KVM: arm64: Introduce VHE-specific kvm_vcpu_run

Christoffer Dall christoffer.dall at linaro.org
Thu Feb 22 01:16:48 PST 2018


On Wed, Feb 21, 2018 at 07:18:32PM +0100, Andrew Jones wrote:
> On Wed, Feb 21, 2018 at 06:43:00PM +0100, Andrew Jones wrote:
> > On Thu, Feb 15, 2018 at 10:03:05PM +0100, Christoffer Dall wrote:
> > > So far this is mostly (see below) a copy of the legacy non-VHE switch
> > > function, but we will start reworking these functions in separate
> > > directions to work on VHE and non-VHE in the most optimal way in later
> > > patches.
> > > 
> > > The only difference after this patch between the VHE and non-VHE run
> > > functions is that we omit the branch-predictor variant-2 hardening for
> > > QC Falkor CPUs, because this workaround is specific to a series of
> > > non-VHE ARMv8.0 CPUs.
> > > 
> > > Reviewed-by: Marc Zyngier <marc.zyngier at arm.com>
> > > Signed-off-by: Christoffer Dall <christoffer.dall at linaro.org>
> > > ---
> > > 
> > > Notes:
> > >     Changes since v3:
> > >      - Added BUG() to 32-bit ARM VHE run function
> > >      - Omitted QC Falkor BP Hardening functionality from VHE-specific
> > >        function
> > >     
> > >     Changes since v2:
> > >      - Reworded commit message
> > >     
> > >     Changes since v1:
> > >      - Rename kvm_vcpu_run to kvm_vcpu_run_vhe and rename __kvm_vcpu_run to
> > >        __kvm_vcpu_run_nvhe
> > >      - Removed stray whitespace line
> > > 
> > >  arch/arm/include/asm/kvm_asm.h   |  5 ++-
> > >  arch/arm/kvm/hyp/switch.c        |  2 +-
> > >  arch/arm64/include/asm/kvm_asm.h |  4 ++-
> > >  arch/arm64/kvm/hyp/switch.c      | 66 +++++++++++++++++++++++++++++++++++++++-
> > >  virt/kvm/arm/arm.c               |  5 ++-
> > >  5 files changed, 77 insertions(+), 5 deletions(-)
> > > 
> > 
> > ...
> > 
> > > diff --git a/virt/kvm/arm/arm.c b/virt/kvm/arm/arm.c
> > > index 2062d9357971..5bd879c78951 100644
> > > --- a/virt/kvm/arm/arm.c
> > > +++ b/virt/kvm/arm/arm.c
> > > @@ -736,7 +736,10 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run)
> > >  		if (has_vhe())
> > >  			kvm_arm_vhe_guest_enter();
> > >  
> > > -		ret = kvm_call_hyp(__kvm_vcpu_run, vcpu);
> > > +		if (has_vhe())
> > > +			ret = kvm_vcpu_run_vhe(vcpu);
> > > +		else
> > > +			ret = kvm_call_hyp(__kvm_vcpu_run_nvhe, vcpu);
> > >  
> > >  		if (has_vhe())
> > >  			kvm_arm_vhe_guest_exit();
> > 
> > We can combine these has_vhe()'s
> > 
> >  if (has_vhe()) {
> >     kvm_arm_vhe_guest_enter();
> >     ret = kvm_vcpu_run_vhe(vcpu);
> >     kvm_arm_vhe_guest_exit();
> >  } else
> >     ret = kvm_call_hyp(__kvm_vcpu_run_nvhe, vcpu);
> 
> Maybe even do a cleanup patch that removes
> kvm_arm_vhe_guest_enter/exit by putting the daif
> masking/restoring directly into kvm_vcpu_run_vhe()?
> 
Yes, indeed.  This is a blind rebasing result on my part.

Thanks,
-Christoffer



More information about the linux-arm-kernel mailing list