[PATCH v3 2/3] KVM/arm/arm64: enable enhanced armv7 fp/simd lazy switch

Christoffer Dall christoffer.dall at linaro.org
Fri Nov 6 03:37:18 PST 2015


On Thu, Nov 05, 2015 at 04:23:41PM -0800, Mario Smarduch wrote:
> 
> 
> On 11/5/2015 6:48 AM, Christoffer Dall wrote:
> > On Fri, Oct 30, 2015 at 02:56:32PM -0700, Mario Smarduch wrote:
> >> This patch tracks vfp/simd hardware state with a vcpu lazy flag. vCPU lazy 
> >> flag is set on guest access and traps to vfp/simd hardware switch handler. On 
> >> vm-enter if lazy flag is set skip trap enable and save host fpexc. On 
> >> vm-exit if flag is set skip hardware context switch and return to host with 
> >> guest context. In vcpu_put check if vcpu lazy flag is set, and execute a 
> >> hardware context switch to restore host.
> >>
> >> Also some arm64 field and empty function are added to compile for arm64.
> >>
> >> Signed-off-by: Mario Smarduch <m.smarduch at samsung.com>
> >> ---
> >>  arch/arm/include/asm/kvm_host.h   |  1 +
> >>  arch/arm/kvm/arm.c                |  6 ++++
> >>  arch/arm/kvm/interrupts.S         | 60 ++++++++++++++++++++++++++++-----------
> >>  arch/arm/kvm/interrupts_head.S    | 14 +++++----
> >>  arch/arm64/include/asm/kvm_host.h |  4 +++
> >>  5 files changed, 63 insertions(+), 22 deletions(-)
> >>
> >> diff --git a/arch/arm/include/asm/kvm_host.h b/arch/arm/include/asm/kvm_host.h
> >> index f1bf551..a9e86e0 100644
> >> --- a/arch/arm/include/asm/kvm_host.h
> >> +++ b/arch/arm/include/asm/kvm_host.h
> >> @@ -227,6 +227,7 @@ int kvm_perf_teardown(void);
> >>  void kvm_mmu_wp_memory_region(struct kvm *kvm, int slot);
> >>  
> >>  struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr);
> >> +void kvm_restore_host_vfp_state(struct kvm_vcpu *);
> >>  
> >>  static inline void kvm_arch_hardware_disable(void) {}
> >>  static inline void kvm_arch_hardware_unsetup(void) {}
> >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c
> >> index dc017ad..11a56fe 100644
> >> --- a/arch/arm/kvm/arm.c
> >> +++ b/arch/arm/kvm/arm.c
> >> @@ -296,6 +296,12 @@ void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
> >>  void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
> >>  {
> >>  	/*
> >> +	 * If fp/simd registers are dirty save guest, restore host before
> > 
> > If the fp/simd registers are dirty, then restore the host state before
> I'd drop 'releasing the cpu', the vcpu thread may be returning to
> user mode.
> > 
> >> +	 * releasing the cpu.
> >> +	 */
> >> +	if (vcpu->arch.vfp_dirty)
> >> +		kvm_restore_host_vfp_state(vcpu);
> >> +	/*
> >>  	 * The arch-generic KVM code expects the cpu field of a vcpu to be -1
> >>  	 * if the vcpu is no longer assigned to a cpu.  This is used for the
> >>  	 * optimized make_all_cpus_request path.
> >> diff --git a/arch/arm/kvm/interrupts.S b/arch/arm/kvm/interrupts.S
> >> index 900ef6d..ca25314 100644
> >> --- a/arch/arm/kvm/interrupts.S
> >> +++ b/arch/arm/kvm/interrupts.S
> >> @@ -28,6 +28,32 @@
> >>  #include "interrupts_head.S"
> >>  
> >>  	.text
> >> +/**
> >> + * void kvm_restore_host_vfp_state(struct vcpu *vcpu) - Executes lazy
> > 
> > nit: Can you move the multi-line description of the function into a
> > separate paragraph?
> Sure.
> > 
> >> + * 	fp/simd switch, saves the guest, restores host. Called from host
> >> + *	mode, placed outside of hyp region start/end.
> > 
> > Put the description in a separate paragraph and get rid of the "executes
> > lazy fp/simd swithch" part, that doesn't help understanding.  Just say
> > that this funciton restores the host state.
> Sure.
> > 
> >> + */
> >> +ENTRY(kvm_restore_host_vfp_state)
> >> +#ifdef CONFIG_VFPv3
> >> +	push    {r4-r7}
> >> +
> >> +	add     r7, vcpu, #VCPU_VFP_GUEST
> >> +	store_vfp_state r7
> >> +
> >> +	add     r7, vcpu, #VCPU_VFP_HOST
> >> +	ldr     r7, [r7]
> >> +	restore_vfp_state r7
> >> +
> >> +	ldr     r3, [vcpu, #VCPU_VFP_HOST_FPEXC]
> >> +	VFPFMXR FPEXC, r3
> >> +
> >> +	mov     r3, #0
> >> +	strb    r3, [vcpu, #VCPU_VFP_DIRTY]
> >> +
> >> +	pop     {r4-r7}
> >> +#endif
> >> +	bx      lr
> >> +ENDPROC(kvm_restore_host_vfp_state)
> >>  
> >>  __kvm_hyp_code_start:
> >>  	.globl __kvm_hyp_code_start
> >> @@ -119,11 +145,16 @@ ENTRY(__kvm_vcpu_run)
> >>  	@ If the host kernel has not been configured with VFPv3 support,
> >>  	@ then it is safer if we deny guests from using it as well.
> >>  #ifdef CONFIG_VFPv3
> >> -	@ Set FPEXC_EN so the guest doesn't trap floating point instructions
> >> +	@ fp/simd register file has already been accessed, so skip host fpexc
> >> +	@ save and access trap enable.
> >> +	vfp_inlazy_mode r7, skip_guest_vfp_trap
> > 
> > So, why do we need to touch this register at all on every CPU exit?
> > 
> > Is it not true that we can only be in one of two state:
> >  1) The register file is not dirty (not touched by the guest) and we
> >     should trap
> >  2) The register file is dirty, and we should not trap to EL2?
> > 
> > Only in the first case do we need to set the FPEXC, and couldn't we just
> > do that on vcpu_load and git rid of all this?  (except HCPTR_TCP which
> > we still need to adjust).
> 
> I'm trying to think what happens if you're preempted after you saved
> the FPEXC and set the FPEXC_EN bit in kvm_arch_vcpu_load(). Could some
> thread pick up a bad FPEXC? May be possible to undo in the vcpu_put().

If you're preempted, vcpu_put will be called.  See kvm_preempt_ops in
virt/kvm/kvm_main.c.

> 
> > 
> >> +
> >>  	VFPFMRX r2, FPEXC		@ VMRS
> >> -	push	{r2}
> >> +	str     r2, [vcpu, #VCPU_VFP_HOST_FPEXC]
> >>  	orr	r2, r2, #FPEXC_EN
> >>  	VFPFMXR FPEXC, r2		@ VMSR
> >> +	set_hcptr vmentry, (HCPTR_TCP(10) | HCPTR_TCP(11))
> >> +skip_guest_vfp_trap:
> >>  #endif
> >>  
> >>  	@ Configure Hyp-role
> >> @@ -131,7 +162,7 @@ ENTRY(__kvm_vcpu_run)
> >>  
> >>  	@ Trap coprocessor CRx accesses
> >>  	set_hstr vmentry
> >> -	set_hcptr vmentry, (HCPTR_TTA | HCPTR_TCP(10) | HCPTR_TCP(11))
> >> +	set_hcptr vmentry, (HCPTR_TTA)
> > 
> > based on the above I think you can rework this to set the mask based on
> > the dirty flag and only hit the HCPTR once.
> 
> Not sure how to do this, tracing always needs to be enabled, and it's
> independent of FP dirty state.

here, you do:

	ldr r4, HCPTR_TTA
	vfp_skip_if_dirty skip_vfp_trap
	orr r4, r4, #(HCPTR_TCP(10) | HCPTR_TCP(11))
skip_vfp_trap:
	set_hcptr vmentry, r4

if that works with the necessary rework of set_hcptr to take a register,
if the orr can be encoded propertly etc.  Maybe it's not worth it, it
just feels weird to touch this registers twice.  Perhaps the nicer fix
is to just rename/refactor set_hcptr to be two functions, set_hcptr_bits
and clear_hcptr_bits.

Thanks,
-Christoffer



More information about the linux-arm-kernel mailing list