[PATCH 09/15] arm64: KVM: Simplify HYP init/teardown

Christoffer Dall christoffer.dall at linaro.org
Tue Jun 28 14:31:41 PDT 2016


On Tue, Jun 07, 2016 at 11:58:29AM +0100, Marc Zyngier wrote:
> Now that we only have the "merged page tables" case to deal with,
> there is a bunch of things we can simplify in the HYP code (both
> at init and teardown time).
> 
> Signed-off-by: Marc Zyngier <marc.zyngier at arm.com>
> ---
>  arch/arm64/include/asm/kvm_host.h | 12 ++------
>  arch/arm64/kvm/hyp-init.S         | 61 +++++----------------------------------
>  arch/arm64/kvm/hyp/entry.S        | 19 ------------
>  arch/arm64/kvm/hyp/hyp-entry.S    | 15 ++++++++++
>  arch/arm64/kvm/reset.c            | 11 -------
>  5 files changed, 26 insertions(+), 92 deletions(-)
> 
> diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h
> index 49095fc..88462c3 100644
> --- a/arch/arm64/include/asm/kvm_host.h
> +++ b/arch/arm64/include/asm/kvm_host.h
> @@ -48,7 +48,6 @@
>  int __attribute_const__ kvm_target_cpu(void);
>  int kvm_reset_vcpu(struct kvm_vcpu *vcpu);
>  int kvm_arch_dev_ioctl_check_extension(long ext);
> -unsigned long kvm_hyp_reset_entry(void);
>  void __extended_idmap_trampoline(phys_addr_t boot_pgd, phys_addr_t idmap_start);
>  
>  struct kvm_arch {
> @@ -357,19 +356,14 @@ static inline void __cpu_init_hyp_mode(phys_addr_t boot_pgd_ptr,
>  	 * Call initialization code, and switch to the full blown
>  	 * HYP code.
>  	 */
> -	__kvm_call_hyp((void *)boot_pgd_ptr, pgd_ptr,
> -		       hyp_stack_ptr, vector_ptr);
> +	__kvm_call_hyp((void *)pgd_ptr, hyp_stack_ptr, vector_ptr);
>  }
>  
> +void __kvm_hyp_teardown(void);
>  static inline void __cpu_reset_hyp_mode(phys_addr_t boot_pgd_ptr,
>  					phys_addr_t phys_idmap_start)
>  {
> -	/*
> -	 * Call reset code, and switch back to stub hyp vectors.
> -	 * Uses __kvm_call_hyp() to avoid kaslr's kvm_ksym_ref() translation.
> -	 */
> -	__kvm_call_hyp((void *)kvm_hyp_reset_entry(),
> -		       boot_pgd_ptr, phys_idmap_start);
> +	kvm_call_hyp(__kvm_hyp_teardown, phys_idmap_start);
>  }
>  
>  static inline void kvm_arch_hardware_unsetup(void) {}
> diff --git a/arch/arm64/kvm/hyp-init.S b/arch/arm64/kvm/hyp-init.S
> index a873a6d..6b29d3d 100644
> --- a/arch/arm64/kvm/hyp-init.S
> +++ b/arch/arm64/kvm/hyp-init.S
> @@ -53,10 +53,9 @@ __invalid:
>  	b	.
>  
>  	/*
> -	 * x0: HYP boot pgd
> -	 * x1: HYP pgd
> -	 * x2: HYP stack
> -	 * x3: HYP vectors
> +	 * x0: HYP pgd
> +	 * x1: HYP stack
> +	 * x2: HYP vectors
>  	 */
>  __do_hyp_init:
>  
> @@ -110,71 +109,27 @@ __do_hyp_init:
>  	msr	sctlr_el2, x4
>  	isb
>  
> -	/* Skip the trampoline dance if we merged the boot and runtime PGDs */
> -	cmp	x0, x1
> -	b.eq	merged
> -
> -	/* MMU is now enabled. Get ready for the trampoline dance */
> -	ldr	x4, =TRAMPOLINE_VA
> -	adr	x5, target
> -	bfi	x4, x5, #0, #PAGE_SHIFT
> -	br	x4
> -
> -target: /* We're now in the trampoline code, switch page tables */
> -	msr	ttbr0_el2, x1
> -	isb
> -
> -	/* Invalidate the old TLBs */
> -	tlbi	alle2
> -	dsb	sy
> -
> -merged:
>  	/* Set the stack and new vectors */
> +	kern_hyp_va	x1
> +	mov	sp, x1
>  	kern_hyp_va	x2
> -	mov	sp, x2
> -	kern_hyp_va	x3
> -	msr	vbar_el2, x3
> +	msr	vbar_el2, x2
>  
>  	/* Hello, World! */
>  	eret
>  ENDPROC(__kvm_hyp_init)
>  
>  	/*
> -	 * Reset kvm back to the hyp stub. This is the trampoline dance in
> -	 * reverse. If kvm used an extended idmap, __extended_idmap_trampoline
> -	 * calls this code directly in the idmap. In this case switching to the
> -	 * boot tables is a no-op.
> -	 *
> -	 * x0: HYP boot pgd
> -	 * x1: HYP phys_idmap_start
> +	 * Reset kvm back to the hyp stub.
>  	 */
>  ENTRY(__kvm_hyp_reset)
> -	/* We're in trampoline code in VA, switch back to boot page tables */
> -	msr	ttbr0_el2, x0
> -	isb
> -
> -	/* Ensure the PA branch doesn't find a stale tlb entry or stale code. */
> -	ic	iallu
> -	tlbi	alle2
> -	dsb	sy
> -	isb
> -
> -	/* Branch into PA space */
> -	adr	x0, 1f
> -	bfi	x1, x0, #0, #PAGE_SHIFT
> -	br	x1
> -
>  	/* We're now in idmap, disable MMU */
> -1:	mrs	x0, sctlr_el2
> +	mrs	x0, sctlr_el2
>  	ldr	x1, =SCTLR_ELx_FLAGS
>  	bic	x0, x0, x1		// Clear SCTL_M and etc
>  	msr	sctlr_el2, x0
>  	isb
>  
> -	/* Invalidate the old TLBs */
> -	tlbi	alle2
> -	dsb	sy
> -

why can we get rid of the above two lines now?

>  	/* Install stub vectors */
>  	adr_l	x0, __hyp_stub_vectors
>  	msr	vbar_el2, x0
> diff --git a/arch/arm64/kvm/hyp/entry.S b/arch/arm64/kvm/hyp/entry.S
> index 70254a6..ce9e5e5 100644
> --- a/arch/arm64/kvm/hyp/entry.S
> +++ b/arch/arm64/kvm/hyp/entry.S
> @@ -164,22 +164,3 @@ alternative_endif
>  
>  	eret
>  ENDPROC(__fpsimd_guest_restore)
> -
> -/*
> - * When using the extended idmap, we don't have a trampoline page we can use
> - * while we switch pages tables during __kvm_hyp_reset. Accessing the idmap
> - * directly would be ideal, but if we're using the extended idmap then the
> - * idmap is located above HYP_PAGE_OFFSET, and the address will be masked by
> - * kvm_call_hyp using kern_hyp_va.
> - *
> - * x0: HYP boot pgd
> - * x1: HYP phys_idmap_start
> - */
> -ENTRY(__extended_idmap_trampoline)
> -	mov	x4, x1
> -	adr_l	x3, __kvm_hyp_reset
> -
> -	/* insert __kvm_hyp_reset()s offset into phys_idmap_start */
> -	bfi	x4, x3, #0, #PAGE_SHIFT
> -	br	x4
> -ENDPROC(__extended_idmap_trampoline)
> diff --git a/arch/arm64/kvm/hyp/hyp-entry.S b/arch/arm64/kvm/hyp/hyp-entry.S
> index 2d87f36..f6d9694 100644
> --- a/arch/arm64/kvm/hyp/hyp-entry.S
> +++ b/arch/arm64/kvm/hyp/hyp-entry.S
> @@ -62,6 +62,21 @@ ENTRY(__vhe_hyp_call)
>  	isb
>  	ret
>  ENDPROC(__vhe_hyp_call)
> +
> +/*
> + * Compute the idmap address of __kvm_hyp_reset based on the idmap
> + * start passed as a parameter, and jump there.
> + *
> + * x0: HYP phys_idmap_start
> + */
> +ENTRY(__kvm_hyp_teardown)
> +	mov	x4, x0
> +	adr_l	x3, __kvm_hyp_reset
> +
> +	/* insert __kvm_hyp_reset()s offset into phys_idmap_start */
> +	bfi	x4, x3, #0, #PAGE_SHIFT
> +	br	x4
> +ENDPROC(__kvm_hyp_teardown)
>  	
>  el1_sync:				// Guest trapped into EL2
>  	save_x0_to_x3
> diff --git a/arch/arm64/kvm/reset.c b/arch/arm64/kvm/reset.c
> index d044ca3..deee1b1 100644
> --- a/arch/arm64/kvm/reset.c
> +++ b/arch/arm64/kvm/reset.c
> @@ -132,14 +132,3 @@ int kvm_reset_vcpu(struct kvm_vcpu *vcpu)
>  	/* Reset timer */
>  	return kvm_timer_vcpu_reset(vcpu, cpu_vtimer_irq);
>  }
> -
> -unsigned long kvm_hyp_reset_entry(void)
> -{
> -	/*
> -	 * KVM is running with merged page tables, which don't have the
> -	 * trampoline page mapped. We know the idmap is still mapped,
> -	 * but can't be called into directly. Use
> -	 * __extended_idmap_trampoline to do the call.
> -	 */
> -	return (unsigned long)kvm_ksym_ref(__extended_idmap_trampoline);
> -}
> -- 
> 2.1.4
> 

I'm not sure I understand why we needed the kvm_hyp_reset_entry
indirection before, but the resulting code here looks good to me.

Thanks,
-Christoffer




More information about the linux-arm-kernel mailing list