[PATCH -next v17 18/20] riscv: KVM: Add vector lazy save/restore support
Anup Patel
anup at brainfault.org
Mon Mar 27 10:36:23 PDT 2023
On Mon, Mar 27, 2023 at 10:21 PM Andy Chiu <andy.chiu at sifive.com> wrote:
>
> From: Vincent Chen <vincent.chen at sifive.com>
>
> This patch adds vector context save/restore for guest VCPUs. To reduce the
> impact on KVM performance, the implementation imitates the FP context
> switch mechanism to lazily store and restore the vector context only when
> the kernel enters/exits the in-kernel run loop and not during the KVM
> world switch.
>
> Signed-off-by: Vincent Chen <vincent.chen at sifive.com>
> Signed-off-by: Greentime Hu <greentime.hu at sifive.com>
> Signed-off-by: Andy Chiu <andy.chiu at sifive.com>
Looks good to me.
Reviewed-by: Anup Patel <anup at brainfault.org>
Acked-by: Anup Patel <anup at brainfault.org>
Regards,
Anup
> ---
> arch/riscv/include/asm/kvm_host.h | 2 +
> arch/riscv/include/asm/kvm_vcpu_vector.h | 82 ++++++++++
> arch/riscv/include/uapi/asm/kvm.h | 7 +
> arch/riscv/kvm/Makefile | 1 +
> arch/riscv/kvm/vcpu.c | 22 +++
> arch/riscv/kvm/vcpu_vector.c | 186 +++++++++++++++++++++++
> 6 files changed, 300 insertions(+)
> create mode 100644 arch/riscv/include/asm/kvm_vcpu_vector.h
> create mode 100644 arch/riscv/kvm/vcpu_vector.c
>
> diff --git a/arch/riscv/include/asm/kvm_host.h b/arch/riscv/include/asm/kvm_host.h
> index cc7da66ee0c0..7e7e23272d32 100644
> --- a/arch/riscv/include/asm/kvm_host.h
> +++ b/arch/riscv/include/asm/kvm_host.h
> @@ -14,6 +14,7 @@
> #include <linux/kvm_types.h>
> #include <linux/spinlock.h>
> #include <asm/hwcap.h>
> +#include <asm/ptrace.h>
> #include <asm/kvm_vcpu_fp.h>
> #include <asm/kvm_vcpu_insn.h>
> #include <asm/kvm_vcpu_sbi.h>
> @@ -141,6 +142,7 @@ struct kvm_cpu_context {
> unsigned long sstatus;
> unsigned long hstatus;
> union __riscv_fp_state fp;
> + struct __riscv_v_ext_state vector;
> };
>
> struct kvm_vcpu_csr {
> diff --git a/arch/riscv/include/asm/kvm_vcpu_vector.h b/arch/riscv/include/asm/kvm_vcpu_vector.h
> new file mode 100644
> index 000000000000..ff994fdd6d0d
> --- /dev/null
> +++ b/arch/riscv/include/asm/kvm_vcpu_vector.h
> @@ -0,0 +1,82 @@
> +/* SPDX-License-Identifier: GPL-2.0-only */
> +/*
> + * Copyright (C) 2022 SiFive
> + *
> + * Authors:
> + * Vincent Chen <vincent.chen at sifive.com>
> + * Greentime Hu <greentime.hu at sifive.com>
> + */
> +
> +#ifndef __KVM_VCPU_RISCV_VECTOR_H
> +#define __KVM_VCPU_RISCV_VECTOR_H
> +
> +#include <linux/types.h>
> +
> +#ifdef CONFIG_RISCV_ISA_V
> +#include <asm/vector.h>
> +#include <asm/kvm_host.h>
> +
> +static __always_inline void __kvm_riscv_vector_save(struct kvm_cpu_context *context)
> +{
> + __riscv_v_vstate_save(&context->vector, context->vector.datap);
> +}
> +
> +static __always_inline void __kvm_riscv_vector_restore(struct kvm_cpu_context *context)
> +{
> + __riscv_v_vstate_restore(&context->vector, context->vector.datap);
> +}
> +
> +void kvm_riscv_vcpu_vector_reset(struct kvm_vcpu *vcpu);
> +void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx,
> + unsigned long *isa);
> +void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx,
> + unsigned long *isa);
> +void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx);
> +void kvm_riscv_vcpu_host_vector_restore(struct kvm_cpu_context *cntx);
> +int kvm_riscv_vcpu_alloc_vector_context(struct kvm_vcpu *vcpu,
> + struct kvm_cpu_context *cntx);
> +void kvm_riscv_vcpu_free_vector_context(struct kvm_vcpu *vcpu);
> +#else
> +
> +struct kvm_cpu_context;
> +
> +static inline void kvm_riscv_vcpu_vector_reset(struct kvm_vcpu *vcpu)
> +{
> +}
> +
> +static inline void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx,
> + unsigned long *isa)
> +{
> +}
> +
> +static inline void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx,
> + unsigned long *isa)
> +{
> +}
> +
> +static inline void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx)
> +{
> +}
> +
> +static inline void kvm_riscv_vcpu_host_vector_restore(struct kvm_cpu_context *cntx)
> +{
> +}
> +
> +static inline int kvm_riscv_vcpu_alloc_vector_context(struct kvm_vcpu *vcpu,
> + struct kvm_cpu_context *cntx)
> +{
> + return 0;
> +}
> +
> +static inline void kvm_riscv_vcpu_free_vector_context(struct kvm_vcpu *vcpu)
> +{
> +}
> +#endif
> +
> +int kvm_riscv_vcpu_get_reg_vector(struct kvm_vcpu *vcpu,
> + const struct kvm_one_reg *reg,
> + unsigned long rtype);
> +int kvm_riscv_vcpu_set_reg_vector(struct kvm_vcpu *vcpu,
> + const struct kvm_one_reg *reg,
> + unsigned long rtype);
> +#endif
> diff --git a/arch/riscv/include/uapi/asm/kvm.h b/arch/riscv/include/uapi/asm/kvm.h
> index d562dcb929ea..0955f9460447 100644
> --- a/arch/riscv/include/uapi/asm/kvm.h
> +++ b/arch/riscv/include/uapi/asm/kvm.h
> @@ -155,6 +155,13 @@ enum KVM_RISCV_ISA_EXT_ID {
> /* ISA Extension registers are mapped as type 7 */
> #define KVM_REG_RISCV_ISA_EXT (0x07 << KVM_REG_RISCV_TYPE_SHIFT)
>
> +/* V extension registers are mapped as type 8 */
> +#define KVM_REG_RISCV_VECTOR (0x08 << KVM_REG_RISCV_TYPE_SHIFT)
> +#define KVM_REG_RISCV_VECTOR_CSR_REG(name) \
> + (offsetof(struct __riscv_v_ext_state, name) / sizeof(unsigned long))
> +#define KVM_REG_RISCV_VECTOR_REG(n) \
> + ((n) + sizeof(struct __riscv_v_ext_state) / sizeof(unsigned long))
> +
> #endif
>
> #endif /* __LINUX_KVM_RISCV_H */
> diff --git a/arch/riscv/kvm/Makefile b/arch/riscv/kvm/Makefile
> index 278e97c06e0a..f29854333cf2 100644
> --- a/arch/riscv/kvm/Makefile
> +++ b/arch/riscv/kvm/Makefile
> @@ -17,6 +17,7 @@ kvm-y += mmu.o
> kvm-y += vcpu.o
> kvm-y += vcpu_exit.o
> kvm-y += vcpu_fp.o
> +kvm-y += vcpu_vector.o
> kvm-y += vcpu_insn.o
> kvm-y += vcpu_switch.o
> kvm-y += vcpu_sbi.o
> diff --git a/arch/riscv/kvm/vcpu.c b/arch/riscv/kvm/vcpu.c
> index bfdd5b73d462..c495ae1a8091 100644
> --- a/arch/riscv/kvm/vcpu.c
> +++ b/arch/riscv/kvm/vcpu.c
> @@ -22,6 +22,8 @@
> #include <asm/cacheflush.h>
> #include <asm/hwcap.h>
> #include <asm/sbi.h>
> +#include <asm/vector.h>
> +#include <asm/kvm_vcpu_vector.h>
>
> const struct _kvm_stats_desc kvm_vcpu_stats_desc[] = {
> KVM_GENERIC_VCPU_STATS(),
> @@ -135,6 +137,8 @@ static void kvm_riscv_reset_vcpu(struct kvm_vcpu *vcpu)
>
> kvm_riscv_vcpu_fp_reset(vcpu);
>
> + kvm_riscv_vcpu_vector_reset(vcpu);
> +
> kvm_riscv_vcpu_timer_reset(vcpu);
>
> WRITE_ONCE(vcpu->arch.irqs_pending, 0);
> @@ -192,6 +196,9 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
> cntx->hstatus |= HSTATUS_SPVP;
> cntx->hstatus |= HSTATUS_SPV;
>
> + if (kvm_riscv_vcpu_alloc_vector_context(vcpu, cntx))
> + return -ENOMEM;
> +
> /* By default, make CY, TM, and IR counters accessible in VU mode */
> reset_csr->scounteren = 0x7;
>
> @@ -227,6 +234,9 @@ void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
>
> /* Free unused pages pre-allocated for G-stage page table mappings */
> kvm_mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
> +
> + /* Free vector context space for host and guest kernel */
> + kvm_riscv_vcpu_free_vector_context(vcpu);
> }
>
> int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
> @@ -610,6 +620,9 @@ static int kvm_riscv_vcpu_set_reg(struct kvm_vcpu *vcpu,
> KVM_REG_RISCV_FP_D);
> case KVM_REG_RISCV_ISA_EXT:
> return kvm_riscv_vcpu_set_reg_isa_ext(vcpu, reg);
> + case KVM_REG_RISCV_VECTOR:
> + return kvm_riscv_vcpu_set_reg_vector(vcpu, reg,
> + KVM_REG_RISCV_VECTOR);
> default:
> break;
> }
> @@ -637,6 +650,9 @@ static int kvm_riscv_vcpu_get_reg(struct kvm_vcpu *vcpu,
> KVM_REG_RISCV_FP_D);
> case KVM_REG_RISCV_ISA_EXT:
> return kvm_riscv_vcpu_get_reg_isa_ext(vcpu, reg);
> + case KVM_REG_RISCV_VECTOR:
> + return kvm_riscv_vcpu_get_reg_vector(vcpu, reg,
> + KVM_REG_RISCV_VECTOR);
> default:
> break;
> }
> @@ -906,6 +922,9 @@ void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
> kvm_riscv_vcpu_host_fp_save(&vcpu->arch.host_context);
> kvm_riscv_vcpu_guest_fp_restore(&vcpu->arch.guest_context,
> vcpu->arch.isa);
> + kvm_riscv_vcpu_host_vector_save(&vcpu->arch.host_context);
> + kvm_riscv_vcpu_guest_vector_restore(&vcpu->arch.guest_context,
> + vcpu->arch.isa);
>
> vcpu->cpu = cpu;
> }
> @@ -921,6 +940,9 @@ void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
> kvm_riscv_vcpu_host_fp_restore(&vcpu->arch.host_context);
>
> kvm_riscv_vcpu_timer_save(vcpu);
> + kvm_riscv_vcpu_guest_vector_save(&vcpu->arch.guest_context,
> + vcpu->arch.isa);
> + kvm_riscv_vcpu_host_vector_restore(&vcpu->arch.host_context);
>
> csr->vsstatus = csr_read(CSR_VSSTATUS);
> csr->vsie = csr_read(CSR_VSIE);
> diff --git a/arch/riscv/kvm/vcpu_vector.c b/arch/riscv/kvm/vcpu_vector.c
> new file mode 100644
> index 000000000000..edd2eecbddc2
> --- /dev/null
> +++ b/arch/riscv/kvm/vcpu_vector.c
> @@ -0,0 +1,186 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2022 SiFive
> + *
> + * Authors:
> + * Vincent Chen <vincent.chen at sifive.com>
> + * Greentime Hu <greentime.hu at sifive.com>
> + */
> +
> +#include <linux/errno.h>
> +#include <linux/err.h>
> +#include <linux/kvm_host.h>
> +#include <linux/uaccess.h>
> +#include <asm/hwcap.h>
> +#include <asm/kvm_vcpu_vector.h>
> +#include <asm/vector.h>
> +
> +#ifdef CONFIG_RISCV_ISA_V
> +void kvm_riscv_vcpu_vector_reset(struct kvm_vcpu *vcpu)
> +{
> + unsigned long *isa = vcpu->arch.isa;
> + struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
> +
> + cntx->sstatus &= ~SR_VS;
> + if (riscv_isa_extension_available(isa, v)) {
> + cntx->sstatus |= SR_VS_INITIAL;
> + WARN_ON(!cntx->vector.datap);
> + memset(cntx->vector.datap, 0, riscv_v_vsize);
> + } else {
> + cntx->sstatus |= SR_VS_OFF;
> + }
> +}
> +
> +static void kvm_riscv_vcpu_vector_clean(struct kvm_cpu_context *cntx)
> +{
> + cntx->sstatus &= ~SR_VS;
> + cntx->sstatus |= SR_VS_CLEAN;
> +}
> +
> +void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx,
> + unsigned long *isa)
> +{
> + if ((cntx->sstatus & SR_VS) == SR_VS_DIRTY) {
> + if (riscv_isa_extension_available(isa, v))
> + __kvm_riscv_vector_save(cntx);
> + kvm_riscv_vcpu_vector_clean(cntx);
> + }
> +}
> +
> +void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx,
> + unsigned long *isa)
> +{
> + if ((cntx->sstatus & SR_VS) != SR_VS_OFF) {
> + if (riscv_isa_extension_available(isa, v))
> + __kvm_riscv_vector_restore(cntx);
> + kvm_riscv_vcpu_vector_clean(cntx);
> + }
> +}
> +
> +void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx)
> +{
> + /* No need to check host sstatus as it can be modified outside */
> + if (riscv_isa_extension_available(NULL, v))
> + __kvm_riscv_vector_save(cntx);
> +}
> +
> +void kvm_riscv_vcpu_host_vector_restore(struct kvm_cpu_context *cntx)
> +{
> + if (riscv_isa_extension_available(NULL, v))
> + __kvm_riscv_vector_restore(cntx);
> +}
> +
> +int kvm_riscv_vcpu_alloc_vector_context(struct kvm_vcpu *vcpu,
> + struct kvm_cpu_context *cntx)
> +{
> + cntx->vector.datap = kmalloc(riscv_v_vsize, GFP_KERNEL);
> + if (!cntx->vector.datap)
> + return -ENOMEM;
> +
> + vcpu->arch.host_context.vector.datap = kzalloc(riscv_v_vsize, GFP_KERNEL);
> + if (!vcpu->arch.host_context.vector.datap)
> + return -ENOMEM;
> +
> + return 0;
> +}
> +
> +void kvm_riscv_vcpu_free_vector_context(struct kvm_vcpu *vcpu)
> +{
> + kfree(vcpu->arch.guest_reset_context.vector.datap);
> + kfree(vcpu->arch.host_context.vector.datap);
> +}
> +#endif
> +
> +static void *kvm_riscv_vcpu_vreg_addr(struct kvm_vcpu *vcpu,
> + unsigned long reg_num,
> + size_t reg_size)
> +{
> + struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
> + void *reg_val;
> + size_t vlenb = riscv_v_vsize / 32;
> +
> + if (reg_num < KVM_REG_RISCV_VECTOR_REG(0)) {
> + if (reg_size != sizeof(unsigned long))
> + return NULL;
> + switch (reg_num) {
> + case KVM_REG_RISCV_VECTOR_CSR_REG(vstart):
> + reg_val = &cntx->vector.vstart;
> + break;
> + case KVM_REG_RISCV_VECTOR_CSR_REG(vl):
> + reg_val = &cntx->vector.vl;
> + break;
> + case KVM_REG_RISCV_VECTOR_CSR_REG(vtype):
> + reg_val = &cntx->vector.vtype;
> + break;
> + case KVM_REG_RISCV_VECTOR_CSR_REG(vcsr):
> + reg_val = &cntx->vector.vcsr;
> + break;
> + case KVM_REG_RISCV_VECTOR_CSR_REG(datap):
> + default:
> + return NULL;
> + }
> + } else if (reg_num <= KVM_REG_RISCV_VECTOR_REG(31)) {
> + if (reg_size != vlenb)
> + return NULL;
> + reg_val = cntx->vector.datap
> + + (reg_num - KVM_REG_RISCV_VECTOR_REG(0)) * vlenb;
> + } else {
> + return NULL;
> + }
> +
> + return reg_val;
> +}
> +
> +int kvm_riscv_vcpu_get_reg_vector(struct kvm_vcpu *vcpu,
> + const struct kvm_one_reg *reg,
> + unsigned long rtype)
> +{
> + unsigned long *isa = vcpu->arch.isa;
> + unsigned long __user *uaddr =
> + (unsigned long __user *)(unsigned long)reg->addr;
> + unsigned long reg_num = reg->id & ~(KVM_REG_ARCH_MASK |
> + KVM_REG_SIZE_MASK |
> + rtype);
> + void *reg_val = NULL;
> + size_t reg_size = KVM_REG_SIZE(reg->id);
> +
> + if (rtype == KVM_REG_RISCV_VECTOR &&
> + riscv_isa_extension_available(isa, v)) {
> + reg_val = kvm_riscv_vcpu_vreg_addr(vcpu, reg_num, reg_size);
> + }
> +
> + if (!reg_val)
> + return -EINVAL;
> +
> + if (copy_to_user(uaddr, reg_val, reg_size))
> + return -EFAULT;
> +
> + return 0;
> +}
> +
> +int kvm_riscv_vcpu_set_reg_vector(struct kvm_vcpu *vcpu,
> + const struct kvm_one_reg *reg,
> + unsigned long rtype)
> +{
> + unsigned long *isa = vcpu->arch.isa;
> + unsigned long __user *uaddr =
> + (unsigned long __user *)(unsigned long)reg->addr;
> + unsigned long reg_num = reg->id & ~(KVM_REG_ARCH_MASK |
> + KVM_REG_SIZE_MASK |
> + rtype);
> + void *reg_val = NULL;
> + size_t reg_size = KVM_REG_SIZE(reg->id);
> +
> + if (rtype == KVM_REG_RISCV_VECTOR &&
> + riscv_isa_extension_available(isa, v)) {
> + reg_val = kvm_riscv_vcpu_vreg_addr(vcpu, reg_num, reg_size);
> + }
> +
> + if (!reg_val)
> + return -EINVAL;
> +
> + if (copy_from_user(reg_val, uaddr, reg_size))
> + return -EFAULT;
> +
> + return 0;
> +}
> --
> 2.17.1
>
More information about the linux-riscv
mailing list