[PATCH v2] ARM/KVM: save and restore generic timer registers

Christoffer Dall christoffer.dall at linaro.org
Thu Jun 20 13:09:23 EDT 2013


On Thu, Jun 20, 2013 at 11:10:48AM +0100, Marc Zyngier wrote:
> On 11/06/13 16:16, Andre Przywara wrote:
> > For migration to work we need to save (and later restore) the state of
> > each core's virtual generic timer.
> > Since this is per VCPU, we can use the [gs]et_one_reg ioctl and export
> > the three needed registers (control, counter, compare value).
> > Though they live in cp15 space, we don't use the existing list, since
> > they need special accessor functions and also the arch timer is
> > optional.
> > 
> > Changes from v1:
> > - move code out of coproc.c and into guest.c and arch_timer.c
> > - present the registers with their native CP15 addresses, but without
> >   using space in the VCPU's cp15 array
> > - do the user space copying in the accessor functions
> > 
> > Signed-off-by: Andre Przywara <andre.przywara at linaro.org>
> > ---
> >  arch/arm/include/asm/kvm_host.h |  5 ++++
> >  arch/arm/include/uapi/asm/kvm.h | 16 ++++++++++
> >  arch/arm/kvm/arch_timer.c       | 65 +++++++++++++++++++++++++++++++++++++++++
> >  arch/arm/kvm/guest.c            | 26 ++++++++++++++++-
> >  4 files changed, 111 insertions(+), 1 deletion(-)
> > 
> > diff --git a/arch/arm/include/asm/kvm_host.h b/arch/arm/include/asm/kvm_host.h
> > index 57cb786..1096e33 100644
> > --- a/arch/arm/include/asm/kvm_host.h
> > +++ b/arch/arm/include/asm/kvm_host.h
> > @@ -224,4 +224,9 @@ static inline int kvm_arch_dev_ioctl_check_extension(long ext)
> >  int kvm_perf_init(void);
> >  int kvm_perf_teardown(void);
> >  
> > +int kvm_arm_num_timer_regs(void);
> > +int kvm_arm_copy_timer_indices(struct kvm_vcpu *, u64 __user *);
> > +int kvm_arm_timer_get_reg(struct kvm_vcpu *, const struct kvm_one_reg *);
> > +int kvm_arm_timer_set_reg(struct kvm_vcpu *, const struct kvm_one_reg *);
> > +
> >  #endif /* __ARM_KVM_HOST_H__ */
> > diff --git a/arch/arm/include/uapi/asm/kvm.h b/arch/arm/include/uapi/asm/kvm.h
> > index c1ee007..e3b0115 100644
> > --- a/arch/arm/include/uapi/asm/kvm.h
> > +++ b/arch/arm/include/uapi/asm/kvm.h
> > @@ -118,6 +118,22 @@ struct kvm_arch_memory_slot {
> >  #define KVM_REG_ARM_32_CRN_MASK		0x0000000000007800
> >  #define KVM_REG_ARM_32_CRN_SHIFT	11
> >  
> > +#define KVM_REG_ARM_32_CP15		(KVM_REG_ARM | KVM_REG_SIZE_U32      | \
> > +					(15ULL << KVM_REG_ARM_COPROC_SHIFT))
> > +#define KVM_REG_ARM_64_CP15		(KVM_REG_ARM | KVM_REG_SIZE_U64      | \
> > +					(15ULL << KVM_REG_ARM_COPROC_SHIFT))
> > +#define KVM_REG_ARM_TIMER_CTL		(KVM_REG_ARM_32_CP15                 | \
> > +					( 3ULL << KVM_REG_ARM_CRM_SHIFT)     | \
> > +					(14ULL << KVM_REG_ARM_32_CRN_SHIFT)  | \
> > +					( 0ULL << KVM_REG_ARM_OPC1_SHIFT)    | \
> > +					( 1ULL << KVM_REG_ARM_32_OPC2_SHIFT))
> > +#define KVM_REG_ARM_TIMER_CNT		(KVM_REG_ARM_64_CP15 | \
> > +					(14ULL << KVM_REG_ARM_CRM_SHIFT)     | \
> > +					( 1ULL << KVM_REG_ARM_OPC1_SHIFT))
> > +#define KVM_REG_ARM_TIMER_CVAL		(KVM_REG_ARM_64_CP15 | \
> > +					(14ULL << KVM_REG_ARM_CRM_SHIFT)     | \
> > +					( 3ULL << KVM_REG_ARM_OPC1_SHIFT))
> > +
> >  /* Normal registers are mapped as coprocessor 16. */
> >  #define KVM_REG_ARM_CORE		(0x0010 << KVM_REG_ARM_COPROC_SHIFT)
> >  #define KVM_REG_ARM_CORE_REG(name)	(offsetof(struct kvm_regs, name) / 4)
> > diff --git a/arch/arm/kvm/arch_timer.c b/arch/arm/kvm/arch_timer.c
> > index c55b608..8d709eb 100644
> > --- a/arch/arm/kvm/arch_timer.c
> > +++ b/arch/arm/kvm/arch_timer.c
> > @@ -18,6 +18,7 @@
> >  
> >  #include <linux/cpu.h>
> >  #include <linux/of_irq.h>
> > +#include <linux/uaccess.h>
> >  #include <linux/kvm.h>
> >  #include <linux/kvm_host.h>
> >  #include <linux/interrupt.h>
> > @@ -171,6 +172,70 @@ static void kvm_timer_init_interrupt(void *info)
> >  	enable_percpu_irq(timer_irq.irq, 0);
> >  }
> >  
> > +int kvm_arm_num_timer_regs(void)
> > +{
> > +	return 3;
> > +}
> > +
> > +int kvm_arm_copy_timer_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
> > +{
> > +	if (put_user(KVM_REG_ARM_TIMER_CTL, uindices))
> > +		return -EFAULT;
> > +	uindices++;
> > +	if (put_user(KVM_REG_ARM_TIMER_CNT, uindices))
> > +		return -EFAULT;
> > +	uindices++;
> > +	if (put_user(KVM_REG_ARM_TIMER_CVAL, uindices))
> > +		return -EFAULT;
> 
> So these macros are going to break arm64. Any chance you could introduce
> them at the same time on both platforms? The rest of the work can be
> delayed, but you shouldn't break arm64 (you'd expect me to say that,
> wouldn't you? ;-).
> 
> Also, I'd like to keep userspace access out of the timer code itself.
> Low level code shouldn't have to know about that. Can you create proper
> accessors instead, and move whole userspace access to coproc.c?
> 

I'm fine with this, coproc.c or guest.c - either way.

> > +	return 0;
> > +}
> > +
> > +int kvm_arm_timer_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
> > +{
> > +	struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu;
> > +	void __user *uaddr = (void __user *)(long)reg->addr;
> > +	u64 val;
> > +	int ret;
> > +
> > +	ret = copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id));
> > +	if (ret != 0)
> > +		return ret;
> > +
> > +	switch (reg->id) {
> > +	case KVM_REG_ARM_TIMER_CTL:
> > +		timer->cntv_ctl = val;
> > +		break;
> > +	case KVM_REG_ARM_TIMER_CNT:
> > +		vcpu->kvm->arch.timer.cntvoff = kvm_phys_timer_read() - val;
> 
> I just realized what bothers me here: You're computing cntvoff on a per
> vcpu basis, while this is a VM property. Which means that as you're
> restoring vcpus, you'll be changing cntvoff - sounds like a bad idea to me.
> 
> The counter is really global. Do we have a way to handle VM-wide
> registers? I think Christoffer was trying to some a similar thing with
> the GIC...
> 

We do have a way, but it requires user space to create a device and keep
track of the device fd just to set/get a single register, which seems
like overkill to me.

I suggest you do one of two things:
 1. Whenever this value is written, make sure it's written across all
    vcpus, so guests always have a consistent view of time (be careful
    about synchronization here).
 2. Move the cntvoff value to the vm struct instead, so there's only one
    offset and a consistent view of time.  This may have an adverse
    effect on the world-switch code performance, but I suspect it would
    completely disappear in the noise.

I dont' feel strongly about either approach.

-Christoffer

> > +		break;
> > +	case KVM_REG_ARM_TIMER_CVAL:
> > +		timer->cntv_cval = val;
> > +		break;
> > +	}
> > +
> > +	return 0;
> > +}
> > +
> > +int kvm_arm_timer_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
> > +{
> > +	struct arch_timer_cpu *timer = &vcpu->arch.timer_cpu;
> > +	void __user *uaddr = (void __user *)(long)reg->addr;
> > +	u64 val;
> > +
> > +	switch (reg->id) {
> > +	case KVM_REG_ARM_TIMER_CTL:
> > +		val = timer->cntv_ctl;
> > +		break;
> > +	case KVM_REG_ARM_TIMER_CNT:
> > +		val = kvm_phys_timer_read() - vcpu->kvm->arch.timer.cntvoff;
> > +		break;
> > +	case KVM_REG_ARM_TIMER_CVAL:
> > +		val = timer->cntv_cval;
> > +		break;
> > +	}
> > +	return copy_to_user(uaddr, &val, KVM_REG_SIZE(reg->id));
> > +}
> >  
> >  static int kvm_timer_cpu_notify(struct notifier_block *self,
> >  				unsigned long action, void *cpu)
> > diff --git a/arch/arm/kvm/guest.c b/arch/arm/kvm/guest.c
> > index 152d036..a50ffb6 100644
> > --- a/arch/arm/kvm/guest.c
> > +++ b/arch/arm/kvm/guest.c
> > @@ -121,7 +121,8 @@ static unsigned long num_core_regs(void)
> >   */
> >  unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu)
> >  {
> > -	return num_core_regs() + kvm_arm_num_coproc_regs(vcpu);
> > +	return num_core_regs() + kvm_arm_num_coproc_regs(vcpu)
> > +		+ kvm_arm_num_timer_regs();
> >  }
> >  
> >  /**
> > @@ -133,6 +134,7 @@ int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
> >  {
> >  	unsigned int i;
> >  	const u64 core_reg = KVM_REG_ARM | KVM_REG_SIZE_U32 | KVM_REG_ARM_CORE;
> > +	int ret;
> >  
> >  	for (i = 0; i < sizeof(struct kvm_regs)/sizeof(u32); i++) {
> >  		if (put_user(core_reg | i, uindices))
> > @@ -140,9 +142,25 @@ int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
> >  		uindices++;
> >  	}
> >  
> > +	ret = kvm_arm_copy_timer_indices(vcpu, uindices);
> > +	if (ret)
> > +		return ret;
> > +	uindices += kvm_arm_num_timer_regs();
> > +
> >  	return kvm_arm_copy_coproc_indices(vcpu, uindices);
> >  }
> >  
> > +static bool is_timer_reg(u64 index)
> > +{
> > +	switch (index) {
> > +	case KVM_REG_ARM_TIMER_CTL:
> > +	case KVM_REG_ARM_TIMER_CNT:
> > +	case KVM_REG_ARM_TIMER_CVAL:
> > +		return true;
> > +	}
> > +	return false;
> > +}
> > +
> >  int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
> >  {
> >  	/* We currently use nothing arch-specific in upper 32 bits */
> > @@ -153,6 +171,9 @@ int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
> >  	if ((reg->id & KVM_REG_ARM_COPROC_MASK) == KVM_REG_ARM_CORE)
> >  		return get_core_reg(vcpu, reg);
> >  
> > +	if (is_timer_reg(reg->id))
> > +		return kvm_arm_timer_get_reg(vcpu, reg);
> > +
> >  	return kvm_arm_coproc_get_reg(vcpu, reg);
> >  }
> >  
> > @@ -166,6 +187,9 @@ int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
> >  	if ((reg->id & KVM_REG_ARM_COPROC_MASK) == KVM_REG_ARM_CORE)
> >  		return set_core_reg(vcpu, reg);
> >  
> > +	if (is_timer_reg(reg->id))
> > +		return kvm_arm_timer_set_reg(vcpu, reg);
> > +
> >  	return kvm_arm_coproc_set_reg(vcpu, reg);
> >  }



More information about the linux-arm-kernel mailing list