[PATCH v3 3/5] KVM: ARM VGIC add kvm_io_bus_ frontend

Eric Auger eric.auger at linaro.org
Tue Jan 27 09:26:05 PST 2015


On 01/27/2015 05:51 PM, Nikolay Nikolaev wrote:
> Hi Andre,
> 
> On Tue, Jan 27, 2015 at 3:31 PM, Andre Przywara <andre.przywara at arm.com> wrote:
>>
>> Hi Nikolay,
>>
>> On 24/01/15 11:59, Nikolay Nikolaev wrote:
>>> In io_mem_abort remove the call to vgic_handle_mmio. The target is to have
>>> a single MMIO handling path - that is through the kvm_io_bus_ API.
>>>
>>> Register a kvm_io_device in kvm_vgic_init on the whole vGIC MMIO region.
>>> Both read and write calls are redirected to vgic_io_dev_access where
>>> kvm_exit_mmio is composed to pass it to vm_ops.handle_mmio.
>>>
>>>
>>> Signed-off-by: Nikolay Nikolaev <n.nikolaev at virtualopensystems.com>
>>> ---
>>>  arch/arm/kvm/mmio.c    |    3 -
>>>  include/kvm/arm_vgic.h |    3 -
>>>  virt/kvm/arm/vgic.c    |  123 ++++++++++++++++++++++++++++++++++++++++++++----
>>>  3 files changed, 114 insertions(+), 15 deletions(-)
>>>
>>> diff --git a/arch/arm/kvm/mmio.c b/arch/arm/kvm/mmio.c
>>> index d852137..8dc2fde 100644
>>> --- a/arch/arm/kvm/mmio.c
>>> +++ b/arch/arm/kvm/mmio.c
>>> @@ -230,9 +230,6 @@ int io_mem_abort(struct kvm_vcpu *vcpu, struct kvm_run *run,
>>>                              fault_ipa, 0);
>>>       }
>>>
>>> -     if (vgic_handle_mmio(vcpu, run, &mmio))
>>> -             return 1;
>>> -
>>
>> Why is this (whole patch) actually needed? Is that just to make it nicer
>> by pulling everything under one umbrella?
> 
> 
> It started from this mail form Christofer:
> https://lkml.org/lkml/2014/3/28/403
Hi Nikolay, Andre,

I also understood that the target was to handle all kernel mmio through
the same API, hence the first patch. This patch shows that at least for
GICv2 it was doable without upheavals in vgic code and it also serves
ioeventd which is good. Andre do you think the price to pay to integrate
missing redistributors and forthcoming components is too high?

Best Regards

Eric


> 
>>
>> For enabling ioeventfd you actually don't need this patch, right?
> Yes, we don't need it.
>> (I am asking because this breaks GICv3 emulation, see below)
>>
>>>       if (handle_kernel_mmio(vcpu, run, &mmio))
>>>               return 1;
>>>
>>> diff --git a/include/kvm/arm_vgic.h b/include/kvm/arm_vgic.h
>>> index 7c55dd5..60639b1 100644
>>> --- a/include/kvm/arm_vgic.h
>>> +++ b/include/kvm/arm_vgic.h
>>> @@ -237,6 +237,7 @@ struct vgic_dist {
>>>       unsigned long           *irq_pending_on_cpu;
>>>
>>>       struct vgic_vm_ops      vm_ops;
>>> +     struct kvm_io_device    *io_dev;
>>>  #endif
>>>  };
>>>
>>> @@ -311,8 +312,6 @@ int kvm_vgic_inject_irq(struct kvm *kvm, int cpuid, unsigned int irq_num,
>>>                       bool level);
>>>  void vgic_v3_dispatch_sgi(struct kvm_vcpu *vcpu, u64 reg);
>>>  int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu);
>>> -bool vgic_handle_mmio(struct kvm_vcpu *vcpu, struct kvm_run *run,
>>> -                   struct kvm_exit_mmio *mmio);
>>>
>>>  #define irqchip_in_kernel(k) (!!((k)->arch.vgic.in_kernel))
>>>  #define vgic_initialized(k)  (!!((k)->arch.vgic.nr_cpus))
>>> diff --git a/virt/kvm/arm/vgic.c b/virt/kvm/arm/vgic.c
>>> index 0cc6ab6..195d2ba 100644
>>> --- a/virt/kvm/arm/vgic.c
>>> +++ b/virt/kvm/arm/vgic.c
>>> @@ -31,6 +31,9 @@
>>>  #include <asm/kvm_emulate.h>
>>>  #include <asm/kvm_arm.h>
>>>  #include <asm/kvm_mmu.h>
>>> +#include <asm/kvm.h>
>>> +
>>> +#include "iodev.h"
>>>
>>>  /*
>>>   * How the whole thing works (courtesy of Christoffer Dall):
>>> @@ -77,6 +80,7 @@
>>>
>>>  #include "vgic.h"
>>>
>>> +static int vgic_register_kvm_io_dev(struct kvm *kvm);
>>>  static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu);
>>>  static void vgic_retire_lr(int lr_nr, int irq, struct kvm_vcpu *vcpu);
>>>  static struct vgic_lr vgic_get_lr(const struct kvm_vcpu *vcpu, int lr);
>>> @@ -97,6 +101,7 @@ static bool queue_sgi(struct kvm_vcpu *vcpu, int irq)
>>>
>>>  int kvm_vgic_map_resources(struct kvm *kvm)
>>>  {
>>> +     vgic_register_kvm_io_dev(kvm);
>>>       return kvm->arch.vgic.vm_ops.map_resources(kvm, vgic);
>>>  }
>>>
>>> @@ -776,27 +781,123 @@ bool vgic_handle_mmio_range(struct kvm_vcpu *vcpu, struct kvm_run *run,
>>>  }
>>>
>>>  /**
>>> - * vgic_handle_mmio - handle an in-kernel MMIO access for the GIC emulation
>>> + * vgic_io_dev_access - handle an in-kernel MMIO access for the GIC emulation
>>>   * @vcpu:      pointer to the vcpu performing the access
>>> - * @run:       pointer to the kvm_run structure
>>> - * @mmio:      pointer to the data describing the access
>>> + * @this:      pointer to the kvm_io_device structure
>>> + * @addr:      the MMIO address being accessed
>>> + * @len:       the length of the accessed data
>>> + * @val:       pointer to the value being written,
>>> + *             or where the read operation will store its result
>>> + * @is_write:  flag to show whether a write access is performed
>>>   *
>>> - * returns true if the MMIO access has been performed in kernel space,
>>> - * and false if it needs to be emulated in user space.
>>> + * returns 0 if the MMIO access has been performed in kernel space,
>>> + * and 1 if it needs to be emulated in user space.
>>>   * Calls the actual handling routine for the selected VGIC model.
>>>   */
>>> -bool vgic_handle_mmio(struct kvm_vcpu *vcpu, struct kvm_run *run,
>>> -                   struct kvm_exit_mmio *mmio)
>>> +static int vgic_io_dev_access(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
>>> +                         gpa_t addr, int len, void *val, bool is_write)
>>>  {
>>> -     if (!irqchip_in_kernel(vcpu->kvm))
>>> -             return false;
>>> +     struct kvm_exit_mmio mmio;
>>> +     bool ret;
>>> +
>>> +     mmio = (struct kvm_exit_mmio) {
>>> +             .phys_addr = addr,
>>> +             .len = len,
>>> +             .is_write = is_write,
>>> +     };
>>> +
>>> +     if (is_write)
>>> +             memcpy(mmio.data, val, len);
>>>
>>>       /*
>>>        * This will currently call either vgic_v2_handle_mmio() or
>>>        * vgic_v3_handle_mmio(), which in turn will call
>>>        * vgic_handle_mmio_range() defined above.
>>>        */
>>> -     return vcpu->kvm->arch.vgic.vm_ops.handle_mmio(vcpu, run, mmio);
>>> +     ret = vcpu->kvm->arch.vgic.vm_ops.handle_mmio(vcpu, vcpu->run, &mmio);
>>> +
>>> +     if (!is_write)
>>> +             memcpy(val, mmio.data, len);
>>> +
>>> +     return ret ? 0 : 1;
>>> +}
>>> +
>>> +static int vgic_io_dev_read(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
>>> +                       gpa_t addr, int len, void *val)
>>> +{
>>> +     return vgic_io_dev_access(vcpu, this, addr, len, val, false);
>>> +}
>>> +
>>> +static int vgic_io_dev_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
>>> +                        gpa_t addr, int len, const void *val)
>>> +{
>>> +     return vgic_io_dev_access(vcpu, this, addr, len, (void *)val, true);
>>> +}
>>> +
>>> +static const struct kvm_io_device_ops vgic_io_dev_ops = {
>>> +     .read       = vgic_io_dev_read,
>>> +     .write      = vgic_io_dev_write,
>>> +};
>>> +
>>> +static int vgic_register_kvm_io_dev(struct kvm *kvm)
>>> +{
>>> +     int len = 0;
>>> +     int ret;
>>> +
>>> +     struct vgic_dist *dist = &kvm->arch.vgic;
>>> +     unsigned long base = dist->vgic_dist_base;
>>> +     u32 type = kvm->arch.vgic.vgic_model;
>>> +     struct kvm_io_device *dev;
>>> +
>>> +     if (IS_VGIC_ADDR_UNDEF(base)) {
>>> +             kvm_err("Need to set vgic distributor address first\n");
>>> +             return -ENXIO;
>>> +     }
>>> +
>>> +     dev = kzalloc(sizeof(struct kvm_io_device), GFP_KERNEL);
>>> +     if (!dev)
>>> +             return -ENOMEM;
>>> +
>>> +     switch (type) {
>>> +     case KVM_DEV_TYPE_ARM_VGIC_V2:
>>> +             len = KVM_VGIC_V2_DIST_SIZE;
>>> +             break;
>>> +#ifdef CONFIG_ARM_GIC_V3
>>> +     case KVM_DEV_TYPE_ARM_VGIC_V3:
>>> +             len = KVM_VGIC_V3_DIST_SIZE;
>>> +             break;
>>> +#endif
>>> +     }
>>
>> But this only registers the GIC distributor, leaving out the
>> redistributor regions introduced by GICv3. To me it looks like this
> I see GICv3 needs more work.
> 
>> kvm_iodevice registration code should be moved into *-emul.c, where each
>> emulated device registers what it needs.
>> Especially in the wake of the upcoming v2M/ITS emulation I think we need
>> a proper solution for this, so I am wondering if we could just leave
>> that patch out (at least for now) and keep the two-line special
>> treatment for the VGIC above in.
>> That should enable ioeventfd without breaking the VGIC.
> Then we're back to the original RFC patch series.
> I have no issues droppin this one (and propably patch 1 in the series)
> and leaving only the eventfd related handling.
> I just need some consensus/confirmation on the mailing list.
> 
> regards,
> Nikolay Nikolaev
> 
>>
>> Cheers,
>> Andre.
>>
>>> +
>>> +     kvm_iodevice_init(dev, &vgic_io_dev_ops);
>>> +
>>> +     mutex_lock(&kvm->slots_lock);
>>> +
>>> +     ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS,
>>> +                     base, len, dev);
>>> +     if (ret < 0)
>>> +             goto out_unlock;
>>> +     mutex_unlock(&kvm->slots_lock);
>>> +
>>> +     kvm->arch.vgic.io_dev = dev;
>>> +
>>> +     return 0;
>>> +
>>> +out_unlock:
>>> +     mutex_unlock(&kvm->slots_lock);
>>> +     kfree(dev);
>>> +     return ret;
>>> +}
>>> +
>>> +static void vgic_unregister_kvm_io_dev(struct kvm *kvm)
>>> +{
>>> +     struct vgic_dist *dist = &kvm->arch.vgic;
>>> +
>>> +     if (dist) {
>>> +             kvm_io_bus_unregister_dev(kvm, KVM_MMIO_BUS, dist->io_dev);
>>> +             kfree(dist->io_dev);
>>> +             dist->io_dev = NULL;
>>> +     }
>>>  }
>>>
>>>  static int vgic_nr_shared_irqs(struct vgic_dist *dist)
>>> @@ -1428,6 +1529,8 @@ void kvm_vgic_destroy(struct kvm *kvm)
>>>       struct kvm_vcpu *vcpu;
>>>       int i;
>>>
>>> +     vgic_unregister_kvm_io_dev(kvm);
>>> +
>>>       kvm_for_each_vcpu(i, vcpu, kvm)
>>>               kvm_vgic_vcpu_destroy(vcpu);
>>>
>>>
>>> _______________________________________________
>>> kvmarm mailing list
>>> kvmarm at lists.cs.columbia.edu
>>> https://lists.cs.columbia.edu/mailman/listinfo/kvmarm
>>>




More information about the linux-arm-kernel mailing list