[PATCH 7/8] KVM: arm/arm64: Avoid accessing GICH registers
Marc Zyngier
marc.zyngier at arm.com
Wed Feb 10 05:34:44 PST 2016
On 10/02/16 12:45, Christoffer Dall wrote:
> On Mon, Feb 08, 2016 at 11:40:21AM +0000, Marc Zyngier wrote:
>> GICv2 registers are *slow*. As in "terrifyingly slow". Which is bad.
>> But we're equaly bad, as we make a point in accessing them even if
>> we don't have any interrupt in flight.
>>
>> A good solution is to first find out if we have anything useful to
>> write into the GIC, and if we don't, to simply not do it. This
>> involves tracking which LRs actually have something valid there.
>>
>> Signed-off-by: Marc Zyngier <marc.zyngier at arm.com>
>> ---
>> arch/arm64/kvm/hyp/vgic-v2-sr.c | 71 ++++++++++++++++++++++++++++-------------
>> include/kvm/arm_vgic.h | 2 ++
>> 2 files changed, 51 insertions(+), 22 deletions(-)
>>
>> diff --git a/arch/arm64/kvm/hyp/vgic-v2-sr.c b/arch/arm64/kvm/hyp/vgic-v2-sr.c
>> index e717612..874a08d 100644
>> --- a/arch/arm64/kvm/hyp/vgic-v2-sr.c
>> +++ b/arch/arm64/kvm/hyp/vgic-v2-sr.c
>> @@ -38,28 +38,40 @@ void __hyp_text __vgic_v2_save_state(struct kvm_vcpu *vcpu)
>>
>> nr_lr = vcpu->arch.vgic_cpu.nr_lr;
>> cpu_if->vgic_vmcr = readl_relaxed(base + GICH_VMCR);
>> - cpu_if->vgic_misr = readl_relaxed(base + GICH_MISR);
>> - eisr0 = readl_relaxed(base + GICH_EISR0);
>> - elrsr0 = readl_relaxed(base + GICH_ELRSR0);
>> - if (unlikely(nr_lr > 32)) {
>> - eisr1 = readl_relaxed(base + GICH_EISR1);
>> - elrsr1 = readl_relaxed(base + GICH_ELRSR1);
>> - } else {
>> - eisr1 = elrsr1 = 0;
>> - }
>> +
>> + if (vcpu->arch.vgic_cpu.live_lrs) {
>> + eisr0 = readl_relaxed(base + GICH_EISR0);
>> + elrsr0 = readl_relaxed(base + GICH_ELRSR0);
>> + cpu_if->vgic_misr = readl_relaxed(base + GICH_MISR);
>> + cpu_if->vgic_apr = readl_relaxed(base + GICH_APR);
>> +
>> + if (unlikely(nr_lr > 32)) {
>> + eisr1 = readl_relaxed(base + GICH_EISR1);
>> + elrsr1 = readl_relaxed(base + GICH_ELRSR1);
>> + } else {
>> + eisr1 = elrsr1 = 0;
>> + }
>> +
>> #ifdef CONFIG_CPU_BIG_ENDIAN
>> - cpu_if->vgic_eisr = ((u64)eisr0 << 32) | eisr1;
>> - cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1;
>> + cpu_if->vgic_eisr = ((u64)eisr0 << 32) | eisr1;
>> + cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1;
>> #else
>> - cpu_if->vgic_eisr = ((u64)eisr1 << 32) | eisr0;
>> - cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0;
>> + cpu_if->vgic_eisr = ((u64)eisr1 << 32) | eisr0;
>> + cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0;
>> #endif
>> - cpu_if->vgic_apr = readl_relaxed(base + GICH_APR);
>>
>> - writel_relaxed(0, base + GICH_HCR);
>> + for (i = 0; i < nr_lr; i++)
>> + if (vcpu->arch.vgic_cpu.live_lrs & (1UL << i))
>> + cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4));
>>
>> - for (i = 0; i < nr_lr; i++)
>> - cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4));
>> + writel_relaxed(0, base + GICH_HCR);
>> +
>> + vcpu->arch.vgic_cpu.live_lrs = 0;
>> + } else {
>> + cpu_if->vgic_eisr = 0;
>> + cpu_if->vgic_elrsr = ~0UL;
>> + cpu_if->vgic_misr = 0;
>> + }
>> }
>>
>> /* vcpu is already in the HYP VA space */
>> @@ -70,15 +82,30 @@ void __hyp_text __vgic_v2_restore_state(struct kvm_vcpu *vcpu)
>> struct vgic_dist *vgic = &kvm->arch.vgic;
>> void __iomem *base = kern_hyp_va(vgic->vctrl_base);
>> int i, nr_lr;
>> + u64 live_lrs = 0;
>>
>> if (!base)
>> return;
>>
>> - writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR);
>> - writel_relaxed(cpu_if->vgic_vmcr, base + GICH_VMCR);
>> - writel_relaxed(cpu_if->vgic_apr, base + GICH_APR);
>> -
>> nr_lr = vcpu->arch.vgic_cpu.nr_lr;
>> +
>> for (i = 0; i < nr_lr; i++)
>> - writel_relaxed(cpu_if->vgic_lr[i], base + GICH_LR0 + (i * 4));
>> + if (cpu_if->vgic_lr[i] & GICH_LR_STATE)
>> + live_lrs |= 1UL << i;
>> +
>> + if (live_lrs) {
>> + writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR);
>> + writel_relaxed(cpu_if->vgic_apr, base + GICH_APR);
>> + for (i = 0; i < nr_lr; i++) {
>> + u32 val = 0;
>> +
>> + if (live_lrs & (1UL << i))
>> + val = cpu_if->vgic_lr[i];
>> +
>> + writel_relaxed(val, base + GICH_LR0 + (i * 4));
>> + }
>> + }
>> +
>> + writel_relaxed(cpu_if->vgic_vmcr, base + GICH_VMCR);
>
> couldn't you optimize this out by storing the last read value and
> compare if anything changed? (you'd have to invalidate the cached value
> on vcpu_put obviously).
Yeah, very good point. Only the guest can update this, so we could even
move it to vcpu_load/vcpu_put entirely, and never save/restore it inside
the run loop.
I'll keep that for a further patch, as it requires a bit of infrastructure.
Thanks,
M.
--
Jazz is not dead. It just smells funny...
More information about the linux-arm-kernel
mailing list