Message ID | 20171012104141.26902-33-christoffer.dall@linaro.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Hi, those last few patches are actually helpful for the Xen port ... On 12/10/17 11:41, Christoffer Dall wrote: > We can program the GICv2 hypervisor control interface logic directly > from the core vgic code and can instead do the save/restore directly > from the flush/sync functions, which can lead to a number of future > optimizations. > > Signed-off-by: Christoffer Dall <christoffer.dall@linaro.org> > --- > arch/arm/kvm/hyp/switch.c | 4 -- > arch/arm64/include/asm/kvm_hyp.h | 2 - > arch/arm64/kvm/hyp/switch.c | 4 -- > virt/kvm/arm/hyp/vgic-v2-sr.c | 83 ------------------------------------ > virt/kvm/arm/vgic/vgic-init.c | 22 ++++++---- > virt/kvm/arm/vgic/vgic-v2.c | 92 ++++++++++++++++++++++++++++++++++++++++ > virt/kvm/arm/vgic/vgic.c | 21 ++++++++- > virt/kvm/arm/vgic/vgic.h | 5 +++ > 8 files changed, 130 insertions(+), 103 deletions(-) > > diff --git a/arch/arm/kvm/hyp/switch.c b/arch/arm/kvm/hyp/switch.c > index c3b9799..0d834f8 100644 > --- a/arch/arm/kvm/hyp/switch.c > +++ b/arch/arm/kvm/hyp/switch.c > @@ -91,16 +91,12 @@ static void __hyp_text __vgic_save_state(struct kvm_vcpu *vcpu) > { > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > __vgic_v3_save_state(vcpu); > - else > - __vgic_v2_save_state(vcpu); > } > > static void __hyp_text __vgic_restore_state(struct kvm_vcpu *vcpu) > { > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > __vgic_v3_restore_state(vcpu); > - else > - __vgic_v2_restore_state(vcpu); > } > > static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) > diff --git a/arch/arm64/include/asm/kvm_hyp.h b/arch/arm64/include/asm/kvm_hyp.h > index 28d5f3c..bd3fe64 100644 > --- a/arch/arm64/include/asm/kvm_hyp.h > +++ b/arch/arm64/include/asm/kvm_hyp.h > @@ -121,8 +121,6 @@ typeof(orig) * __hyp_text fname(void) \ > return val; \ > } > > -void __vgic_v2_save_state(struct kvm_vcpu *vcpu); > -void __vgic_v2_restore_state(struct kvm_vcpu *vcpu); > int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu); > > void __vgic_v3_save_state(struct kvm_vcpu *vcpu); > diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c > index 5692aa0..90da506 100644 > --- a/arch/arm64/kvm/hyp/switch.c > +++ b/arch/arm64/kvm/hyp/switch.c > @@ -186,16 +186,12 @@ static void __hyp_text __vgic_save_state(struct kvm_vcpu *vcpu) > { > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > __vgic_v3_save_state(vcpu); > - else > - __vgic_v2_save_state(vcpu); > } > > static void __hyp_text __vgic_restore_state(struct kvm_vcpu *vcpu) > { > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > __vgic_v3_restore_state(vcpu); > - else > - __vgic_v2_restore_state(vcpu); > } > > static bool __hyp_text __true_value(void) > diff --git a/virt/kvm/arm/hyp/vgic-v2-sr.c b/virt/kvm/arm/hyp/vgic-v2-sr.c > index a3f18d3..b433257 100644 > --- a/virt/kvm/arm/hyp/vgic-v2-sr.c > +++ b/virt/kvm/arm/hyp/vgic-v2-sr.c > @@ -22,89 +22,6 @@ > #include <asm/kvm_emulate.h> > #include <asm/kvm_hyp.h> > > -static void __hyp_text save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) > -{ > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > - int nr_lr = (kern_hyp_va(&kvm_vgic_global_state))->nr_lr; > - u32 elrsr0, elrsr1; > - > - elrsr0 = readl_relaxed(base + GICH_ELRSR0); > - if (unlikely(nr_lr > 32)) > - elrsr1 = readl_relaxed(base + GICH_ELRSR1); > - else > - elrsr1 = 0; > - > -#ifdef CONFIG_CPU_BIG_ENDIAN > - cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; > -#else > - cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; > -#endif > -} > - > -static void __hyp_text save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > -{ > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > - int i; > - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > - > - for (i = 0; i < used_lrs; i++) { > - if (cpu_if->vgic_elrsr & (1UL << i)) > - cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > - else > - cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > - > - writel_relaxed(0, base + GICH_LR0 + (i * 4)); > - } > -} > - > -/* vcpu is already in the HYP VA space */ > -void __hyp_text __vgic_v2_save_state(struct kvm_vcpu *vcpu) > -{ > - struct kvm *kvm = kern_hyp_va(vcpu->kvm); > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > - struct vgic_dist *vgic = &kvm->arch.vgic; > - void __iomem *base = kern_hyp_va(vgic->vctrl_base); > - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > - > - if (!base) > - return; > - > - if (used_lrs) { > - cpu_if->vgic_apr = readl_relaxed(base + GICH_APR); > - > - save_elrsr(vcpu, base); > - save_lrs(vcpu, base); > - > - writel_relaxed(0, base + GICH_HCR); > - } else { > - cpu_if->vgic_elrsr = ~0UL; > - cpu_if->vgic_apr = 0; > - } > -} > - > -/* vcpu is already in the HYP VA space */ > -void __hyp_text __vgic_v2_restore_state(struct kvm_vcpu *vcpu) > -{ > - struct kvm *kvm = kern_hyp_va(vcpu->kvm); > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > - struct vgic_dist *vgic = &kvm->arch.vgic; > - void __iomem *base = kern_hyp_va(vgic->vctrl_base); > - int i; > - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > - > - if (!base) > - return; > - > - if (used_lrs) { > - writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); > - writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); > - for (i = 0; i < used_lrs; i++) { > - writel_relaxed(cpu_if->vgic_lr[i], > - base + GICH_LR0 + (i * 4)); > - } > - } > -} > - > #ifdef CONFIG_ARM64 > /* > * __vgic_v2_perform_cpuif_access -- perform a GICV access on behalf of the > diff --git a/virt/kvm/arm/vgic/vgic-init.c b/virt/kvm/arm/vgic/vgic-init.c > index 5801261f..fa2b565 100644 > --- a/virt/kvm/arm/vgic/vgic-init.c > +++ b/virt/kvm/arm/vgic/vgic-init.c > @@ -425,14 +425,16 @@ static int vgic_init_cpu_dying(unsigned int cpu) > return 0; > } > > -static irqreturn_t vgic_maintenance_handler(int irq, void *data) > +static irqreturn_t vgic_v3_maintenance_handler(int irq, void *data) > { > - /* > - * We cannot rely on the vgic maintenance interrupt to be > - * delivered synchronously. This means we can only use it to > - * exit the VM, and we perform the handling of EOIed > - * interrupts on the exit path (see vgic_process_maintenance). > - */ I always found this comment quite enlightening, especially as it points out that we need to deviate somewhat from the architectural idea here. I see that you have shortened it below. Is it no longer true? Can we keep the more elaborate version? If not here, then below? > + BUG(); /* Not implemented lazy save/restore on GICv3 */ > + return IRQ_HANDLED; > +} > + > +static irqreturn_t vgic_v2_maintenance_handler(int irq, void *data) > +{ > + struct kvm_vcpu *vcpu = (struct kvm_vcpu *)data; I think you need an empty line here, to separate variable declaration from code. > + vgic_v2_handle_maintenance(vcpu); > return IRQ_HANDLED; > } > > @@ -464,6 +466,7 @@ void kvm_vgic_init_cpu_hardware(void) > int kvm_vgic_hyp_init(void) > { > const struct gic_kvm_info *gic_kvm_info; > + irqreturn_t (*handler)(int irq, void *data); > int ret; > > gic_kvm_info = gic_get_kvm_info(); > @@ -478,6 +481,7 @@ int kvm_vgic_hyp_init(void) > switch (gic_kvm_info->type) { > case GIC_V2: > ret = vgic_v2_probe(gic_kvm_info); > + handler = vgic_v2_maintenance_handler; > break; 1> case GIC_V3: > ret = vgic_v3_probe(gic_kvm_info); > @@ -485,6 +489,7 @@ int kvm_vgic_hyp_init(void) > static_branch_enable(&kvm_vgic_global_state.gicv3_cpuif); > kvm_info("GIC system register CPU interface enabled\n"); > } > + handler = vgic_v3_maintenance_handler; > break; > default: > ret = -ENODEV; > @@ -494,8 +499,7 @@ int kvm_vgic_hyp_init(void) > return ret; > > kvm_vgic_global_state.maint_irq = gic_kvm_info->maint_irq; > - ret = request_percpu_irq(kvm_vgic_global_state.maint_irq, > - vgic_maintenance_handler, > + ret = request_percpu_irq(kvm_vgic_global_state.maint_irq, handler, > "vgic", kvm_get_running_vcpus()); > if (ret) { > kvm_err("Cannot register interrupt %d\n", > diff --git a/virt/kvm/arm/vgic/vgic-v2.c b/virt/kvm/arm/vgic/vgic-v2.c > index 8089710..259079b 100644 > --- a/virt/kvm/arm/vgic/vgic-v2.c > +++ b/virt/kvm/arm/vgic/vgic-v2.c > @@ -37,6 +37,17 @@ void vgic_v2_init_lrs(void) > vgic_v2_write_lr(i, 0); > } > > +void vgic_v2_handle_maintenance(struct kvm_vcpu *vcpu) > +{ > + void __iomem *base = kvm_vgic_global_state.vctrl_base; > + > + /* > + * Disable maintenance interrupt as we only use it to generate an exit > + * from the VM. > + */ Isn't that comment a bit misleading, as writing 0 to HCR not only disables all interrupt sources, but also the whole GICV interface altogether (bit 0: EN)? I see that it gets enabled later on when writing ->vgic_hcr into the register, but this function here looks a bit surprising to me. In general these changes to the interrupt handling leave me a bit puzzled. Should this be a separate patch? Or explained in the commit message? > + writel_relaxed(0, base + GICH_HCR); > +} > + > void vgic_v2_set_underflow(struct kvm_vcpu *vcpu) > { > struct vgic_v2_cpu_if *cpuif = &vcpu->arch.vgic_cpu.vgic_v2; > @@ -393,6 +404,87 @@ int vgic_v2_probe(const struct gic_kvm_info *info) > return ret; > } > > +static void save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) > +{ > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > + int nr_lr = kvm_vgic_global_state.nr_lr; > + u32 elrsr0, elrsr1; > + > + elrsr0 = readl_relaxed(base + GICH_ELRSR0); > + if (unlikely(nr_lr > 32)) > + elrsr1 = readl_relaxed(base + GICH_ELRSR1); > + else > + elrsr1 = 0; > + > +#ifdef CONFIG_CPU_BIG_ENDIAN > + cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; > +#else > + cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; > +#endif I have some gut feeling that this is really broken, since we mix up endian *byte* ordering with *bit* ordering here, don't we? I understand it's just copied and gets removed later on, so I was wondering if you could actually move patch 35/37 ("Get rid of vgic_elrsr") before this patch here, to avoid copying bogus code around? Or does 35/37 depend on 34/37 to be correct? > +} > + > +static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > +{ > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > + int i; > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > + > + for (i = 0; i < used_lrs; i++) { > + if (cpu_if->vgic_elrsr & (1UL << i)) > + cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > + else > + cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > + > + writel_relaxed(0, base + GICH_LR0 + (i * 4)); > + } > +} > + > +void vgic_v2_save_state(struct kvm_vcpu *vcpu) > +{ > + struct kvm *kvm = vcpu->kvm; > + struct vgic_dist *vgic = &kvm->arch.vgic; > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > + void __iomem *base = vgic->vctrl_base; > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > + > + if (!base) > + return; > + > + if (used_lrs) { > + cpu_if->vgic_apr = readl_relaxed(base + GICH_APR); > + > + save_elrsr(vcpu, base); > + save_lrs(vcpu, base); > + > + writel_relaxed(0, base + GICH_HCR); > + } else { > + cpu_if->vgic_elrsr = ~0UL; > + cpu_if->vgic_apr = 0; > + } > +} > + > +void vgic_v2_restore_state(struct kvm_vcpu *vcpu) > +{ > + struct kvm *kvm = vcpu->kvm; > + struct vgic_dist *vgic = &kvm->arch.vgic; > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > + void __iomem *base = vgic->vctrl_base; > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > + int i; > + > + if (!base) > + return; > + > + if (used_lrs) { > + writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); > + writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); > + for (i = 0; i < used_lrs; i++) { > + writel_relaxed(cpu_if->vgic_lr[i], > + base + GICH_LR0 + (i * 4)); > + } > + } > +} > + > void vgic_v2_load(struct kvm_vcpu *vcpu) > { > struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > diff --git a/virt/kvm/arm/vgic/vgic.c b/virt/kvm/arm/vgic/vgic.c > index b1bd238..790fd66 100644 > --- a/virt/kvm/arm/vgic/vgic.c > +++ b/virt/kvm/arm/vgic/vgic.c > @@ -18,6 +18,8 @@ > #include <linux/kvm_host.h> > #include <linux/list_sort.h> > > +#include <asm/kvm_hyp.h> Why do you need that? Commenting this out seems to compile anyway for me. > + > #include "vgic.h" > > #define CREATE_TRACE_POINTS > @@ -683,11 +685,19 @@ static void vgic_flush_lr_state(struct kvm_vcpu *vcpu) > vgic_clear_lr(vcpu, count); > } > > +static inline void vgic_save_state(struct kvm_vcpu *vcpu) Isn't "inline" frowned upon in .c files? > +{ > + if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > + vgic_v2_save_state(vcpu); > +} > + > /* Sync back the hardware VGIC state into our emulation after a guest's run. */ > void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu) > { > struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; > > + vgic_save_state(vcpu); > + > /* An empty ap_list_head implies used_lrs == 0 */ > if (list_empty(&vcpu->arch.vgic_cpu.ap_list_head)) > return; > @@ -697,6 +707,12 @@ void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu) > vgic_prune_ap_list(vcpu); > } > > +static inline void vgic_restore_state(struct kvm_vcpu *vcpu) Same here. Apart from those rather cosmetic issues I have at least verified that the code is actually moved from vgic-v2-sr.c to vgic-v2.c, plus/minus the required changes when moving this from HYP to vGIC/EL1. Cheers, Andre. > +{ > + if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > + vgic_v2_restore_state(vcpu); > +} > + > /* Flush our emulation state into the GIC hardware before entering the guest. */ > void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu) > { > @@ -710,13 +726,16 @@ void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu) > * this. > */ > if (list_empty(&vcpu->arch.vgic_cpu.ap_list_head)) > - return; > + goto out; > > DEBUG_SPINLOCK_BUG_ON(!irqs_disabled()); > > spin_lock(&vcpu->arch.vgic_cpu.ap_list_lock); > vgic_flush_lr_state(vcpu); > spin_unlock(&vcpu->arch.vgic_cpu.ap_list_lock); > + > +out: > + vgic_restore_state(vcpu); > } > > void kvm_vgic_load(struct kvm_vcpu *vcpu) > diff --git a/virt/kvm/arm/vgic/vgic.h b/virt/kvm/arm/vgic/vgic.h > index 4f8aecb..9895396 100644 > --- a/virt/kvm/arm/vgic/vgic.h > +++ b/virt/kvm/arm/vgic/vgic.h > @@ -167,6 +167,11 @@ int vgic_register_dist_iodev(struct kvm *kvm, gpa_t dist_base_address, > void vgic_v2_init_lrs(void); > void vgic_v2_load(struct kvm_vcpu *vcpu); > void vgic_v2_put(struct kvm_vcpu *vcpu); > +void vgic_v2_handle_maintenance(struct kvm_vcpu *vcpu); > + > +void vgic_v2_save_state(struct kvm_vcpu *vcpu); > +void vgic_v2_restore_state(struct kvm_vcpu *vcpu); > + > > static inline void vgic_get_irq_kref(struct vgic_irq *irq) > { >
On Wed, Nov 15, 2017 at 05:50:07PM +0000, Andre Przywara wrote: > Hi, > > those last few patches are actually helpful for the Xen port ... [...] > > +static void save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) > > +{ > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > + int nr_lr = kvm_vgic_global_state.nr_lr; > > + u32 elrsr0, elrsr1; > > + > > + elrsr0 = readl_relaxed(base + GICH_ELRSR0); > > + if (unlikely(nr_lr > 32)) > > + elrsr1 = readl_relaxed(base + GICH_ELRSR1); > > + else > > + elrsr1 = 0; > > + > > +#ifdef CONFIG_CPU_BIG_ENDIAN > > + cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; > > +#else > > + cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; > > +#endif > > I have some gut feeling that this is really broken, since we mix up > endian *byte* ordering with *bit* ordering here, don't we? Good feeling indeed. :) We have bitmap_{from,to)_u32array for things like this. But it was considered bad-designed, and I proposed new bitmap_{from,to)_arr32(). https://lkml.org/lkml/2017/11/15/592 What else I have in mind, to introduce something like bitmap_{from,to}_pair_32() as most of current users of bitmap_{from,to)_u32array(), (and those who should use it but don't, like this one) have only 2 32-bit halfwords to be copied from/to bitmap. Also, it will be complementary to bitmap_from_u64(). More reading about bitmap/array conversion is in comment to BITMAP_FROM_U64 macro. > I understand it's just copied and gets removed later on, so I was > wondering if you could actually move patch 35/37 ("Get rid of > vgic_elrsr") before this patch here, to avoid copying bogus code around? > Or does 35/37 depend on 34/37 to be correct? > > > +} > > + > > +static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > > +{ > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > + int i; > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > + > > + for (i = 0; i < used_lrs; i++) { > > + if (cpu_if->vgic_elrsr & (1UL << i)) So, the vgic_elrsr is naturally bitmap, and bitmap API is preferred if no other considerations: if (test_bit(i, cpu_if->vgic_elrsr)) > > + cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > > + else > > + cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > > + > > + writel_relaxed(0, base + GICH_LR0 + (i * 4)); > > + } > > +} I'd also headscratch about using for_each_clear_bit() here: /* * Setup default vgic_lr values somewhere earlier. * Not needed at all if you take my suggestion for * vgic_v2_restore_state() below */ for (i = 0; i < used_lrs; i++) cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) { [...] for_each_clear_bit (i, cpu_if->vgic_elrsr, used_lrs) cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); for (i = 0; i < used_lrs; i++) writel_relaxed(0, base + GICH_LR0 + (i * 4)); } Not sure how performance-critical this path is, but sometimes things get really faster with bitmaps. [...] > > +void vgic_v2_restore_state(struct kvm_vcpu *vcpu) > > +{ > > + struct kvm *kvm = vcpu->kvm; > > + struct vgic_dist *vgic = &kvm->arch.vgic; > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > + void __iomem *base = vgic->vctrl_base; > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > + int i; > > + > > + if (!base) > > + return; > > + > > + if (used_lrs) { > > + writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); > > + writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); > > + for (i = 0; i < used_lrs; i++) { > > + writel_relaxed(cpu_if->vgic_lr[i], > > + base + GICH_LR0 + (i * 4)); > > + } > > + } > > +} The alternative approach would be: for (i = 0; i < used_lrs; i++) { if (test_bit(i, cpu_if->vgic_elrsr)) writel_relaxed(~GICH_LR_STATE, base + GICH_LR0 + (i * 4)); else writel_relaxed(cpu_if->vgic_lr[i], base + GICH_LR0 + (i * 4)); } If cpu_if->vgic_elrsr is untouched in-between of course. It will make save_lrs() simpler and this function more verbose. Thanks, Yury
On Wed, Nov 15, 2017 at 05:50:07PM +0000, Andre Przywara wrote: > Hi, > > those last few patches are actually helpful for the Xen port ... cool! > > On 12/10/17 11:41, Christoffer Dall wrote: > > We can program the GICv2 hypervisor control interface logic directly > > from the core vgic code and can instead do the save/restore directly > > from the flush/sync functions, which can lead to a number of future > > optimizations. > > > > Signed-off-by: Christoffer Dall <christoffer.dall@linaro.org> > > --- > > arch/arm/kvm/hyp/switch.c | 4 -- > > arch/arm64/include/asm/kvm_hyp.h | 2 - > > arch/arm64/kvm/hyp/switch.c | 4 -- > > virt/kvm/arm/hyp/vgic-v2-sr.c | 83 ------------------------------------ > > virt/kvm/arm/vgic/vgic-init.c | 22 ++++++---- > > virt/kvm/arm/vgic/vgic-v2.c | 92 ++++++++++++++++++++++++++++++++++++++++ > > virt/kvm/arm/vgic/vgic.c | 21 ++++++++- > > virt/kvm/arm/vgic/vgic.h | 5 +++ > > 8 files changed, 130 insertions(+), 103 deletions(-) > > > > diff --git a/arch/arm/kvm/hyp/switch.c b/arch/arm/kvm/hyp/switch.c > > index c3b9799..0d834f8 100644 > > --- a/arch/arm/kvm/hyp/switch.c > > +++ b/arch/arm/kvm/hyp/switch.c > > @@ -91,16 +91,12 @@ static void __hyp_text __vgic_save_state(struct kvm_vcpu *vcpu) > > { > > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > > __vgic_v3_save_state(vcpu); > > - else > > - __vgic_v2_save_state(vcpu); > > } > > > > static void __hyp_text __vgic_restore_state(struct kvm_vcpu *vcpu) > > { > > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > > __vgic_v3_restore_state(vcpu); > > - else > > - __vgic_v2_restore_state(vcpu); > > } > > > > static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) > > diff --git a/arch/arm64/include/asm/kvm_hyp.h b/arch/arm64/include/asm/kvm_hyp.h > > index 28d5f3c..bd3fe64 100644 > > --- a/arch/arm64/include/asm/kvm_hyp.h > > +++ b/arch/arm64/include/asm/kvm_hyp.h > > @@ -121,8 +121,6 @@ typeof(orig) * __hyp_text fname(void) \ > > return val; \ > > } > > > > -void __vgic_v2_save_state(struct kvm_vcpu *vcpu); > > -void __vgic_v2_restore_state(struct kvm_vcpu *vcpu); > > int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu); > > > > void __vgic_v3_save_state(struct kvm_vcpu *vcpu); > > diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c > > index 5692aa0..90da506 100644 > > --- a/arch/arm64/kvm/hyp/switch.c > > +++ b/arch/arm64/kvm/hyp/switch.c > > @@ -186,16 +186,12 @@ static void __hyp_text __vgic_save_state(struct kvm_vcpu *vcpu) > > { > > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > > __vgic_v3_save_state(vcpu); > > - else > > - __vgic_v2_save_state(vcpu); > > } > > > > static void __hyp_text __vgic_restore_state(struct kvm_vcpu *vcpu) > > { > > if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > > __vgic_v3_restore_state(vcpu); > > - else > > - __vgic_v2_restore_state(vcpu); > > } > > > > static bool __hyp_text __true_value(void) > > diff --git a/virt/kvm/arm/hyp/vgic-v2-sr.c b/virt/kvm/arm/hyp/vgic-v2-sr.c > > index a3f18d3..b433257 100644 > > --- a/virt/kvm/arm/hyp/vgic-v2-sr.c > > +++ b/virt/kvm/arm/hyp/vgic-v2-sr.c > > @@ -22,89 +22,6 @@ > > #include <asm/kvm_emulate.h> > > #include <asm/kvm_hyp.h> > > > > -static void __hyp_text save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) > > -{ > > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > - int nr_lr = (kern_hyp_va(&kvm_vgic_global_state))->nr_lr; > > - u32 elrsr0, elrsr1; > > - > > - elrsr0 = readl_relaxed(base + GICH_ELRSR0); > > - if (unlikely(nr_lr > 32)) > > - elrsr1 = readl_relaxed(base + GICH_ELRSR1); > > - else > > - elrsr1 = 0; > > - > > -#ifdef CONFIG_CPU_BIG_ENDIAN > > - cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; > > -#else > > - cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; > > -#endif > > -} > > - > > -static void __hyp_text save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > > -{ > > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > - int i; > > - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > - > > - for (i = 0; i < used_lrs; i++) { > > - if (cpu_if->vgic_elrsr & (1UL << i)) > > - cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > > - else > > - cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > > - > > - writel_relaxed(0, base + GICH_LR0 + (i * 4)); > > - } > > -} > > - > > -/* vcpu is already in the HYP VA space */ > > -void __hyp_text __vgic_v2_save_state(struct kvm_vcpu *vcpu) > > -{ > > - struct kvm *kvm = kern_hyp_va(vcpu->kvm); > > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > - struct vgic_dist *vgic = &kvm->arch.vgic; > > - void __iomem *base = kern_hyp_va(vgic->vctrl_base); > > - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > - > > - if (!base) > > - return; > > - > > - if (used_lrs) { > > - cpu_if->vgic_apr = readl_relaxed(base + GICH_APR); > > - > > - save_elrsr(vcpu, base); > > - save_lrs(vcpu, base); > > - > > - writel_relaxed(0, base + GICH_HCR); > > - } else { > > - cpu_if->vgic_elrsr = ~0UL; > > - cpu_if->vgic_apr = 0; > > - } > > -} > > - > > -/* vcpu is already in the HYP VA space */ > > -void __hyp_text __vgic_v2_restore_state(struct kvm_vcpu *vcpu) > > -{ > > - struct kvm *kvm = kern_hyp_va(vcpu->kvm); > > - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > - struct vgic_dist *vgic = &kvm->arch.vgic; > > - void __iomem *base = kern_hyp_va(vgic->vctrl_base); > > - int i; > > - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > - > > - if (!base) > > - return; > > - > > - if (used_lrs) { > > - writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); > > - writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); > > - for (i = 0; i < used_lrs; i++) { > > - writel_relaxed(cpu_if->vgic_lr[i], > > - base + GICH_LR0 + (i * 4)); > > - } > > - } > > -} > > - > > #ifdef CONFIG_ARM64 > > /* > > * __vgic_v2_perform_cpuif_access -- perform a GICV access on behalf of the > > diff --git a/virt/kvm/arm/vgic/vgic-init.c b/virt/kvm/arm/vgic/vgic-init.c > > index 5801261f..fa2b565 100644 > > --- a/virt/kvm/arm/vgic/vgic-init.c > > +++ b/virt/kvm/arm/vgic/vgic-init.c > > @@ -425,14 +425,16 @@ static int vgic_init_cpu_dying(unsigned int cpu) > > return 0; > > } > > > > -static irqreturn_t vgic_maintenance_handler(int irq, void *data) > > +static irqreturn_t vgic_v3_maintenance_handler(int irq, void *data) > > { > > - /* > > - * We cannot rely on the vgic maintenance interrupt to be > > - * delivered synchronously. This means we can only use it to > > - * exit the VM, and we perform the handling of EOIed > > - * interrupts on the exit path (see vgic_process_maintenance). > > - */ > > I always found this comment quite enlightening, especially as it points > out that we need to deviate somewhat from the architectural idea here. > I see that you have shortened it below. Is it no longer true? Can we > keep the more elaborate version? If not here, then below? > I think I found it a bit misleading or harder to understand now that we may actually do some work in the maintenance handler (disable the vgic). But if you think it's better to keep it, I'm happy to make that change for v2. > > + BUG(); /* Not implemented lazy save/restore on GICv3 */ > > + return IRQ_HANDLED; > > +} > > + > > +static irqreturn_t vgic_v2_maintenance_handler(int irq, void *data) > > +{ > > + struct kvm_vcpu *vcpu = (struct kvm_vcpu *)data; > > I think you need an empty line here, to separate variable declaration > from code. > sure > > + vgic_v2_handle_maintenance(vcpu); > > > return IRQ_HANDLED; > > } > > > > @@ -464,6 +466,7 @@ void kvm_vgic_init_cpu_hardware(void) > > int kvm_vgic_hyp_init(void) > > { > > const struct gic_kvm_info *gic_kvm_info; > > + irqreturn_t (*handler)(int irq, void *data); > > int ret; > > > > gic_kvm_info = gic_get_kvm_info(); > > @@ -478,6 +481,7 @@ int kvm_vgic_hyp_init(void) > > switch (gic_kvm_info->type) { > > case GIC_V2: > > ret = vgic_v2_probe(gic_kvm_info); > > + handler = vgic_v2_maintenance_handler; > > break; > 1> case GIC_V3: > > ret = vgic_v3_probe(gic_kvm_info); > > @@ -485,6 +489,7 @@ int kvm_vgic_hyp_init(void) > > static_branch_enable(&kvm_vgic_global_state.gicv3_cpuif); > > kvm_info("GIC system register CPU interface enabled\n"); > > } > > + handler = vgic_v3_maintenance_handler; > > break; > > default: > > ret = -ENODEV; > > @@ -494,8 +499,7 @@ int kvm_vgic_hyp_init(void) > > return ret; > > > > kvm_vgic_global_state.maint_irq = gic_kvm_info->maint_irq; > > - ret = request_percpu_irq(kvm_vgic_global_state.maint_irq, > > - vgic_maintenance_handler, > > + ret = request_percpu_irq(kvm_vgic_global_state.maint_irq, handler, > > "vgic", kvm_get_running_vcpus()); > > if (ret) { > > kvm_err("Cannot register interrupt %d\n", > > diff --git a/virt/kvm/arm/vgic/vgic-v2.c b/virt/kvm/arm/vgic/vgic-v2.c > > index 8089710..259079b 100644 > > --- a/virt/kvm/arm/vgic/vgic-v2.c > > +++ b/virt/kvm/arm/vgic/vgic-v2.c > > @@ -37,6 +37,17 @@ void vgic_v2_init_lrs(void) > > vgic_v2_write_lr(i, 0); > > } > > > > +void vgic_v2_handle_maintenance(struct kvm_vcpu *vcpu) > > +{ > > + void __iomem *base = kvm_vgic_global_state.vctrl_base; > > + > > + /* > > + * Disable maintenance interrupt as we only use it to generate an exit > > + * from the VM. > > + */ > > Isn't that comment a bit misleading, as writing 0 to HCR not only > disables all interrupt sources, but also the whole GICV interface > altogether (bit 0: EN)? > I see that it gets enabled later on when writing ->vgic_hcr into the > register, but this function here looks a bit surprising to me. I don't think the comment is misleading and I'm not trying to explain how the whole hardware works, but the intention behind this code. I'll try to expand this comment and the commit message. The point is that we want to get back out of the hyp code (or just back to the vgic sync function on VHE) so that we in the case of no interrupts in flight, only check a single variable, and then in the case of having interrupts in flight, manage the hardware then. That means that we leave the VGIC on, which in turn means that we may now actually see the maintenance interrupt firing, and in that case, we want to disable the VGIC to prevent continuous maintenance interrupts from firing, but we still only handle the actual maintenance (resample the line) in the sync function. I'll try to find some nice way of explaining this. > > In general these changes to the interrupt handling leave me a bit > puzzled. Should this be a separate patch? Or explained in the commit > message? > Should definitely be explained. I tried splitting the logic, but couldn't find a nice working way to do that. I'll have another look. > > + writel_relaxed(0, base + GICH_HCR); > > +} > > + > > void vgic_v2_set_underflow(struct kvm_vcpu *vcpu) > > { > > struct vgic_v2_cpu_if *cpuif = &vcpu->arch.vgic_cpu.vgic_v2; > > @@ -393,6 +404,87 @@ int vgic_v2_probe(const struct gic_kvm_info *info) > > return ret; > > } > > > > +static void save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) > > +{ > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > + int nr_lr = kvm_vgic_global_state.nr_lr; > > + u32 elrsr0, elrsr1; > > + > > + elrsr0 = readl_relaxed(base + GICH_ELRSR0); > > + if (unlikely(nr_lr > 32)) > > + elrsr1 = readl_relaxed(base + GICH_ELRSR1); > > + else > > + elrsr1 = 0; > > + > > +#ifdef CONFIG_CPU_BIG_ENDIAN > > + cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; > > +#else > > + cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; > > +#endif > > I have some gut feeling that this is really broken, since we mix up > endian *byte* ordering with *bit* ordering here, don't we? Yes, it's broken, as I also noted in the commit message later. > > I understand it's just copied and gets removed later on, so I was > wondering if you could actually move patch 35/37 ("Get rid of > vgic_elrsr") before this patch here, to avoid copying bogus code around? > Or does 35/37 depend on 34/37 to be correct? > I don't remember. I'll give it a try. > > +} > > + > > +static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > > +{ > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > + int i; > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > + > > + for (i = 0; i < used_lrs; i++) { > > + if (cpu_if->vgic_elrsr & (1UL << i)) > > + cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > > + else > > + cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > > + > > + writel_relaxed(0, base + GICH_LR0 + (i * 4)); > > + } > > +} > > + > > +void vgic_v2_save_state(struct kvm_vcpu *vcpu) > > +{ > > + struct kvm *kvm = vcpu->kvm; > > + struct vgic_dist *vgic = &kvm->arch.vgic; > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > + void __iomem *base = vgic->vctrl_base; > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > + > > + if (!base) > > + return; > > + > > + if (used_lrs) { > > + cpu_if->vgic_apr = readl_relaxed(base + GICH_APR); > > + > > + save_elrsr(vcpu, base); > > + save_lrs(vcpu, base); > > + > > + writel_relaxed(0, base + GICH_HCR); > > + } else { > > + cpu_if->vgic_elrsr = ~0UL; > > + cpu_if->vgic_apr = 0; > > + } > > +} > > + > > +void vgic_v2_restore_state(struct kvm_vcpu *vcpu) > > +{ > > + struct kvm *kvm = vcpu->kvm; > > + struct vgic_dist *vgic = &kvm->arch.vgic; > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > + void __iomem *base = vgic->vctrl_base; > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > + int i; > > + > > + if (!base) > > + return; > > + > > + if (used_lrs) { > > + writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); > > + writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); > > + for (i = 0; i < used_lrs; i++) { > > + writel_relaxed(cpu_if->vgic_lr[i], > > + base + GICH_LR0 + (i * 4)); > > + } > > + } > > +} > > + > > void vgic_v2_load(struct kvm_vcpu *vcpu) > > { > > struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > diff --git a/virt/kvm/arm/vgic/vgic.c b/virt/kvm/arm/vgic/vgic.c > > index b1bd238..790fd66 100644 > > --- a/virt/kvm/arm/vgic/vgic.c > > +++ b/virt/kvm/arm/vgic/vgic.c > > @@ -18,6 +18,8 @@ > > #include <linux/kvm_host.h> > > #include <linux/list_sort.h> > > > > +#include <asm/kvm_hyp.h> > > Why do you need that? Commenting this out seems to compile anyway for me. > Probably while I was rearranging patches this got in there somehow. > > + > > #include "vgic.h" > > > > #define CREATE_TRACE_POINTS > > @@ -683,11 +685,19 @@ static void vgic_flush_lr_state(struct kvm_vcpu *vcpu) > > vgic_clear_lr(vcpu, count); > > } > > > > +static inline void vgic_save_state(struct kvm_vcpu *vcpu) > > Isn't "inline" frowned upon in .c files? > Opinions seem to vary on this one. I actually observed this making a difference in the compilation for some of these smaller functions with my compiler so I included it, and there are certainly examples of this in the kernel. If there's a policy I missed or someone has a strong argument that I have to remove the inline word here, then let me know. > > +{ > > + if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) > > + vgic_v2_save_state(vcpu); > > +} > > + > > /* Sync back the hardware VGIC state into our emulation after a guest's run. */ > > void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu) > > { > > struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; > > > > + vgic_save_state(vcpu); > > + > > /* An empty ap_list_head implies used_lrs == 0 */ > > if (list_empty(&vcpu->arch.vgic_cpu.ap_list_head)) > > return; > > @@ -697,6 +707,12 @@ void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu) > > vgic_prune_ap_list(vcpu); > > } > > > > +static inline void vgic_restore_state(struct kvm_vcpu *vcpu) > > Same here. > > Apart from those rather cosmetic issues I have at least verified that > the code is actually moved from vgic-v2-sr.c to vgic-v2.c, plus/minus > the required changes when moving this from HYP to vGIC/EL1. > Thanks for having a look! -Christoffer
On Sun, Nov 26, 2017 at 01:29:30PM +0300, Yury Norov wrote: > On Wed, Nov 15, 2017 at 05:50:07PM +0000, Andre Przywara wrote: > > Hi, > > > > those last few patches are actually helpful for the Xen port ... > > [...] > > > > +static void save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) > > > +{ > > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > > + int nr_lr = kvm_vgic_global_state.nr_lr; > > > + u32 elrsr0, elrsr1; > > > + > > > + elrsr0 = readl_relaxed(base + GICH_ELRSR0); > > > + if (unlikely(nr_lr > 32)) > > > + elrsr1 = readl_relaxed(base + GICH_ELRSR1); > > > + else > > > + elrsr1 = 0; > > > + > > > +#ifdef CONFIG_CPU_BIG_ENDIAN > > > + cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; > > > +#else > > > + cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; > > > +#endif > > > > I have some gut feeling that this is really broken, since we mix up > > endian *byte* ordering with *bit* ordering here, don't we? > > Good feeling indeed. :) > > We have bitmap_{from,to)_u32array for things like this. But it was > considered bad-designed, and I proposed new bitmap_{from,to)_arr32(). > > https://lkml.org/lkml/2017/11/15/592 > > What else I have in mind, to introduce something like bitmap_{from,to}_pair_32() > as most of current users of bitmap_{from,to)_u32array(), (and those who should > use it but don't, like this one) have only 2 32-bit halfwords to be copied > from/to bitmap. > > Also, it will be complementary to bitmap_from_u64(). > > More reading about bitmap/array conversion is in comment to BITMAP_FROM_U64 > macro. > I have no idea what you want to introduce here. If you have an idea on how to improve the code, patches are welcome. Please keep in mind, that the purpose of this patch is to move code around to improve the GIC handling performance, not changing the lower-level details of the code. > > I understand it's just copied and gets removed later on, so I was > > wondering if you could actually move patch 35/37 ("Get rid of > > vgic_elrsr") before this patch here, to avoid copying bogus code around? > > Or does 35/37 depend on 34/37 to be correct? > > > > > +} > > > + > > > +static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > > > +{ > > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > > + int i; > > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > > + > > > + for (i = 0; i < used_lrs; i++) { > > > + if (cpu_if->vgic_elrsr & (1UL << i)) > > So, the vgic_elrsr is naturally bitmap, and bitmap API is preferred if no > other considerations: > if (test_bit(i, cpu_if->vgic_elrsr)) > > > > + cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > > > + else > > > + cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > > > + > > > + writel_relaxed(0, base + GICH_LR0 + (i * 4)); > > > + } > > > +} > > I'd also headscratch about using for_each_clear_bit() here: > > /* > * Setup default vgic_lr values somewhere earlier. Not sure what the 'default' values are. > * Not needed at all if you take my suggestion for > * vgic_v2_restore_state() below > */ > for (i = 0; i < used_lrs; i++) > cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > > static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > { > [...] > > for_each_clear_bit (i, cpu_if->vgic_elrsr, used_lrs) > cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > > for (i = 0; i < used_lrs; i++) > writel_relaxed(0, base + GICH_LR0 + (i * 4)); > } > > Not sure how performance-critical this path is, but sometimes things > get really faster with bitmaps. > Your suggestion below would require us to maintain elrsr when we setup list registers, and I don't really see the benefit. > [...] > > > > +void vgic_v2_restore_state(struct kvm_vcpu *vcpu) > > > +{ > > > + struct kvm *kvm = vcpu->kvm; > > > + struct vgic_dist *vgic = &kvm->arch.vgic; > > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > > + void __iomem *base = vgic->vctrl_base; > > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > > + int i; > > > + > > > + if (!base) > > > + return; > > > + > > > + if (used_lrs) { > > > + writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); > > > + writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); > > > + for (i = 0; i < used_lrs; i++) { > > > + writel_relaxed(cpu_if->vgic_lr[i], > > > + base + GICH_LR0 + (i * 4)); > > > + } > > > + } > > > +} > > The alternative approach would be: > for (i = 0; i < used_lrs; i++) { > if (test_bit(i, cpu_if->vgic_elrsr)) > writel_relaxed(~GICH_LR_STATE, base + GICH_LR0 + (i * 4)); > else > writel_relaxed(cpu_if->vgic_lr[i], base + GICH_LR0 + (i * 4)); > } > > If cpu_if->vgic_elrsr is untouched in-between of course. It will make > save_lrs() simpler and this function more verbose. > I don't understand your suggestion. As you will see later, we will get rid of storing the elrsr completely with a measureable performance improvement. If you think you can improve the code beyond that, a follow-up patch would be most welcome. Note that on all the implementations I'm familiar with, the maximum number of LRs is four, so we're not wading through massive bitmaps in practice here. Thanks, -Christoffer
On Sun, Nov 26, 2017 at 08:46:41PM +0100, Christoffer Dall wrote: > On Sun, Nov 26, 2017 at 01:29:30PM +0300, Yury Norov wrote: > > On Wed, Nov 15, 2017 at 05:50:07PM +0000, Andre Przywara wrote: > > > Hi, > > > > > > those last few patches are actually helpful for the Xen port ... > > > > [...] > > > > > > +static void save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) > > > > +{ > > > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > > > + int nr_lr = kvm_vgic_global_state.nr_lr; > > > > + u32 elrsr0, elrsr1; > > > > + > > > > + elrsr0 = readl_relaxed(base + GICH_ELRSR0); > > > > + if (unlikely(nr_lr > 32)) > > > > + elrsr1 = readl_relaxed(base + GICH_ELRSR1); > > > > + else > > > > + elrsr1 = 0; > > > > + > > > > +#ifdef CONFIG_CPU_BIG_ENDIAN > > > > + cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; > > > > +#else > > > > + cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; > > > > +#endif > > > > > > I have some gut feeling that this is really broken, since we mix up > > > endian *byte* ordering with *bit* ordering here, don't we? > > > > Good feeling indeed. :) > > > > We have bitmap_{from,to)_u32array for things like this. But it was > > considered bad-designed, and I proposed new bitmap_{from,to)_arr32(). > > > > https://lkml.org/lkml/2017/11/15/592 > > > > What else I have in mind, to introduce something like bitmap_{from,to}_pair_32() > > as most of current users of bitmap_{from,to)_u32array(), (and those who should > > use it but don't, like this one) have only 2 32-bit halfwords to be copied > > from/to bitmap. > > > > Also, it will be complementary to bitmap_from_u64(). > > > > More reading about bitmap/array conversion is in comment to BITMAP_FROM_U64 > > macro. > > > > I have no idea what you want to introduce here. If you have an idea on > how to improve the code, patches are welcome. That's about Andre's gut feeling, not about your patch. I have some ideas related to it, and just want to share it to him - that's all. > Please keep in mind, that the purpose of this patch is to move code > around to improve the GIC handling performance, not changing the > lower-level details of the code. > > > > I understand it's just copied and gets removed later on, so I was > > > wondering if you could actually move patch 35/37 ("Get rid of > > > vgic_elrsr") before this patch here, to avoid copying bogus code around? > > > Or does 35/37 depend on 34/37 to be correct? > > > > > > > +} > > > > + > > > > +static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > > > > +{ > > > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > > > + int i; > > > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > > > + > > > > + for (i = 0; i < used_lrs; i++) { > > > > + if (cpu_if->vgic_elrsr & (1UL << i)) > > > > So, the vgic_elrsr is naturally bitmap, and bitmap API is preferred if no > > other considerations: > > if (test_bit(i, cpu_if->vgic_elrsr)) > > > > > > + cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > > > > + else > > > > + cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > > > > + > > > > + writel_relaxed(0, base + GICH_LR0 + (i * 4)); > > > > + } > > > > +} > > > > I'd also headscratch about using for_each_clear_bit() here: > > > > /* > > * Setup default vgic_lr values somewhere earlier. > > Not sure what the 'default' values are. > > > * Not needed at all if you take my suggestion for > > * vgic_v2_restore_state() below > > */ > > for (i = 0; i < used_lrs; i++) > > cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; > > > > static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) > > { > > [...] > > > > for_each_clear_bit (i, cpu_if->vgic_elrsr, used_lrs) > > cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); > > > > for (i = 0; i < used_lrs; i++) > > writel_relaxed(0, base + GICH_LR0 + (i * 4)); > > } > > > > Not sure how performance-critical this path is, but sometimes things > > get really faster with bitmaps. > > > > Your suggestion below would require us to maintain elrsr when we setup > list registers, and I don't really see the benefit. That's what I asked - is it maintained or not. If not then it will not work. > > [...] > > > > > > +void vgic_v2_restore_state(struct kvm_vcpu *vcpu) > > > > +{ > > > > + struct kvm *kvm = vcpu->kvm; > > > > + struct vgic_dist *vgic = &kvm->arch.vgic; > > > > + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; > > > > + void __iomem *base = vgic->vctrl_base; > > > > + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; > > > > + int i; > > > > + > > > > + if (!base) > > > > + return; > > > > + > > > > + if (used_lrs) { > > > > + writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); > > > > + writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); > > > > + for (i = 0; i < used_lrs; i++) { > > > > + writel_relaxed(cpu_if->vgic_lr[i], > > > > + base + GICH_LR0 + (i * 4)); > > > > + } > > > > + } > > > > +} > > > > The alternative approach would be: > > for (i = 0; i < used_lrs; i++) { > > if (test_bit(i, cpu_if->vgic_elrsr)) > > writel_relaxed(~GICH_LR_STATE, base + GICH_LR0 + (i * 4)); > > else > > writel_relaxed(cpu_if->vgic_lr[i], base + GICH_LR0 + (i * 4)); > > } > > > > If cpu_if->vgic_elrsr is untouched in-between of course. It will make > > save_lrs() simpler and this function more verbose. > > > I don't understand your suggestion. As you will see later, we will get > rid of storing the elrsr completely with a measureable performance > improvement. OK, now I see. Sorry for stupid questions - I just start learning codebase. By the way, can you share the technique that you use to measure performance? It would be great if I can reproduce your results. > If you think you can improve the code beyond that, a follow-up patch > would be most welcome. > > Note that on all the implementations I'm familiar with, the maximum > number of LRs is four, so we're not wading through massive bitmaps in > practice here. > > Thanks, > -Christoffer
diff --git a/arch/arm/kvm/hyp/switch.c b/arch/arm/kvm/hyp/switch.c index c3b9799..0d834f8 100644 --- a/arch/arm/kvm/hyp/switch.c +++ b/arch/arm/kvm/hyp/switch.c @@ -91,16 +91,12 @@ static void __hyp_text __vgic_save_state(struct kvm_vcpu *vcpu) { if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) __vgic_v3_save_state(vcpu); - else - __vgic_v2_save_state(vcpu); } static void __hyp_text __vgic_restore_state(struct kvm_vcpu *vcpu) { if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) __vgic_v3_restore_state(vcpu); - else - __vgic_v2_restore_state(vcpu); } static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) diff --git a/arch/arm64/include/asm/kvm_hyp.h b/arch/arm64/include/asm/kvm_hyp.h index 28d5f3c..bd3fe64 100644 --- a/arch/arm64/include/asm/kvm_hyp.h +++ b/arch/arm64/include/asm/kvm_hyp.h @@ -121,8 +121,6 @@ typeof(orig) * __hyp_text fname(void) \ return val; \ } -void __vgic_v2_save_state(struct kvm_vcpu *vcpu); -void __vgic_v2_restore_state(struct kvm_vcpu *vcpu); int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu); void __vgic_v3_save_state(struct kvm_vcpu *vcpu); diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c index 5692aa0..90da506 100644 --- a/arch/arm64/kvm/hyp/switch.c +++ b/arch/arm64/kvm/hyp/switch.c @@ -186,16 +186,12 @@ static void __hyp_text __vgic_save_state(struct kvm_vcpu *vcpu) { if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) __vgic_v3_save_state(vcpu); - else - __vgic_v2_save_state(vcpu); } static void __hyp_text __vgic_restore_state(struct kvm_vcpu *vcpu) { if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) __vgic_v3_restore_state(vcpu); - else - __vgic_v2_restore_state(vcpu); } static bool __hyp_text __true_value(void) diff --git a/virt/kvm/arm/hyp/vgic-v2-sr.c b/virt/kvm/arm/hyp/vgic-v2-sr.c index a3f18d3..b433257 100644 --- a/virt/kvm/arm/hyp/vgic-v2-sr.c +++ b/virt/kvm/arm/hyp/vgic-v2-sr.c @@ -22,89 +22,6 @@ #include <asm/kvm_emulate.h> #include <asm/kvm_hyp.h> -static void __hyp_text save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) -{ - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; - int nr_lr = (kern_hyp_va(&kvm_vgic_global_state))->nr_lr; - u32 elrsr0, elrsr1; - - elrsr0 = readl_relaxed(base + GICH_ELRSR0); - if (unlikely(nr_lr > 32)) - elrsr1 = readl_relaxed(base + GICH_ELRSR1); - else - elrsr1 = 0; - -#ifdef CONFIG_CPU_BIG_ENDIAN - cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; -#else - cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; -#endif -} - -static void __hyp_text save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) -{ - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; - int i; - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; - - for (i = 0; i < used_lrs; i++) { - if (cpu_if->vgic_elrsr & (1UL << i)) - cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; - else - cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); - - writel_relaxed(0, base + GICH_LR0 + (i * 4)); - } -} - -/* vcpu is already in the HYP VA space */ -void __hyp_text __vgic_v2_save_state(struct kvm_vcpu *vcpu) -{ - struct kvm *kvm = kern_hyp_va(vcpu->kvm); - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; - struct vgic_dist *vgic = &kvm->arch.vgic; - void __iomem *base = kern_hyp_va(vgic->vctrl_base); - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; - - if (!base) - return; - - if (used_lrs) { - cpu_if->vgic_apr = readl_relaxed(base + GICH_APR); - - save_elrsr(vcpu, base); - save_lrs(vcpu, base); - - writel_relaxed(0, base + GICH_HCR); - } else { - cpu_if->vgic_elrsr = ~0UL; - cpu_if->vgic_apr = 0; - } -} - -/* vcpu is already in the HYP VA space */ -void __hyp_text __vgic_v2_restore_state(struct kvm_vcpu *vcpu) -{ - struct kvm *kvm = kern_hyp_va(vcpu->kvm); - struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; - struct vgic_dist *vgic = &kvm->arch.vgic; - void __iomem *base = kern_hyp_va(vgic->vctrl_base); - int i; - u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; - - if (!base) - return; - - if (used_lrs) { - writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); - writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); - for (i = 0; i < used_lrs; i++) { - writel_relaxed(cpu_if->vgic_lr[i], - base + GICH_LR0 + (i * 4)); - } - } -} - #ifdef CONFIG_ARM64 /* * __vgic_v2_perform_cpuif_access -- perform a GICV access on behalf of the diff --git a/virt/kvm/arm/vgic/vgic-init.c b/virt/kvm/arm/vgic/vgic-init.c index 5801261f..fa2b565 100644 --- a/virt/kvm/arm/vgic/vgic-init.c +++ b/virt/kvm/arm/vgic/vgic-init.c @@ -425,14 +425,16 @@ static int vgic_init_cpu_dying(unsigned int cpu) return 0; } -static irqreturn_t vgic_maintenance_handler(int irq, void *data) +static irqreturn_t vgic_v3_maintenance_handler(int irq, void *data) { - /* - * We cannot rely on the vgic maintenance interrupt to be - * delivered synchronously. This means we can only use it to - * exit the VM, and we perform the handling of EOIed - * interrupts on the exit path (see vgic_process_maintenance). - */ + BUG(); /* Not implemented lazy save/restore on GICv3 */ + return IRQ_HANDLED; +} + +static irqreturn_t vgic_v2_maintenance_handler(int irq, void *data) +{ + struct kvm_vcpu *vcpu = (struct kvm_vcpu *)data; + vgic_v2_handle_maintenance(vcpu); return IRQ_HANDLED; } @@ -464,6 +466,7 @@ void kvm_vgic_init_cpu_hardware(void) int kvm_vgic_hyp_init(void) { const struct gic_kvm_info *gic_kvm_info; + irqreturn_t (*handler)(int irq, void *data); int ret; gic_kvm_info = gic_get_kvm_info(); @@ -478,6 +481,7 @@ int kvm_vgic_hyp_init(void) switch (gic_kvm_info->type) { case GIC_V2: ret = vgic_v2_probe(gic_kvm_info); + handler = vgic_v2_maintenance_handler; break; case GIC_V3: ret = vgic_v3_probe(gic_kvm_info); @@ -485,6 +489,7 @@ int kvm_vgic_hyp_init(void) static_branch_enable(&kvm_vgic_global_state.gicv3_cpuif); kvm_info("GIC system register CPU interface enabled\n"); } + handler = vgic_v3_maintenance_handler; break; default: ret = -ENODEV; @@ -494,8 +499,7 @@ int kvm_vgic_hyp_init(void) return ret; kvm_vgic_global_state.maint_irq = gic_kvm_info->maint_irq; - ret = request_percpu_irq(kvm_vgic_global_state.maint_irq, - vgic_maintenance_handler, + ret = request_percpu_irq(kvm_vgic_global_state.maint_irq, handler, "vgic", kvm_get_running_vcpus()); if (ret) { kvm_err("Cannot register interrupt %d\n", diff --git a/virt/kvm/arm/vgic/vgic-v2.c b/virt/kvm/arm/vgic/vgic-v2.c index 8089710..259079b 100644 --- a/virt/kvm/arm/vgic/vgic-v2.c +++ b/virt/kvm/arm/vgic/vgic-v2.c @@ -37,6 +37,17 @@ void vgic_v2_init_lrs(void) vgic_v2_write_lr(i, 0); } +void vgic_v2_handle_maintenance(struct kvm_vcpu *vcpu) +{ + void __iomem *base = kvm_vgic_global_state.vctrl_base; + + /* + * Disable maintenance interrupt as we only use it to generate an exit + * from the VM. + */ + writel_relaxed(0, base + GICH_HCR); +} + void vgic_v2_set_underflow(struct kvm_vcpu *vcpu) { struct vgic_v2_cpu_if *cpuif = &vcpu->arch.vgic_cpu.vgic_v2; @@ -393,6 +404,87 @@ int vgic_v2_probe(const struct gic_kvm_info *info) return ret; } +static void save_elrsr(struct kvm_vcpu *vcpu, void __iomem *base) +{ + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; + int nr_lr = kvm_vgic_global_state.nr_lr; + u32 elrsr0, elrsr1; + + elrsr0 = readl_relaxed(base + GICH_ELRSR0); + if (unlikely(nr_lr > 32)) + elrsr1 = readl_relaxed(base + GICH_ELRSR1); + else + elrsr1 = 0; + +#ifdef CONFIG_CPU_BIG_ENDIAN + cpu_if->vgic_elrsr = ((u64)elrsr0 << 32) | elrsr1; +#else + cpu_if->vgic_elrsr = ((u64)elrsr1 << 32) | elrsr0; +#endif +} + +static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base) +{ + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; + int i; + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; + + for (i = 0; i < used_lrs; i++) { + if (cpu_if->vgic_elrsr & (1UL << i)) + cpu_if->vgic_lr[i] &= ~GICH_LR_STATE; + else + cpu_if->vgic_lr[i] = readl_relaxed(base + GICH_LR0 + (i * 4)); + + writel_relaxed(0, base + GICH_LR0 + (i * 4)); + } +} + +void vgic_v2_save_state(struct kvm_vcpu *vcpu) +{ + struct kvm *kvm = vcpu->kvm; + struct vgic_dist *vgic = &kvm->arch.vgic; + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; + void __iomem *base = vgic->vctrl_base; + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; + + if (!base) + return; + + if (used_lrs) { + cpu_if->vgic_apr = readl_relaxed(base + GICH_APR); + + save_elrsr(vcpu, base); + save_lrs(vcpu, base); + + writel_relaxed(0, base + GICH_HCR); + } else { + cpu_if->vgic_elrsr = ~0UL; + cpu_if->vgic_apr = 0; + } +} + +void vgic_v2_restore_state(struct kvm_vcpu *vcpu) +{ + struct kvm *kvm = vcpu->kvm; + struct vgic_dist *vgic = &kvm->arch.vgic; + struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; + void __iomem *base = vgic->vctrl_base; + u64 used_lrs = vcpu->arch.vgic_cpu.used_lrs; + int i; + + if (!base) + return; + + if (used_lrs) { + writel_relaxed(cpu_if->vgic_hcr, base + GICH_HCR); + writel_relaxed(cpu_if->vgic_apr, base + GICH_APR); + for (i = 0; i < used_lrs; i++) { + writel_relaxed(cpu_if->vgic_lr[i], + base + GICH_LR0 + (i * 4)); + } + } +} + void vgic_v2_load(struct kvm_vcpu *vcpu) { struct vgic_v2_cpu_if *cpu_if = &vcpu->arch.vgic_cpu.vgic_v2; diff --git a/virt/kvm/arm/vgic/vgic.c b/virt/kvm/arm/vgic/vgic.c index b1bd238..790fd66 100644 --- a/virt/kvm/arm/vgic/vgic.c +++ b/virt/kvm/arm/vgic/vgic.c @@ -18,6 +18,8 @@ #include <linux/kvm_host.h> #include <linux/list_sort.h> +#include <asm/kvm_hyp.h> + #include "vgic.h" #define CREATE_TRACE_POINTS @@ -683,11 +685,19 @@ static void vgic_flush_lr_state(struct kvm_vcpu *vcpu) vgic_clear_lr(vcpu, count); } +static inline void vgic_save_state(struct kvm_vcpu *vcpu) +{ + if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) + vgic_v2_save_state(vcpu); +} + /* Sync back the hardware VGIC state into our emulation after a guest's run. */ void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu) { struct vgic_cpu *vgic_cpu = &vcpu->arch.vgic_cpu; + vgic_save_state(vcpu); + /* An empty ap_list_head implies used_lrs == 0 */ if (list_empty(&vcpu->arch.vgic_cpu.ap_list_head)) return; @@ -697,6 +707,12 @@ void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu) vgic_prune_ap_list(vcpu); } +static inline void vgic_restore_state(struct kvm_vcpu *vcpu) +{ + if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) + vgic_v2_restore_state(vcpu); +} + /* Flush our emulation state into the GIC hardware before entering the guest. */ void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu) { @@ -710,13 +726,16 @@ void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu) * this. */ if (list_empty(&vcpu->arch.vgic_cpu.ap_list_head)) - return; + goto out; DEBUG_SPINLOCK_BUG_ON(!irqs_disabled()); spin_lock(&vcpu->arch.vgic_cpu.ap_list_lock); vgic_flush_lr_state(vcpu); spin_unlock(&vcpu->arch.vgic_cpu.ap_list_lock); + +out: + vgic_restore_state(vcpu); } void kvm_vgic_load(struct kvm_vcpu *vcpu) diff --git a/virt/kvm/arm/vgic/vgic.h b/virt/kvm/arm/vgic/vgic.h index 4f8aecb..9895396 100644 --- a/virt/kvm/arm/vgic/vgic.h +++ b/virt/kvm/arm/vgic/vgic.h @@ -167,6 +167,11 @@ int vgic_register_dist_iodev(struct kvm *kvm, gpa_t dist_base_address, void vgic_v2_init_lrs(void); void vgic_v2_load(struct kvm_vcpu *vcpu); void vgic_v2_put(struct kvm_vcpu *vcpu); +void vgic_v2_handle_maintenance(struct kvm_vcpu *vcpu); + +void vgic_v2_save_state(struct kvm_vcpu *vcpu); +void vgic_v2_restore_state(struct kvm_vcpu *vcpu); + static inline void vgic_get_irq_kref(struct vgic_irq *irq) {
We can program the GICv2 hypervisor control interface logic directly from the core vgic code and can instead do the save/restore directly from the flush/sync functions, which can lead to a number of future optimizations. Signed-off-by: Christoffer Dall <christoffer.dall@linaro.org> --- arch/arm/kvm/hyp/switch.c | 4 -- arch/arm64/include/asm/kvm_hyp.h | 2 - arch/arm64/kvm/hyp/switch.c | 4 -- virt/kvm/arm/hyp/vgic-v2-sr.c | 83 ------------------------------------ virt/kvm/arm/vgic/vgic-init.c | 22 ++++++---- virt/kvm/arm/vgic/vgic-v2.c | 92 ++++++++++++++++++++++++++++++++++++++++ virt/kvm/arm/vgic/vgic.c | 21 ++++++++- virt/kvm/arm/vgic/vgic.h | 5 +++ 8 files changed, 130 insertions(+), 103 deletions(-)