Message ID | 20121110154625.3274.39369.stgit@chazy-air (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Sat, Nov 10, 2012 at 03:46:25PM +0000, Christoffer Dall wrote: > From: Marc Zyngier <marc.zyngier@arm.com> > > Do the necessary save/restore dance for the timers in the world > switch code. In the process, allow the guest to read the physical > counter, which is useful for its own clock_event_device. > > Signed-off-by: Marc Zyngier <marc.zyngier@arm.com> > Signed-off-by: Christoffer Dall <c.dall@virtualopensystems.com> > --- > arch/arm/kernel/asm-offsets.c | 8 ++++++++ > arch/arm/kvm/arm.c | 3 +++ > arch/arm/kvm/interrupts_head.S | 41 ++++++++++++++++++++++++++++++++++++++++ > 3 files changed, 52 insertions(+) > > diff --git a/arch/arm/kernel/asm-offsets.c b/arch/arm/kernel/asm-offsets.c > index 39b6221..50df318 100644 > --- a/arch/arm/kernel/asm-offsets.c > +++ b/arch/arm/kernel/asm-offsets.c > @@ -177,6 +177,14 @@ int main(void) > DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr)); > DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr)); > DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr)); > +#ifdef CONFIG_KVM_ARM_TIMER > + DEFINE(VCPU_TIMER_CNTV_CTL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_ctl)); > + DEFINE(VCPU_TIMER_CNTV_CVALH, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.high)); > + DEFINE(VCPU_TIMER_CNTV_CVALL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.low)); > + DEFINE(KVM_TIMER_CNTVOFF_H, offsetof(struct kvm, arch.timer.cntvoff32.high)); > + DEFINE(KVM_TIMER_CNTVOFF_L, offsetof(struct kvm, arch.timer.cntvoff32.low)); > + DEFINE(KVM_TIMER_ENABLED, offsetof(struct kvm, arch.timer.enabled)); > +#endif > DEFINE(KVM_VGIC_VCTRL, offsetof(struct kvm, arch.vgic.vctrl_base)); > #endif > DEFINE(KVM_VTTBR, offsetof(struct kvm, arch.vttbr)); > diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c > index 1716f12..8cdef69 100644 > --- a/arch/arm/kvm/arm.c > +++ b/arch/arm/kvm/arm.c > @@ -661,6 +661,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) > update_vttbr(vcpu->kvm); > > kvm_vgic_sync_to_cpu(vcpu); > + kvm_timer_sync_to_cpu(vcpu); > > local_irq_disable(); > > @@ -674,6 +675,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) > > if (ret <= 0 || need_new_vmid_gen(vcpu->kvm)) { > local_irq_enable(); > + kvm_timer_sync_from_cpu(vcpu); > kvm_vgic_sync_from_cpu(vcpu); > continue; > } > @@ -713,6 +715,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) > * Back from guest > *************************************************************/ > > + kvm_timer_sync_from_cpu(vcpu); > kvm_vgic_sync_from_cpu(vcpu); > > ret = handle_exit(vcpu, run, ret); > diff --git a/arch/arm/kvm/interrupts_head.S b/arch/arm/kvm/interrupts_head.S > index 0003aab..ece84d1 100644 > --- a/arch/arm/kvm/interrupts_head.S > +++ b/arch/arm/kvm/interrupts_head.S > @@ -422,6 +422,25 @@ > #define CNTHCTL_PL1PCEN (1 << 1) > > .macro save_timer_state vcpup > +#ifdef CONFIG_KVM_ARM_TIMER > + ldr r4, [\vcpup, #VCPU_KVM] > + ldr r2, [r4, #KVM_TIMER_ENABLED] > + cmp r2, #0 > + beq 1f > + > + mrc p15, 0, r2, c14, c3, 1 @ CNTV_CTL > + and r2, #3 Why do you need this and (you do the masking on the restore path)? > + str r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] > + bic r2, #1 @ Clear ENABLE > + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL > + isb > + > + mrrc p15, 3, r2, r3, c14 @ CNTV_CVAL > + str r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] > + str r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] > + > +1: > +#endif > @ Allow physical timer/counter access for the host > mrc p15, 4, r2, c14, c1, 0 @ CNTHCTL > orr r2, r2, #(CNTHCTL_PL1PCEN | CNTHCTL_PL1PCTEN) > @@ -435,6 +454,28 @@ > orr r2, r2, #CNTHCTL_PL1PCTEN > bic r2, r2, #CNTHCTL_PL1PCEN > mcr p15, 4, r2, c14, c1, 0 @ CNTHCTL > + > +#ifdef CONFIG_KVM_ARM_TIMER > + ldr r4, [\vcpup, #VCPU_KVM] > + ldr r2, [r4, #KVM_TIMER_ENABLED] > + cmp r2, #0 > + beq 1f > + > + ldr r3, [r4, #KVM_TIMER_CNTVOFF_H] > + ldr r2, [r4, #KVM_TIMER_CNTVOFF_L] > + mcrr p15, 4, r2, r3, c14 @ CNTVOFF > + isb > + > + ldr r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] > + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] > + mcrr p15, 3, r2, r3, c14 @ CNTV_CVAL > + > + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] > + and r2, #3 Please use the long form syntax for these instructions (and r2, r2, #3) as toolchains have historically been picky about accepting these when not targetting Thumb and it also makes it consistent with the rest of arch/arm/. > + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL > + isb > +1: > +#endif It's pretty horrible to take a macro argument (vcpup) and then open-code all the other register numbers without moving the argument somewhere nice. Are callers just expected to understand that you clobber {r2,r3,r4} and avoid them? If so, why not just use the PCS so that people don't have to learn a new convention? Will -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On 23/11/12 16:30, Will Deacon wrote: > On Sat, Nov 10, 2012 at 03:46:25PM +0000, Christoffer Dall wrote: >> From: Marc Zyngier <marc.zyngier@arm.com> >> >> Do the necessary save/restore dance for the timers in the world >> switch code. In the process, allow the guest to read the physical >> counter, which is useful for its own clock_event_device. >> >> Signed-off-by: Marc Zyngier <marc.zyngier@arm.com> >> Signed-off-by: Christoffer Dall <c.dall@virtualopensystems.com> >> --- >> arch/arm/kernel/asm-offsets.c | 8 ++++++++ >> arch/arm/kvm/arm.c | 3 +++ >> arch/arm/kvm/interrupts_head.S | 41 ++++++++++++++++++++++++++++++++++++++++ >> 3 files changed, 52 insertions(+) >> >> diff --git a/arch/arm/kernel/asm-offsets.c b/arch/arm/kernel/asm-offsets.c >> index 39b6221..50df318 100644 >> --- a/arch/arm/kernel/asm-offsets.c >> +++ b/arch/arm/kernel/asm-offsets.c >> @@ -177,6 +177,14 @@ int main(void) >> DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr)); >> DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr)); >> DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr)); >> +#ifdef CONFIG_KVM_ARM_TIMER >> + DEFINE(VCPU_TIMER_CNTV_CTL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_ctl)); >> + DEFINE(VCPU_TIMER_CNTV_CVALH, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.high)); >> + DEFINE(VCPU_TIMER_CNTV_CVALL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.low)); >> + DEFINE(KVM_TIMER_CNTVOFF_H, offsetof(struct kvm, arch.timer.cntvoff32.high)); >> + DEFINE(KVM_TIMER_CNTVOFF_L, offsetof(struct kvm, arch.timer.cntvoff32.low)); >> + DEFINE(KVM_TIMER_ENABLED, offsetof(struct kvm, arch.timer.enabled)); >> +#endif >> DEFINE(KVM_VGIC_VCTRL, offsetof(struct kvm, arch.vgic.vctrl_base)); >> #endif >> DEFINE(KVM_VTTBR, offsetof(struct kvm, arch.vttbr)); >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c >> index 1716f12..8cdef69 100644 >> --- a/arch/arm/kvm/arm.c >> +++ b/arch/arm/kvm/arm.c >> @@ -661,6 +661,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) >> update_vttbr(vcpu->kvm); >> >> kvm_vgic_sync_to_cpu(vcpu); >> + kvm_timer_sync_to_cpu(vcpu); >> >> local_irq_disable(); >> >> @@ -674,6 +675,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) >> >> if (ret <= 0 || need_new_vmid_gen(vcpu->kvm)) { >> local_irq_enable(); >> + kvm_timer_sync_from_cpu(vcpu); >> kvm_vgic_sync_from_cpu(vcpu); >> continue; >> } >> @@ -713,6 +715,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) >> * Back from guest >> *************************************************************/ >> >> + kvm_timer_sync_from_cpu(vcpu); >> kvm_vgic_sync_from_cpu(vcpu); >> >> ret = handle_exit(vcpu, run, ret); >> diff --git a/arch/arm/kvm/interrupts_head.S b/arch/arm/kvm/interrupts_head.S >> index 0003aab..ece84d1 100644 >> --- a/arch/arm/kvm/interrupts_head.S >> +++ b/arch/arm/kvm/interrupts_head.S >> @@ -422,6 +422,25 @@ >> #define CNTHCTL_PL1PCEN (1 << 1) >> >> .macro save_timer_state vcpup >> +#ifdef CONFIG_KVM_ARM_TIMER >> + ldr r4, [\vcpup, #VCPU_KVM] >> + ldr r2, [r4, #KVM_TIMER_ENABLED] >> + cmp r2, #0 >> + beq 1f >> + >> + mrc p15, 0, r2, c14, c3, 1 @ CNTV_CTL >> + and r2, #3 > > Why do you need this and (you do the masking on the restore path)? Probably not necessary. We could just clear the enable below. > >> + str r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] >> + bic r2, #1 @ Clear ENABLE >> + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL >> + isb >> + >> + mrrc p15, 3, r2, r3, c14 @ CNTV_CVAL >> + str r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] >> + str r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] >> + >> +1: >> +#endif >> @ Allow physical timer/counter access for the host >> mrc p15, 4, r2, c14, c1, 0 @ CNTHCTL >> orr r2, r2, #(CNTHCTL_PL1PCEN | CNTHCTL_PL1PCTEN) >> @@ -435,6 +454,28 @@ >> orr r2, r2, #CNTHCTL_PL1PCTEN >> bic r2, r2, #CNTHCTL_PL1PCEN >> mcr p15, 4, r2, c14, c1, 0 @ CNTHCTL >> + >> +#ifdef CONFIG_KVM_ARM_TIMER >> + ldr r4, [\vcpup, #VCPU_KVM] >> + ldr r2, [r4, #KVM_TIMER_ENABLED] >> + cmp r2, #0 >> + beq 1f >> + >> + ldr r3, [r4, #KVM_TIMER_CNTVOFF_H] >> + ldr r2, [r4, #KVM_TIMER_CNTVOFF_L] >> + mcrr p15, 4, r2, r3, c14 @ CNTVOFF >> + isb >> + >> + ldr r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] >> + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] >> + mcrr p15, 3, r2, r3, c14 @ CNTV_CVAL >> + >> + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] >> + and r2, #3 > > Please use the long form syntax for these instructions (and r2, r2, #3) as > toolchains have historically been picky about accepting these when not > targetting Thumb and it also makes it consistent with the rest of arch/arm/. OK. > >> + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL >> + isb >> +1: >> +#endif > > It's pretty horrible to take a macro argument (vcpup) and then open-code all > the other register numbers without moving the argument somewhere nice. Are > callers just expected to understand that you clobber {r2,r3,r4} and avoid > them? If so, why not just use the PCS so that people don't have to learn a > new convention? I'm happy to remove this vcpup and stick to r0/r1 (depending on the path). All macros have to know about the side effects anyway, as we switch all the registers. M.
On Fri, Nov 23, 2012 at 12:04 PM, Marc Zyngier <marc.zyngier@arm.com> wrote: > On 23/11/12 16:30, Will Deacon wrote: >> On Sat, Nov 10, 2012 at 03:46:25PM +0000, Christoffer Dall wrote: >>> From: Marc Zyngier <marc.zyngier@arm.com> >>> >>> Do the necessary save/restore dance for the timers in the world >>> switch code. In the process, allow the guest to read the physical >>> counter, which is useful for its own clock_event_device. >>> >>> Signed-off-by: Marc Zyngier <marc.zyngier@arm.com> >>> Signed-off-by: Christoffer Dall <c.dall@virtualopensystems.com> >>> --- >>> arch/arm/kernel/asm-offsets.c | 8 ++++++++ >>> arch/arm/kvm/arm.c | 3 +++ >>> arch/arm/kvm/interrupts_head.S | 41 ++++++++++++++++++++++++++++++++++++++++ >>> 3 files changed, 52 insertions(+) >>> >>> diff --git a/arch/arm/kernel/asm-offsets.c b/arch/arm/kernel/asm-offsets.c >>> index 39b6221..50df318 100644 >>> --- a/arch/arm/kernel/asm-offsets.c >>> +++ b/arch/arm/kernel/asm-offsets.c >>> @@ -177,6 +177,14 @@ int main(void) >>> DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr)); >>> DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr)); >>> DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr)); >>> +#ifdef CONFIG_KVM_ARM_TIMER >>> + DEFINE(VCPU_TIMER_CNTV_CTL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_ctl)); >>> + DEFINE(VCPU_TIMER_CNTV_CVALH, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.high)); >>> + DEFINE(VCPU_TIMER_CNTV_CVALL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.low)); >>> + DEFINE(KVM_TIMER_CNTVOFF_H, offsetof(struct kvm, arch.timer.cntvoff32.high)); >>> + DEFINE(KVM_TIMER_CNTVOFF_L, offsetof(struct kvm, arch.timer.cntvoff32.low)); >>> + DEFINE(KVM_TIMER_ENABLED, offsetof(struct kvm, arch.timer.enabled)); >>> +#endif >>> DEFINE(KVM_VGIC_VCTRL, offsetof(struct kvm, arch.vgic.vctrl_base)); >>> #endif >>> DEFINE(KVM_VTTBR, offsetof(struct kvm, arch.vttbr)); >>> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c >>> index 1716f12..8cdef69 100644 >>> --- a/arch/arm/kvm/arm.c >>> +++ b/arch/arm/kvm/arm.c >>> @@ -661,6 +661,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) >>> update_vttbr(vcpu->kvm); >>> >>> kvm_vgic_sync_to_cpu(vcpu); >>> + kvm_timer_sync_to_cpu(vcpu); >>> >>> local_irq_disable(); >>> >>> @@ -674,6 +675,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) >>> >>> if (ret <= 0 || need_new_vmid_gen(vcpu->kvm)) { >>> local_irq_enable(); >>> + kvm_timer_sync_from_cpu(vcpu); >>> kvm_vgic_sync_from_cpu(vcpu); >>> continue; >>> } >>> @@ -713,6 +715,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) >>> * Back from guest >>> *************************************************************/ >>> >>> + kvm_timer_sync_from_cpu(vcpu); >>> kvm_vgic_sync_from_cpu(vcpu); >>> >>> ret = handle_exit(vcpu, run, ret); >>> diff --git a/arch/arm/kvm/interrupts_head.S b/arch/arm/kvm/interrupts_head.S >>> index 0003aab..ece84d1 100644 >>> --- a/arch/arm/kvm/interrupts_head.S >>> +++ b/arch/arm/kvm/interrupts_head.S >>> @@ -422,6 +422,25 @@ >>> #define CNTHCTL_PL1PCEN (1 << 1) >>> >>> .macro save_timer_state vcpup >>> +#ifdef CONFIG_KVM_ARM_TIMER >>> + ldr r4, [\vcpup, #VCPU_KVM] >>> + ldr r2, [r4, #KVM_TIMER_ENABLED] >>> + cmp r2, #0 >>> + beq 1f >>> + >>> + mrc p15, 0, r2, c14, c3, 1 @ CNTV_CTL >>> + and r2, #3 >> >> Why do you need this and (you do the masking on the restore path)? > > Probably not necessary. We could just clear the enable below. > >> >>> + str r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] >>> + bic r2, #1 @ Clear ENABLE >>> + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL >>> + isb >>> + >>> + mrrc p15, 3, r2, r3, c14 @ CNTV_CVAL >>> + str r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] >>> + str r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] >>> + >>> +1: >>> +#endif >>> @ Allow physical timer/counter access for the host >>> mrc p15, 4, r2, c14, c1, 0 @ CNTHCTL >>> orr r2, r2, #(CNTHCTL_PL1PCEN | CNTHCTL_PL1PCTEN) >>> @@ -435,6 +454,28 @@ >>> orr r2, r2, #CNTHCTL_PL1PCTEN >>> bic r2, r2, #CNTHCTL_PL1PCEN >>> mcr p15, 4, r2, c14, c1, 0 @ CNTHCTL >>> + >>> +#ifdef CONFIG_KVM_ARM_TIMER >>> + ldr r4, [\vcpup, #VCPU_KVM] >>> + ldr r2, [r4, #KVM_TIMER_ENABLED] >>> + cmp r2, #0 >>> + beq 1f >>> + >>> + ldr r3, [r4, #KVM_TIMER_CNTVOFF_H] >>> + ldr r2, [r4, #KVM_TIMER_CNTVOFF_L] >>> + mcrr p15, 4, r2, r3, c14 @ CNTVOFF >>> + isb >>> + >>> + ldr r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] >>> + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] >>> + mcrr p15, 3, r2, r3, c14 @ CNTV_CVAL >>> + >>> + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] >>> + and r2, #3 >> >> Please use the long form syntax for these instructions (and r2, r2, #3) as >> toolchains have historically been picky about accepting these when not >> targetting Thumb and it also makes it consistent with the rest of arch/arm/. > > OK. > >> >>> + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL >>> + isb >>> +1: >>> +#endif >> >> It's pretty horrible to take a macro argument (vcpup) and then open-code all >> the other register numbers without moving the argument somewhere nice. Are >> callers just expected to understand that you clobber {r2,r3,r4} and avoid >> them? If so, why not just use the PCS so that people don't have to learn a >> new convention? > > I'm happy to remove this vcpup and stick to r0/r1 (depending on the > path). All macros have to know about the side effects anyway, as we > switch all the registers. > Using the PCS is not going to help us, because we need a lot of registers for a lot of these operations and we don't want to push/pop extra stuff to the stack here - this is in the super critical path. The nature of this code is not one where you have random callers that need to work under certain assumptions - you need to scrutinize every line here. What we could do is pass all temporary variables to the macros, but it becomes ugly when they take 5+ temporary regs, and it doesn't make sense for register save/restore that also makes assumptions about which register is used for the vcpu pointer. Alternatively, and more attainable, we could make it a convention that all the macros in this file take r0 as the vcpu pointer. Actually, let's use r0 for the cpu pointer all over. Why not. Let me craft a patch and send that out right away! -Christoffer -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/arch/arm/kernel/asm-offsets.c b/arch/arm/kernel/asm-offsets.c index 39b6221..50df318 100644 --- a/arch/arm/kernel/asm-offsets.c +++ b/arch/arm/kernel/asm-offsets.c @@ -177,6 +177,14 @@ int main(void) DEFINE(VGIC_CPU_APR, offsetof(struct vgic_cpu, vgic_apr)); DEFINE(VGIC_CPU_LR, offsetof(struct vgic_cpu, vgic_lr)); DEFINE(VGIC_CPU_NR_LR, offsetof(struct vgic_cpu, nr_lr)); +#ifdef CONFIG_KVM_ARM_TIMER + DEFINE(VCPU_TIMER_CNTV_CTL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_ctl)); + DEFINE(VCPU_TIMER_CNTV_CVALH, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.high)); + DEFINE(VCPU_TIMER_CNTV_CVALL, offsetof(struct kvm_vcpu, arch.timer_cpu.cntv_cval32.low)); + DEFINE(KVM_TIMER_CNTVOFF_H, offsetof(struct kvm, arch.timer.cntvoff32.high)); + DEFINE(KVM_TIMER_CNTVOFF_L, offsetof(struct kvm, arch.timer.cntvoff32.low)); + DEFINE(KVM_TIMER_ENABLED, offsetof(struct kvm, arch.timer.enabled)); +#endif DEFINE(KVM_VGIC_VCTRL, offsetof(struct kvm, arch.vgic.vctrl_base)); #endif DEFINE(KVM_VTTBR, offsetof(struct kvm, arch.vttbr)); diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c index 1716f12..8cdef69 100644 --- a/arch/arm/kvm/arm.c +++ b/arch/arm/kvm/arm.c @@ -661,6 +661,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) update_vttbr(vcpu->kvm); kvm_vgic_sync_to_cpu(vcpu); + kvm_timer_sync_to_cpu(vcpu); local_irq_disable(); @@ -674,6 +675,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) if (ret <= 0 || need_new_vmid_gen(vcpu->kvm)) { local_irq_enable(); + kvm_timer_sync_from_cpu(vcpu); kvm_vgic_sync_from_cpu(vcpu); continue; } @@ -713,6 +715,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *run) * Back from guest *************************************************************/ + kvm_timer_sync_from_cpu(vcpu); kvm_vgic_sync_from_cpu(vcpu); ret = handle_exit(vcpu, run, ret); diff --git a/arch/arm/kvm/interrupts_head.S b/arch/arm/kvm/interrupts_head.S index 0003aab..ece84d1 100644 --- a/arch/arm/kvm/interrupts_head.S +++ b/arch/arm/kvm/interrupts_head.S @@ -422,6 +422,25 @@ #define CNTHCTL_PL1PCEN (1 << 1) .macro save_timer_state vcpup +#ifdef CONFIG_KVM_ARM_TIMER + ldr r4, [\vcpup, #VCPU_KVM] + ldr r2, [r4, #KVM_TIMER_ENABLED] + cmp r2, #0 + beq 1f + + mrc p15, 0, r2, c14, c3, 1 @ CNTV_CTL + and r2, #3 + str r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] + bic r2, #1 @ Clear ENABLE + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL + isb + + mrrc p15, 3, r2, r3, c14 @ CNTV_CVAL + str r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] + str r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] + +1: +#endif @ Allow physical timer/counter access for the host mrc p15, 4, r2, c14, c1, 0 @ CNTHCTL orr r2, r2, #(CNTHCTL_PL1PCEN | CNTHCTL_PL1PCTEN) @@ -435,6 +454,28 @@ orr r2, r2, #CNTHCTL_PL1PCTEN bic r2, r2, #CNTHCTL_PL1PCEN mcr p15, 4, r2, c14, c1, 0 @ CNTHCTL + +#ifdef CONFIG_KVM_ARM_TIMER + ldr r4, [\vcpup, #VCPU_KVM] + ldr r2, [r4, #KVM_TIMER_ENABLED] + cmp r2, #0 + beq 1f + + ldr r3, [r4, #KVM_TIMER_CNTVOFF_H] + ldr r2, [r4, #KVM_TIMER_CNTVOFF_L] + mcrr p15, 4, r2, r3, c14 @ CNTVOFF + isb + + ldr r3, [\vcpup, #VCPU_TIMER_CNTV_CVALH] + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CVALL] + mcrr p15, 3, r2, r3, c14 @ CNTV_CVAL + + ldr r2, [\vcpup, #VCPU_TIMER_CNTV_CTL] + and r2, #3 + mcr p15, 0, r2, c14, c3, 1 @ CNTV_CTL + isb +1: +#endif .endm /* Configures the HSTR (Hyp System Trap Register) on entry/return