diff mbox series

KVM: X86: Don't load/put guest FPU context for sleeping AP

Message ID 1583823679-17648-1-git-send-email-wanpengli@tencent.com (mailing list archive)
State New, archived
Headers show
Series KVM: X86: Don't load/put guest FPU context for sleeping AP | expand

Commit Message

Wanpeng Li March 10, 2020, 7:01 a.m. UTC
From: Wanpeng Li <wanpengli@tencent.com>

kvm_load_guest_fpu() and kvm_put_guest_fpu() each consume more than 14us 
observed by ftrace, the qemu userspace FPU is swapped out for the guest 
FPU context for the duration of the KVM_RUN ioctl even if sleeping AP, 
we shouldn't load/put guest FPU context for this case especially for 
serverless scenario which sensitives to boot time.

Signed-off-by: Wanpeng Li <wanpengli@tencent.com>
---
 arch/x86/kvm/x86.c | 8 +++++---
 1 file changed, 5 insertions(+), 3 deletions(-)

Comments

Sean Christopherson March 10, 2020, 4:01 p.m. UTC | #1
On Tue, Mar 10, 2020 at 03:01:19PM +0800, Wanpeng Li wrote:
> From: Wanpeng Li <wanpengli@tencent.com>
> 
> kvm_load_guest_fpu() and kvm_put_guest_fpu() each consume more than 14us 
> observed by ftrace, the qemu userspace FPU is swapped out for the guest 
> FPU context for the duration of the KVM_RUN ioctl even if sleeping AP, 
> we shouldn't load/put guest FPU context for this case especially for 
> serverless scenario which sensitives to boot time.
> 
> Signed-off-by: Wanpeng Li <wanpengli@tencent.com>
> ---
>  arch/x86/kvm/x86.c | 8 +++++---
>  1 file changed, 5 insertions(+), 3 deletions(-)
> 
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> index 5de2006..080ffa4 100644
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -8680,7 +8680,6 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  
>  	vcpu_load(vcpu);
>  	kvm_sigset_activate(vcpu);
> -	kvm_load_guest_fpu(vcpu);
>  
>  	if (unlikely(vcpu->arch.mp_state == KVM_MP_STATE_UNINITIALIZED)) {
>  		if (kvm_run->immediate_exit) {
> @@ -8718,12 +8717,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  		}
>  	}
>  
> +	kvm_load_guest_fpu(vcpu);

Ugh, so this isn't safe on MPX capable CPUs, kvm_apic_accept_events() can
trigger kvm_vcpu_reset() with @init_event=true and try to unload guest_fpu.

We could hack around that issue, but it'd be ugly, and I'm also concerned
that calling vmx_vcpu_reset() without guest_fpu loaded will be problematic
in the future with all the things that are getting managed by XSAVE.

> +
>  	if (unlikely(vcpu->arch.complete_userspace_io)) {
>  		int (*cui)(struct kvm_vcpu *) = vcpu->arch.complete_userspace_io;
>  		vcpu->arch.complete_userspace_io = NULL;
>  		r = cui(vcpu);
>  		if (r <= 0)
> -			goto out;
> +			goto out_fpu;
>  	} else
>  		WARN_ON(vcpu->arch.pio.count || vcpu->mmio_needed);
>  
> @@ -8732,8 +8733,9 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>  	else
>  		r = vcpu_run(vcpu);
>  
> -out:
> +out_fpu:
>  	kvm_put_guest_fpu(vcpu);
> +out:
>  	if (vcpu->run->kvm_valid_regs)
>  		store_regs(vcpu);
>  	post_kvm_run_save(vcpu);
> -- 
> 2.7.4
>
Wanpeng Li March 11, 2020, 1:12 a.m. UTC | #2
On Wed, 11 Mar 2020 at 00:01, Sean Christopherson
<sean.j.christopherson@intel.com> wrote:
>
> On Tue, Mar 10, 2020 at 03:01:19PM +0800, Wanpeng Li wrote:
> > From: Wanpeng Li <wanpengli@tencent.com>
> >
> > kvm_load_guest_fpu() and kvm_put_guest_fpu() each consume more than 14us
> > observed by ftrace, the qemu userspace FPU is swapped out for the guest
> > FPU context for the duration of the KVM_RUN ioctl even if sleeping AP,
> > we shouldn't load/put guest FPU context for this case especially for
> > serverless scenario which sensitives to boot time.
> >
> > Signed-off-by: Wanpeng Li <wanpengli@tencent.com>
> > ---
> >  arch/x86/kvm/x86.c | 8 +++++---
> >  1 file changed, 5 insertions(+), 3 deletions(-)
> >
> > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> > index 5de2006..080ffa4 100644
> > --- a/arch/x86/kvm/x86.c
> > +++ b/arch/x86/kvm/x86.c
> > @@ -8680,7 +8680,6 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> >
> >       vcpu_load(vcpu);
> >       kvm_sigset_activate(vcpu);
> > -     kvm_load_guest_fpu(vcpu);
> >
> >       if (unlikely(vcpu->arch.mp_state == KVM_MP_STATE_UNINITIALIZED)) {
> >               if (kvm_run->immediate_exit) {
> > @@ -8718,12 +8717,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> >               }
> >       }
> >
> > +     kvm_load_guest_fpu(vcpu);
>
> Ugh, so this isn't safe on MPX capable CPUs, kvm_apic_accept_events() can
> trigger kvm_vcpu_reset() with @init_event=true and try to unload guest_fpu.

Right.

>
> We could hack around that issue, but it'd be ugly, and I'm also concerned
> that calling vmx_vcpu_reset() without guest_fpu loaded will be problematic
> in the future with all the things that are getting managed by XSAVE.
>
> > +
> >       if (unlikely(vcpu->arch.complete_userspace_io)) {
> >               int (*cui)(struct kvm_vcpu *) = vcpu->arch.complete_userspace_io;
> >               vcpu->arch.complete_userspace_io = NULL;
> >               r = cui(vcpu);
> >               if (r <= 0)
> > -                     goto out;
> > +                     goto out_fpu;
> >       } else
> >               WARN_ON(vcpu->arch.pio.count || vcpu->mmio_needed);
> >
> > @@ -8732,8 +8733,9 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> >       else
> >               r = vcpu_run(vcpu);
> >
> > -out:
> > +out_fpu:
> >       kvm_put_guest_fpu(vcpu);
> > +out:
> >       if (vcpu->run->kvm_valid_regs)
> >               store_regs(vcpu);
> >       post_kvm_run_save(vcpu);
> > --
> > 2.7.4
> >
diff mbox series

Patch

diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 5de2006..080ffa4 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -8680,7 +8680,6 @@  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 
 	vcpu_load(vcpu);
 	kvm_sigset_activate(vcpu);
-	kvm_load_guest_fpu(vcpu);
 
 	if (unlikely(vcpu->arch.mp_state == KVM_MP_STATE_UNINITIALIZED)) {
 		if (kvm_run->immediate_exit) {
@@ -8718,12 +8717,14 @@  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 		}
 	}
 
+	kvm_load_guest_fpu(vcpu);
+
 	if (unlikely(vcpu->arch.complete_userspace_io)) {
 		int (*cui)(struct kvm_vcpu *) = vcpu->arch.complete_userspace_io;
 		vcpu->arch.complete_userspace_io = NULL;
 		r = cui(vcpu);
 		if (r <= 0)
-			goto out;
+			goto out_fpu;
 	} else
 		WARN_ON(vcpu->arch.pio.count || vcpu->mmio_needed);
 
@@ -8732,8 +8733,9 @@  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
 	else
 		r = vcpu_run(vcpu);
 
-out:
+out_fpu:
 	kvm_put_guest_fpu(vcpu);
+out:
 	if (vcpu->run->kvm_valid_regs)
 		store_regs(vcpu);
 	post_kvm_run_save(vcpu);