Message ID | 20191011194032.240572-3-aaronlewis@google.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Add support for XSAVES on AMD and unify it with Intel | expand |
On Fri, Oct 11, 2019 at 12:40:29PM -0700, Aaron Lewis wrote: > Set IA32_XSS for the guest and host during VM Enter and VM Exit > transitions rather than by using the MSR-load areas. > > By moving away from using the MSR-load area we can have a unified > solution for both AMD and Intel. > > Reviewed-by: Jim Mattson <jmattson@google.com> > Signed-off-by: Aaron Lewis <aaronlewis@google.com> > --- > arch/x86/include/asm/kvm_host.h | 1 + > arch/x86/kvm/svm.c | 7 +++++-- > arch/x86/kvm/vmx/vmx.c | 22 ++++++++++------------ > arch/x86/kvm/x86.c | 23 +++++++++++++++++++---- > arch/x86/kvm/x86.h | 4 ++-- > 5 files changed, 37 insertions(+), 20 deletions(-) > > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > index 50eb430b0ad8..634c2598e389 100644 > --- a/arch/x86/include/asm/kvm_host.h > +++ b/arch/x86/include/asm/kvm_host.h > @@ -562,6 +562,7 @@ struct kvm_vcpu_arch { > u64 smbase; > u64 smi_count; > bool tpr_access_reporting; > + bool xsaves_enabled; > u64 ia32_xss; > u64 microcode_version; > u64 arch_capabilities; > diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > index f8ecb6df5106..da69e95beb4d 100644 > --- a/arch/x86/kvm/svm.c > +++ b/arch/x86/kvm/svm.c > @@ -5628,7 +5628,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > svm->vmcb->save.cr2 = vcpu->arch.cr2; > > clgi(); > - kvm_load_guest_xcr0(vcpu); > + kvm_load_guest_xsave_controls(vcpu); > > if (lapic_in_kernel(vcpu) && > vcpu->arch.apic->lapic_timer.timer_advance_ns) > @@ -5778,7 +5778,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI)) > kvm_before_interrupt(&svm->vcpu); > > - kvm_put_guest_xcr0(vcpu); > + kvm_load_host_xsave_controls(vcpu); > stgi(); > > /* Any pending NMI will happen here */ > @@ -5887,6 +5887,9 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu) > { > struct vcpu_svm *svm = to_svm(vcpu); > > + vcpu->arch.xsaves_enabled = guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > + boot_cpu_has(X86_FEATURE_XSAVES); This looks very much like a functional change to SVM, which feels wrong for a patch with a subject of "KVM: VMX: Use wrmsr for switching between guest and host IA32_XSS" feels wrong. Shouldn't this be unconditionally set false in this patch, and then enabled in " kvm: svm: Add support for XSAVES on AMD"? > + > /* Update nrips enabled cache */ > svm->nrips_enabled = !!guest_cpuid_has(&svm->vcpu, X86_FEATURE_NRIPS); > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > index 409e9a7323f1..ce3020914c69 100644 > --- a/arch/x86/kvm/vmx/vmx.c > +++ b/arch/x86/kvm/vmx/vmx.c > @@ -106,8 +106,6 @@ module_param(enable_apicv, bool, S_IRUGO); > static bool __read_mostly nested = 1; > module_param(nested, bool, S_IRUGO); > > -static u64 __read_mostly host_xss; > - > bool __read_mostly enable_pml = 1; > module_param_named(pml, enable_pml, bool, S_IRUGO); > > @@ -2074,11 +2072,6 @@ static int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) > if (data != 0) > return 1; > vcpu->arch.ia32_xss = data; > - if (vcpu->arch.ia32_xss != host_xss) > - add_atomic_switch_msr(vmx, MSR_IA32_XSS, > - vcpu->arch.ia32_xss, host_xss, false); > - else > - clear_atomic_switch_msr(vmx, MSR_IA32_XSS); I'm pretty sure the "vmx_xsaves_supported()" check in this case statement can be removed after this patch. The host support check was necessary when the MSR load/save list was being used, but moving to xsaves_enabled means theres no risk of writing MSR_IA32_XSS when it's not supported. > break; > case MSR_IA32_RTIT_CTL: > if ((pt_mode != PT_MODE_HOST_GUEST) || > @@ -4038,6 +4031,8 @@ static void vmx_compute_secondary_exec_control(struct vcpu_vmx *vmx) > guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > guest_cpuid_has(vcpu, X86_FEATURE_XSAVES); > > + vcpu->arch.xsaves_enabled = xsaves_enabled; Doesn't this conflict with the direction of the previous patch, "KVM: VMX: Remove unneeded check for X86_FEATURE_XSAVE"? xsaves_enabled is %true iff both XSAVE and XSAVES are exposed to the guest. Alternatively, couldn't KVM check arch.xsaves_enabled in the MSR handling? > + > if (!xsaves_enabled) > exec_control &= ~SECONDARY_EXEC_XSAVES; > > @@ -6540,7 +6535,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu) > if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP) > vmx_set_interrupt_shadow(vcpu, 0); > > - kvm_load_guest_xcr0(vcpu); > + kvm_load_guest_xsave_controls(vcpu); > > if (static_cpu_has(X86_FEATURE_PKU) && > kvm_read_cr4_bits(vcpu, X86_CR4_PKE) && > @@ -6647,7 +6642,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu) > __write_pkru(vmx->host_pkru); > } > > - kvm_put_guest_xcr0(vcpu); > + kvm_load_host_xsave_controls(vcpu); > > vmx->nested.nested_run_pending = 0; > vmx->idt_vectoring_info = 0; > @@ -7091,6 +7086,12 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > { > struct vcpu_vmx *vmx = to_vmx(vcpu); > > + /* > + * This will be set back to true in vmx_compute_secondary_exec_control() > + * if it should be true, so setting it to false here is okay. Or simply: /* xsaves_enabled is recomputed in vmx_compute_secondary_exec_control(). */ > + */ > + vcpu->arch.xsaves_enabled = false; > + > if (cpu_has_secondary_exec_ctrls()) { > vmx_compute_secondary_exec_control(vmx); > vmcs_set_secondary_exec_control(vmx); > @@ -7599,9 +7600,6 @@ static __init int hardware_setup(void) > WARN_ONCE(host_bndcfgs, "KVM: BNDCFGS in host will be lost"); > } > > - if (boot_cpu_has(X86_FEATURE_XSAVES)) > - rdmsrl(MSR_IA32_XSS, host_xss); > - > if (!cpu_has_vmx_vpid() || !cpu_has_vmx_invvpid() || > !(cpu_has_vmx_invvpid_single() || cpu_has_vmx_invvpid_global())) > enable_vpid = 0; > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index 661e2bf38526..a61570d7034b 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -176,6 +176,8 @@ struct kvm_shared_msrs { > static struct kvm_shared_msrs_global __read_mostly shared_msrs_global; > static struct kvm_shared_msrs __percpu *shared_msrs; > > +static u64 __read_mostly host_xss; > + > struct kvm_stats_debugfs_item debugfs_entries[] = { > { "pf_fixed", VCPU_STAT(pf_fixed) }, > { "pf_guest", VCPU_STAT(pf_guest) }, > @@ -812,27 +814,37 @@ void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw) > } > EXPORT_SYMBOL_GPL(kvm_lmsw); > > -void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu) > +void kvm_load_guest_xsave_controls(struct kvm_vcpu *vcpu) > { > if (kvm_read_cr4_bits(vcpu, X86_CR4_OSXSAVE) && > !vcpu->guest_xcr0_loaded) { > /* kvm_set_xcr() also depends on this */ > if (vcpu->arch.xcr0 != host_xcr0) > xsetbv(XCR_XFEATURE_ENABLED_MASK, vcpu->arch.xcr0); > + > + if (vcpu->arch.xsaves_enabled && > + vcpu->arch.ia32_xss != host_xss) > + wrmsrl(MSR_IA32_XSS, vcpu->arch.ia32_xss); > + > vcpu->guest_xcr0_loaded = 1; > } > } > -EXPORT_SYMBOL_GPL(kvm_load_guest_xcr0); > +EXPORT_SYMBOL_GPL(kvm_load_guest_xsave_controls); > > -void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu) > +void kvm_load_host_xsave_controls(struct kvm_vcpu *vcpu) > { > if (vcpu->guest_xcr0_loaded) { > if (vcpu->arch.xcr0 != host_xcr0) > xsetbv(XCR_XFEATURE_ENABLED_MASK, host_xcr0); > + > + if (vcpu->arch.xsaves_enabled && > + vcpu->arch.ia32_xss != host_xss) > + wrmsrl(MSR_IA32_XSS, host_xss); > + > vcpu->guest_xcr0_loaded = 0; > } > } > -EXPORT_SYMBOL_GPL(kvm_put_guest_xcr0); > +EXPORT_SYMBOL_GPL(kvm_load_host_xsave_controls); > > static int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr) > { > @@ -9293,6 +9305,9 @@ int kvm_arch_hardware_setup(void) > kvm_default_tsc_scaling_ratio = 1ULL << kvm_tsc_scaling_ratio_frac_bits; > } > > + if (boot_cpu_has(X86_FEATURE_XSAVES)) > + rdmsrl(MSR_IA32_XSS, host_xss); More code that goes beyond the subject of the patch, as host_xss is now being read and potentially switched on SVM platforms. I don't mind the change itself being part of this patch (unlike setting xsaves_enabled), but IMO the subject should make it more clear that this is a common x86 change. Arguably this should be two patches, one to do what this subject says and a second to move the code to common x86, but that seems like overkill. > kvm_init_msr_list(); > return 0; > } > diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h > index dbf7442a822b..0d04e865665b 100644 > --- a/arch/x86/kvm/x86.h > +++ b/arch/x86/kvm/x86.h > @@ -366,7 +366,7 @@ static inline bool kvm_pat_valid(u64 data) > return (data | ((data & 0x0202020202020202ull) << 1)) == data; > } > > -void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu); > -void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu); > +void kvm_load_guest_xsave_controls(struct kvm_vcpu *vcpu); > +void kvm_load_host_xsave_controls(struct kvm_vcpu *vcpu); > > #endif > -- > 2.23.0.700.g56cf767bdb-goog >
On Fri, Oct 11, 2019 at 5:18 PM Sean Christopherson <sean.j.christopherson@intel.com> wrote: > > On Fri, Oct 11, 2019 at 12:40:29PM -0700, Aaron Lewis wrote: > > Set IA32_XSS for the guest and host during VM Enter and VM Exit > > transitions rather than by using the MSR-load areas. > > > > By moving away from using the MSR-load area we can have a unified > > solution for both AMD and Intel. > > > > Reviewed-by: Jim Mattson <jmattson@google.com> > > Signed-off-by: Aaron Lewis <aaronlewis@google.com> > > --- > > arch/x86/include/asm/kvm_host.h | 1 + > > arch/x86/kvm/svm.c | 7 +++++-- > > arch/x86/kvm/vmx/vmx.c | 22 ++++++++++------------ > > arch/x86/kvm/x86.c | 23 +++++++++++++++++++---- > > arch/x86/kvm/x86.h | 4 ++-- > > 5 files changed, 37 insertions(+), 20 deletions(-) > > > > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > > index 50eb430b0ad8..634c2598e389 100644 > > --- a/arch/x86/include/asm/kvm_host.h > > +++ b/arch/x86/include/asm/kvm_host.h > > @@ -562,6 +562,7 @@ struct kvm_vcpu_arch { > > u64 smbase; > > u64 smi_count; > > bool tpr_access_reporting; > > + bool xsaves_enabled; > > u64 ia32_xss; > > u64 microcode_version; > > u64 arch_capabilities; > > diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > > index f8ecb6df5106..da69e95beb4d 100644 > > --- a/arch/x86/kvm/svm.c > > +++ b/arch/x86/kvm/svm.c > > @@ -5628,7 +5628,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > > svm->vmcb->save.cr2 = vcpu->arch.cr2; > > > > clgi(); > > - kvm_load_guest_xcr0(vcpu); > > + kvm_load_guest_xsave_controls(vcpu); > > > > if (lapic_in_kernel(vcpu) && > > vcpu->arch.apic->lapic_timer.timer_advance_ns) > > @@ -5778,7 +5778,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > > if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI)) > > kvm_before_interrupt(&svm->vcpu); > > > > - kvm_put_guest_xcr0(vcpu); > > + kvm_load_host_xsave_controls(vcpu); > > stgi(); > > > > /* Any pending NMI will happen here */ > > @@ -5887,6 +5887,9 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu) > > { > > struct vcpu_svm *svm = to_svm(vcpu); > > > > + vcpu->arch.xsaves_enabled = guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > > + boot_cpu_has(X86_FEATURE_XSAVES); > > This looks very much like a functional change to SVM, which feels wrong > for a patch with a subject of "KVM: VMX: Use wrmsr for switching between > guest and host IA32_XSS" feels wrong. Shouldn't this be unconditionally > set false in this patch, and then enabled in " kvm: svm: Add support for > XSAVES on AMD"? Nothing is being enabled here. Vcpu->arch.xsaves_enabled simply tells us whether or not the guest can execute the XSAVES instruction. Any guest with the ability to set CR4.OSXSAVE on an AMD host that supports XSAVES can use the instruction. > > + > > /* Update nrips enabled cache */ > > svm->nrips_enabled = !!guest_cpuid_has(&svm->vcpu, X86_FEATURE_NRIPS); > > > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > > index 409e9a7323f1..ce3020914c69 100644 > > --- a/arch/x86/kvm/vmx/vmx.c > > +++ b/arch/x86/kvm/vmx/vmx.c > > @@ -106,8 +106,6 @@ module_param(enable_apicv, bool, S_IRUGO); > > static bool __read_mostly nested = 1; > > module_param(nested, bool, S_IRUGO); > > > > -static u64 __read_mostly host_xss; > > - > > bool __read_mostly enable_pml = 1; > > module_param_named(pml, enable_pml, bool, S_IRUGO); > > > > @@ -2074,11 +2072,6 @@ static int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) > > if (data != 0) > > return 1; > > vcpu->arch.ia32_xss = data; > > - if (vcpu->arch.ia32_xss != host_xss) > > - add_atomic_switch_msr(vmx, MSR_IA32_XSS, > > - vcpu->arch.ia32_xss, host_xss, false); > > - else > > - clear_atomic_switch_msr(vmx, MSR_IA32_XSS); > > I'm pretty sure the "vmx_xsaves_supported()" check in this case statement > can be removed after this patch. The host support check was necessary > when the MSR load/save list was being used, but moving to xsaves_enabled > means theres no risk of writing MSR_IA32_XSS when it's not supported. I agree. vmx_xsaves_supported() only tells us if the XSAVES instruction can be enabled. It does not tell us anything about the existence of IA32_XSS in the guest. > > break; > > case MSR_IA32_RTIT_CTL: > > if ((pt_mode != PT_MODE_HOST_GUEST) || > > @@ -4038,6 +4031,8 @@ static void vmx_compute_secondary_exec_control(struct vcpu_vmx *vmx) > > guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > > guest_cpuid_has(vcpu, X86_FEATURE_XSAVES); > > > > + vcpu->arch.xsaves_enabled = xsaves_enabled; > > Doesn't this conflict with the direction of the previous patch, "KVM: VMX: > Remove unneeded check for X86_FEATURE_XSAVE"? xsaves_enabled is %true iff > both XSAVE and XSAVES are exposed to the guest. There's no conflict, because the predicates are different. The previous patch removed the test for XSAVE in the emulation of RDMSR/WRMSR(IA32_XSS), because the presence of the IA32_XSS MSR depends only on CPUID.(EAX=0DH, ECX=1):EAX.XSS[bit 3]. However, architecturally, the XSAVES instruction will raise #UD if either CPUID.01H:ECX.XSAVE[bit 26] = 0 or CPUID.(EAX=0DH,ECX=1):EAX.XSS[bit 3] = 0. (Unfortunately, AMD does not allow us to enforce architectural behavior without breaking the existing ABI, as Paolo pointed out previously. This puts us in the bizarre situation where an AMD guest may be able to execute XSAVES even though it cannot read or write IA32_XSS.) > Alternatively, couldn't KVM check arch.xsaves_enabled in the MSR handling? No. The existence of the IA32_XSS MSR is not the same as the ability of the guest to execute XSAVES. > > + > > if (!xsaves_enabled) > > exec_control &= ~SECONDARY_EXEC_XSAVES; > > > > @@ -6540,7 +6535,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu) > > if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP) > > vmx_set_interrupt_shadow(vcpu, 0); > > > > - kvm_load_guest_xcr0(vcpu); > > + kvm_load_guest_xsave_controls(vcpu); > > > > if (static_cpu_has(X86_FEATURE_PKU) && > > kvm_read_cr4_bits(vcpu, X86_CR4_PKE) && > > @@ -6647,7 +6642,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu) > > __write_pkru(vmx->host_pkru); > > } > > > > - kvm_put_guest_xcr0(vcpu); > > + kvm_load_host_xsave_controls(vcpu); > > > > vmx->nested.nested_run_pending = 0; > > vmx->idt_vectoring_info = 0; > > @@ -7091,6 +7086,12 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > { > > struct vcpu_vmx *vmx = to_vmx(vcpu); > > > > + /* > > + * This will be set back to true in vmx_compute_secondary_exec_control() > > + * if it should be true, so setting it to false here is okay. > > Or simply: > > /* xsaves_enabled is recomputed in vmx_compute_secondary_exec_control(). */ > > > + */ > > + vcpu->arch.xsaves_enabled = false; > > + > > if (cpu_has_secondary_exec_ctrls()) { > > vmx_compute_secondary_exec_control(vmx); > > vmcs_set_secondary_exec_control(vmx); > > @@ -7599,9 +7600,6 @@ static __init int hardware_setup(void) > > WARN_ONCE(host_bndcfgs, "KVM: BNDCFGS in host will be lost"); > > } > > > > - if (boot_cpu_has(X86_FEATURE_XSAVES)) > > - rdmsrl(MSR_IA32_XSS, host_xss); > > - > > if (!cpu_has_vmx_vpid() || !cpu_has_vmx_invvpid() || > > !(cpu_has_vmx_invvpid_single() || cpu_has_vmx_invvpid_global())) > > enable_vpid = 0; > > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > > index 661e2bf38526..a61570d7034b 100644 > > --- a/arch/x86/kvm/x86.c > > +++ b/arch/x86/kvm/x86.c > > @@ -176,6 +176,8 @@ struct kvm_shared_msrs { > > static struct kvm_shared_msrs_global __read_mostly shared_msrs_global; > > static struct kvm_shared_msrs __percpu *shared_msrs; > > > > +static u64 __read_mostly host_xss; > > + > > struct kvm_stats_debugfs_item debugfs_entries[] = { > > { "pf_fixed", VCPU_STAT(pf_fixed) }, > > { "pf_guest", VCPU_STAT(pf_guest) }, > > @@ -812,27 +814,37 @@ void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw) > > } > > EXPORT_SYMBOL_GPL(kvm_lmsw); > > > > -void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu) > > +void kvm_load_guest_xsave_controls(struct kvm_vcpu *vcpu) > > { > > if (kvm_read_cr4_bits(vcpu, X86_CR4_OSXSAVE) && > > !vcpu->guest_xcr0_loaded) { > > /* kvm_set_xcr() also depends on this */ > > if (vcpu->arch.xcr0 != host_xcr0) > > xsetbv(XCR_XFEATURE_ENABLED_MASK, vcpu->arch.xcr0); > > + > > + if (vcpu->arch.xsaves_enabled && > > + vcpu->arch.ia32_xss != host_xss) > > + wrmsrl(MSR_IA32_XSS, vcpu->arch.ia32_xss); > > + > > vcpu->guest_xcr0_loaded = 1; > > } > > } > > -EXPORT_SYMBOL_GPL(kvm_load_guest_xcr0); > > +EXPORT_SYMBOL_GPL(kvm_load_guest_xsave_controls); > > > > -void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu) > > +void kvm_load_host_xsave_controls(struct kvm_vcpu *vcpu) > > { > > if (vcpu->guest_xcr0_loaded) { > > if (vcpu->arch.xcr0 != host_xcr0) > > xsetbv(XCR_XFEATURE_ENABLED_MASK, host_xcr0); > > + > > + if (vcpu->arch.xsaves_enabled && > > + vcpu->arch.ia32_xss != host_xss) > > + wrmsrl(MSR_IA32_XSS, host_xss); > > + > > vcpu->guest_xcr0_loaded = 0; > > } > > } > > -EXPORT_SYMBOL_GPL(kvm_put_guest_xcr0); > > +EXPORT_SYMBOL_GPL(kvm_load_host_xsave_controls); > > > > static int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr) > > { > > @@ -9293,6 +9305,9 @@ int kvm_arch_hardware_setup(void) > > kvm_default_tsc_scaling_ratio = 1ULL << kvm_tsc_scaling_ratio_frac_bits; > > } > > > > + if (boot_cpu_has(X86_FEATURE_XSAVES)) > > + rdmsrl(MSR_IA32_XSS, host_xss); > > More code that goes beyond the subject of the patch, as host_xss is now > being read and potentially switched on SVM platforms. I don't mind the > change itself being part of this patch (unlike setting xsaves_enabled), > but IMO the subject should make it more clear that this is a common x86 > change. Arguably this should be two patches, one to do what this subject > says and a second to move the code to common x86, but that seems like > overkill. > > > kvm_init_msr_list(); > > return 0; > > } > > diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h > > index dbf7442a822b..0d04e865665b 100644 > > --- a/arch/x86/kvm/x86.h > > +++ b/arch/x86/kvm/x86.h > > @@ -366,7 +366,7 @@ static inline bool kvm_pat_valid(u64 data) > > return (data | ((data & 0x0202020202020202ull) << 1)) == data; > > } > > > > -void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu); > > -void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu); > > +void kvm_load_guest_xsave_controls(struct kvm_vcpu *vcpu); > > +void kvm_load_host_xsave_controls(struct kvm_vcpu *vcpu); > > > > #endif > > -- > > 2.23.0.700.g56cf767bdb-goog > >
On Sat, Oct 12, 2019 at 10:36:15AM -0700, Jim Mattson wrote: > On Fri, Oct 11, 2019 at 5:18 PM Sean Christopherson > <sean.j.christopherson@intel.com> wrote: > > > > On Fri, Oct 11, 2019 at 12:40:29PM -0700, Aaron Lewis wrote: > > > Set IA32_XSS for the guest and host during VM Enter and VM Exit > > > transitions rather than by using the MSR-load areas. > > > > > > By moving away from using the MSR-load area we can have a unified > > > solution for both AMD and Intel. > > > > > > Reviewed-by: Jim Mattson <jmattson@google.com> > > > Signed-off-by: Aaron Lewis <aaronlewis@google.com> > > > --- > > > arch/x86/include/asm/kvm_host.h | 1 + > > > arch/x86/kvm/svm.c | 7 +++++-- > > > arch/x86/kvm/vmx/vmx.c | 22 ++++++++++------------ > > > arch/x86/kvm/x86.c | 23 +++++++++++++++++++---- > > > arch/x86/kvm/x86.h | 4 ++-- > > > 5 files changed, 37 insertions(+), 20 deletions(-) > > > > > > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > > > index 50eb430b0ad8..634c2598e389 100644 > > > --- a/arch/x86/include/asm/kvm_host.h > > > +++ b/arch/x86/include/asm/kvm_host.h > > > @@ -562,6 +562,7 @@ struct kvm_vcpu_arch { > > > u64 smbase; > > > u64 smi_count; > > > bool tpr_access_reporting; > > > + bool xsaves_enabled; > > > u64 ia32_xss; > > > u64 microcode_version; > > > u64 arch_capabilities; > > > diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > > > index f8ecb6df5106..da69e95beb4d 100644 > > > --- a/arch/x86/kvm/svm.c > > > +++ b/arch/x86/kvm/svm.c > > > @@ -5628,7 +5628,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > > > svm->vmcb->save.cr2 = vcpu->arch.cr2; > > > > > > clgi(); > > > - kvm_load_guest_xcr0(vcpu); > > > + kvm_load_guest_xsave_controls(vcpu); > > > > > > if (lapic_in_kernel(vcpu) && > > > vcpu->arch.apic->lapic_timer.timer_advance_ns) > > > @@ -5778,7 +5778,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > > > if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI)) > > > kvm_before_interrupt(&svm->vcpu); > > > > > > - kvm_put_guest_xcr0(vcpu); > > > + kvm_load_host_xsave_controls(vcpu); > > > stgi(); > > > > > > /* Any pending NMI will happen here */ > > > @@ -5887,6 +5887,9 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu) > > > { > > > struct vcpu_svm *svm = to_svm(vcpu); > > > > > > + vcpu->arch.xsaves_enabled = guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > > > + boot_cpu_has(X86_FEATURE_XSAVES); > > > > This looks very much like a functional change to SVM, which feels wrong > > for a patch with a subject of "KVM: VMX: Use wrmsr for switching between > > guest and host IA32_XSS" feels wrong. Shouldn't this be unconditionally > > set false in this patch, and then enabled in " kvm: svm: Add support for > > XSAVES on AMD"? > > Nothing is being enabled here. Vcpu->arch.xsaves_enabled simply tells > us whether or not the guest can execute the XSAVES instruction. Any > guest with the ability to set CR4.OSXSAVE on an AMD host that supports > XSAVES can use the instruction. Not enabling per se, but it's a functional change as it means MSR_IA32_XSS will be written in kvm_load_{guest,host}_xsave_controls() if host_xss!=0. > > > + > > > /* Update nrips enabled cache */ > > > svm->nrips_enabled = !!guest_cpuid_has(&svm->vcpu, X86_FEATURE_NRIPS); > > > > > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > > > index 409e9a7323f1..ce3020914c69 100644 > > > --- a/arch/x86/kvm/vmx/vmx.c > > > +++ b/arch/x86/kvm/vmx/vmx.c > > > @@ -106,8 +106,6 @@ module_param(enable_apicv, bool, S_IRUGO); > > > static bool __read_mostly nested = 1; > > > module_param(nested, bool, S_IRUGO); > > > > > > -static u64 __read_mostly host_xss; > > > - > > > bool __read_mostly enable_pml = 1; > > > module_param_named(pml, enable_pml, bool, S_IRUGO); > > > > > > @@ -2074,11 +2072,6 @@ static int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) > > > if (data != 0) > > > return 1; > > > vcpu->arch.ia32_xss = data; > > > - if (vcpu->arch.ia32_xss != host_xss) > > > - add_atomic_switch_msr(vmx, MSR_IA32_XSS, > > > - vcpu->arch.ia32_xss, host_xss, false); > > > - else > > > - clear_atomic_switch_msr(vmx, MSR_IA32_XSS); > > > > I'm pretty sure the "vmx_xsaves_supported()" check in this case statement > > can be removed after this patch. The host support check was necessary > > when the MSR load/save list was being used, but moving to xsaves_enabled > > means theres no risk of writing MSR_IA32_XSS when it's not supported. > > I agree. vmx_xsaves_supported() only tells us if the XSAVES > instruction can be enabled. It does not tell us anything about the > existence of IA32_XSS in the guest. > > > > break; > > > case MSR_IA32_RTIT_CTL: > > > if ((pt_mode != PT_MODE_HOST_GUEST) || > > > @@ -4038,6 +4031,8 @@ static void vmx_compute_secondary_exec_control(struct vcpu_vmx *vmx) > > > guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > > > guest_cpuid_has(vcpu, X86_FEATURE_XSAVES); > > > > > > + vcpu->arch.xsaves_enabled = xsaves_enabled; > > > > Doesn't this conflict with the direction of the previous patch, "KVM: VMX: > > Remove unneeded check for X86_FEATURE_XSAVE"? xsaves_enabled is %true iff > > both XSAVE and XSAVES are exposed to the guest. > > There's no conflict, because the predicates are different. The > previous patch removed the test for XSAVE in the emulation of > RDMSR/WRMSR(IA32_XSS), because the presence of the IA32_XSS MSR > depends only on CPUID.(EAX=0DH, ECX=1):EAX.XSS[bit 3]. However, > architecturally, the XSAVES instruction will raise #UD if either > CPUID.01H:ECX.XSAVE[bit 26] = 0 or CPUID.(EAX=0DH,ECX=1):EAX.XSS[bit > 3] = 0. (Unfortunately, AMD does not allow us to enforce architectural > behavior without breaking the existing ABI, as Paolo pointed out > previously. This puts us in the bizarre situation where an AMD guest > may be able to execute XSAVES even though it cannot read or write > IA32_XSS.) > > > Alternatively, couldn't KVM check arch.xsaves_enabled in the MSR handling? > > No. The existence of the IA32_XSS MSR is not the same as the ability > of the guest to execute XSAVES. Ugh, that's obnoxious.
On Mon, Oct 14, 2019 at 12:05 PM Sean Christopherson <sean.j.christopherson@intel.com> wrote: > > On Sat, Oct 12, 2019 at 10:36:15AM -0700, Jim Mattson wrote: > > On Fri, Oct 11, 2019 at 5:18 PM Sean Christopherson > > <sean.j.christopherson@intel.com> wrote: > > > > > > On Fri, Oct 11, 2019 at 12:40:29PM -0700, Aaron Lewis wrote: > > > > Set IA32_XSS for the guest and host during VM Enter and VM Exit > > > > transitions rather than by using the MSR-load areas. > > > > > > > > By moving away from using the MSR-load area we can have a unified > > > > solution for both AMD and Intel. > > > > > > > > Reviewed-by: Jim Mattson <jmattson@google.com> > > > > Signed-off-by: Aaron Lewis <aaronlewis@google.com> > > > > --- > > > > arch/x86/include/asm/kvm_host.h | 1 + > > > > arch/x86/kvm/svm.c | 7 +++++-- > > > > arch/x86/kvm/vmx/vmx.c | 22 ++++++++++------------ > > > > arch/x86/kvm/x86.c | 23 +++++++++++++++++++---- > > > > arch/x86/kvm/x86.h | 4 ++-- > > > > 5 files changed, 37 insertions(+), 20 deletions(-) > > > > > > > > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > > > > index 50eb430b0ad8..634c2598e389 100644 > > > > --- a/arch/x86/include/asm/kvm_host.h > > > > +++ b/arch/x86/include/asm/kvm_host.h > > > > @@ -562,6 +562,7 @@ struct kvm_vcpu_arch { > > > > u64 smbase; > > > > u64 smi_count; > > > > bool tpr_access_reporting; > > > > + bool xsaves_enabled; > > > > u64 ia32_xss; > > > > u64 microcode_version; > > > > u64 arch_capabilities; > > > > diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > > > > index f8ecb6df5106..da69e95beb4d 100644 > > > > --- a/arch/x86/kvm/svm.c > > > > +++ b/arch/x86/kvm/svm.c > > > > @@ -5628,7 +5628,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > > > > svm->vmcb->save.cr2 = vcpu->arch.cr2; > > > > > > > > clgi(); > > > > - kvm_load_guest_xcr0(vcpu); > > > > + kvm_load_guest_xsave_controls(vcpu); > > > > > > > > if (lapic_in_kernel(vcpu) && > > > > vcpu->arch.apic->lapic_timer.timer_advance_ns) > > > > @@ -5778,7 +5778,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) > > > > if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI)) > > > > kvm_before_interrupt(&svm->vcpu); > > > > > > > > - kvm_put_guest_xcr0(vcpu); > > > > + kvm_load_host_xsave_controls(vcpu); > > > > stgi(); > > > > > > > > /* Any pending NMI will happen here */ > > > > @@ -5887,6 +5887,9 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu) > > > > { > > > > struct vcpu_svm *svm = to_svm(vcpu); > > > > > > > > + vcpu->arch.xsaves_enabled = guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > > > > + boot_cpu_has(X86_FEATURE_XSAVES); > > > > > > This looks very much like a functional change to SVM, which feels wrong > > > for a patch with a subject of "KVM: VMX: Use wrmsr for switching between > > > guest and host IA32_XSS" feels wrong. Shouldn't this be unconditionally > > > set false in this patch, and then enabled in " kvm: svm: Add support for > > > XSAVES on AMD"? > > > > Nothing is being enabled here. Vcpu->arch.xsaves_enabled simply tells > > us whether or not the guest can execute the XSAVES instruction. Any > > guest with the ability to set CR4.OSXSAVE on an AMD host that supports > > XSAVES can use the instruction. > > Not enabling per se, but it's a functional change as it means MSR_IA32_XSS > will be written in kvm_load_{guest,host}_xsave_controls() if host_xss!=0. Fortunately, host_xss is guaranteed to be zero for the nonce. :-) Perhaps the commit message just needs to be updated? The only alternatives I see are: 1. Deliberately introducing buggy code to be removed later in the series, or 2. Introducing SVM-specific code first, to be removed later in the series.
On Mon, Oct 14, 2019 at 02:01:41PM -0700, Jim Mattson wrote: > On Mon, Oct 14, 2019 at 12:05 PM Sean Christopherson > <sean.j.christopherson@intel.com> wrote: > > > > On Sat, Oct 12, 2019 at 10:36:15AM -0700, Jim Mattson wrote: > > > On Fri, Oct 11, 2019 at 5:18 PM Sean Christopherson > > > <sean.j.christopherson@intel.com> wrote: > > > > > > > > On Fri, Oct 11, 2019 at 12:40:29PM -0700, Aaron Lewis wrote: > > > > > @@ -5887,6 +5887,9 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu) > > > > > { > > > > > struct vcpu_svm *svm = to_svm(vcpu); > > > > > > > > > > + vcpu->arch.xsaves_enabled = guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && > > > > > + boot_cpu_has(X86_FEATURE_XSAVES); > > > > > > > > This looks very much like a functional change to SVM, which feels wrong > > > > for a patch with a subject of "KVM: VMX: Use wrmsr for switching between > > > > guest and host IA32_XSS" feels wrong. Shouldn't this be unconditionally > > > > set false in this patch, and then enabled in " kvm: svm: Add support for > > > > XSAVES on AMD"? > > > > > > Nothing is being enabled here. Vcpu->arch.xsaves_enabled simply tells > > > us whether or not the guest can execute the XSAVES instruction. Any > > > guest with the ability to set CR4.OSXSAVE on an AMD host that supports > > > XSAVES can use the instruction. > > > > Not enabling per se, but it's a functional change as it means MSR_IA32_XSS > > will be written in kvm_load_{guest,host}_xsave_controls() if host_xss!=0. > > Fortunately, host_xss is guaranteed to be zero for the nonce. :-) And then someone backports something and things go sideways. > Perhaps the commit message just needs to be updated? The only > alternatives I see are: > 1. Deliberately introducing buggy code to be removed later in the series, or > 2. Introducing SVM-specific code first, to be removed later in the series. I'd go with: 2a. Make the change actually local to VMX as the subject implies, and introduce SVM-specific code in a separate patch, to be removed later. Then the series looks like: KVM: VMX: Remove unneeded check for X86_FEATURE_XSAVE KVM: VMX: Use wrmsr for switching between guest and host IA32_XSS kvm: svm: Add support for XSAVES on AMD KVM: x86: Move IA32_XSS handling to common x86 code KVM: x86: Add IA32_XSS to the emulated_msrs list It'd mean introducing vcpu->arch.xsaves_enabled in the VMX patch and temporarily having it unused by SVM, but that's minor and can't backfire if host_xss suddently is non-zero. Side topic, I'm pretty sure vcpu->guest_xcr0_loaded is a waste of an 'int' and a conditional branch. VMX and SVM are the only users, and both unconditionally pair kvm_load_guest_xcr0() with kvm_put_guest_xcr0(). Removing the useless flag as another prerequisite patch would make the intermediate VMX and SVM patches slightly less awkward as they wouldn't have to decide between adding a kludgy check on guest_xcr0_loaded and ignoring it altogether.
diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h index 50eb430b0ad8..634c2598e389 100644 --- a/arch/x86/include/asm/kvm_host.h +++ b/arch/x86/include/asm/kvm_host.h @@ -562,6 +562,7 @@ struct kvm_vcpu_arch { u64 smbase; u64 smi_count; bool tpr_access_reporting; + bool xsaves_enabled; u64 ia32_xss; u64 microcode_version; u64 arch_capabilities; diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c index f8ecb6df5106..da69e95beb4d 100644 --- a/arch/x86/kvm/svm.c +++ b/arch/x86/kvm/svm.c @@ -5628,7 +5628,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) svm->vmcb->save.cr2 = vcpu->arch.cr2; clgi(); - kvm_load_guest_xcr0(vcpu); + kvm_load_guest_xsave_controls(vcpu); if (lapic_in_kernel(vcpu) && vcpu->arch.apic->lapic_timer.timer_advance_ns) @@ -5778,7 +5778,7 @@ static void svm_vcpu_run(struct kvm_vcpu *vcpu) if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI)) kvm_before_interrupt(&svm->vcpu); - kvm_put_guest_xcr0(vcpu); + kvm_load_host_xsave_controls(vcpu); stgi(); /* Any pending NMI will happen here */ @@ -5887,6 +5887,9 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu) { struct vcpu_svm *svm = to_svm(vcpu); + vcpu->arch.xsaves_enabled = guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && + boot_cpu_has(X86_FEATURE_XSAVES); + /* Update nrips enabled cache */ svm->nrips_enabled = !!guest_cpuid_has(&svm->vcpu, X86_FEATURE_NRIPS); diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c index 409e9a7323f1..ce3020914c69 100644 --- a/arch/x86/kvm/vmx/vmx.c +++ b/arch/x86/kvm/vmx/vmx.c @@ -106,8 +106,6 @@ module_param(enable_apicv, bool, S_IRUGO); static bool __read_mostly nested = 1; module_param(nested, bool, S_IRUGO); -static u64 __read_mostly host_xss; - bool __read_mostly enable_pml = 1; module_param_named(pml, enable_pml, bool, S_IRUGO); @@ -2074,11 +2072,6 @@ static int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info) if (data != 0) return 1; vcpu->arch.ia32_xss = data; - if (vcpu->arch.ia32_xss != host_xss) - add_atomic_switch_msr(vmx, MSR_IA32_XSS, - vcpu->arch.ia32_xss, host_xss, false); - else - clear_atomic_switch_msr(vmx, MSR_IA32_XSS); break; case MSR_IA32_RTIT_CTL: if ((pt_mode != PT_MODE_HOST_GUEST) || @@ -4038,6 +4031,8 @@ static void vmx_compute_secondary_exec_control(struct vcpu_vmx *vmx) guest_cpuid_has(vcpu, X86_FEATURE_XSAVE) && guest_cpuid_has(vcpu, X86_FEATURE_XSAVES); + vcpu->arch.xsaves_enabled = xsaves_enabled; + if (!xsaves_enabled) exec_control &= ~SECONDARY_EXEC_XSAVES; @@ -6540,7 +6535,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu) if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP) vmx_set_interrupt_shadow(vcpu, 0); - kvm_load_guest_xcr0(vcpu); + kvm_load_guest_xsave_controls(vcpu); if (static_cpu_has(X86_FEATURE_PKU) && kvm_read_cr4_bits(vcpu, X86_CR4_PKE) && @@ -6647,7 +6642,7 @@ static void vmx_vcpu_run(struct kvm_vcpu *vcpu) __write_pkru(vmx->host_pkru); } - kvm_put_guest_xcr0(vcpu); + kvm_load_host_xsave_controls(vcpu); vmx->nested.nested_run_pending = 0; vmx->idt_vectoring_info = 0; @@ -7091,6 +7086,12 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) { struct vcpu_vmx *vmx = to_vmx(vcpu); + /* + * This will be set back to true in vmx_compute_secondary_exec_control() + * if it should be true, so setting it to false here is okay. + */ + vcpu->arch.xsaves_enabled = false; + if (cpu_has_secondary_exec_ctrls()) { vmx_compute_secondary_exec_control(vmx); vmcs_set_secondary_exec_control(vmx); @@ -7599,9 +7600,6 @@ static __init int hardware_setup(void) WARN_ONCE(host_bndcfgs, "KVM: BNDCFGS in host will be lost"); } - if (boot_cpu_has(X86_FEATURE_XSAVES)) - rdmsrl(MSR_IA32_XSS, host_xss); - if (!cpu_has_vmx_vpid() || !cpu_has_vmx_invvpid() || !(cpu_has_vmx_invvpid_single() || cpu_has_vmx_invvpid_global())) enable_vpid = 0; diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index 661e2bf38526..a61570d7034b 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -176,6 +176,8 @@ struct kvm_shared_msrs { static struct kvm_shared_msrs_global __read_mostly shared_msrs_global; static struct kvm_shared_msrs __percpu *shared_msrs; +static u64 __read_mostly host_xss; + struct kvm_stats_debugfs_item debugfs_entries[] = { { "pf_fixed", VCPU_STAT(pf_fixed) }, { "pf_guest", VCPU_STAT(pf_guest) }, @@ -812,27 +814,37 @@ void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw) } EXPORT_SYMBOL_GPL(kvm_lmsw); -void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu) +void kvm_load_guest_xsave_controls(struct kvm_vcpu *vcpu) { if (kvm_read_cr4_bits(vcpu, X86_CR4_OSXSAVE) && !vcpu->guest_xcr0_loaded) { /* kvm_set_xcr() also depends on this */ if (vcpu->arch.xcr0 != host_xcr0) xsetbv(XCR_XFEATURE_ENABLED_MASK, vcpu->arch.xcr0); + + if (vcpu->arch.xsaves_enabled && + vcpu->arch.ia32_xss != host_xss) + wrmsrl(MSR_IA32_XSS, vcpu->arch.ia32_xss); + vcpu->guest_xcr0_loaded = 1; } } -EXPORT_SYMBOL_GPL(kvm_load_guest_xcr0); +EXPORT_SYMBOL_GPL(kvm_load_guest_xsave_controls); -void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu) +void kvm_load_host_xsave_controls(struct kvm_vcpu *vcpu) { if (vcpu->guest_xcr0_loaded) { if (vcpu->arch.xcr0 != host_xcr0) xsetbv(XCR_XFEATURE_ENABLED_MASK, host_xcr0); + + if (vcpu->arch.xsaves_enabled && + vcpu->arch.ia32_xss != host_xss) + wrmsrl(MSR_IA32_XSS, host_xss); + vcpu->guest_xcr0_loaded = 0; } } -EXPORT_SYMBOL_GPL(kvm_put_guest_xcr0); +EXPORT_SYMBOL_GPL(kvm_load_host_xsave_controls); static int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr) { @@ -9293,6 +9305,9 @@ int kvm_arch_hardware_setup(void) kvm_default_tsc_scaling_ratio = 1ULL << kvm_tsc_scaling_ratio_frac_bits; } + if (boot_cpu_has(X86_FEATURE_XSAVES)) + rdmsrl(MSR_IA32_XSS, host_xss); + kvm_init_msr_list(); return 0; } diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h index dbf7442a822b..0d04e865665b 100644 --- a/arch/x86/kvm/x86.h +++ b/arch/x86/kvm/x86.h @@ -366,7 +366,7 @@ static inline bool kvm_pat_valid(u64 data) return (data | ((data & 0x0202020202020202ull) << 1)) == data; } -void kvm_load_guest_xcr0(struct kvm_vcpu *vcpu); -void kvm_put_guest_xcr0(struct kvm_vcpu *vcpu); +void kvm_load_guest_xsave_controls(struct kvm_vcpu *vcpu); +void kvm_load_host_xsave_controls(struct kvm_vcpu *vcpu); #endif