@@ -524,6 +524,7 @@ struct kvm_x86_ops {
int (*get_tdp_level)(void);
u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
bool (*gb_page_enable)(void);
+ void (*emulate_vmexit)(struct kvm_vcpu *vcpu);
const struct trace_print_flags *exit_reasons_str;
};
@@ -3610,6 +3610,8 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
r = 0;
goto out;
}
+ if (test_and_clear_bit(KVM_REQ_VMEXIT, &vcpu->requests))
+ kvm_x86_ops->emulate_vmexit(vcpu);
}
preempt_disable();
@@ -3638,6 +3640,21 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
else if (kvm_cpu_has_interrupt(vcpu) || req_int_win)
kvm_x86_ops->enable_irq_window(vcpu);
+ /*
+ * With nested KVM the enable_irq_window() function may cause an
+ * #vmexit if the vcpu is running in guest mode. A #vmexit may sleep
+ * and can't be executed at this stage. So we use the request field to
+ * tell KVM that a #vmexit has to be done before we can enter the guest
+ * again. The code below checks for this request.
+ */
+ if (vcpu->requests) {
+ local_irq_enable();
+ preempt_enable();
+ r = 1;
+ goto out;
+ }
+
+
if (kvm_lapic_enabled(vcpu)) {
update_cr8_intercept(vcpu);
kvm_lapic_sync_to_vapic(vcpu);
@@ -38,6 +38,7 @@
#define KVM_REQ_MMU_SYNC 7
#define KVM_REQ_KVMCLOCK_UPDATE 8
#define KVM_REQ_KICK 9
+#define KVM_REQ_VMEXIT 10
#define KVM_USERSPACE_IRQ_SOURCE_ID 0