@@ -7,6 +7,21 @@
#include <linux/module.h>
#include <linux/moduleparam.h>
+#include <linux/printk.h>
+
+#include <asm/mmu_context.h>
+#include <asm/pgalloc.h>
+
+#include "isolation.h"
+
+struct mm_struct kvm_mm = {
+ .mm_rb = RB_ROOT,
+ .mm_users = ATOMIC_INIT(2),
+ .mm_count = ATOMIC_INIT(1),
+ .mmap_sem = __RWSEM_INITIALIZER(kvm_mm.mmap_sem),
+ .page_table_lock = __SPIN_LOCK_UNLOCKED(kvm_mm.page_table_lock),
+ .mmlist = LIST_HEAD_INIT(kvm_mm.mmlist),
+};
/*
* When set to true, KVM #VMExit handlers run in isolated address space
@@ -24,3 +39,83 @@
*/
static bool __read_mostly address_space_isolation;
module_param(address_space_isolation, bool, 0444);
+
+static int kvm_isolation_init_mm(void)
+{
+ pgd_t *kvm_pgd;
+ gfp_t gfp_mask;
+
+ gfp_mask = GFP_KERNEL | __GFP_ZERO;
+ kvm_pgd = (pgd_t *)__get_free_pages(gfp_mask, PGD_ALLOCATION_ORDER);
+ if (!kvm_pgd)
+ return -ENOMEM;
+
+#ifdef CONFIG_PAGE_TABLE_ISOLATION
+ /*
+ * With PTI, we have two PGDs: one the kernel page table, and one
+ * for the user page table. The PGD with the kernel page table has
+ * to be the entire kernel address space because paranoid faults
+ * will unconditionally use it. So we define the KVM address space
+ * in the user table space, although it will be used in the kernel.
+ */
+
+ /* initialize the kernel page table */
+ memcpy(kvm_pgd, current->active_mm->pgd, sizeof(pgd_t) * PTRS_PER_PGD);
+
+ /* define kvm_mm with the user page table */
+ kvm_mm.pgd = kernel_to_user_pgdp(kvm_pgd);
+#else /* CONFIG_PAGE_TABLE_ISOLATION */
+ kvm_mm.pgd = kvm_pgd;
+#endif /* CONFIG_PAGE_TABLE_ISOLATION */
+ mm_init_cpumask(&kvm_mm);
+ init_new_context(NULL, &kvm_mm);
+
+ return 0;
+}
+
+static void kvm_isolation_uninit_mm(void)
+{
+ pgd_t *kvm_pgd;
+
+ BUG_ON(current->active_mm == &kvm_mm);
+
+ destroy_context(&kvm_mm);
+
+#ifdef CONFIG_PAGE_TABLE_ISOLATION
+ /*
+ * With PTI, the KVM address space is defined in the user
+ * page table space, but the full PGD starts with the kernel
+ * page table space.
+ */
+ kvm_pgd = user_to_kernel_pgdp(kvm_pgd);
+#else /* CONFIG_PAGE_TABLE_ISOLATION */
+ kvm_pgd = kvm_mm.pgd;
+#endif /* CONFIG_PAGE_TABLE_ISOLATION */
+ kvm_mm.pgd = NULL;
+ free_pages((unsigned long)kvm_pgd, PGD_ALLOCATION_ORDER);
+}
+
+int kvm_isolation_init(void)
+{
+ int r;
+
+ if (!address_space_isolation)
+ return 0;
+
+ r = kvm_isolation_init_mm();
+ if (r)
+ return r;
+
+ pr_info("KVM: x86: Running with isolated address space\n");
+
+ return 0;
+}
+
+void kvm_isolation_uninit(void)
+{
+ if (!address_space_isolation)
+ return;
+
+ kvm_isolation_uninit_mm();
+ pr_info("KVM: x86: End of isolated address space\n");
+}
new file mode 100644
@@ -0,0 +1,8 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+#ifndef ARCH_X86_KVM_ISOLATION_H
+#define ARCH_X86_KVM_ISOLATION_H
+
+extern int kvm_isolation_init(void);
+extern void kvm_isolation_uninit(void);
+
+#endif
@@ -29,6 +29,7 @@
#include "cpuid.h"
#include "pmu.h"
#include "hyperv.h"
+#include "isolation.h"
#include <linux/clocksource.h>
#include <linux/interrupt.h>
@@ -6972,10 +6973,14 @@ int kvm_arch_init(void *opaque)
goto out_free_x86_fpu_cache;
}
- r = kvm_mmu_module_init();
+ r = kvm_isolation_init();
if (r)
goto out_free_percpu;
+ r = kvm_mmu_module_init();
+ if (r)
+ goto out_uninit_isolation;
+
kvm_set_mmio_spte_mask();
kvm_x86_ops = ops;
@@ -7000,6 +7005,8 @@ int kvm_arch_init(void *opaque)
return 0;
+out_uninit_isolation:
+ kvm_isolation_uninit();
out_free_percpu:
free_percpu(shared_msrs);
out_free_x86_fpu_cache:
@@ -7024,6 +7031,7 @@ void kvm_arch_exit(void)
#ifdef CONFIG_X86_64
pvclock_gtod_unregister_notifier(&pvclock_gtod_notifier);
#endif
+ kvm_isolation_uninit();
kvm_x86_ops = NULL;
kvm_mmu_module_exit();
free_percpu(shared_msrs);