Message ID | alpine.DEB.2.02.1304151506560.2551@ahonig-virtual-machine (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Il 16/04/2013 00:10, Andrew Honig ha scritto: > > The motivation for this patch is to fix a 20KB leak of memory in vmx.c > when a VM is created and destroyed. > > On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for > architecture specific reasons. It currently allocates the pages on behalf > of user space, but has no way of cleanly freeing that memory while the > user space process is still running. For user space processes that want > more control over that memory, this patch allows user space to provide the > memory that KVM uses. Isn't the bug simply that kvm_arch_free_memslot needs to free the memory that was allocated in kvm_arch_prepare_memory_region? Then closing /dev/kvm will fix it. The conditions for doing vm_mmap and vm_munmap in kvm_arch_prepare_memory_region/kvm_arch_commit_memory_region perhaps can also be simplified. The vm_munmamp in kvm_arch_commit_memory_region probably can go completely, the test on kvm_arch_prepare_memory_region simplified to npages != old.npages, and kvm_arch_free_memslot can check something like "if (free->user_alloc && (!dont || !dont->user_alloc || free->userspace_addr != dont->userspace_addr))". Paolo > Signed-off-by: Andrew Honig <ahonig@google.com> > --- > Documentation/virtual/kvm/api.txt | 8 +++++++ > arch/arm/kvm/arm.c | 6 +++++ > arch/ia64/kvm/kvm-ia64.c | 6 +++++ > arch/powerpc/kvm/powerpc.c | 6 +++++ > arch/s390/kvm/kvm-s390.c | 6 +++++ > arch/x86/include/asm/kvm_host.h | 7 ++++++ > arch/x86/kvm/svm.c | 8 +++++++ > arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- > arch/x86/kvm/x86.c | 12 ++++++++-- > include/linux/kvm_host.h | 2 ++ > virt/kvm/kvm_main.c | 2 ++ > 11 files changed, 105 insertions(+), 5 deletions(-) > > diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt > index 119358d..aa18cac 100644 > --- a/Documentation/virtual/kvm/api.txt > +++ b/Documentation/virtual/kvm/api.txt > @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr > be identical. This allows large pages in the guest to be backed by large > pages in the host. > > +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture > +specific reasons. Calling this ioctl with the special memslot > +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for > +that memory. If that memslot is not set before creating VCPUs for a VM then > +kvm will allocate the memory on behalf of user space, but userspace will not > +be able to free that memory. User space should treat this memory as opaque > +and not modify it. > + > The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and > KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of > writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to > diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c > index 5a93698..ac52f14 100644 > --- a/arch/arm/kvm/arm.c > +++ b/arch/arm/kvm/arm.c > @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, > return 0; > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c > index ad3126a..570dd97 100644 > --- a/arch/ia64/kvm/kvm-ia64.c > +++ b/arch/ia64/kvm/kvm-ia64.c > @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > return 0; > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c > index 934413c..6e3843b 100644 > --- a/arch/powerpc/kvm/powerpc.c > +++ b/arch/powerpc/kvm/powerpc.c > @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > return kvmppc_core_create_memslot(slot, npages); > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c > index 4cf35a0..a97f495 100644 > --- a/arch/s390/kvm/kvm-s390.c > +++ b/arch/s390/kvm/kvm-s390.c > @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > return 0; > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > /* Section: memory related */ > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > index 4979778..7215817 100644 > --- a/arch/x86/include/asm/kvm_host.h > +++ b/arch/x86/include/asm/kvm_host.h > @@ -37,6 +37,7 @@ > /* memory slots that are not exposed to userspace */ > #define KVM_PRIVATE_MEM_SLOTS 3 > #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) > +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 > > #define KVM_MMIO_SIZE 16 > > @@ -553,6 +554,9 @@ struct kvm_arch { > struct page *ept_identity_pagetable; > bool ept_identity_pagetable_done; > gpa_t ept_identity_map_addr; > + unsigned long ept_ptr; > + unsigned long apic_ptr; > + unsigned long tss_ptr; > > unsigned long irq_sourcesq_bitmap; > s64 kvmclock_offset; > @@ -640,6 +644,9 @@ struct kvm_x86_ops { > bool (*cpu_has_accelerated_tpr)(void); > void (*cpuid_update)(struct kvm_vcpu *vcpu); > > + int (*set_private_memory)(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem); > + > /* Create, but do not attach this VCPU */ > struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); > void (*vcpu_free)(struct kvm_vcpu *vcpu); > diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > index e1b1ce2..3cc4e56 100644 > --- a/arch/x86/kvm/svm.c > +++ b/arch/x86/kvm/svm.c > @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) > return 0; > } > > +static int svm_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) > { > struct vcpu_svm *svm; > @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { > .hardware_disable = svm_hardware_disable, > .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, > > + .set_private_memory = svm_set_private_memory, > + > .vcpu_create = svm_create_vcpu, > .vcpu_free = svm_free_vcpu, > .vcpu_reset = svm_vcpu_reset, > diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c > index 6667042..796ac07 100644 > --- a/arch/x86/kvm/vmx.c > +++ b/arch/x86/kvm/vmx.c > @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) > kvm_userspace_mem.flags = 0; > kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; > kvm_userspace_mem.memory_size = PAGE_SIZE; > - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + if (kvm->arch.apic_ptr) { > + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); > + } else { > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + } > + > if (r) > goto out; > > @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) > kvm_userspace_mem.guest_phys_addr = > kvm->arch.ept_identity_map_addr; > kvm_userspace_mem.memory_size = PAGE_SIZE; > - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + if (kvm->arch.ept_ptr) { > + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); > + } else { > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + } > + > if (r) > goto out; > > @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) > .flags = 0, > }; > > - ret = kvm_set_memory_region(kvm, &tss_mem, false); > + if (kvm->arch.tss_ptr) { > + tss_mem.userspace_addr = kvm->arch.tss_ptr; > + ret = kvm_set_memory_region(kvm, &tss_mem, true); > + } else { > + ret = kvm_set_memory_region(kvm, &tss_mem, false); > + } > + > if (ret) > return ret; > kvm->arch.tss_addr = addr; > @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) > vmx_complete_interrupts(vmx); > } > > +static int vmx_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + /* > + * Early sanity checking so userspace gets an error message during > + * memory setup and not when trying to use this memory. > + * Checks to see if the memory is valid are performed later when > + * the memory is used. > + */ > + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || > + mem->memory_size & (PAGE_SIZE - 1) || > + mem->memory_size < PAGE_SIZE * 5) > + return -EINVAL; > + > + kvm->arch.ept_ptr = mem->userspace_addr; > + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; > + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; > + > + return 0; > +} > + > static void vmx_free_vcpu(struct kvm_vcpu *vcpu) > { > struct vcpu_vmx *vmx = to_vmx(vcpu); > @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { > .hardware_disable = hardware_disable, > .cpu_has_accelerated_tpr = report_flexpriority, > > + .set_private_memory = vmx_set_private_memory, > + > .vcpu_create = vmx_create_vcpu, > .vcpu_free = vmx_free_vcpu, > .vcpu_reset = vmx_vcpu_reset, > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index e172132..7045d0a 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) > kvm_free_pit(kvm); > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return kvm_x86_ops->set_private_memory(kvm, mem); > +} > + > void kvm_arch_destroy_vm(struct kvm *kvm) > { > kvm_iommu_unmap_guest(kvm); > @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, > * Only private memory slots need to be mapped here since > * KVM_SET_MEMORY_REGION ioctl is no longer supported. > */ > - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { > + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && > + !user_alloc) { > unsigned long userspace_addr; > > /* > @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, > > int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; > > - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { > + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && > + !user_alloc) { > int ret; > > ret = vm_munmap(old.userspace_addr, > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index c139582..f441d1f 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, > void kvm_arch_free_memslot(struct kvm_memory_slot *free, > struct kvm_memory_slot *dont); > int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem); > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index f18013f..5372225 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, > kvm_userspace_memory_region *mem, > bool user_alloc) > { > + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) > + return kvm_arch_set_private_memory(kvm, mem); > if (mem->slot >= KVM_USER_MEM_SLOTS) > return -EINVAL; > return kvm_set_memory_region(kvm, mem, user_alloc); > -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Mon, Apr 15, 2013 at 03:10:32PM -0700, Andrew Honig wrote: > > The motivation for this patch is to fix a 20KB leak of memory in vmx.c > when a VM is created and destroyed. > > On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for > architecture specific reasons. It currently allocates the pages on behalf > of user space, but has no way of cleanly freeing that memory while the > user space process is still running. For user space processes that want > more control over that memory, this patch allows user space to provide the > memory that KVM uses. > > Signed-off-by: Andrew Honig <ahonig@google.com> I thought we agreed on not adding new API and using __kvm_set_memory_region() to unregister private memory regions. > --- > Documentation/virtual/kvm/api.txt | 8 +++++++ > arch/arm/kvm/arm.c | 6 +++++ > arch/ia64/kvm/kvm-ia64.c | 6 +++++ > arch/powerpc/kvm/powerpc.c | 6 +++++ > arch/s390/kvm/kvm-s390.c | 6 +++++ > arch/x86/include/asm/kvm_host.h | 7 ++++++ > arch/x86/kvm/svm.c | 8 +++++++ > arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- > arch/x86/kvm/x86.c | 12 ++++++++-- > include/linux/kvm_host.h | 2 ++ > virt/kvm/kvm_main.c | 2 ++ > 11 files changed, 105 insertions(+), 5 deletions(-) > > diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt > index 119358d..aa18cac 100644 > --- a/Documentation/virtual/kvm/api.txt > +++ b/Documentation/virtual/kvm/api.txt > @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr > be identical. This allows large pages in the guest to be backed by large > pages in the host. > > +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture > +specific reasons. Calling this ioctl with the special memslot > +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for > +that memory. If that memslot is not set before creating VCPUs for a VM then > +kvm will allocate the memory on behalf of user space, but userspace will not > +be able to free that memory. User space should treat this memory as opaque > +and not modify it. > + > The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and > KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of > writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to > diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c > index 5a93698..ac52f14 100644 > --- a/arch/arm/kvm/arm.c > +++ b/arch/arm/kvm/arm.c > @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, > return 0; > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c > index ad3126a..570dd97 100644 > --- a/arch/ia64/kvm/kvm-ia64.c > +++ b/arch/ia64/kvm/kvm-ia64.c > @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > return 0; > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c > index 934413c..6e3843b 100644 > --- a/arch/powerpc/kvm/powerpc.c > +++ b/arch/powerpc/kvm/powerpc.c > @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > return kvmppc_core_create_memslot(slot, npages); > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c > index 4cf35a0..a97f495 100644 > --- a/arch/s390/kvm/kvm-s390.c > +++ b/arch/s390/kvm/kvm-s390.c > @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > return 0; > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > /* Section: memory related */ > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > index 4979778..7215817 100644 > --- a/arch/x86/include/asm/kvm_host.h > +++ b/arch/x86/include/asm/kvm_host.h > @@ -37,6 +37,7 @@ > /* memory slots that are not exposed to userspace */ > #define KVM_PRIVATE_MEM_SLOTS 3 > #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) > +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 > > #define KVM_MMIO_SIZE 16 > > @@ -553,6 +554,9 @@ struct kvm_arch { > struct page *ept_identity_pagetable; > bool ept_identity_pagetable_done; > gpa_t ept_identity_map_addr; > + unsigned long ept_ptr; > + unsigned long apic_ptr; > + unsigned long tss_ptr; > > unsigned long irq_sources_bitmap; > s64 kvmclock_offset; > @@ -640,6 +644,9 @@ struct kvm_x86_ops { > bool (*cpu_has_accelerated_tpr)(void); > void (*cpuid_update)(struct kvm_vcpu *vcpu); > > + int (*set_private_memory)(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem); > + > /* Create, but do not attach this VCPU */ > struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); > void (*vcpu_free)(struct kvm_vcpu *vcpu); > diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > index e1b1ce2..3cc4e56 100644 > --- a/arch/x86/kvm/svm.c > +++ b/arch/x86/kvm/svm.c > @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) > return 0; > } > > +static int svm_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return 0; > +} > + > static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) > { > struct vcpu_svm *svm; > @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { > .hardware_disable = svm_hardware_disable, > .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, > > + .set_private_memory = svm_set_private_memory, > + > .vcpu_create = svm_create_vcpu, > .vcpu_free = svm_free_vcpu, > .vcpu_reset = svm_vcpu_reset, > diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c > index 6667042..796ac07 100644 > --- a/arch/x86/kvm/vmx.c > +++ b/arch/x86/kvm/vmx.c > @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) > kvm_userspace_mem.flags = 0; > kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; > kvm_userspace_mem.memory_size = PAGE_SIZE; > - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + if (kvm->arch.apic_ptr) { > + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); > + } else { > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + } > + > if (r) > goto out; > > @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) > kvm_userspace_mem.guest_phys_addr = > kvm->arch.ept_identity_map_addr; > kvm_userspace_mem.memory_size = PAGE_SIZE; > - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + if (kvm->arch.ept_ptr) { > + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); > + } else { > + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > + } > + > if (r) > goto out; > > @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) > .flags = 0, > }; > > - ret = kvm_set_memory_region(kvm, &tss_mem, false); > + if (kvm->arch.tss_ptr) { > + tss_mem.userspace_addr = kvm->arch.tss_ptr; > + ret = kvm_set_memory_region(kvm, &tss_mem, true); > + } else { > + ret = kvm_set_memory_region(kvm, &tss_mem, false); > + } > + > if (ret) > return ret; > kvm->arch.tss_addr = addr; > @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) > vmx_complete_interrupts(vmx); > } > > +static int vmx_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + /* > + * Early sanity checking so userspace gets an error message during > + * memory setup and not when trying to use this memory. > + * Checks to see if the memory is valid are performed later when > + * the memory is used. > + */ > + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || > + mem->memory_size & (PAGE_SIZE - 1) || > + mem->memory_size < PAGE_SIZE * 5) > + return -EINVAL; > + > + kvm->arch.ept_ptr = mem->userspace_addr; > + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; > + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; > + > + return 0; > +} > + > static void vmx_free_vcpu(struct kvm_vcpu *vcpu) > { > struct vcpu_vmx *vmx = to_vmx(vcpu); > @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { > .hardware_disable = hardware_disable, > .cpu_has_accelerated_tpr = report_flexpriority, > > + .set_private_memory = vmx_set_private_memory, > + > .vcpu_create = vmx_create_vcpu, > .vcpu_free = vmx_free_vcpu, > .vcpu_reset = vmx_vcpu_reset, > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index e172132..7045d0a 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) > kvm_free_pit(kvm); > } > > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem) > +{ > + return kvm_x86_ops->set_private_memory(kvm, mem); > +} > + > void kvm_arch_destroy_vm(struct kvm *kvm) > { > kvm_iommu_unmap_guest(kvm); > @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, > * Only private memory slots need to be mapped here since > * KVM_SET_MEMORY_REGION ioctl is no longer supported. > */ > - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { > + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && > + !user_alloc) { > unsigned long userspace_addr; > > /* > @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, > > int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; > > - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { > + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && > + !user_alloc) { > int ret; > > ret = vm_munmap(old.userspace_addr, > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index c139582..f441d1f 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, > void kvm_arch_free_memslot(struct kvm_memory_slot *free, > struct kvm_memory_slot *dont); > int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); > +int kvm_arch_set_private_memory(struct kvm *kvm, > + struct kvm_userspace_memory_region *mem); > int kvm_arch_prepare_memory_region(struct kvm *kvm, > struct kvm_memory_slot *memslot, > struct kvm_memory_slot old, > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index f18013f..5372225 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, > kvm_userspace_memory_region *mem, > bool user_alloc) > { > + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) > + return kvm_arch_set_private_memory(kvm, mem); > if (mem->slot >= KVM_USER_MEM_SLOTS) > return -EINVAL; > return kvm_set_memory_region(kvm, mem, user_alloc); > -- > 1.7.10.4 > > -- > To unsubscribe from this list: send the line "unsubscribe kvm" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Gleb. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Freeing the memory in kvm_arch_free_memslot is as good as anywhere else in KVM. The problem is that this memory is in the user space process mm. This codepath could be called after the mm is destroyed in the case of an process exit without closing the fd, which will result in a panic on vm_munmap when it tries to access the mm. There's also the possibility that another process closes the fd and messing with that processes memory map seems like it's asking for trouble. On Wed, Apr 17, 2013 at 4:42 AM, Paolo Bonzini <pbonzini@redhat.com> wrote: > Il 16/04/2013 00:10, Andrew Honig ha scritto: >> >> The motivation for this patch is to fix a 20KB leak of memory in vmx.c >> when a VM is created and destroyed. >> >> On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for >> architecture specific reasons. It currently allocates the pages on behalf >> of user space, but has no way of cleanly freeing that memory while the >> user space process is still running. For user space processes that want >> more control over that memory, this patch allows user space to provide the >> memory that KVM uses. > > Isn't the bug simply that kvm_arch_free_memslot needs to free the memory > that was allocated in kvm_arch_prepare_memory_region? Then closing > /dev/kvm will fix it. > > The conditions for doing vm_mmap and vm_munmap in > kvm_arch_prepare_memory_region/kvm_arch_commit_memory_region perhaps can > also be simplified. > > The vm_munmamp in kvm_arch_commit_memory_region probably can go > completely, the test on kvm_arch_prepare_memory_region simplified to > npages != old.npages, and kvm_arch_free_memslot can check something like > "if (free->user_alloc && (!dont || !dont->user_alloc || > free->userspace_addr != dont->userspace_addr))". > > Paolo > >> Signed-off-by: Andrew Honig <ahonig@google.com> >> --- >> Documentation/virtual/kvm/api.txt | 8 +++++++ >> arch/arm/kvm/arm.c | 6 +++++ >> arch/ia64/kvm/kvm-ia64.c | 6 +++++ >> arch/powerpc/kvm/powerpc.c | 6 +++++ >> arch/s390/kvm/kvm-s390.c | 6 +++++ >> arch/x86/include/asm/kvm_host.h | 7 ++++++ >> arch/x86/kvm/svm.c | 8 +++++++ >> arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- >> arch/x86/kvm/x86.c | 12 ++++++++-- >> include/linux/kvm_host.h | 2 ++ >> virt/kvm/kvm_main.c | 2 ++ >> 11 files changed, 105 insertions(+), 5 deletions(-) >> >> diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt >> index 119358d..aa18cac 100644 >> --- a/Documentation/virtual/kvm/api.txt >> +++ b/Documentation/virtual/kvm/api.txt >> @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr >> be identical. This allows large pages in the guest to be backed by large >> pages in the host. >> >> +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture >> +specific reasons. Calling this ioctl with the special memslot >> +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for >> +that memory. If that memslot is not set before creating VCPUs for a VM then >> +kvm will allocate the memory on behalf of user space, but userspace will not >> +be able to free that memory. User space should treat this memory as opaque >> +and not modify it. >> + >> The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and >> KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of >> writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c >> index 5a93698..ac52f14 100644 >> --- a/arch/arm/kvm/arm.c >> +++ b/arch/arm/kvm/arm.c >> @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, >> return 0; >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c >> index ad3126a..570dd97 100644 >> --- a/arch/ia64/kvm/kvm-ia64.c >> +++ b/arch/ia64/kvm/kvm-ia64.c >> @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> return 0; >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c >> index 934413c..6e3843b 100644 >> --- a/arch/powerpc/kvm/powerpc.c >> +++ b/arch/powerpc/kvm/powerpc.c >> @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> return kvmppc_core_create_memslot(slot, npages); >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c >> index 4cf35a0..a97f495 100644 >> --- a/arch/s390/kvm/kvm-s390.c >> +++ b/arch/s390/kvm/kvm-s390.c >> @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> return 0; >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> /* Section: memory related */ >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h >> index 4979778..7215817 100644 >> --- a/arch/x86/include/asm/kvm_host.h >> +++ b/arch/x86/include/asm/kvm_host.h >> @@ -37,6 +37,7 @@ >> /* memory slots that are not exposed to userspace */ >> #define KVM_PRIVATE_MEM_SLOTS 3 >> #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) >> +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 >> >> #define KVM_MMIO_SIZE 16 >> >> @@ -553,6 +554,9 @@ struct kvm_arch { >> struct page *ept_identity_pagetable; >> bool ept_identity_pagetable_done; >> gpa_t ept_identity_map_addr; >> + unsigned long ept_ptr; >> + unsigned long apic_ptr; >> + unsigned long tss_ptr; >> >> unsigned long irq_sourcesq_bitmap; >> s64 kvmclock_offset; >> @@ -640,6 +644,9 @@ struct kvm_x86_ops { >> bool (*cpu_has_accelerated_tpr)(void); >> void (*cpuid_update)(struct kvm_vcpu *vcpu); >> >> + int (*set_private_memory)(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem); >> + >> /* Create, but do not attach this VCPU */ >> struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); >> void (*vcpu_free)(struct kvm_vcpu *vcpu); >> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c >> index e1b1ce2..3cc4e56 100644 >> --- a/arch/x86/kvm/svm.c >> +++ b/arch/x86/kvm/svm.c >> @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) >> return 0; >> } >> >> +static int svm_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) >> { >> struct vcpu_svm *svm; >> @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { >> .hardware_disable = svm_hardware_disable, >> .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, >> >> + .set_private_memory = svm_set_private_memory, >> + >> .vcpu_create = svm_create_vcpu, >> .vcpu_free = svm_free_vcpu, >> .vcpu_reset = svm_vcpu_reset, >> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c >> index 6667042..796ac07 100644 >> --- a/arch/x86/kvm/vmx.c >> +++ b/arch/x86/kvm/vmx.c >> @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) >> kvm_userspace_mem.flags = 0; >> kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + if (kvm->arch.apic_ptr) { >> + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> + } else { >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + } >> + >> if (r) >> goto out; >> >> @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) >> kvm_userspace_mem.guest_phys_addr = >> kvm->arch.ept_identity_map_addr; >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + if (kvm->arch.ept_ptr) { >> + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> + } else { >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + } >> + >> if (r) >> goto out; >> >> @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) >> .flags = 0, >> }; >> >> - ret = kvm_set_memory_region(kvm, &tss_mem, false); >> + if (kvm->arch.tss_ptr) { >> + tss_mem.userspace_addr = kvm->arch.tss_ptr; >> + ret = kvm_set_memory_region(kvm, &tss_mem, true); >> + } else { >> + ret = kvm_set_memory_region(kvm, &tss_mem, false); >> + } >> + >> if (ret) >> return ret; >> kvm->arch.tss_addr = addr; >> @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) >> vmx_complete_interrupts(vmx); >> } >> >> +static int vmx_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + /* >> + * Early sanity checking so userspace gets an error message during >> + * memory setup and not when trying to use this memory. >> + * Checks to see if the memory is valid are performed later when >> + * the memory is used. >> + */ >> + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || >> + mem->memory_size & (PAGE_SIZE - 1) || >> + mem->memory_size < PAGE_SIZE * 5) >> + return -EINVAL; >> + >> + kvm->arch.ept_ptr = mem->userspace_addr; >> + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; >> + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; >> + >> + return 0; >> +} >> + >> static void vmx_free_vcpu(struct kvm_vcpu *vcpu) >> { >> struct vcpu_vmx *vmx = to_vmx(vcpu); >> @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { >> .hardware_disable = hardware_disable, >> .cpu_has_accelerated_tpr = report_flexpriority, >> >> + .set_private_memory = vmx_set_private_memory, >> + >> .vcpu_create = vmx_create_vcpu, >> .vcpu_free = vmx_free_vcpu, >> .vcpu_reset = vmx_vcpu_reset, >> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c >> index e172132..7045d0a 100644 >> --- a/arch/x86/kvm/x86.c >> +++ b/arch/x86/kvm/x86.c >> @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) >> kvm_free_pit(kvm); >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return kvm_x86_ops->set_private_memory(kvm, mem); >> +} >> + >> void kvm_arch_destroy_vm(struct kvm *kvm) >> { >> kvm_iommu_unmap_guest(kvm); >> @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, >> * Only private memory slots need to be mapped here since >> * KVM_SET_MEMORY_REGION ioctl is no longer supported. >> */ >> - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { >> + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && >> + !user_alloc) { >> unsigned long userspace_addr; >> >> /* >> @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, >> >> int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; >> >> - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { >> + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && >> + !user_alloc) { >> int ret; >> >> ret = vm_munmap(old.userspace_addr, >> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h >> index c139582..f441d1f 100644 >> --- a/include/linux/kvm_host.h >> +++ b/include/linux/kvm_host.h >> @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, >> void kvm_arch_free_memslot(struct kvm_memory_slot *free, >> struct kvm_memory_slot *dont); >> int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem); >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c >> index f18013f..5372225 100644 >> --- a/virt/kvm/kvm_main.c >> +++ b/virt/kvm/kvm_main.c >> @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, >> kvm_userspace_memory_region *mem, >> bool user_alloc) >> { >> + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) >> + return kvm_arch_set_private_memory(kvm, mem); >> if (mem->slot >= KVM_USER_MEM_SLOTS) >> return -EINVAL; >> return kvm_set_memory_region(kvm, mem, user_alloc); >> > -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
I'm happy to not add a new API and use __kvm_set_memory_region to unregister private memory regions, but I thought chaning the API was the approach you asked for when I sent a previous patch. See the end of: http://article.gmane.org/gmane.comp.emulators.kvm.devel/107753 Did I misunderstand your comment from 8 April? On Wed, Apr 17, 2013 at 6:10 AM, Gleb Natapov <gleb@redhat.com> wrote: > On Mon, Apr 15, 2013 at 03:10:32PM -0700, Andrew Honig wrote: >> >> The motivation for this patch is to fix a 20KB leak of memory in vmx.c >> when a VM is created and destroyed. >> >> On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for >> architecture specific reasons. It currently allocates the pages on behalf >> of user space, but has no way of cleanly freeing that memory while the >> user space process is still running. For user space processes that want >> more control over that memory, this patch allows user space to provide the >> memory that KVM uses. >> >> Signed-off-by: Andrew Honig <ahonig@google.com> > I thought we agreed on not adding new API and using > __kvm_set_memory_region() to unregister private memory regions. > > >> --- >> Documentation/virtual/kvm/api.txt | 8 +++++++ >> arch/arm/kvm/arm.c | 6 +++++ >> arch/ia64/kvm/kvm-ia64.c | 6 +++++ >> arch/powerpc/kvm/powerpc.c | 6 +++++ >> arch/s390/kvm/kvm-s390.c | 6 +++++ >> arch/x86/include/asm/kvm_host.h | 7 ++++++ >> arch/x86/kvm/svm.c | 8 +++++++ >> arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- >> arch/x86/kvm/x86.c | 12 ++++++++-- >> include/linux/kvm_host.h | 2 ++ >> virt/kvm/kvm_main.c | 2 ++ >> 11 files changed, 105 insertions(+), 5 deletions(-) >> >> diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt >> index 119358d..aa18cac 100644 >> --- a/Documentation/virtual/kvm/api.txt >> +++ b/Documentation/virtual/kvm/api.txt >> @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr >> be identical. This allows large pages in the guest to be backed by large >> pages in the host. >> >> +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture >> +specific reasons. Calling this ioctl with the special memslot >> +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for >> +that memory. If that memslot is not set before creating VCPUs for a VM then >> +kvm will allocate the memory on behalf of user space, but userspace will not >> +be able to free that memory. User space should treat this memory as opaque >> +and not modify it. >> + >> The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and >> KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of >> writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c >> index 5a93698..ac52f14 100644 >> --- a/arch/arm/kvm/arm.c >> +++ b/arch/arm/kvm/arm.c >> @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, >> return 0; >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c >> index ad3126a..570dd97 100644 >> --- a/arch/ia64/kvm/kvm-ia64.c >> +++ b/arch/ia64/kvm/kvm-ia64.c >> @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> return 0; >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c >> index 934413c..6e3843b 100644 >> --- a/arch/powerpc/kvm/powerpc.c >> +++ b/arch/powerpc/kvm/powerpc.c >> @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> return kvmppc_core_create_memslot(slot, npages); >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c >> index 4cf35a0..a97f495 100644 >> --- a/arch/s390/kvm/kvm-s390.c >> +++ b/arch/s390/kvm/kvm-s390.c >> @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> return 0; >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> /* Section: memory related */ >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h >> index 4979778..7215817 100644 >> --- a/arch/x86/include/asm/kvm_host.h >> +++ b/arch/x86/include/asm/kvm_host.h >> @@ -37,6 +37,7 @@ >> /* memory slots that are not exposed to userspace */ >> #define KVM_PRIVATE_MEM_SLOTS 3 >> #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) >> +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 >> >> #define KVM_MMIO_SIZE 16 >> >> @@ -553,6 +554,9 @@ struct kvm_arch { >> struct page *ept_identity_pagetable; >> bool ept_identity_pagetable_done; >> gpa_t ept_identity_map_addr; >> + unsigned long ept_ptr; >> + unsigned long apic_ptr; >> + unsigned long tss_ptr; >> >> unsigned long irq_sources_bitmap; >> s64 kvmclock_offset; >> @@ -640,6 +644,9 @@ struct kvm_x86_ops { >> bool (*cpu_has_accelerated_tpr)(void); >> void (*cpuid_update)(struct kvm_vcpu *vcpu); >> >> + int (*set_private_memory)(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem); >> + >> /* Create, but do not attach this VCPU */ >> struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); >> void (*vcpu_free)(struct kvm_vcpu *vcpu); >> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c >> index e1b1ce2..3cc4e56 100644 >> --- a/arch/x86/kvm/svm.c >> +++ b/arch/x86/kvm/svm.c >> @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) >> return 0; >> } >> >> +static int svm_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return 0; >> +} >> + >> static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) >> { >> struct vcpu_svm *svm; >> @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { >> .hardware_disable = svm_hardware_disable, >> .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, >> >> + .set_private_memory = svm_set_private_memory, >> + >> .vcpu_create = svm_create_vcpu, >> .vcpu_free = svm_free_vcpu, >> .vcpu_reset = svm_vcpu_reset, >> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c >> index 6667042..796ac07 100644 >> --- a/arch/x86/kvm/vmx.c >> +++ b/arch/x86/kvm/vmx.c >> @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) >> kvm_userspace_mem.flags = 0; >> kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + if (kvm->arch.apic_ptr) { >> + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> + } else { >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + } >> + >> if (r) >> goto out; >> >> @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) >> kvm_userspace_mem.guest_phys_addr = >> kvm->arch.ept_identity_map_addr; >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + if (kvm->arch.ept_ptr) { >> + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> + } else { >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> + } >> + >> if (r) >> goto out; >> >> @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) >> .flags = 0, >> }; >> >> - ret = kvm_set_memory_region(kvm, &tss_mem, false); >> + if (kvm->arch.tss_ptr) { >> + tss_mem.userspace_addr = kvm->arch.tss_ptr; >> + ret = kvm_set_memory_region(kvm, &tss_mem, true); >> + } else { >> + ret = kvm_set_memory_region(kvm, &tss_mem, false); >> + } >> + >> if (ret) >> return ret; >> kvm->arch.tss_addr = addr; >> @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) >> vmx_complete_interrupts(vmx); >> } >> >> +static int vmx_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + /* >> + * Early sanity checking so userspace gets an error message during >> + * memory setup and not when trying to use this memory. >> + * Checks to see if the memory is valid are performed later when >> + * the memory is used. >> + */ >> + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || >> + mem->memory_size & (PAGE_SIZE - 1) || >> + mem->memory_size < PAGE_SIZE * 5) >> + return -EINVAL; >> + >> + kvm->arch.ept_ptr = mem->userspace_addr; >> + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; >> + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; >> + >> + return 0; >> +} >> + >> static void vmx_free_vcpu(struct kvm_vcpu *vcpu) >> { >> struct vcpu_vmx *vmx = to_vmx(vcpu); >> @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { >> .hardware_disable = hardware_disable, >> .cpu_has_accelerated_tpr = report_flexpriority, >> >> + .set_private_memory = vmx_set_private_memory, >> + >> .vcpu_create = vmx_create_vcpu, >> .vcpu_free = vmx_free_vcpu, >> .vcpu_reset = vmx_vcpu_reset, >> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c >> index e172132..7045d0a 100644 >> --- a/arch/x86/kvm/x86.c >> +++ b/arch/x86/kvm/x86.c >> @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) >> kvm_free_pit(kvm); >> } >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem) >> +{ >> + return kvm_x86_ops->set_private_memory(kvm, mem); >> +} >> + >> void kvm_arch_destroy_vm(struct kvm *kvm) >> { >> kvm_iommu_unmap_guest(kvm); >> @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, >> * Only private memory slots need to be mapped here since >> * KVM_SET_MEMORY_REGION ioctl is no longer supported. >> */ >> - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { >> + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && >> + !user_alloc) { >> unsigned long userspace_addr; >> >> /* >> @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, >> >> int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; >> >> - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { >> + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && >> + !user_alloc) { >> int ret; >> >> ret = vm_munmap(old.userspace_addr, >> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h >> index c139582..f441d1f 100644 >> --- a/include/linux/kvm_host.h >> +++ b/include/linux/kvm_host.h >> @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, >> void kvm_arch_free_memslot(struct kvm_memory_slot *free, >> struct kvm_memory_slot *dont); >> int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> + struct kvm_userspace_memory_region *mem); >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> struct kvm_memory_slot *memslot, >> struct kvm_memory_slot old, >> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c >> index f18013f..5372225 100644 >> --- a/virt/kvm/kvm_main.c >> +++ b/virt/kvm/kvm_main.c >> @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, >> kvm_userspace_memory_region *mem, >> bool user_alloc) >> { >> + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) >> + return kvm_arch_set_private_memory(kvm, mem); >> if (mem->slot >= KVM_USER_MEM_SLOTS) >> return -EINVAL; >> return kvm_set_memory_region(kvm, mem, user_alloc); >> -- >> 1.7.10.4 >> >> -- >> To unsubscribe from this list: send the line "unsubscribe kvm" in >> the body of a message to majordomo@vger.kernel.org >> More majordomo info at http://vger.kernel.org/majordomo-info.html > > -- > Gleb. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Wed, Apr 17, 2013 at 08:24:21AM -0700, Andrew Honig wrote: > I'm happy to not add a new API and use __kvm_set_memory_region to > unregister private memory regions, but I thought chaning the API was > the approach you asked for when I sent a previous patch. See the end > of: http://article.gmane.org/gmane.comp.emulators.kvm.devel/107753 > > Did I misunderstand your comment from 8 April? > Ugh, yes. My "Please send a second version" was in response to your question "or should I send a second version of this current patch?" Where second version is the same as the first one but uses __kvm_set_memory_region() for slot deletion. I am sorry about the confusion :( > On Wed, Apr 17, 2013 at 6:10 AM, Gleb Natapov <gleb@redhat.com> wrote: > > On Mon, Apr 15, 2013 at 03:10:32PM -0700, Andrew Honig wrote: > >> > >> The motivation for this patch is to fix a 20KB leak of memory in vmx.c > >> when a VM is created and destroyed. > >> > >> On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for > >> architecture specific reasons. It currently allocates the pages on behalf > >> of user space, but has no way of cleanly freeing that memory while the > >> user space process is still running. For user space processes that want > >> more control over that memory, this patch allows user space to provide the > >> memory that KVM uses. > >> > >> Signed-off-by: Andrew Honig <ahonig@google.com> > > I thought we agreed on not adding new API and using > > __kvm_set_memory_region() to unregister private memory regions. > > > > > >> --- > >> Documentation/virtual/kvm/api.txt | 8 +++++++ > >> arch/arm/kvm/arm.c | 6 +++++ > >> arch/ia64/kvm/kvm-ia64.c | 6 +++++ > >> arch/powerpc/kvm/powerpc.c | 6 +++++ > >> arch/s390/kvm/kvm-s390.c | 6 +++++ > >> arch/x86/include/asm/kvm_host.h | 7 ++++++ > >> arch/x86/kvm/svm.c | 8 +++++++ > >> arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- > >> arch/x86/kvm/x86.c | 12 ++++++++-- > >> include/linux/kvm_host.h | 2 ++ > >> virt/kvm/kvm_main.c | 2 ++ > >> 11 files changed, 105 insertions(+), 5 deletions(-) > >> > >> diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt > >> index 119358d..aa18cac 100644 > >> --- a/Documentation/virtual/kvm/api.txt > >> +++ b/Documentation/virtual/kvm/api.txt > >> @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr > >> be identical. This allows large pages in the guest to be backed by large > >> pages in the host. > >> > >> +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture > >> +specific reasons. Calling this ioctl with the special memslot > >> +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for > >> +that memory. If that memslot is not set before creating VCPUs for a VM then > >> +kvm will allocate the memory on behalf of user space, but userspace will not > >> +be able to free that memory. User space should treat this memory as opaque > >> +and not modify it. > >> + > >> The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and > >> KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of > >> writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to > >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c > >> index 5a93698..ac52f14 100644 > >> --- a/arch/arm/kvm/arm.c > >> +++ b/arch/arm/kvm/arm.c > >> @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, > >> return 0; > >> } > >> > >> +int kvm_arch_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem) > >> +{ > >> + return 0; > >> +} > >> + > >> int kvm_arch_prepare_memory_region(struct kvm *kvm, > >> struct kvm_memory_slot *memslot, > >> struct kvm_memory_slot old, > >> diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c > >> index ad3126a..570dd97 100644 > >> --- a/arch/ia64/kvm/kvm-ia64.c > >> +++ b/arch/ia64/kvm/kvm-ia64.c > >> @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > >> return 0; > >> } > >> > >> +int kvm_arch_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem) > >> +{ > >> + return 0; > >> +} > >> + > >> int kvm_arch_prepare_memory_region(struct kvm *kvm, > >> struct kvm_memory_slot *memslot, > >> struct kvm_memory_slot old, > >> diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c > >> index 934413c..6e3843b 100644 > >> --- a/arch/powerpc/kvm/powerpc.c > >> +++ b/arch/powerpc/kvm/powerpc.c > >> @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > >> return kvmppc_core_create_memslot(slot, npages); > >> } > >> > >> +int kvm_arch_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem) > >> +{ > >> + return 0; > >> +} > >> + > >> int kvm_arch_prepare_memory_region(struct kvm *kvm, > >> struct kvm_memory_slot *memslot, > >> struct kvm_memory_slot old, > >> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c > >> index 4cf35a0..a97f495 100644 > >> --- a/arch/s390/kvm/kvm-s390.c > >> +++ b/arch/s390/kvm/kvm-s390.c > >> @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) > >> return 0; > >> } > >> > >> +int kvm_arch_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem) > >> +{ > >> + return 0; > >> +} > >> + > >> /* Section: memory related */ > >> int kvm_arch_prepare_memory_region(struct kvm *kvm, > >> struct kvm_memory_slot *memslot, > >> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > >> index 4979778..7215817 100644 > >> --- a/arch/x86/include/asm/kvm_host.h > >> +++ b/arch/x86/include/asm/kvm_host.h > >> @@ -37,6 +37,7 @@ > >> /* memory slots that are not exposed to userspace */ > >> #define KVM_PRIVATE_MEM_SLOTS 3 > >> #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) > >> +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 > >> > >> #define KVM_MMIO_SIZE 16 > >> > >> @@ -553,6 +554,9 @@ struct kvm_arch { > >> struct page *ept_identity_pagetable; > >> bool ept_identity_pagetable_done; > >> gpa_t ept_identity_map_addr; > >> + unsigned long ept_ptr; > >> + unsigned long apic_ptr; > >> + unsigned long tss_ptr; > >> > >> unsigned long irq_sources_bitmap; > >> s64 kvmclock_offset; > >> @@ -640,6 +644,9 @@ struct kvm_x86_ops { > >> bool (*cpu_has_accelerated_tpr)(void); > >> void (*cpuid_update)(struct kvm_vcpu *vcpu); > >> > >> + int (*set_private_memory)(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem); > >> + > >> /* Create, but do not attach this VCPU */ > >> struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); > >> void (*vcpu_free)(struct kvm_vcpu *vcpu); > >> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c > >> index e1b1ce2..3cc4e56 100644 > >> --- a/arch/x86/kvm/svm.c > >> +++ b/arch/x86/kvm/svm.c > >> @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) > >> return 0; > >> } > >> > >> +static int svm_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem) > >> +{ > >> + return 0; > >> +} > >> + > >> static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) > >> { > >> struct vcpu_svm *svm; > >> @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { > >> .hardware_disable = svm_hardware_disable, > >> .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, > >> > >> + .set_private_memory = svm_set_private_memory, > >> + > >> .vcpu_create = svm_create_vcpu, > >> .vcpu_free = svm_free_vcpu, > >> .vcpu_reset = svm_vcpu_reset, > >> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c > >> index 6667042..796ac07 100644 > >> --- a/arch/x86/kvm/vmx.c > >> +++ b/arch/x86/kvm/vmx.c > >> @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) > >> kvm_userspace_mem.flags = 0; > >> kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; > >> kvm_userspace_mem.memory_size = PAGE_SIZE; > >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > >> + if (kvm->arch.apic_ptr) { > >> + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; > >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); > >> + } else { > >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > >> + } > >> + > >> if (r) > >> goto out; > >> > >> @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) > >> kvm_userspace_mem.guest_phys_addr = > >> kvm->arch.ept_identity_map_addr; > >> kvm_userspace_mem.memory_size = PAGE_SIZE; > >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > >> + if (kvm->arch.ept_ptr) { > >> + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; > >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); > >> + } else { > >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); > >> + } > >> + > >> if (r) > >> goto out; > >> > >> @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) > >> .flags = 0, > >> }; > >> > >> - ret = kvm_set_memory_region(kvm, &tss_mem, false); > >> + if (kvm->arch.tss_ptr) { > >> + tss_mem.userspace_addr = kvm->arch.tss_ptr; > >> + ret = kvm_set_memory_region(kvm, &tss_mem, true); > >> + } else { > >> + ret = kvm_set_memory_region(kvm, &tss_mem, false); > >> + } > >> + > >> if (ret) > >> return ret; > >> kvm->arch.tss_addr = addr; > >> @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) > >> vmx_complete_interrupts(vmx); > >> } > >> > >> +static int vmx_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem) > >> +{ > >> + /* > >> + * Early sanity checking so userspace gets an error message during > >> + * memory setup and not when trying to use this memory. > >> + * Checks to see if the memory is valid are performed later when > >> + * the memory is used. > >> + */ > >> + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || > >> + mem->memory_size & (PAGE_SIZE - 1) || > >> + mem->memory_size < PAGE_SIZE * 5) > >> + return -EINVAL; > >> + > >> + kvm->arch.ept_ptr = mem->userspace_addr; > >> + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; > >> + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; > >> + > >> + return 0; > >> +} > >> + > >> static void vmx_free_vcpu(struct kvm_vcpu *vcpu) > >> { > >> struct vcpu_vmx *vmx = to_vmx(vcpu); > >> @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { > >> .hardware_disable = hardware_disable, > >> .cpu_has_accelerated_tpr = report_flexpriority, > >> > >> + .set_private_memory = vmx_set_private_memory, > >> + > >> .vcpu_create = vmx_create_vcpu, > >> .vcpu_free = vmx_free_vcpu, > >> .vcpu_reset = vmx_vcpu_reset, > >> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > >> index e172132..7045d0a 100644 > >> --- a/arch/x86/kvm/x86.c > >> +++ b/arch/x86/kvm/x86.c > >> @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) > >> kvm_free_pit(kvm); > >> } > >> > >> +int kvm_arch_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem) > >> +{ > >> + return kvm_x86_ops->set_private_memory(kvm, mem); > >> +} > >> + > >> void kvm_arch_destroy_vm(struct kvm *kvm) > >> { > >> kvm_iommu_unmap_guest(kvm); > >> @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, > >> * Only private memory slots need to be mapped here since > >> * KVM_SET_MEMORY_REGION ioctl is no longer supported. > >> */ > >> - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { > >> + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && > >> + !user_alloc) { > >> unsigned long userspace_addr; > >> > >> /* > >> @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, > >> > >> int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; > >> > >> - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { > >> + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && > >> + !user_alloc) { > >> int ret; > >> > >> ret = vm_munmap(old.userspace_addr, > >> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > >> index c139582..f441d1f 100644 > >> --- a/include/linux/kvm_host.h > >> +++ b/include/linux/kvm_host.h > >> @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, > >> void kvm_arch_free_memslot(struct kvm_memory_slot *free, > >> struct kvm_memory_slot *dont); > >> int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); > >> +int kvm_arch_set_private_memory(struct kvm *kvm, > >> + struct kvm_userspace_memory_region *mem); > >> int kvm_arch_prepare_memory_region(struct kvm *kvm, > >> struct kvm_memory_slot *memslot, > >> struct kvm_memory_slot old, > >> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > >> index f18013f..5372225 100644 > >> --- a/virt/kvm/kvm_main.c > >> +++ b/virt/kvm/kvm_main.c > >> @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, > >> kvm_userspace_memory_region *mem, > >> bool user_alloc) > >> { > >> + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) > >> + return kvm_arch_set_private_memory(kvm, mem); > >> if (mem->slot >= KVM_USER_MEM_SLOTS) > >> return -EINVAL; > >> return kvm_set_memory_region(kvm, mem, user_alloc); > >> -- > >> 1.7.10.4 > >> > >> -- > >> To unsubscribe from this list: send the line "unsubscribe kvm" in > >> the body of a message to majordomo@vger.kernel.org > >> More majordomo info at http://vger.kernel.org/majordomo-info.html > > > > -- > > Gleb. -- Gleb. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Alright, I'll send that soon. Sorry for the mixup. On Wed, Apr 17, 2013 at 8:30 AM, Gleb Natapov <gleb@redhat.com> wrote: > On Wed, Apr 17, 2013 at 08:24:21AM -0700, Andrew Honig wrote: >> I'm happy to not add a new API and use __kvm_set_memory_region to >> unregister private memory regions, but I thought chaning the API was >> the approach you asked for when I sent a previous patch. See the end >> of: http://article.gmane.org/gmane.comp.emulators.kvm.devel/107753 >> >> Did I misunderstand your comment from 8 April? >> > Ugh, yes. My "Please send a second version" was in response to your > question "or should I send a second version of this current patch?" > Where second version is the same as the first one but uses > __kvm_set_memory_region() for slot deletion. I am sorry about the > confusion :( > >> On Wed, Apr 17, 2013 at 6:10 AM, Gleb Natapov <gleb@redhat.com> wrote: >> > On Mon, Apr 15, 2013 at 03:10:32PM -0700, Andrew Honig wrote: >> >> >> >> The motivation for this patch is to fix a 20KB leak of memory in vmx.c >> >> when a VM is created and destroyed. >> >> >> >> On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for >> >> architecture specific reasons. It currently allocates the pages on behalf >> >> of user space, but has no way of cleanly freeing that memory while the >> >> user space process is still running. For user space processes that want >> >> more control over that memory, this patch allows user space to provide the >> >> memory that KVM uses. >> >> >> >> Signed-off-by: Andrew Honig <ahonig@google.com> >> > I thought we agreed on not adding new API and using >> > __kvm_set_memory_region() to unregister private memory regions. >> > >> > >> >> --- >> >> Documentation/virtual/kvm/api.txt | 8 +++++++ >> >> arch/arm/kvm/arm.c | 6 +++++ >> >> arch/ia64/kvm/kvm-ia64.c | 6 +++++ >> >> arch/powerpc/kvm/powerpc.c | 6 +++++ >> >> arch/s390/kvm/kvm-s390.c | 6 +++++ >> >> arch/x86/include/asm/kvm_host.h | 7 ++++++ >> >> arch/x86/kvm/svm.c | 8 +++++++ >> >> arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- >> >> arch/x86/kvm/x86.c | 12 ++++++++-- >> >> include/linux/kvm_host.h | 2 ++ >> >> virt/kvm/kvm_main.c | 2 ++ >> >> 11 files changed, 105 insertions(+), 5 deletions(-) >> >> >> >> diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt >> >> index 119358d..aa18cac 100644 >> >> --- a/Documentation/virtual/kvm/api.txt >> >> +++ b/Documentation/virtual/kvm/api.txt >> >> @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr >> >> be identical. This allows large pages in the guest to be backed by large >> >> pages in the host. >> >> >> >> +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture >> >> +specific reasons. Calling this ioctl with the special memslot >> >> +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for >> >> +that memory. If that memslot is not set before creating VCPUs for a VM then >> >> +kvm will allocate the memory on behalf of user space, but userspace will not >> >> +be able to free that memory. User space should treat this memory as opaque >> >> +and not modify it. >> >> + >> >> The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and >> >> KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of >> >> writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to >> >> diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c >> >> index 5a93698..ac52f14 100644 >> >> --- a/arch/arm/kvm/arm.c >> >> +++ b/arch/arm/kvm/arm.c >> >> @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, >> >> return 0; >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c >> >> index ad3126a..570dd97 100644 >> >> --- a/arch/ia64/kvm/kvm-ia64.c >> >> +++ b/arch/ia64/kvm/kvm-ia64.c >> >> @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> >> return 0; >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c >> >> index 934413c..6e3843b 100644 >> >> --- a/arch/powerpc/kvm/powerpc.c >> >> +++ b/arch/powerpc/kvm/powerpc.c >> >> @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> >> return kvmppc_core_create_memslot(slot, npages); >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c >> >> index 4cf35a0..a97f495 100644 >> >> --- a/arch/s390/kvm/kvm-s390.c >> >> +++ b/arch/s390/kvm/kvm-s390.c >> >> @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) >> >> return 0; >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> /* Section: memory related */ >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h >> >> index 4979778..7215817 100644 >> >> --- a/arch/x86/include/asm/kvm_host.h >> >> +++ b/arch/x86/include/asm/kvm_host.h >> >> @@ -37,6 +37,7 @@ >> >> /* memory slots that are not exposed to userspace */ >> >> #define KVM_PRIVATE_MEM_SLOTS 3 >> >> #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) >> >> +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 >> >> >> >> #define KVM_MMIO_SIZE 16 >> >> >> >> @@ -553,6 +554,9 @@ struct kvm_arch { >> >> struct page *ept_identity_pagetable; >> >> bool ept_identity_pagetable_done; >> >> gpa_t ept_identity_map_addr; >> >> + unsigned long ept_ptr; >> >> + unsigned long apic_ptr; >> >> + unsigned long tss_ptr; >> >> >> >> unsigned long irq_sources_bitmap; >> >> s64 kvmclock_offset; >> >> @@ -640,6 +644,9 @@ struct kvm_x86_ops { >> >> bool (*cpu_has_accelerated_tpr)(void); >> >> void (*cpuid_update)(struct kvm_vcpu *vcpu); >> >> >> >> + int (*set_private_memory)(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem); >> >> + >> >> /* Create, but do not attach this VCPU */ >> >> struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); >> >> void (*vcpu_free)(struct kvm_vcpu *vcpu); >> >> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c >> >> index e1b1ce2..3cc4e56 100644 >> >> --- a/arch/x86/kvm/svm.c >> >> +++ b/arch/x86/kvm/svm.c >> >> @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) >> >> return 0; >> >> } >> >> >> >> +static int svm_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return 0; >> >> +} >> >> + >> >> static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) >> >> { >> >> struct vcpu_svm *svm; >> >> @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { >> >> .hardware_disable = svm_hardware_disable, >> >> .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, >> >> >> >> + .set_private_memory = svm_set_private_memory, >> >> + >> >> .vcpu_create = svm_create_vcpu, >> >> .vcpu_free = svm_free_vcpu, >> >> .vcpu_reset = svm_vcpu_reset, >> >> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c >> >> index 6667042..796ac07 100644 >> >> --- a/arch/x86/kvm/vmx.c >> >> +++ b/arch/x86/kvm/vmx.c >> >> @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) >> >> kvm_userspace_mem.flags = 0; >> >> kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; >> >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + if (kvm->arch.apic_ptr) { >> >> + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> >> + } else { >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + } >> >> + >> >> if (r) >> >> goto out; >> >> >> >> @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) >> >> kvm_userspace_mem.guest_phys_addr = >> >> kvm->arch.ept_identity_map_addr; >> >> kvm_userspace_mem.memory_size = PAGE_SIZE; >> >> - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + if (kvm->arch.ept_ptr) { >> >> + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); >> >> + } else { >> >> + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); >> >> + } >> >> + >> >> if (r) >> >> goto out; >> >> >> >> @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) >> >> .flags = 0, >> >> }; >> >> >> >> - ret = kvm_set_memory_region(kvm, &tss_mem, false); >> >> + if (kvm->arch.tss_ptr) { >> >> + tss_mem.userspace_addr = kvm->arch.tss_ptr; >> >> + ret = kvm_set_memory_region(kvm, &tss_mem, true); >> >> + } else { >> >> + ret = kvm_set_memory_region(kvm, &tss_mem, false); >> >> + } >> >> + >> >> if (ret) >> >> return ret; >> >> kvm->arch.tss_addr = addr; >> >> @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) >> >> vmx_complete_interrupts(vmx); >> >> } >> >> >> >> +static int vmx_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + /* >> >> + * Early sanity checking so userspace gets an error message during >> >> + * memory setup and not when trying to use this memory. >> >> + * Checks to see if the memory is valid are performed later when >> >> + * the memory is used. >> >> + */ >> >> + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || >> >> + mem->memory_size & (PAGE_SIZE - 1) || >> >> + mem->memory_size < PAGE_SIZE * 5) >> >> + return -EINVAL; >> >> + >> >> + kvm->arch.ept_ptr = mem->userspace_addr; >> >> + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; >> >> + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; >> >> + >> >> + return 0; >> >> +} >> >> + >> >> static void vmx_free_vcpu(struct kvm_vcpu *vcpu) >> >> { >> >> struct vcpu_vmx *vmx = to_vmx(vcpu); >> >> @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { >> >> .hardware_disable = hardware_disable, >> >> .cpu_has_accelerated_tpr = report_flexpriority, >> >> >> >> + .set_private_memory = vmx_set_private_memory, >> >> + >> >> .vcpu_create = vmx_create_vcpu, >> >> .vcpu_free = vmx_free_vcpu, >> >> .vcpu_reset = vmx_vcpu_reset, >> >> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c >> >> index e172132..7045d0a 100644 >> >> --- a/arch/x86/kvm/x86.c >> >> +++ b/arch/x86/kvm/x86.c >> >> @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) >> >> kvm_free_pit(kvm); >> >> } >> >> >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem) >> >> +{ >> >> + return kvm_x86_ops->set_private_memory(kvm, mem); >> >> +} >> >> + >> >> void kvm_arch_destroy_vm(struct kvm *kvm) >> >> { >> >> kvm_iommu_unmap_guest(kvm); >> >> @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> * Only private memory slots need to be mapped here since >> >> * KVM_SET_MEMORY_REGION ioctl is no longer supported. >> >> */ >> >> - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { >> >> + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && >> >> + !user_alloc) { >> >> unsigned long userspace_addr; >> >> >> >> /* >> >> @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, >> >> >> >> int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; >> >> >> >> - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { >> >> + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && >> >> + !user_alloc) { >> >> int ret; >> >> >> >> ret = vm_munmap(old.userspace_addr, >> >> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h >> >> index c139582..f441d1f 100644 >> >> --- a/include/linux/kvm_host.h >> >> +++ b/include/linux/kvm_host.h >> >> @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, >> >> void kvm_arch_free_memslot(struct kvm_memory_slot *free, >> >> struct kvm_memory_slot *dont); >> >> int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); >> >> +int kvm_arch_set_private_memory(struct kvm *kvm, >> >> + struct kvm_userspace_memory_region *mem); >> >> int kvm_arch_prepare_memory_region(struct kvm *kvm, >> >> struct kvm_memory_slot *memslot, >> >> struct kvm_memory_slot old, >> >> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c >> >> index f18013f..5372225 100644 >> >> --- a/virt/kvm/kvm_main.c >> >> +++ b/virt/kvm/kvm_main.c >> >> @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, >> >> kvm_userspace_memory_region *mem, >> >> bool user_alloc) >> >> { >> >> + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) >> >> + return kvm_arch_set_private_memory(kvm, mem); >> >> if (mem->slot >= KVM_USER_MEM_SLOTS) >> >> return -EINVAL; >> >> return kvm_set_memory_region(kvm, mem, user_alloc); >> >> -- >> >> 1.7.10.4 >> >> >> >> -- >> >> To unsubscribe from this list: send the line "unsubscribe kvm" in >> >> the body of a message to majordomo@vger.kernel.org >> >> More majordomo info at http://vger.kernel.org/majordomo-info.html >> > >> > -- >> > Gleb. > > -- > Gleb. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Il 17/04/2013 17:19, Andrew Honig ha scritto: > Freeing the memory in kvm_arch_free_memslot is as good as anywhere > else in KVM. The problem is that this memory is in the user space > process mm. This codepath could be called after the mm is destroyed > in the case of an process exit without closing the fd, which will > result in a panic on vm_munmap when it tries to access the mm. I think that's not a problem, the KVM file descriptor keeps the mm alive. > There's also the possibility that another process closes the fd and > messing with that processes memory map seems like it's asking for > trouble. You can check that current->mm == kvm->mm and leak the memory if they don't match. Paolo -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/Documentation/virtual/kvm/api.txt b/Documentation/virtual/kvm/api.txt index 119358d..aa18cac 100644 --- a/Documentation/virtual/kvm/api.txt +++ b/Documentation/virtual/kvm/api.txt @@ -879,6 +879,14 @@ It is recommended that the lower 21 bits of guest_phys_addr and userspace_addr be identical. This allows large pages in the guest to be backed by large pages in the host. +On x86/vmx architectures KVM needs 5 pages of user space memory for architecture +specific reasons. Calling this ioctl with the special memslot +KVM_PRIVATE_MEMORY_MEMSLOT will tell kvm which user space memory to use for +that memory. If that memslot is not set before creating VCPUs for a VM then +kvm will allocate the memory on behalf of user space, but userspace will not +be able to free that memory. User space should treat this memory as opaque +and not modify it. + The flags field supports two flags: KVM_MEM_LOG_DIRTY_PAGES and KVM_MEM_READONLY. The former can be set to instruct KVM to keep track of writes to memory within the slot. See KVM_GET_DIRTY_LOG ioctl to know how to diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c index 5a93698..ac52f14 100644 --- a/arch/arm/kvm/arm.c +++ b/arch/arm/kvm/arm.c @@ -228,6 +228,12 @@ int kvm_arch_set_memory_region(struct kvm *kvm, return 0; } +int kvm_arch_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem) +{ + return 0; +} + int kvm_arch_prepare_memory_region(struct kvm *kvm, struct kvm_memory_slot *memslot, struct kvm_memory_slot old, diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c index ad3126a..570dd97 100644 --- a/arch/ia64/kvm/kvm-ia64.c +++ b/arch/ia64/kvm/kvm-ia64.c @@ -1576,6 +1576,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) return 0; } +int kvm_arch_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem) +{ + return 0; +} + int kvm_arch_prepare_memory_region(struct kvm *kvm, struct kvm_memory_slot *memslot, struct kvm_memory_slot old, diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c index 934413c..6e3843b 100644 --- a/arch/powerpc/kvm/powerpc.c +++ b/arch/powerpc/kvm/powerpc.c @@ -410,6 +410,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) return kvmppc_core_create_memslot(slot, npages); } +int kvm_arch_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem) +{ + return 0; +} + int kvm_arch_prepare_memory_region(struct kvm *kvm, struct kvm_memory_slot *memslot, struct kvm_memory_slot old, diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c index 4cf35a0..a97f495 100644 --- a/arch/s390/kvm/kvm-s390.c +++ b/arch/s390/kvm/kvm-s390.c @@ -971,6 +971,12 @@ int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages) return 0; } +int kvm_arch_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem) +{ + return 0; +} + /* Section: memory related */ int kvm_arch_prepare_memory_region(struct kvm *kvm, struct kvm_memory_slot *memslot, diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h index 4979778..7215817 100644 --- a/arch/x86/include/asm/kvm_host.h +++ b/arch/x86/include/asm/kvm_host.h @@ -37,6 +37,7 @@ /* memory slots that are not exposed to userspace */ #define KVM_PRIVATE_MEM_SLOTS 3 #define KVM_MEM_SLOTS_NUM (KVM_USER_MEM_SLOTS + KVM_PRIVATE_MEM_SLOTS) +#define KVM_PRIVATE_MEMORY_MEMSLOT 0x80000001 #define KVM_MMIO_SIZE 16 @@ -553,6 +554,9 @@ struct kvm_arch { struct page *ept_identity_pagetable; bool ept_identity_pagetable_done; gpa_t ept_identity_map_addr; + unsigned long ept_ptr; + unsigned long apic_ptr; + unsigned long tss_ptr; unsigned long irq_sources_bitmap; s64 kvmclock_offset; @@ -640,6 +644,9 @@ struct kvm_x86_ops { bool (*cpu_has_accelerated_tpr)(void); void (*cpuid_update)(struct kvm_vcpu *vcpu); + int (*set_private_memory)(struct kvm *kvm, + struct kvm_userspace_memory_region *mem); + /* Create, but do not attach this VCPU */ struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); void (*vcpu_free)(struct kvm_vcpu *vcpu); diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c index e1b1ce2..3cc4e56 100644 --- a/arch/x86/kvm/svm.c +++ b/arch/x86/kvm/svm.c @@ -1211,6 +1211,12 @@ static int svm_vcpu_reset(struct kvm_vcpu *vcpu) return 0; } +static int svm_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem) +{ + return 0; +} + static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) { struct vcpu_svm *svm; @@ -4257,6 +4263,8 @@ static struct kvm_x86_ops svm_x86_ops = { .hardware_disable = svm_hardware_disable, .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr, + .set_private_memory = svm_set_private_memory, + .vcpu_create = svm_create_vcpu, .vcpu_free = svm_free_vcpu, .vcpu_reset = svm_vcpu_reset, diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c index 6667042..796ac07 100644 --- a/arch/x86/kvm/vmx.c +++ b/arch/x86/kvm/vmx.c @@ -3692,7 +3692,13 @@ static int alloc_apic_access_page(struct kvm *kvm) kvm_userspace_mem.flags = 0; kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL; kvm_userspace_mem.memory_size = PAGE_SIZE; - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); + if (kvm->arch.apic_ptr) { + kvm_userspace_mem.userspace_addr = kvm->arch.apic_ptr; + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); + } else { + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); + } + if (r) goto out; @@ -3722,7 +3728,13 @@ static int alloc_identity_pagetable(struct kvm *kvm) kvm_userspace_mem.guest_phys_addr = kvm->arch.ept_identity_map_addr; kvm_userspace_mem.memory_size = PAGE_SIZE; - r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); + if (kvm->arch.ept_ptr) { + kvm_userspace_mem.userspace_addr = kvm->arch.ept_ptr; + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, true); + } else { + r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, false); + } + if (r) goto out; @@ -4362,7 +4374,13 @@ static int vmx_set_tss_addr(struct kvm *kvm, unsigned int addr) .flags = 0, }; - ret = kvm_set_memory_region(kvm, &tss_mem, false); + if (kvm->arch.tss_ptr) { + tss_mem.userspace_addr = kvm->arch.tss_ptr; + ret = kvm_set_memory_region(kvm, &tss_mem, true); + } else { + ret = kvm_set_memory_region(kvm, &tss_mem, false); + } + if (ret) return ret; kvm->arch.tss_addr = addr; @@ -6683,6 +6701,27 @@ static void __noclone vmx_vcpu_run(struct kvm_vcpu *vcpu) vmx_complete_interrupts(vmx); } +static int vmx_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem) +{ + /* + * Early sanity checking so userspace gets an error message during + * memory setup and not when trying to use this memory. + * Checks to see if the memory is valid are performed later when + * the memory is used. + */ + if (!mem->userspace_addr || mem->userspace_addr & (PAGE_SIZE - 1) || + mem->memory_size & (PAGE_SIZE - 1) || + mem->memory_size < PAGE_SIZE * 5) + return -EINVAL; + + kvm->arch.ept_ptr = mem->userspace_addr; + kvm->arch.apic_ptr = mem->userspace_addr + PAGE_SIZE; + kvm->arch.tss_ptr = mem->userspace_addr + PAGE_SIZE * 2; + + return 0; +} + static void vmx_free_vcpu(struct kvm_vcpu *vcpu) { struct vcpu_vmx *vmx = to_vmx(vcpu); @@ -7532,6 +7571,8 @@ static struct kvm_x86_ops vmx_x86_ops = { .hardware_disable = hardware_disable, .cpu_has_accelerated_tpr = report_flexpriority, + .set_private_memory = vmx_set_private_memory, + .vcpu_create = vmx_create_vcpu, .vcpu_free = vmx_free_vcpu, .vcpu_reset = vmx_vcpu_reset, diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index e172132..7045d0a 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -6809,6 +6809,12 @@ void kvm_arch_sync_events(struct kvm *kvm) kvm_free_pit(kvm); } +int kvm_arch_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem) +{ + return kvm_x86_ops->set_private_memory(kvm, mem); +} + void kvm_arch_destroy_vm(struct kvm *kvm) { kvm_iommu_unmap_guest(kvm); @@ -6913,7 +6919,8 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, * Only private memory slots need to be mapped here since * KVM_SET_MEMORY_REGION ioctl is no longer supported. */ - if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages) { + if ((memslot->id >= KVM_USER_MEM_SLOTS) && npages && !old.npages && + !user_alloc) { unsigned long userspace_addr; /* @@ -6941,7 +6948,8 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, int nr_mmu_pages = 0, npages = mem->memory_size >> PAGE_SHIFT; - if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages) { + if ((mem->slot >= KVM_USER_MEM_SLOTS) && old.npages && !npages && + !user_alloc) { int ret; ret = vm_munmap(old.userspace_addr, diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h index c139582..f441d1f 100644 --- a/include/linux/kvm_host.h +++ b/include/linux/kvm_host.h @@ -461,6 +461,8 @@ int __kvm_set_memory_region(struct kvm *kvm, void kvm_arch_free_memslot(struct kvm_memory_slot *free, struct kvm_memory_slot *dont); int kvm_arch_create_memslot(struct kvm_memory_slot *slot, unsigned long npages); +int kvm_arch_set_private_memory(struct kvm *kvm, + struct kvm_userspace_memory_region *mem); int kvm_arch_prepare_memory_region(struct kvm *kvm, struct kvm_memory_slot *memslot, struct kvm_memory_slot old, diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index f18013f..5372225 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -949,6 +949,8 @@ int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, kvm_userspace_memory_region *mem, bool user_alloc) { + if (mem->slot == KVM_PRIVATE_MEMORY_MEMSLOT) + return kvm_arch_set_private_memory(kvm, mem); if (mem->slot >= KVM_USER_MEM_SLOTS) return -EINVAL; return kvm_set_memory_region(kvm, mem, user_alloc);
The motivation for this patch is to fix a 20KB leak of memory in vmx.c when a VM is created and destroyed. On x86/vmx platforms KVM needs 5 pages of userspace memory per VM for architecture specific reasons. It currently allocates the pages on behalf of user space, but has no way of cleanly freeing that memory while the user space process is still running. For user space processes that want more control over that memory, this patch allows user space to provide the memory that KVM uses. Signed-off-by: Andrew Honig <ahonig@google.com> --- Documentation/virtual/kvm/api.txt | 8 +++++++ arch/arm/kvm/arm.c | 6 +++++ arch/ia64/kvm/kvm-ia64.c | 6 +++++ arch/powerpc/kvm/powerpc.c | 6 +++++ arch/s390/kvm/kvm-s390.c | 6 +++++ arch/x86/include/asm/kvm_host.h | 7 ++++++ arch/x86/kvm/svm.c | 8 +++++++ arch/x86/kvm/vmx.c | 47 ++++++++++++++++++++++++++++++++++--- arch/x86/kvm/x86.c | 12 ++++++++-- include/linux/kvm_host.h | 2 ++ virt/kvm/kvm_main.c | 2 ++ 11 files changed, 105 insertions(+), 5 deletions(-)