From patchwork Tue May 4 13:07:02 2010 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Takuya Yoshikawa X-Patchwork-Id: 96766 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by demeter.kernel.org (8.14.3/8.14.3) with ESMTP id o44D7vus004877 for ; Tue, 4 May 2010 13:07:57 GMT Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1759917Ab0EDNHI (ORCPT ); Tue, 4 May 2010 09:07:08 -0400 Received: from mail-pw0-f46.google.com ([209.85.160.46]:58090 "EHLO mail-pw0-f46.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1758583Ab0EDNHE (ORCPT ); Tue, 4 May 2010 09:07:04 -0400 Received: by pwi5 with SMTP id 5so148849pwi.19 for ; Tue, 04 May 2010 06:07:04 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:date:from:to:cc:subject :message-id:in-reply-to:references:x-mailer:mime-version :content-type:content-transfer-encoding; bh=ioN8wB/+2UOi4qyWqNjardcUWr3e94aYX1vo9qiVa78=; b=ZT9jAq20OMwcD8ffESe+YxLVGLv4/z/P+7CUYjNyQq6a+gigm4e1AKu+YpAvGF1o7z vCvzDNnfpNw70i3BuxMLg5Th8uhFZ1cIC2em6s4i1mXjPqD02ZGq0HSZ4yIcYF8LEgX2 FDFR/8Z1bp+AeSwaVFoZAtnQhfVT14h+MbUXI= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=date:from:to:cc:subject:message-id:in-reply-to:references:x-mailer :mime-version:content-type:content-transfer-encoding; b=c6oYmsw3xVJrL75Up56kEt8UDflckuUDE5BmA4Hs8UwOScmKLxXUXV3NuamktMB2R0 o1X3hIgxjgtFc+y67hq2jZj8rzANAIGvxhE2l5Dq07Qmofrmkg4hpg0Ifr0MZNyvw1tP PhWxHGsFNCsit8Av194fScf9LzCL5PXfL3u9Y= Received: by 10.140.57.21 with SMTP id f21mr4518630rva.165.1272978423985; Tue, 04 May 2010 06:07:03 -0700 (PDT) Received: from stein (v079161.dynamic.ppp.asahi-net.or.jp [124.155.79.161]) by mx.google.com with ESMTPS id l29sm1910065rvb.4.2010.05.04.06.07.00 (version=SSLv3 cipher=RC4-MD5); Tue, 04 May 2010 06:07:03 -0700 (PDT) Date: Tue, 4 May 2010 22:07:02 +0900 From: Takuya Yoshikawa To: Takuya Yoshikawa Cc: avi@redhat.com, mtosatti@redhat.com, agraf@suse.de, yoshikawa.takuya@oss.ntt.co.jp, fernando@oss.ntt.co.jp, kvm@vger.kernel.org, kvm-ppc@vger.kernel.org, kvm-ia64@vger.kernel.org, tglx@linutronix.de, mingo@redhat.com, hpa@zytor.com, x86@kernel.org, benh@kernel.crashing.org, paulus@samba.org, linuxppc-dev@ozlabs.org, arnd@arndb.de, linux-arch@vger.kernel.org, linux-kernel@vger.kernel.org Subject: [RFC][PATCH RFC 10/12] KVM: move dirty bitmaps to user space Message-Id: <20100504220702.f8ba6ccc.takuya.yoshikawa@gmail.com> In-Reply-To: <20100504215645.6448af8f.takuya.yoshikawa@gmail.com> References: <20100504215645.6448af8f.takuya.yoshikawa@gmail.com> X-Mailer: Sylpheed 3.0.2 (GTK+ 2.20.0; i486-pc-linux-gnu) Mime-Version: 1.0 Sender: kvm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org X-Greylist: IP, sender and recipient auto-whitelisted, not delayed by milter-greylist-4.2.3 (demeter.kernel.org [140.211.167.41]); Tue, 04 May 2010 13:07:58 +0000 (UTC) diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c index 17fd65c..03503e6 100644 --- a/arch/ia64/kvm/kvm-ia64.c +++ b/arch/ia64/kvm/kvm-ia64.c @@ -1823,11 +1823,19 @@ static int kvm_ia64_sync_dirty_log(struct kvm *kvm, n = kvm_dirty_bitmap_bytes(memslot); base = memslot->base_gfn / BITS_PER_LONG; + r = -EFAULT; + if (!access_ok(VERIFY_WRITE, memslot->dirty_bitmap, n)) + goto out; + for (i = 0; i < n/sizeof(long); ++i) { if (dirty_bitmap[base + i]) memslot->is_dirty = true; - memslot->dirty_bitmap[i] = dirty_bitmap[base + i]; + if (__put_user(dirty_bitmap[base + i], + &memslot->dirty_bitmap[i])) { + r = -EFAULT; + goto out; + } dirty_bitmap[base + i] = 0; } r = 0; @@ -1858,7 +1866,10 @@ int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, if (memslot->is_dirty) { kvm_flush_remote_tlbs(kvm); n = kvm_dirty_bitmap_bytes(memslot); - memset(memslot->dirty_bitmap, 0, n); + if (clear_user(memslot->dirty_bitmap, n)) { + r = -EFAULT; + goto out; + } memslot->is_dirty = false; } r = 0; diff --git a/arch/powerpc/kvm/book3s.c b/arch/powerpc/kvm/book3s.c index 4b074f1..2a31d2f 100644 --- a/arch/powerpc/kvm/book3s.c +++ b/arch/powerpc/kvm/book3s.c @@ -1210,7 +1210,10 @@ int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, kvmppc_mmu_pte_pflush(vcpu, ga, ga_end); n = kvm_dirty_bitmap_bytes(memslot); - memset(memslot->dirty_bitmap, 0, n); + if (clear_user(memslot->dirty_bitmap, n)) { + r = -EFAULT; + goto out; + } memslot->is_dirty = false; } diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index 023c7f8..32a3d94 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -2760,40 +2760,37 @@ int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, /* If nothing is dirty, don't bother messing with page tables. */ if (memslot->is_dirty) { struct kvm_memslots *slots, *old_slots; - unsigned long *dirty_bitmap; + unsigned long __user *dirty_bitmap; + unsigned long __user *dirty_bitmap_old; spin_lock(&kvm->mmu_lock); kvm_mmu_slot_remove_write_access(kvm, log->slot); spin_unlock(&kvm->mmu_lock); - r = -ENOMEM; - dirty_bitmap = vmalloc(n); - if (!dirty_bitmap) + dirty_bitmap = memslot->dirty_bitmap; + dirty_bitmap_old = memslot->dirty_bitmap_old; + r = -EFAULT; + if (clear_user(dirty_bitmap_old, n)) goto out; - memset(dirty_bitmap, 0, n); r = -ENOMEM; slots = kzalloc(sizeof(struct kvm_memslots), GFP_KERNEL); - if (!slots) { - vfree(dirty_bitmap); + if (!slots) goto out; - } + memcpy(slots, kvm->memslots, sizeof(struct kvm_memslots)); - slots->memslots[log->slot].dirty_bitmap = dirty_bitmap; + slots->memslots[log->slot].dirty_bitmap = dirty_bitmap_old; + slots->memslots[log->slot].dirty_bitmap_old = dirty_bitmap; slots->memslots[log->slot].is_dirty = false; old_slots = kvm->memslots; rcu_assign_pointer(kvm->memslots, slots); synchronize_srcu_expedited(&kvm->srcu); - dirty_bitmap = old_slots->memslots[log->slot].dirty_bitmap; kfree(old_slots); r = -EFAULT; - if (copy_to_user(log->dirty_bitmap, dirty_bitmap, n)) { - vfree(dirty_bitmap); + if (copy_in_user(log->dirty_bitmap, dirty_bitmap, n)) goto out; - } - vfree(dirty_bitmap); } else { r = -EFAULT; if (clear_user(log->dirty_bitmap, n)) diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h index 0aa6ecb..c95e2b7 100644 --- a/include/linux/kvm_host.h +++ b/include/linux/kvm_host.h @@ -116,7 +116,8 @@ struct kvm_memory_slot { unsigned long npages; unsigned long flags; unsigned long *rmap; - unsigned long *dirty_bitmap; + unsigned long __user *dirty_bitmap; + unsigned long __user *dirty_bitmap_old; bool is_dirty; struct { unsigned long rmap_pde; diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 3e3acad..ddcf65a 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -437,8 +437,20 @@ out_err_nodisable: static void kvm_destroy_dirty_bitmap(struct kvm_memory_slot *memslot) { - vfree(memslot->dirty_bitmap); + unsigned long user_addr; + unsigned long n = kvm_dirty_bitmap_bytes(memslot); + + if (!memslot->dirty_bitmap) + return; + + user_addr = min((unsigned long)memslot->dirty_bitmap, + (unsigned long)memslot->dirty_bitmap_old); + down_write(¤t->mm->mmap_sem); + do_munmap(current->mm, user_addr, 2 * n); + up_write(¤t->mm->mmap_sem); + memslot->dirty_bitmap = NULL; + memslot->dirty_bitmap_old = NULL; } /* @@ -472,8 +484,12 @@ void kvm_free_physmem(struct kvm *kvm) int i; struct kvm_memslots *slots = kvm->memslots; - for (i = 0; i < slots->nmemslots; ++i) + for (i = 0; i < slots->nmemslots; ++i) { + /* VM process will exit: we don't unmap by ourselves. */ + slots->memslots[i].dirty_bitmap = NULL; + slots->memslots[i].dirty_bitmap_old = NULL; kvm_free_physmem_slot(&slots->memslots[i], NULL); + } kfree(kvm->memslots); } @@ -527,14 +543,35 @@ static int kvm_vm_release(struct inode *inode, struct file *filp) static int kvm_create_dirty_bitmap(struct kvm_memory_slot *memslot) { - unsigned long dirty_bytes = kvm_dirty_bitmap_bytes(memslot); + int err; + unsigned long user_addr; + unsigned long n = kvm_dirty_bitmap_bytes(memslot); - memslot->dirty_bitmap = vmalloc(dirty_bytes); - if (!memslot->dirty_bitmap) - return -ENOMEM; + down_write(¤t->mm->mmap_sem); + user_addr = do_mmap(NULL, 0, 2 * n, + PROT_READ | PROT_WRITE, + MAP_PRIVATE | MAP_ANONYMOUS, 0); + up_write(¤t->mm->mmap_sem); + + if (IS_ERR((void *)user_addr)) { + err = PTR_ERR((void *)user_addr); + goto out; + } + + memslot->dirty_bitmap = (unsigned long __user *)user_addr; + memslot->dirty_bitmap_old = (unsigned long __user *)(user_addr + n); + if (clear_user(memslot->dirty_bitmap, 2 * n)) { + err = -EFAULT; + goto out_unmap; + } - memset(memslot->dirty_bitmap, 0, dirty_bytes); return 0; +out_unmap: + down_write(¤t->mm->mmap_sem); + do_munmap(current->mm, user_addr, 2 * n); + up_write(¤t->mm->mmap_sem); +out: + return err; } /* @@ -799,7 +836,7 @@ int kvm_get_dirty_log(struct kvm *kvm, struct kvm_dirty_log *log) n = kvm_dirty_bitmap_bytes(memslot); r = -EFAULT; - if (copy_to_user(log->dirty_bitmap, memslot->dirty_bitmap, n)) + if (copy_in_user(log->dirty_bitmap, memslot->dirty_bitmap, n)) goto out; r = 0; @@ -1195,11 +1232,16 @@ void mark_page_dirty(struct kvm *kvm, gfn_t gfn) gfn = unalias_gfn(kvm, gfn); memslot = gfn_to_memslot_unaliased(kvm, gfn); if (memslot && memslot->dirty_bitmap) { - unsigned long rel_gfn = gfn - memslot->base_gfn; + int nr = generic_le_bit_offset(gfn - memslot->base_gfn); - generic___set_le_bit(rel_gfn, memslot->dirty_bitmap); + if (kvm_set_bit_user(nr, memslot->dirty_bitmap)) + goto out_fault; memslot->is_dirty = true; } + return; + +out_fault: + printk(KERN_WARNING "%s: kvm_set_bit_user failed.\n", __func__); } /*