From patchwork Tue Nov 28 20:03:18 2017 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Nitesh Lal X-Patchwork-Id: 10080981 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork.web.codeaurora.org (Postfix) with ESMTP id 83351602BC for ; Tue, 28 Nov 2017 20:03:39 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 6E6AB2965F for ; Tue, 28 Nov 2017 20:03:39 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 62DCA29669; Tue, 28 Nov 2017 20:03:39 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-6.9 required=2.0 tests=BAYES_00,RCVD_IN_DNSWL_HI autolearn=ham version=3.3.1 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 047922965F for ; Tue, 28 Nov 2017 20:03:39 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754240AbdK1UDg (ORCPT ); Tue, 28 Nov 2017 15:03:36 -0500 Received: from mx1.redhat.com ([209.132.183.28]:36862 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754208AbdK1UDe (ORCPT ); Tue, 28 Nov 2017 15:03:34 -0500 Received: from smtp.corp.redhat.com (int-mx05.intmail.prod.int.phx2.redhat.com [10.5.11.15]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id A197E64049; Tue, 28 Nov 2017 20:03:34 +0000 (UTC) Received: from Dungeon.bos.redhat.com (dhcp-17-15.bos.redhat.com [10.18.17.15]) by smtp.corp.redhat.com (Postfix) with ESMTP id 22D2F5EDE5; Tue, 28 Nov 2017 20:03:33 +0000 (UTC) From: nilal@redhat.com To: kvm@vger.kernel.org, pbonzini@redhat.com, pagupta@redhat.com, wei.w.wang@intel.com, yang.zhang.wz@gmail.com, riel@redhat.com, david@redhat.com, mst@redhat.com, dodgen@google.com, konrad.wilk@oracle.com Subject: [Patch v5 1/7] KVM: Support for guest page hinting Date: Tue, 28 Nov 2017 15:03:18 -0500 Message-Id: <20171128200324.4432-2-nilal@redhat.com> In-Reply-To: <20171128200324.4432-1-nilal@redhat.com> References: <20171128200324.4432-1-nilal@redhat.com> X-Scanned-By: MIMEDefang 2.79 on 10.5.11.15 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.25]); Tue, 28 Nov 2017 20:03:34 +0000 (UTC) Sender: kvm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP From: Nitesh Narayan Lal This patch includes the following: 1. Basic skeleton for the support 2. Enablement of x86 platform to use the same Signed-off-by: Nitesh Narayan Lal --- arch/x86/Kbuild | 2 +- arch/x86/kvm/Makefile | 2 ++ include/linux/gfp.h | 7 +++++++ virt/kvm/Kconfig | 6 ++++++ virt/kvm/page_hinting.c | 46 ++++++++++++++++++++++++++++++++++++++++++++++ 5 files changed, 62 insertions(+), 1 deletion(-) create mode 100644 virt/kvm/page_hinting.c diff --git a/arch/x86/Kbuild b/arch/x86/Kbuild index 0038a2d..7d39d7d 100644 --- a/arch/x86/Kbuild +++ b/arch/x86/Kbuild @@ -2,7 +2,7 @@ obj-y += entry/ obj-$(CONFIG_PERF_EVENTS) += events/ -obj-$(CONFIG_KVM) += kvm/ +obj-$(subst m,y,$(CONFIG_KVM)) += kvm/ # Xen paravirtualization support obj-$(CONFIG_XEN) += xen/ diff --git a/arch/x86/kvm/Makefile b/arch/x86/kvm/Makefile index dc4f2fd..866dd56 100644 --- a/arch/x86/kvm/Makefile +++ b/arch/x86/kvm/Makefile @@ -16,6 +16,8 @@ kvm-y += x86.o mmu.o emulate.o i8259.o irq.o lapic.o \ i8254.o ioapic.o irq_comm.o cpuid.o pmu.o mtrr.o \ hyperv.o page_track.o debugfs.o +obj-$(CONFIG_KVM_FREE_PAGE_HINTING) += $(KVM)/page_hinting.o + kvm-intel-y += vmx.o pmu_intel.o kvm-amd-y += svm.o pmu_amd.o diff --git a/include/linux/gfp.h b/include/linux/gfp.h index 1a4582b..e02369b 100644 --- a/include/linux/gfp.h +++ b/include/linux/gfp.h @@ -439,6 +439,13 @@ static inline struct zonelist *node_zonelist(int nid, gfp_t flags) return NODE_DATA(nid)->node_zonelists + gfp_zonelist(flags); } +#ifdef CONFIG_KVM_FREE_PAGE_HINTING +#define HAVE_ARCH_ALLOC_PAGE +#define HAVE_ARCH_FREE_PAGE +void arch_free_page(struct page *page, int order); +void arch_alloc_page(struct page *page, int order); +#endif + #ifndef HAVE_ARCH_FREE_PAGE static inline void arch_free_page(struct page *page, int order) { } #endif diff --git a/virt/kvm/Kconfig b/virt/kvm/Kconfig index 70691c0..f7bfbb3 100644 --- a/virt/kvm/Kconfig +++ b/virt/kvm/Kconfig @@ -51,3 +51,9 @@ config KVM_COMPAT config HAVE_KVM_IRQ_BYPASS bool + +config KVM_FREE_PAGE_HINTING + def_bool y + depends on KVM + select VIRTIO + select VIRTIO_BALLOON diff --git a/virt/kvm/page_hinting.c b/virt/kvm/page_hinting.c new file mode 100644 index 0000000..39d2b1d --- /dev/null +++ b/virt/kvm/page_hinting.c @@ -0,0 +1,46 @@ +#include +#include +#include +#include +#include + +#include + +#define MAX_FGPT_ENTRIES 1000 +#define HYPERLIST_THRESHOLD 500 +/* + * struct kvm_free_pages - Tracks the pages which are freed by the guest. + * @pfn - page frame number for the page which is to be freed + * @pages - number of pages which are supposed to be freed. + * A global array object is used to hold the list of pfn and number of pages + * which are freed by the guest. This list may also have fragmentated pages so + * defragmentation is a must prior to the hypercall. + */ +struct kvm_free_pages { + unsigned long pfn; + unsigned int pages; +}; + +/* + * hypervisor_pages - It is a dummy structure passed with the hypercall. + * @pfn - page frame number for the page which is to be freed. + * @pages - number of pages which are supposed to be freed. + * A global array object is used to to hold the list of pfn and pages and is + * passed as part of the hypercall. + */ +struct hypervisor_pages { + unsigned long pfn; + unsigned int pages; +}; + +DEFINE_PER_CPU(struct kvm_free_pages [MAX_FGPT_ENTRIES], kvm_pt); +DEFINE_PER_CPU(int, kvm_pt_idx); +struct hypervisor_pages hypervisor_pagelist[MAX_FGPT_ENTRIES]; + +void arch_alloc_page(struct page *page, int order) +{ +} + +void arch_free_page(struct page *page, int order) +{ +}