From patchwork Sun Mar 19 08:22:22 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Binbin Wu X-Patchwork-Id: 13180255 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 91955C7618A for ; Sun, 19 Mar 2023 08:22:44 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230158AbjCSIWm (ORCPT ); Sun, 19 Mar 2023 04:22:42 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54660 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229561AbjCSIWl (ORCPT ); Sun, 19 Mar 2023 04:22:41 -0400 Received: from mga03.intel.com (mga03.intel.com [134.134.136.65]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 112D422C95 for ; Sun, 19 Mar 2023 01:22:39 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1679214159; x=1710750159; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=QlvC8KjyOonSyQURFfiETkdw4gYK4IaymmJRQr/dpF8=; b=ey4VF4wu1pAo7TAc0In55P83hCf3WW7L93m0qgqrwkFtq378zIjiKUDa X3JfXE5QHXVICGTiQK8QerNSQ1bB/SAPT0GnaYEjdYXH8y6k4mFXvJmjo Q8nlRew26Bejo4fgi20Z1A2R/GZ90DLm2aUcCeRq6fuEKv5QCepszyHy7 OPYD6r20uD4TrjcxfOXC0xCFKGQZfuwTiTZB3DA4XpMtevuM6fyhNSrQu 7rxIpLFyC8712kAIFxeoXTw7rHKlkS1xOZlX92E7vkQsxD+AEYyuJDbXp byT137jRpJqB3A9nvgYedonk/9V8eAO1KtqUkgD8D0qgPaycwi1ihICJc Q==; X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="340849320" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="340849320" Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:38 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="749741102" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="749741102" Received: from binbinwu-mobl.ccr.corp.intel.com ([10.254.209.111]) by fmsmga004-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:37 -0700 From: Binbin Wu To: kvm@vger.kernel.org, seanjc@google.com, pbonzini@redhat.com Cc: chao.gao@intel.com, robert.hu@linux.intel.com, binbin.wu@linux.intel.com Subject: [PATCH v2 1/4] x86: Allow setting of CR3 LAM bits if LAM supported Date: Sun, 19 Mar 2023 16:22:22 +0800 Message-Id: <20230319082225.14302-2-binbin.wu@linux.intel.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20230319082225.14302-1-binbin.wu@linux.intel.com> References: <20230319082225.14302-1-binbin.wu@linux.intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org If LAM is supported, VM entry allows CR3.LAM_U48 (bit 62) and CR3.LAM_U57 (bit 61) to be set in CR3 field. Change the test result expectations when setting CR3.LAM_U48 or CR3.LAM_U57 on vmlaunch tests when LAM is supported. Signed-off-by: Binbin Wu --- lib/x86/processor.h | 2 ++ x86/vmx_tests.c | 6 +++++- 2 files changed, 7 insertions(+), 1 deletion(-) diff --git a/lib/x86/processor.h b/lib/x86/processor.h index 3d58ef7..8373bbe 100644 --- a/lib/x86/processor.h +++ b/lib/x86/processor.h @@ -55,6 +55,8 @@ #define X86_CR0_PG BIT(X86_CR0_PG_BIT) #define X86_CR3_PCID_MASK GENMASK(11, 0) +#define X86_CR3_LAM_U57_BIT (61) +#define X86_CR3_LAM_U48_BIT (62) #define X86_CR4_VME_BIT (0) #define X86_CR4_VME BIT(X86_CR4_VME_BIT) diff --git a/x86/vmx_tests.c b/x86/vmx_tests.c index 7bba816..1be22ac 100644 --- a/x86/vmx_tests.c +++ b/x86/vmx_tests.c @@ -7000,7 +7000,11 @@ static void test_host_ctl_regs(void) cr3 = cr3_saved | (1ul << i); vmcs_write(HOST_CR3, cr3); report_prefix_pushf("HOST_CR3 %lx", cr3); - test_vmx_vmlaunch(VMXERR_ENTRY_INVALID_HOST_STATE_FIELD); + if (this_cpu_has(X86_FEATURE_LAM) && + ((i==X86_CR3_LAM_U57_BIT) || (i==X86_CR3_LAM_U48_BIT))) + test_vmx_vmlaunch(0); + else + test_vmx_vmlaunch(VMXERR_ENTRY_INVALID_HOST_STATE_FIELD); report_prefix_pop(); } From patchwork Sun Mar 19 08:22:23 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Binbin Wu X-Patchwork-Id: 13180256 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id D8982C6FD1F for ; Sun, 19 Mar 2023 08:22:46 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230035AbjCSIWp (ORCPT ); Sun, 19 Mar 2023 04:22:45 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54684 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229649AbjCSIWo (ORCPT ); Sun, 19 Mar 2023 04:22:44 -0400 Received: from mga03.intel.com (mga03.intel.com [134.134.136.65]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 31A2322DDC for ; Sun, 19 Mar 2023 01:22:42 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1679214162; x=1710750162; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=tSJFwXefhT7C1mfo+0Zad2ESiY3x2waCLfng0ARiCQs=; b=Jto116cInYwolVXUvbE3XirfCxheFF2tVhR09Cfrb4A4KPqzj6qZCbog rGVVACc3DYicBsGSTlOXwTbLhGeeHJcR6E6HlYztSOL87uuaz4TgEhTqm bJpcYL8ldTwZHQ/gs1UMEqvxNrFrMdps/qHIodHuKtx3cOBinVx7RZSW5 za8JaJz7KLARsuOehxfC1ALFRwxB4PaT3jrRgzKT5hPoKwjfc64setQR7 xxRtwsH1Mef0Z/KByBQelfdR0nyuQj8gZMGDfttB7lLfrgo52Sg+YEPx1 EK2oiYEKpv9cLwlIvvHdc8N+0F3ZTOAASHeTxPxPv2IxZ0xCezY9pRn2a w==; X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="340849324" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="340849324" Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:41 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="749741164" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="749741164" Received: from binbinwu-mobl.ccr.corp.intel.com ([10.254.209.111]) by fmsmga004-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:39 -0700 From: Binbin Wu To: kvm@vger.kernel.org, seanjc@google.com, pbonzini@redhat.com Cc: chao.gao@intel.com, robert.hu@linux.intel.com, binbin.wu@linux.intel.com Subject: [PATCH v2 2/4] x86: Add test case for LAM_SUP Date: Sun, 19 Mar 2023 16:22:23 +0800 Message-Id: <20230319082225.14302-3-binbin.wu@linux.intel.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20230319082225.14302-1-binbin.wu@linux.intel.com> References: <20230319082225.14302-1-binbin.wu@linux.intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org From: Robert Hoo This unit test covers: 1. CR4.LAM_SUP toggle has expected behavior according to LAM status. 2. Memory access (here is strcpy() for test example) with supervisor mode address containing LAM meta data, behave as expected per LAM status. 3. MMIO memory access with supervisor mode address containing LAM meta data, behave as expected per LAM status. 4. INVLPG memory operand doens't contain LAM meta data, if the address is non-canonical form then the INVLPG is the same as a NOP (no #GP). 5. INVPCID memory operand (descriptor pointer) could contain LAM meta data, however, the address in the descriptor should be canonical. In x86/unittests.cfg, add 2 test cases/guest conf, with and without LAM. LAM feature spec: https://cdrdv2.intel.com/v1/dl/getContent/671368, Chap 7 LINEAR ADDRESS MASKING (LAM) Signed-off-by: Robert Hoo Co-developed-by: Binbin Wu Signed-off-by: Binbin Wu --- lib/x86/processor.h | 3 + x86/Makefile.x86_64 | 1 + x86/lam.c | 296 ++++++++++++++++++++++++++++++++++++++++++++ x86/unittests.cfg | 10 ++ 4 files changed, 310 insertions(+) create mode 100644 x86/lam.c diff --git a/lib/x86/processor.h b/lib/x86/processor.h index 8373bbe..4bb8cd7 100644 --- a/lib/x86/processor.h +++ b/lib/x86/processor.h @@ -107,6 +107,8 @@ #define X86_CR4_CET BIT(X86_CR4_CET_BIT) #define X86_CR4_PKS_BIT (24) #define X86_CR4_PKS BIT(X86_CR4_PKS_BIT) +#define X86_CR4_LAM_SUP_BIT (28) +#define X86_CR4_LAM_SUP BIT(X86_CR4_LAM_SUP_BIT) #define X86_EFLAGS_CF_BIT (0) #define X86_EFLAGS_CF BIT(X86_EFLAGS_CF_BIT) @@ -250,6 +252,7 @@ static inline bool is_intel(void) #define X86_FEATURE_SPEC_CTRL (CPUID(0x7, 0, EDX, 26)) #define X86_FEATURE_ARCH_CAPABILITIES (CPUID(0x7, 0, EDX, 29)) #define X86_FEATURE_PKS (CPUID(0x7, 0, ECX, 31)) +#define X86_FEATURE_LAM (CPUID(0x7, 1, EAX, 26)) /* * Extended Leafs, a.k.a. AMD defined diff --git a/x86/Makefile.x86_64 b/x86/Makefile.x86_64 index f483dea..fa11eb3 100644 --- a/x86/Makefile.x86_64 +++ b/x86/Makefile.x86_64 @@ -34,6 +34,7 @@ tests += $(TEST_DIR)/rdpru.$(exe) tests += $(TEST_DIR)/pks.$(exe) tests += $(TEST_DIR)/pmu_lbr.$(exe) tests += $(TEST_DIR)/pmu_pebs.$(exe) +tests += $(TEST_DIR)/lam.$(exe) ifeq ($(CONFIG_EFI),y) tests += $(TEST_DIR)/amd_sev.$(exe) diff --git a/x86/lam.c b/x86/lam.c new file mode 100644 index 0000000..a5f4e51 --- /dev/null +++ b/x86/lam.c @@ -0,0 +1,296 @@ +/* + * Intel LAM_SUP unit test + * + * Copyright (C) 2023 Intel + * + * Author: Robert Hoo + * Binbin Wu + * + * This work is licensed under the terms of the GNU LGPL, version 2 or + * later. + */ + +#include "libcflat.h" +#include "processor.h" +#include "desc.h" +#include "vmalloc.h" +#include "alloc_page.h" +#include "vm.h" +#include "asm/io.h" +#include "ioram.h" + +#define LAM57_BITS 6 +#define LAM48_BITS 15 +#define LAM57_MASK GENMASK_ULL(62, 57) +#define LAM48_MASK GENMASK_ULL(62, 48) + +struct invpcid_desc { + u64 pcid : 12; + u64 rsv : 52; + u64 addr : 64; +}; + +static int get_sup_lam_bits(void) +{ + if (this_cpu_has(X86_FEATURE_LA57) && read_cr4() & X86_CR4_LA57) + return LAM57_BITS; + else + return LAM48_BITS; +} + +/* According to LAM mode, set metadata in high bits */ +static u64 set_metadata(u64 src, unsigned long lam) +{ + u64 metadata; + + switch (lam) { + case LAM57_BITS: /* Set metadata in bits 62:57 */ + metadata = (NONCANONICAL & ((1UL << LAM57_BITS) - 1)) << 57; + metadata |= (src & ~(LAM57_MASK)); + break; + case LAM48_BITS: /* Set metadata in bits 62:48 */ + metadata = (NONCANONICAL & ((1UL << LAM48_BITS) - 1)) << 48; + metadata |= (src & ~(LAM48_MASK)); + break; + default: + metadata = src; + break; + } + + return metadata; +} + +static void cr4_set_lam_sup(void *data) +{ + unsigned long cr4; + + cr4 = read_cr4(); + write_cr4_safe(cr4 | X86_CR4_LAM_SUP); +} + +static void cr4_clear_lam_sup(void *data) +{ + unsigned long cr4; + + cr4 = read_cr4(); + write_cr4_safe(cr4 & ~X86_CR4_LAM_SUP); +} + +static void test_cr4_lam_set_clear(bool lam_enumerated) +{ + bool fault; + + fault = test_for_exception(GP_VECTOR, &cr4_set_lam_sup, NULL); + if (lam_enumerated) + report(!fault && (read_cr4() & X86_CR4_LAM_SUP), + "Set CR4.LAM_SUP"); + else + report(fault, "Set CR4.LAM_SUP causes #GP"); + + fault = test_for_exception(GP_VECTOR, &cr4_clear_lam_sup, NULL); + report(!fault, "Clear CR4.LAM_SUP"); +} + +static void do_strcpy(void *mem) +{ + strcpy((char *)mem, "LAM SUP Test string."); +} + +static inline uint64_t test_tagged_ptr(uint64_t arg1, uint64_t arg2, + uint64_t arg3, uint64_t arg4) +{ + bool lam_enumerated = !!arg1; + int lam_bits = (int)arg2; + u64 *ptr = (u64 *)arg3; + bool la_57 = !!arg4; + bool fault; + + fault = test_for_exception(GP_VECTOR, do_strcpy, ptr); + report(!fault, "strcpy to untagged addr"); + + ptr = (u64 *)set_metadata((u64)ptr, lam_bits); + fault = test_for_exception(GP_VECTOR, do_strcpy, ptr); + if (lam_enumerated) + report(!fault, "strcpy to tagged addr"); + else + report(fault, "strcpy to tagged addr causes #GP"); + + if (lam_enumerated && (lam_bits==LAM57_BITS) && !la_57) { + ptr = (u64 *)set_metadata((u64)ptr, LAM48_BITS); + fault = test_for_exception(GP_VECTOR, do_strcpy, ptr); + report(fault, "strcpy to non-LAM-canonical addr causes #GP"); + } + + return 0; +} + +/* Refer to emulator.c */ +static void do_mov_mmio(void *mem) +{ + unsigned long t1, t2; + + // test mov reg, r/m and mov r/m, reg + t1 = 0x123456789abcdefull & -1ul; + asm volatile("mov %[t1], (%[mem])\n\t" + "mov (%[mem]), %[t2]" + : [t2]"=r"(t2) + : [t1]"r"(t1), [mem]"r"(mem) + : "memory"); +} + +static inline uint64_t test_tagged_mmio_ptr(uint64_t arg1, uint64_t arg2, + uint64_t arg3, uint64_t arg4) +{ + bool lam_enumerated = !!arg1; + int lam_bits = (int)arg2; + u64 *ptr = (u64 *)arg3; + bool la_57 = !!arg4; + bool fault; + + fault = test_for_exception(GP_VECTOR, do_mov_mmio, ptr); + report(!fault, "Access MMIO with untagged addr"); + + ptr = (u64 *)set_metadata((u64)ptr, lam_bits); + fault = test_for_exception(GP_VECTOR, do_mov_mmio, ptr); + if (lam_enumerated) + report(!fault, "Access MMIO with tagged addr"); + else + report(fault, "Access MMIO with tagged addr causes #GP"); + + if (lam_enumerated && (lam_bits==LAM57_BITS) && !la_57) { + ptr = (u64 *)set_metadata((u64)ptr, LAM48_BITS); + fault = test_for_exception(GP_VECTOR, do_mov_mmio, ptr); + report(fault, "Access MMIO with non-LAM-canonical addr" + " causes #GP"); + } + + return 0; +} + +static void do_invlpg(void *mem) +{ + invlpg(mem); +} + +static void do_invlpg_fep(void *mem) +{ + asm volatile(KVM_FEP "invlpg (%0)" ::"r" (mem) : "memory"); +} + +/* invlpg with tagged address is same as NOP, no #GP */ +static void test_invlpg(void *va, bool fep) +{ + bool fault; + u64 *ptr; + + ptr = (u64 *)set_metadata((u64)va, get_sup_lam_bits()); + if (fep) + fault = test_for_exception(GP_VECTOR, do_invlpg_fep, ptr); + else + fault = test_for_exception(GP_VECTOR, do_invlpg, ptr); + + report(!fault, "%sINVLPG with tagged addr", fep?"fep: ":""); +} + +static void do_invpcid(void *desc) +{ + unsigned long type = 0; + struct invpcid_desc *desc_ptr = (struct invpcid_desc *)desc; + + asm volatile("invpcid %0, %1" : + : "m" (*desc_ptr), "r" (type) + : "memory"); +} + +static void test_invpcid(bool lam_enumerated, void *data) +{ + struct invpcid_desc *desc_ptr = (struct invpcid_desc *) data; + int lam_bits = get_sup_lam_bits(); + bool fault; + + if (!this_cpu_has(X86_FEATURE_PCID) || + !this_cpu_has(X86_FEATURE_INVPCID)) { + report_skip("INVPCID not supported"); + return; + } + + memset(desc_ptr, 0, sizeof(struct invpcid_desc)); + desc_ptr->addr = (u64)data + 16; + + fault = test_for_exception(GP_VECTOR, do_invpcid, desc_ptr); + report(!fault, "INVPCID: untagged pointer + untagged addr"); + + desc_ptr->addr = set_metadata(desc_ptr->addr, lam_bits); + fault = test_for_exception(GP_VECTOR, do_invpcid, desc_ptr); + report(fault, "INVPCID: untagged pointer + tagged addr causes #GP"); + + desc_ptr->addr = (u64)data + 16; + desc_ptr = (struct invpcid_desc *)set_metadata((u64)desc_ptr, lam_bits); + fault = test_for_exception(GP_VECTOR, do_invpcid, desc_ptr); + if (lam_enumerated && (read_cr4() & X86_CR4_LAM_SUP)) + report(!fault, "INVPCID: tagged pointer + untagged addr"); + else + report(fault, "INVPCID: tagged pointer + untagged addr" + " causes #GP"); + + desc_ptr = (struct invpcid_desc *)data; + desc_ptr->addr = (u64)data + 16; + desc_ptr->addr = set_metadata(desc_ptr->addr, lam_bits); + desc_ptr = (struct invpcid_desc *)set_metadata((u64)desc_ptr, lam_bits); + fault = test_for_exception(GP_VECTOR, do_invpcid, desc_ptr); + report(fault, "INVPCID: tagged pointer + tagged addr causes #GP"); +} + +static void test_lam_sup(bool lam_enumerated, bool fep_available) +{ + void *vaddr, *vaddr_mmio; + phys_addr_t paddr; + bool fault; + bool la_57 = read_cr4() & X86_CR4_LA57; + int lam_bits = get_sup_lam_bits(); + + vaddr = alloc_vpage(); + vaddr_mmio = alloc_vpage(); + paddr = virt_to_phys(alloc_page()); + install_page(current_page_table(), paddr, vaddr); + install_page(current_page_table(), IORAM_BASE_PHYS, vaddr_mmio); + + test_cr4_lam_set_clear(lam_enumerated); + + /* Set for the following LAM_SUP tests */ + if (lam_enumerated) { + fault = test_for_exception(GP_VECTOR, &cr4_set_lam_sup, NULL); + report(!fault && (read_cr4() & X86_CR4_LAM_SUP), + "Set CR4.LAM_SUP"); + } + + test_tagged_ptr(lam_enumerated, lam_bits, (u64)vaddr, la_57); + test_tagged_mmio_ptr(lam_enumerated, lam_bits, (u64)vaddr_mmio, la_57); + test_invlpg(vaddr, false); + test_invpcid(lam_enumerated, vaddr); + + if (fep_available) + test_invlpg(vaddr, true); +} + +int main(int ac, char **av) +{ + bool lam_enumerated; + bool fep_available = is_fep_available(); + + setup_vm(); + + lam_enumerated = this_cpu_has(X86_FEATURE_LAM); + if (!lam_enumerated) + report_info("This CPU doesn't support LAM feature\n"); + else + report_info("This CPU supports LAM feature\n"); + + if (!fep_available) + report_skip("Skipping tests the forced emulation, " + "use kvm.force_emulation_prefix=1 to enable\n"); + + test_lam_sup(lam_enumerated, fep_available); + + return report_summary(); +} diff --git a/x86/unittests.cfg b/x86/unittests.cfg index f324e32..34b09eb 100644 --- a/x86/unittests.cfg +++ b/x86/unittests.cfg @@ -478,3 +478,13 @@ file = cet.flat arch = x86_64 smp = 2 extra_params = -enable-kvm -m 2048 -cpu host + +[intel-lam] +file = lam.flat +arch = x86_64 +extra_params = -enable-kvm -cpu host + +[intel-no-lam] +file = lam.flat +arch = x86_64 +extra_params = -enable-kvm -cpu host,-lam From patchwork Sun Mar 19 08:22:24 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Binbin Wu X-Patchwork-Id: 13180257 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8FF9FC7618E for ; Sun, 19 Mar 2023 08:22:48 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230163AbjCSIWr (ORCPT ); Sun, 19 Mar 2023 04:22:47 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54716 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229649AbjCSIWq (ORCPT ); Sun, 19 Mar 2023 04:22:46 -0400 Received: from mga03.intel.com (mga03.intel.com [134.134.136.65]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 1A87322DF5 for ; Sun, 19 Mar 2023 01:22:45 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1679214165; x=1710750165; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=HFxmk2rD1obiF2XNrKsMVvrSeRRxD3JtzbNcGpqCW/M=; b=Nc8DGTGYWOGwG5VjgbziBS4nHkCdgHwymjyqHktJeg0GOWZv4kfxlVUR sNPaM2HTV80ORlveYjCJop3MCwCWfL5to7Ad6jBYiedRPBd/vAiF3qQIB g8zD/U/QRUu30ZMTRkV29eEkI4vowQ6gPrOOk3oJNiIM2JDMS0xutb1N2 SYYL4NgUXXRzd0+SFizw3KiQQ+WBgDUInsuxGyvzkGOFZJdERuPRNqK90 pATXxZ9sOSan4+V4+vMbJdERPpu9AOBR80CNO72e7keF80YuZm6d2tM3k oCZmT9LlWK9fYyZdZb+OpMiApE0kYYHjjWFcaLZCTfT/k7DBhJh1ubfBR Q==; X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="340849329" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="340849329" Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:44 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="749741205" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="749741205" Received: from binbinwu-mobl.ccr.corp.intel.com ([10.254.209.111]) by fmsmga004-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:43 -0700 From: Binbin Wu To: kvm@vger.kernel.org, seanjc@google.com, pbonzini@redhat.com Cc: chao.gao@intel.com, robert.hu@linux.intel.com, binbin.wu@linux.intel.com Subject: [PATCH v2 3/4] x86: Add test cases for LAM_{U48,U57} Date: Sun, 19 Mar 2023 16:22:24 +0800 Message-Id: <20230319082225.14302-4-binbin.wu@linux.intel.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20230319082225.14302-1-binbin.wu@linux.intel.com> References: <20230319082225.14302-1-binbin.wu@linux.intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org This unit test covers: 1. CR3 LAM bits toggle has expected behavior according to LAM status. 2. Memory access using strcpy() with user mode address containing LAM meta data, behave as expected per LAM status. 3. MMIO memory access with user mode address containing LAM meta data, behave as expected per LAM status. Signed-off-by: Binbin Wu --- lib/x86/processor.h | 2 ++ x86/lam.c | 46 ++++++++++++++++++++++++++++++++++++++++++++- 2 files changed, 47 insertions(+), 1 deletion(-) diff --git a/lib/x86/processor.h b/lib/x86/processor.h index 4bb8cd7..a181e0b 100644 --- a/lib/x86/processor.h +++ b/lib/x86/processor.h @@ -56,7 +56,9 @@ #define X86_CR3_PCID_MASK GENMASK(11, 0) #define X86_CR3_LAM_U57_BIT (61) +#define X86_CR3_LAM_U57 BIT_ULL(X86_CR3_LAM_U57_BIT) #define X86_CR3_LAM_U48_BIT (62) +#define X86_CR3_LAM_U48 BIT_ULL(X86_CR3_LAM_U48_BIT) #define X86_CR4_VME_BIT (0) #define X86_CR4_VME BIT(X86_CR4_VME_BIT) diff --git a/x86/lam.c b/x86/lam.c index a5f4e51..8945440 100644 --- a/x86/lam.c +++ b/x86/lam.c @@ -1,5 +1,5 @@ /* - * Intel LAM_SUP unit test + * Intel LAM unit test * * Copyright (C) 2023 Intel * @@ -18,11 +18,13 @@ #include "vm.h" #include "asm/io.h" #include "ioram.h" +#include "usermode.h" #define LAM57_BITS 6 #define LAM48_BITS 15 #define LAM57_MASK GENMASK_ULL(62, 57) #define LAM48_MASK GENMASK_ULL(62, 48) +#define CR3_LAM_BITS_MASK (X86_CR3_LAM_U48 | X86_CR3_LAM_U57) struct invpcid_desc { u64 pcid : 12; @@ -273,6 +275,47 @@ static void test_lam_sup(bool lam_enumerated, bool fep_available) test_invlpg(vaddr, true); } +static void test_lam_user(bool lam_enumerated) +{ + unsigned long cr3; + bool is_la57; + unsigned r; + bool raised_vector = false; + phys_addr_t paddr; + + paddr = virt_to_phys(alloc_page()); + install_page((void *)(read_cr3()& ~CR3_LAM_BITS_MASK), paddr, (void *)paddr); + install_page((void *)(read_cr3()& ~CR3_LAM_BITS_MASK), IORAM_BASE_PHYS, + (void *)IORAM_BASE_PHYS); + + cr3 = read_cr3(); + is_la57 = !!(read_cr4() & X86_CR4_LA57); + + /* Test LAM_U48 */ + if(lam_enumerated) { + r = write_cr3_safe((cr3 & ~X86_CR3_LAM_U57) | X86_CR3_LAM_U48); + report(r==0 && ((read_cr3() & CR3_LAM_BITS_MASK) == X86_CR3_LAM_U48), + "Set LAM_U48"); + } + + run_in_user((usermode_func)test_tagged_ptr, GP_VECTOR, lam_enumerated, + LAM48_BITS, paddr, is_la57, &raised_vector); + run_in_user((usermode_func)test_tagged_mmio_ptr, GP_VECTOR, lam_enumerated, + LAM48_BITS, IORAM_BASE_PHYS, is_la57, &raised_vector); + + + /* Test LAM_U57 */ + if(lam_enumerated) { + r = write_cr3_safe(cr3 | X86_CR3_LAM_U57); + report(r==0 && (read_cr3() & X86_CR3_LAM_U57), "Set LAM_U57"); + } + + run_in_user((usermode_func)test_tagged_ptr, GP_VECTOR, lam_enumerated, + LAM57_BITS, paddr, is_la57, &raised_vector); + run_in_user((usermode_func)test_tagged_mmio_ptr, GP_VECTOR, lam_enumerated, + LAM57_BITS, IORAM_BASE_PHYS, is_la57, &raised_vector); +} + int main(int ac, char **av) { bool lam_enumerated; @@ -291,6 +334,7 @@ int main(int ac, char **av) "use kvm.force_emulation_prefix=1 to enable\n"); test_lam_sup(lam_enumerated, fep_available); + test_lam_user(lam_enumerated); return report_summary(); } From patchwork Sun Mar 19 08:22:25 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Binbin Wu X-Patchwork-Id: 13180258 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 6D7C2C6FD1F for ; Sun, 19 Mar 2023 08:22:52 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230174AbjCSIWv (ORCPT ); Sun, 19 Mar 2023 04:22:51 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54818 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230167AbjCSIWt (ORCPT ); Sun, 19 Mar 2023 04:22:49 -0400 Received: from mga03.intel.com (mga03.intel.com [134.134.136.65]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 050D324C85 for ; Sun, 19 Mar 2023 01:22:47 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1679214168; x=1710750168; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=AA1LhgMGpiTDo/qs/Zr+Cp2CNmiv1WbX0CpLXr5ylTI=; b=CxB4EhLNNt7285Ese2haACRGTkhG7yn/g/FdTlpGD4mBPnpA5nd3W/U1 GrQh/A19ONCnXYr7MUrdjZ3pSg9sRFKR0IjRluXV8N95B7jk8cz90jH6i VNealFeD4RjULBXnj3o/3F3zg8valwsBW8STCsFRHup/xq4yjklk+FmKq mEZ3o0jHUnH6Vm56XtMbOTsSNYa36Q+RSnoej2l5JXUmuFlkZgTc6ziJK yq1rSESb9I3VeYhdcVi1vkqAFDvBgzl63+NsE0e9GvzVGwkyvNMTqpYvH pLCwF85m8y27cPurfvJmjwl+LMFQTte+Z8Ap8/d+d+or2mVQmki5JSZjz w==; X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="340849334" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="340849334" Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:47 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10653"; a="749741235" X-IronPort-AV: E=Sophos;i="5.98,273,1673942400"; d="scan'208";a="749741235" Received: from binbinwu-mobl.ccr.corp.intel.com ([10.254.209.111]) by fmsmga004-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Mar 2023 01:22:46 -0700 From: Binbin Wu To: kvm@vger.kernel.org, seanjc@google.com, pbonzini@redhat.com Cc: chao.gao@intel.com, robert.hu@linux.intel.com, binbin.wu@linux.intel.com Subject: [PATCH v2 4/4] x86: Add test case for INVVPID with LAM Date: Sun, 19 Mar 2023 16:22:25 +0800 Message-Id: <20230319082225.14302-5-binbin.wu@linux.intel.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20230319082225.14302-1-binbin.wu@linux.intel.com> References: <20230319082225.14302-1-binbin.wu@linux.intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org When LAM is on, the linear address of INVVPID operand can contain metadata, and the linear address in the INVVPID descriptor can caontain metadata. The added cases use tagged descriptor address or/and tagged target invalidation address to make sure the behaviors are expected when LAM is on. Also, INVVPID cases can be used as the common test cases for VMX instruction vmexits. Signed-off-by: Binbin Wu --- x86/vmx_tests.c | 73 +++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 73 insertions(+) diff --git a/x86/vmx_tests.c b/x86/vmx_tests.c index 1be22ac..9e9589f 100644 --- a/x86/vmx_tests.c +++ b/x86/vmx_tests.c @@ -3225,6 +3225,78 @@ static void invvpid_test_not_in_vmx_operation(void) TEST_ASSERT(!vmx_on()); } +#define LAM57_BITS 6 +#define LAM48_BITS 15 +#define LAM57_MASK GENMASK_ULL(62, 57) +#define LAM48_MASK GENMASK_ULL(62, 48) + +/* According to LAM mode, set metadata in high bits */ +static u64 set_metadata(u64 src, unsigned long lam) +{ + u64 metadata; + + switch (lam) { + case LAM57_BITS: /* Set metadata in bits 62:57 */ + metadata = (NONCANONICAL & ((1UL << LAM57_BITS) - 1)) << 57; + metadata |= (src & ~(LAM57_MASK)); + break; + case LAM48_BITS: /* Set metadata in bits 62:48 */ + metadata = (NONCANONICAL & ((1UL << LAM48_BITS) - 1)) << 48; + metadata |= (src & ~(LAM48_MASK)); + break; + default: + metadata = src; + break; + } + + return metadata; +} + +static void invvpid_test_lam(void) +{ + void *vaddr; + struct invvpid_operand *operand; + int lam_bits = LAM48_BITS; + bool fault; + + if (!this_cpu_has(X86_FEATURE_LAM)) { + report_skip("LAM is not supported, skip INVVPID with LAM"); + return; + } + + if (this_cpu_has(X86_FEATURE_LA57) && read_cr4() & X86_CR4_LA57) + lam_bits = LAM57_BITS; + + vaddr = alloc_vpage(); + install_page(current_page_table(), virt_to_phys(alloc_page()), vaddr); + operand = (struct invvpid_operand *)vaddr; + operand->vpid = 0xffff; + operand->gla = (u64)vaddr; + + write_cr4_safe(read_cr4() | X86_CR4_LAM_SUP); + if (!(read_cr4() & X86_CR4_LAM_SUP)) { + report_skip("Failed to enable LAM_SUP"); + return; + } + + operand = (struct invvpid_operand *)vaddr; + operand->gla = set_metadata(operand->gla, lam_bits); + fault = test_for_exception(GP_VECTOR, ds_invvpid, operand); + report(!fault, "INVVPID (LAM on): untagged pointer + tagged addr"); + + operand = (struct invvpid_operand *)set_metadata((u64)operand, lam_bits); + operand->gla = (u64)vaddr; + fault = test_for_exception(GP_VECTOR, ds_invvpid, operand); + report(!fault, "INVVPID (LAM on): tagged pointer + untagged addr"); + + operand = (struct invvpid_operand *)set_metadata((u64)operand, lam_bits); + operand->gla = set_metadata(operand->gla, lam_bits); + fault = test_for_exception(GP_VECTOR, ds_invvpid, operand); + report(!fault, "INVVPID (LAM on): tagged pointer + tagged addr"); + + write_cr4_safe(read_cr4() & ~X86_CR4_LAM_SUP); +} + /* * This does not test real-address mode, virtual-8086 mode, protected mode, * or CPL > 0. @@ -3282,6 +3354,7 @@ static void invvpid_test(void) invvpid_test_pf(); invvpid_test_compatibility_mode(); invvpid_test_not_in_vmx_operation(); + invvpid_test_lam(); } /*