From patchwork Tue May 3 05:01:36 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Kyle Huey X-Patchwork-Id: 12834996 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 50420C433EF for ; Tue, 3 May 2022 05:01:46 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231419AbiECFFP (ORCPT ); Tue, 3 May 2022 01:05:15 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:58534 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231406AbiECFFM (ORCPT ); Tue, 3 May 2022 01:05:12 -0400 Received: from mail-pl1-x630.google.com (mail-pl1-x630.google.com [IPv6:2607:f8b0:4864:20::630]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D806E3EABF for ; Mon, 2 May 2022 22:01:39 -0700 (PDT) Received: by mail-pl1-x630.google.com with SMTP id d17so1107415plg.0 for ; Mon, 02 May 2022 22:01:39 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=kylehuey.com; s=google; h=from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=EnEoLxNESFUTI46rXZ49NiSZqFRgdh8+36EAvyHGQns=; b=gd+RFqIrUmVYueZ4Bw6793FFqgcHp4mHcQbLos/DbvQXnZz8eUFvcHxvIkvGnUGo74 /cnoBUkC53CVEEz6O/hV9GLnxN6IM2eyZA7oZ6pCGT7cnPzdacr9X7afPAgMe+3UbzKq oQD6FpyOTUMVZOhQ8/gXizwELs7An1vrq32aK/PdpIEvgE2pvUjsJB7iph3PToDaYsjt A2wQkEU9zPSAjUl7Yr20MEKVr7fNJYZQ7flnKP12PGnI27d2E5wJK2oQhkcCKbx1d+Tf DI3zZbZ/YV0Nbo1XUI3uEBcFYCLZuARDBXf6OxRInIKL/7r4LQcwKbjVbQhL7WXxDS+K MbLw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=EnEoLxNESFUTI46rXZ49NiSZqFRgdh8+36EAvyHGQns=; b=YWvqrMIYJv0mkiiIPD+GECS5dr6NEqn/ENN0YQ8hOcowAqYF9GiLozaOl4Uy45ocCX oEDVNdUZWInDQbZ5XTh72buHax7LddPxNO7JvIbLvVunMnBVx4xd33A6E1ZlQ9kj7Q6X AZY5Aa6CCvkaW1f4eLEDRfBPlv5Dypm5Shv/ENmb9Kvpr/1KqnVvWTChXWEmCMQMYmGN zALTqNp5jrC9mVXM1lIjTmZbvtY+acrXZXVIOuW67uhATGAtfwB9LfF7D9YbSYvCD1rB mtwfibTbdRrI19iOi5tWkInmXA5CFwf/3sV7S60oFQmF16njZqLEniDWnjAeZ0LV91N/ HO5w== X-Gm-Message-State: AOAM533SQugFuxb06Qy+gHMPTWXFO3c7xqAdyiJR5VH0JOY12eaG+Ufe oYVFLmiJNJL4+V8//ze9VqI4fA== X-Google-Smtp-Source: ABdhPJxRPzN58QziFP/ngC1wLLm0Rp2I/CIWsVaWLrSIL+s77G48MTmB7eQ5OUnmhLU0KxaMgROmzA== X-Received: by 2002:a17:903:189:b0:15e:9584:fbe7 with SMTP id z9-20020a170903018900b0015e9584fbe7mr12153547plg.65.1651554099135; Mon, 02 May 2022 22:01:39 -0700 (PDT) Received: from minbar.home.kylehuey.com (c-71-198-251-229.hsd1.ca.comcast.net. [71.198.251.229]) by smtp.gmail.com with ESMTPSA id 3-20020aa79143000000b0050dc76281e6sm5519416pfi.192.2022.05.02.22.01.37 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 02 May 2022 22:01:38 -0700 (PDT) From: Kyle Huey X-Google-Original-From: Kyle Huey To: linux-kernel@vger.kernel.org Cc: Paolo Bonzini , Sean Christopherson , Vitaly Kuznetsov , Wanpeng Li , Jim Mattson , Joerg Roedel , Thomas Gleixner , Ingo Molnar , Borislav Petkov , Dave Hansen , x86@kernel.org, "H. Peter Anvin" , kvm@vger.kernel.org, Robert O'Callahan , Keno Fischer , Kyle Huey Subject: [PATCH] KVM: x86/svm: Account for family 17h event renumberings in amd_pmc_perf_hw_id Date: Mon, 2 May 2022 22:01:36 -0700 Message-Id: <20220503050136.86298-1-khuey@kylehuey.com> X-Mailer: git-send-email 2.36.0 MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org From: Kyle Huey Zen renumbered some of the performance counters that correspond to the well known events in perf_hw_id. This code in KVM was never updated for that, so guest that attempt to use counters on Zen that correspond to the pre-Zen perf_hw_id values will silently receive the wrong values. This has been observed in the wild with rr[0] when running in Zen 3 guests. rr uses the retired conditional branch counter 00d1 which is incorrectly recognized by KVM as PERF_COUNT_HW_STALLED_CYCLES_BACKEND. [0] https://rr-project.org/ Signed-off-by: Kyle Huey --- arch/x86/kvm/svm/pmu.c | 28 +++++++++++++++++++++++++--- 1 file changed, 25 insertions(+), 3 deletions(-) diff --git a/arch/x86/kvm/svm/pmu.c b/arch/x86/kvm/svm/pmu.c index b14860863c39..43d24746c0a7 100644 --- a/arch/x86/kvm/svm/pmu.c +++ b/arch/x86/kvm/svm/pmu.c @@ -45,6 +45,22 @@ static struct kvm_event_hw_type_mapping amd_event_mapping[] = { [7] = { 0xd1, 0x00, PERF_COUNT_HW_STALLED_CYCLES_BACKEND }, }; +/* duplicated from amd_f17h_perfmon_event_map. */ +static struct kvm_event_hw_type_mapping amd_f17h_event_mapping[] = { + [0] = { 0x76, 0x00, PERF_COUNT_HW_CPU_CYCLES }, + [1] = { 0xc0, 0x00, PERF_COUNT_HW_INSTRUCTIONS }, + [2] = { 0x60, 0xff, PERF_COUNT_HW_CACHE_REFERENCES }, + [3] = { 0x64, 0x09, PERF_COUNT_HW_CACHE_MISSES }, + [4] = { 0xc2, 0x00, PERF_COUNT_HW_BRANCH_INSTRUCTIONS }, + [5] = { 0xc3, 0x00, PERF_COUNT_HW_BRANCH_MISSES }, + [6] = { 0x87, 0x02, PERF_COUNT_HW_STALLED_CYCLES_FRONTEND }, + [7] = { 0x87, 0x01, PERF_COUNT_HW_STALLED_CYCLES_BACKEND }, +}; + +/* amd_pmc_perf_hw_id depends on these being the same size */ +static_assert(ARRAY_SIZE(amd_event_mapping) == + ARRAY_SIZE(amd_f17h_event_mapping)); + static unsigned int get_msr_base(struct kvm_pmu *pmu, enum pmu_type type) { struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); @@ -140,6 +156,7 @@ static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr, static unsigned int amd_pmc_perf_hw_id(struct kvm_pmc *pmc) { + struct kvm_event_hw_type_mapping *event_mapping; u8 event_select = pmc->eventsel & ARCH_PERFMON_EVENTSEL_EVENT; u8 unit_mask = (pmc->eventsel & ARCH_PERFMON_EVENTSEL_UMASK) >> 8; int i; @@ -148,15 +165,20 @@ static unsigned int amd_pmc_perf_hw_id(struct kvm_pmc *pmc) if (WARN_ON(pmc_is_fixed(pmc))) return PERF_COUNT_HW_MAX; + if (boot_cpu_data.x86 >= 0x17) + event_mapping = amd_f17h_event_mapping; + else + event_mapping = amd_event_mapping; + for (i = 0; i < ARRAY_SIZE(amd_event_mapping); i++) - if (amd_event_mapping[i].eventsel == event_select - && amd_event_mapping[i].unit_mask == unit_mask) + if (event_mapping[i].eventsel == event_select + && event_mapping[i].unit_mask == unit_mask) break; if (i == ARRAY_SIZE(amd_event_mapping)) return PERF_COUNT_HW_MAX; - return amd_event_mapping[i].event_type; + return event_mapping[i].event_type; } /* check if a PMC is enabled by comparing it against global_ctrl bits. Because