From patchwork Mon Oct 9 23:08:48 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Raghavendra Rao Ananta X-Patchwork-Id: 13414649 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id D81A1E9413E for ; Mon, 9 Oct 2023 23:09:43 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=bombadil.20210309; h=Sender: Content-Transfer-Encoding:Content-Type:List-Subscribe:List-Help:List-Post: List-Archive:List-Unsubscribe:List-Id:Cc:To:From:Subject:Message-ID: References:Mime-Version:In-Reply-To:Date:Reply-To:Content-ID: Content-Description:Resent-Date:Resent-From:Resent-Sender:Resent-To:Resent-Cc :Resent-Message-ID:List-Owner; bh=uFwtbv3HivxZ7A0X5Zec6IZqmuPtWXLYTcxXHShJL+M=; b=B8lWwSxqs2p8wQs9UFbDJAVit7 o6f3DecqeWgSYiewVlvrHTvxkjTwuSvy/y7Hx6NvP1cjyRiGMvSjrC/1i4bDRcbtYPM+HHfBYwTg/ xppcd7jI/Vc9Me6sfQSvtkRV4l9fSmuoATbCndLBZBA9akF4cn6dy3NoOxwN6qd7vdQZwF39pA6qa MzQrGpcdfFVPz2lchJFp0rKs5Ge4Xn9MAiUuZcp67zB8ZNF0Nxbipzn38VBhirw1jJ6DCg86eh880 h8MSZASbidhPOiBUiZd7gqWehFJN1Ho7qZte5ZIRdw2gb8UYlC7+DxF2S1ijC/w0/pa2qEKpTnZHx 7DdTMbHA==; Received: from localhost ([::1] helo=bombadil.infradead.org) by bombadil.infradead.org with esmtp (Exim 4.96 #2 (Red Hat Linux)) id 1qpzNC-00Bmmg-0M; Mon, 09 Oct 2023 23:09:14 +0000 Received: from mail-yb1-xb49.google.com ([2607:f8b0:4864:20::b49]) by bombadil.infradead.org with esmtps (Exim 4.96 #2 (Red Hat Linux)) id 1qpzN5-00BmjD-0s for linux-arm-kernel@lists.infradead.org; Mon, 09 Oct 2023 23:09:09 +0000 Received: by mail-yb1-xb49.google.com with SMTP id 3f1490d57ef6-d81e9981ff4so6623715276.3 for ; Mon, 09 Oct 2023 16:09:06 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1696892945; x=1697497745; darn=lists.infradead.org; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:from:to:cc:subject:date:message-id:reply-to; bh=z/EXS3E0QingI/dhMmhRcelks2qlBAd4MtKbKAc3JtE=; b=BaG/RmrJT5KIjYKqWx28p9bLxgBHePUhN+QxWRTR0CL8Wz5BWhN9frMT96dvK/PcRl 4bex2zLs2yP2kOlrWWTSYpwV+5gxRSY53v0Qwz33w5ztp+R/B7DJ2IfqyBORMDHBOANX nKPEJz7IMHP640v17qMI0SjjLP0s/4qWnHc2c+SC90bmW6HAbW0jqWp5YQSvOhVNrXfs UP1q84TRHwP5TTYnEHXjV2XcdtC8qLdwznYeDCVg2RJhltZZzSh6NuPYf/l5ruKfok/O tqBWACC37PsP76zCNY1AaanH+xMl8JzuUCtbnsgRpEwwzk9r36boAAoDQ132q0ub3Ga7 cN+w== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1696892945; x=1697497745; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=z/EXS3E0QingI/dhMmhRcelks2qlBAd4MtKbKAc3JtE=; b=AUOxFXIdBiYPcyKRwtLetSGmq8MYGtaTky2DuQnGYUTgAyvsl2pXCIsAG439ZUPs5w k4OSntrlodEsQVaEiby84uckIb0EDYBcCDbHkyEOyvqeLv03Ew5H1Hj+1Ab5yDappS9n FgcjoHpVRpQObVwJRtvz6wwjqgQFq+x+JBn3Was1wekQrfxs398CPl1LoXvP+C5dOitg /cYyCdZTKWw1krfn0dtb1IjHtfqSjkmHkRLrdWawLWLvvrV37xpSHOrkOxL/AbVToPpC nJsJJ6oQv9/G8VRgHBC1eYshQ4FzKMi3CfoVrJVnkaFIR3pfc6KBXde1eFhUBu1Jtvts i9oA== X-Gm-Message-State: AOJu0YwUiqhe9n8QfV5eVJAVK9quVn1GBh8sa7CIH0zpadoCqZoKTEKP GluwXBRQnAt5iTHZwT6hpOmBOpeLJXKU X-Google-Smtp-Source: AGHT+IFKsr7JXVWu1qWTpXDJfo7L0hMPkdhBsdaOWv/PZufLpBw0bG7I4Sg+xG29iJZKC5s8M0y0vdlKLtJf X-Received: from rananta-linux.c.googlers.com ([fda3:e722:ac3:cc00:2b:ff92:c0a8:20a1]) (user=rananta job=sendgmr) by 2002:a25:c583:0:b0:d86:56f4:e4a3 with SMTP id v125-20020a25c583000000b00d8656f4e4a3mr248890ybe.13.1696892945384; Mon, 09 Oct 2023 16:09:05 -0700 (PDT) Date: Mon, 9 Oct 2023 23:08:48 +0000 In-Reply-To: <20231009230858.3444834-1-rananta@google.com> Mime-Version: 1.0 References: <20231009230858.3444834-1-rananta@google.com> X-Mailer: git-send-email 2.42.0.609.gbb76f46606-goog Message-ID: <20231009230858.3444834-3-rananta@google.com> Subject: [PATCH v7 02/12] KVM: arm64: PMU: Set the default PMU for the guest before vCPU reset From: Raghavendra Rao Ananta To: Oliver Upton , Marc Zyngier Cc: Alexandru Elisei , James Morse , Suzuki K Poulose , Paolo Bonzini , Zenghui Yu , Shaoqin Huang , Jing Zhang , Reiji Watanabe , Colton Lewis , Raghavendra Rao Anata , linux-arm-kernel@lists.infradead.org, kvmarm@lists.linux.dev, linux-kernel@vger.kernel.org, kvm@vger.kernel.org X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20231009_160907_308213_C9FF3FAF X-CRM114-Status: GOOD ( 21.30 ) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.34 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org From: Reiji Watanabe The following patches will use the number of counters information from the arm_pmu and use this to set the PMCR.N for the guest during vCPU reset. However, since the guest is not associated with any arm_pmu until userspace configures the vPMU device attributes, and a reset can happen before this event, assign a default PMU to the guest just before doing the reset. Signed-off-by: Reiji Watanabe Signed-off-by: Raghavendra Rao Ananta --- arch/arm64/kvm/arm.c | 20 ++++++++++++++++++++ arch/arm64/kvm/pmu-emul.c | 12 ++---------- include/kvm/arm_pmu.h | 6 ++++++ 3 files changed, 28 insertions(+), 10 deletions(-) diff --git a/arch/arm64/kvm/arm.c b/arch/arm64/kvm/arm.c index 78b0970eb8e6..708a53b70a7b 100644 --- a/arch/arm64/kvm/arm.c +++ b/arch/arm64/kvm/arm.c @@ -1313,6 +1313,23 @@ static bool kvm_vcpu_init_changed(struct kvm_vcpu *vcpu, KVM_VCPU_MAX_FEATURES); } +static int kvm_vcpu_set_pmu(struct kvm_vcpu *vcpu) +{ + struct kvm *kvm = vcpu->kvm; + + if (!kvm_arm_support_pmu_v3()) + return -EINVAL; + + /* + * When the vCPU has a PMU, but no PMU is set for the guest + * yet, set the default one. + */ + if (unlikely(!kvm->arch.arm_pmu)) + return kvm_arm_set_default_pmu(kvm); + + return 0; +} + static int __kvm_vcpu_set_target(struct kvm_vcpu *vcpu, const struct kvm_vcpu_init *init) { @@ -1328,6 +1345,9 @@ static int __kvm_vcpu_set_target(struct kvm_vcpu *vcpu, bitmap_copy(kvm->arch.vcpu_features, &features, KVM_VCPU_MAX_FEATURES); + if (kvm_vcpu_has_pmu(vcpu) && kvm_vcpu_set_pmu(vcpu)) + goto out_unlock; + /* Now we know what it is, we can reset it. */ kvm_reset_vcpu(vcpu); diff --git a/arch/arm64/kvm/pmu-emul.c b/arch/arm64/kvm/pmu-emul.c index eb5dcb12dafe..cc30c246c010 100644 --- a/arch/arm64/kvm/pmu-emul.c +++ b/arch/arm64/kvm/pmu-emul.c @@ -717,8 +717,7 @@ static struct arm_pmu *kvm_pmu_probe_armpmu(void) * It is still necessary to get a valid cpu, though, to probe for the * default PMU instance as userspace is not required to specify a PMU * type. In order to uphold the preexisting behavior KVM selects the - * PMU instance for the core where the first call to the - * KVM_ARM_VCPU_PMU_V3_CTRL attribute group occurs. A dependent use case + * PMU instance for the core during the vcpu reset. A dependent use case * would be a user with disdain of all things big.LITTLE that affines * the VMM to a particular cluster of cores. * @@ -893,7 +892,7 @@ static void kvm_arm_set_pmu(struct kvm *kvm, struct arm_pmu *arm_pmu) * where vCPUs can be scheduled on any core but the guest * counters could stop working. */ -static int kvm_arm_set_default_pmu(struct kvm *kvm) +int kvm_arm_set_default_pmu(struct kvm *kvm) { struct arm_pmu *arm_pmu = kvm_pmu_probe_armpmu(); @@ -946,13 +945,6 @@ int kvm_arm_pmu_v3_set_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr) if (vcpu->arch.pmu.created) return -EBUSY; - if (!kvm->arch.arm_pmu) { - int ret = kvm_arm_set_default_pmu(kvm); - - if (ret) - return ret; - } - switch (attr->attr) { case KVM_ARM_VCPU_PMU_V3_IRQ: { int __user *uaddr = (int __user *)(long)attr->addr; diff --git a/include/kvm/arm_pmu.h b/include/kvm/arm_pmu.h index 3546ebc469ad..858ed9ce828a 100644 --- a/include/kvm/arm_pmu.h +++ b/include/kvm/arm_pmu.h @@ -101,6 +101,7 @@ void kvm_vcpu_pmu_resync_el0(void); }) u8 kvm_arm_pmu_get_pmuver_limit(void); +int kvm_arm_set_default_pmu(struct kvm *kvm); #else struct kvm_pmu { @@ -174,6 +175,11 @@ static inline u8 kvm_arm_pmu_get_pmuver_limit(void) } static inline void kvm_vcpu_pmu_resync_el0(void) {} +static inline int kvm_arm_set_default_pmu(struct kvm *kvm) +{ + return -ENODEV; +} + #endif #endif