From patchwork Tue Mar 18 15:30:21 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Andrii Sultanov X-Patchwork-Id: 14021171 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id 7C908C28B28 for ; Tue, 18 Mar 2025 15:32:06 +0000 (UTC) Received: from list by lists.xenproject.org with outflank-mailman.919331.1323823 (Exim 4.92) (envelope-from ) id 1tuYv8-0002ik-5M; Tue, 18 Mar 2025 15:31:58 +0000 X-Outflank-Mailman: Message body and most headers restored to incoming version Received: by outflank-mailman (output) from mailman id 919331.1323823; Tue, 18 Mar 2025 15:31:58 +0000 Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1tuYv8-0002ib-2R; Tue, 18 Mar 2025 15:31:58 +0000 Received: by outflank-mailman (input) for mailman id 919331; Tue, 18 Mar 2025 15:31:56 +0000 Received: from se1-gles-flk1-in.inumbo.com ([94.247.172.50] helo=se1-gles-flk1.inumbo.com) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1tuYv6-0001ML-IP for xen-devel@lists.xenproject.org; Tue, 18 Mar 2025 15:31:56 +0000 Received: from mail-wm1-x330.google.com (mail-wm1-x330.google.com [2a00:1450:4864:20::330]) by se1-gles-flk1.inumbo.com (Halon) with ESMTPS id 1f151085-040e-11f0-9899-31a8f345e629; Tue, 18 Mar 2025 16:31:54 +0100 (CET) Received: by mail-wm1-x330.google.com with SMTP id 5b1f17b1804b1-43cef035a3bso25529785e9.1 for ; Tue, 18 Mar 2025 08:31:54 -0700 (PDT) Received: from localhost.localdomain (cpc92320-cmbg19-2-0-cust1786.5-4.cable.virginm.net. [82.13.70.251]) by smtp.gmail.com with ESMTPSA id 5b1f17b1804b1-43d1ffbcf12sm135384415e9.12.2025.03.18.08.31.52 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 18 Mar 2025 08:31:52 -0700 (PDT) X-BeenThere: xen-devel@lists.xenproject.org List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Errors-To: xen-devel-bounces@lists.xenproject.org Precedence: list Sender: "Xen-devel" X-Inumbo-ID: 1f151085-040e-11f0-9899-31a8f345e629 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20230601; t=1742311913; x=1742916713; darn=lists.xenproject.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=byyrGrfPPMb7yp0dDoy8Ot8gd6oeoDPxj7m5qTsv1dM=; b=i3sJBMYVk3B8jUCjM7Dm+JlGvlpVmzeJKjXbGFprNSghaL5h4ZBrFojbR+mu8LJhhV MLy9lkmJjhuhuzvvQVtifeuhjsmqgAMwMMzafgvZGYN6ZfOOtD3SNijAbPJVTRWQo1m6 hZ7znMGO/11ObJ4+EZ7ryX2FBrU3nW869ZwbyhEsL/+XKi8AJ0I9LYXoGcs/d5JaT0mP CEe6CTEV7VX66jwmZ0o1GvNbURAwONawkCew+ZyOfXhkDMylk1sqXXgDgEI67RyqnyxP uvRcYp0oQyfThHXgc6OFxpR/Ia0rzC0RBoRsSfiEZDs/R3XJ66Z2JviKr8/3hk1/8xwu Hr1w== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1742311913; x=1742916713; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=byyrGrfPPMb7yp0dDoy8Ot8gd6oeoDPxj7m5qTsv1dM=; b=tNsjA89TowSJZ+4BeOdrsveMnNc38ff9GM4JmMdMZ4XgIChvBuehCcLZqeuXHAtDJO tWP4z8E26B67lLLejuY4Yd8idOTn/9m3eGS2Bm1R9nz4BMezMexFTAE2v6zJAozNJUcO 6YfVvHjrHh9vbIpNZ+oxXJ8eLN4mxe1by3JlZliFt2niQmjXugkxRS0ABudHbGscG2SS xq/jilo/L1e8EcGx+7OM4qiEi5pSylsdzxJAUcWcYbZx0pHf336gv2xNz73vp58lKK/P GuTC5ZeIrRsQadAQMw1zCe+ERUSr3khDS0He6CcVmXyzLnkeTe5OjZylqGj01e38zszr LHCw== X-Gm-Message-State: AOJu0YyiTzZt+XhGnbENa2NuAZxTC+850yBkNJ+n0y2Jb0/saGvIlPKJ CbyYCutQzDCY6DMv91o3801QzfPhPKxS7/0kt7PXGP226zsAthOIWqIyubnM X-Gm-Gg: ASbGncs0qXypbddwDkb1rBl01UsHFW10Kfu8do9AO9LJJKbajXr2dWbVstsLDKf2zV+ GFnkn0axVZB6HD1MdkZ0xwj1NvfcAXOHQKGf/iwbAIctMA0RBswybplmHgKqMkHmTjTSDz7kfUm Gq+cXa7qO3WKqb3mhVO+ANqMRtzO4Ch/0OtCuSMoL8U4U2uTQ6Cfx+IcVS+Z2FSAIs0KDMey0OZ TTewf/JIbEyiQnznB2jMplmtdlIfVxqDqQlTg7khNZI5XQIXbXmKkB/qhd1dCjtloQBSX2+2dAY M5BNG6wsXne1pILExceNEKS4bOqEF4vCLufo358Ty6Hh2CgN3K76fExShkSrqD+wcQzPWeXIRKd wtfiap30nhdgJfvU7K+sgMryyWDVXByIV0ByzPoXrglkt X-Google-Smtp-Source: AGHT+IG28ysCqoO/y1XOvhRFRRgBbqZQe7qLWS1Xwb+hlcTmlENpXrnZaaOnsuCV+ZMaf287LqSPtA== X-Received: by 2002:a05:600c:1d2a:b0:43c:e2dd:98f3 with SMTP id 5b1f17b1804b1-43d3b9ce1ddmr26437345e9.21.1742311912942; Tue, 18 Mar 2025 08:31:52 -0700 (PDT) From: Andrii Sultanov To: xen-devel@lists.xenproject.org Cc: Andrii Sultanov , Jan Beulich , Andrew Cooper , =?utf-8?q?Roger_Pau_Monn=C3=A9?= Subject: [PATCH v3 1/3] drivers: Change amd_iommu struct to contain pci_sbdf_t, simplify code Date: Tue, 18 Mar 2025 15:30:21 +0000 Message-ID: <0d578d1bd063f5d58f3dbaf2ab0e21143333a50e.1742311401.git.sultanovandriy@gmail.com> X-Mailer: git-send-email 2.49.0 In-Reply-To: References: MIME-Version: 1.0 Following on from 250d87dc3ff9 ("x86/msi: Change __msi_set_enable() to take pci_sbdf_t"), struct amd_iommu has its seg and bdf fields backwards with relation to pci_sbdf_t. Instead of regenerating sbdf_t from seg+bdf, make the struct contain pci_sbdf_t directly, which simplifies code. I've avoided aliasing pci_sbdf_t's fields here, so all the usages of amd_iommu->seg|bdf dropped down a namespace to amd_iommu->sbdf.seg|bdf and required renaming. Bloat-o-meter reports: add/remove: 0/0 grow/shrink: 6/11 up/down: 135/-327 (-192) Function old new delta _einittext 22028 22092 +64 amd_iommu_prepare 853 897 +44 _hvm_dpci_msi_eoi 157 168 +11 __mon_lengths 2928 2936 +8 _invalidate_all_devices 133 138 +5 amd_iommu_get_reserved_device_memory 521 524 +3 amd_iommu_domain_destroy 46 43 -3 build_info 752 744 -8 amd_iommu_add_device 856 844 -12 amd_iommu_msi_enable 33 20 -13 update_intremap_entry_from_msi_msg 879 859 -20 amd_iommu_get_supported_ivhd_type 86 54 -32 amd_iommu_detect_one_acpi 918 886 -32 iterate_ivrs_mappings 169 129 -40 flush_command_buffer 460 417 -43 set_iommu_interrupt_handler 421 377 -44 enable_iommu 1745 1665 -80 Resolves: https://gitlab.com/xen-project/xen/-/issues/198 Reported-by: Andrew Cooper Signed-off-by: Andrii Sultanov --- Changes in V3: * Dropped the union with seg+bdf/pci_sbdf_t to avoid aliasing, renamed all users appropriately Changes in V2: * Split single commit into several patches * Added the commit title of the referenced patch * Dropped brackets around &(iommu->sbdf) and &(sbdf) --- xen/drivers/passthrough/amd/iommu.h | 4 +-- xen/drivers/passthrough/amd/iommu_acpi.c | 16 +++++----- xen/drivers/passthrough/amd/iommu_cmd.c | 8 ++--- xen/drivers/passthrough/amd/iommu_detect.c | 18 +++++------ xen/drivers/passthrough/amd/iommu_init.c | 35 ++++++++++----------- xen/drivers/passthrough/amd/iommu_intr.c | 26 +++++++-------- xen/drivers/passthrough/amd/iommu_map.c | 6 ++-- xen/drivers/passthrough/amd/pci_amd_iommu.c | 22 ++++++------- 8 files changed, 66 insertions(+), 69 deletions(-) diff --git a/xen/drivers/passthrough/amd/iommu.h b/xen/drivers/passthrough/amd/iommu.h index 00e81b4b2a..ba541f7943 100644 --- a/xen/drivers/passthrough/amd/iommu.h +++ b/xen/drivers/passthrough/amd/iommu.h @@ -77,8 +77,8 @@ struct amd_iommu { struct list_head list; spinlock_t lock; /* protect iommu */ - u16 seg; - u16 bdf; + pci_sbdf_t sbdf; + struct msi_desc msi; u16 cap_offset; diff --git a/xen/drivers/passthrough/amd/iommu_acpi.c b/xen/drivers/passthrough/amd/iommu_acpi.c index 5bdbfb5ba8..025d9be40f 100644 --- a/xen/drivers/passthrough/amd/iommu_acpi.c +++ b/xen/drivers/passthrough/amd/iommu_acpi.c @@ -58,7 +58,7 @@ static void __init add_ivrs_mapping_entry( uint16_t bdf, uint16_t alias_id, uint8_t flags, unsigned int ext_flags, bool alloc_irt, struct amd_iommu *iommu) { - struct ivrs_mappings *ivrs_mappings = get_ivrs_mappings(iommu->seg); + struct ivrs_mappings *ivrs_mappings = get_ivrs_mappings(iommu->sbdf.seg); ASSERT( ivrs_mappings != NULL ); @@ -70,7 +70,7 @@ static void __init add_ivrs_mapping_entry( ivrs_mappings[bdf].device_flags = flags; /* Don't map an IOMMU by itself. */ - if ( iommu->bdf == bdf ) + if ( iommu->sbdf.bdf == bdf ) return; /* Allocate interrupt remapping table if needed. */ @@ -96,7 +96,7 @@ static void __init add_ivrs_mapping_entry( if ( !ivrs_mappings[alias_id].intremap_table ) panic("No memory for %pp's IRT\n", - &PCI_SBDF(iommu->seg, alias_id)); + &PCI_SBDF(iommu->sbdf.seg, alias_id)); } } @@ -112,7 +112,7 @@ static struct amd_iommu * __init find_iommu_from_bdf_cap( struct amd_iommu *iommu; for_each_amd_iommu ( iommu ) - if ( (iommu->seg == seg) && (iommu->bdf == bdf) && + if ( (iommu->sbdf.seg == seg) && (iommu->sbdf.bdf == bdf) && (iommu->cap_offset == cap_offset) ) return iommu; @@ -297,13 +297,13 @@ static int __init register_range_for_iommu_devices( /* reserve unity-mapped page entries for devices */ for ( bdf = rc = 0; !rc && bdf < ivrs_bdf_entries; bdf++ ) { - if ( iommu != find_iommu_for_device(iommu->seg, bdf) ) + if ( iommu != find_iommu_for_device(iommu->sbdf.seg, bdf) ) continue; - req = get_ivrs_mappings(iommu->seg)[bdf].dte_requestor_id; - rc = reserve_unity_map_for_device(iommu->seg, bdf, base, length, + req = get_ivrs_mappings(iommu->sbdf.seg)[bdf].dte_requestor_id; + rc = reserve_unity_map_for_device(iommu->sbdf.seg, bdf, base, length, iw, ir, false) ?: - reserve_unity_map_for_device(iommu->seg, req, base, length, + reserve_unity_map_for_device(iommu->sbdf.seg, req, base, length, iw, ir, false); } diff --git a/xen/drivers/passthrough/amd/iommu_cmd.c b/xen/drivers/passthrough/amd/iommu_cmd.c index 83c525b84f..4defa0a44d 100644 --- a/xen/drivers/passthrough/amd/iommu_cmd.c +++ b/xen/drivers/passthrough/amd/iommu_cmd.c @@ -40,7 +40,7 @@ static void send_iommu_command(struct amd_iommu *iommu, IOMMU_RING_BUFFER_PTR_MASK) ) { printk_once(XENLOG_ERR "AMD IOMMU %pp: no cmd slot available\n", - &PCI_SBDF(iommu->seg, iommu->bdf)); + &PCI_SBDF(iommu->sbdf.seg, iommu->sbdf.bdf)); cpu_relax(); } @@ -85,7 +85,7 @@ static void flush_command_buffer(struct amd_iommu *iommu, threshold |= threshold << 1; printk(XENLOG_WARNING "AMD IOMMU %pp: %scompletion wait taking too long\n", - &PCI_SBDF(iommu->seg, iommu->bdf), + &iommu->sbdf, timeout_base ? "iotlb " : ""); timeout = 0; } @@ -95,7 +95,7 @@ static void flush_command_buffer(struct amd_iommu *iommu, if ( !timeout ) printk(XENLOG_WARNING "AMD IOMMU %pp: %scompletion wait took %lums\n", - &PCI_SBDF(iommu->seg, iommu->bdf), + &iommu->sbdf, timeout_base ? "iotlb " : "", (NOW() - start) / 10000000); } @@ -300,7 +300,7 @@ void amd_iommu_flush_iotlb(u8 devfn, const struct pci_dev *pdev, if ( !iommu_has_cap(iommu, PCI_CAP_IOTLB_SHIFT) ) return; - req_id = get_dma_requestor_id(iommu->seg, PCI_BDF(pdev->bus, devfn)); + req_id = get_dma_requestor_id(iommu->sbdf.seg, PCI_BDF(pdev->bus, devfn)); queueid = req_id; maxpend = pdev->ats.queue_depth & 0xff; diff --git a/xen/drivers/passthrough/amd/iommu_detect.c b/xen/drivers/passthrough/amd/iommu_detect.c index cede44e651..b9ab09b7c1 100644 --- a/xen/drivers/passthrough/amd/iommu_detect.c +++ b/xen/drivers/passthrough/amd/iommu_detect.c @@ -162,8 +162,8 @@ int __init amd_iommu_detect_one_acpi( spin_lock_init(&iommu->lock); INIT_LIST_HEAD(&iommu->ats_devices); - iommu->seg = ivhd_block->pci_segment_group; - iommu->bdf = ivhd_block->header.device_id; + iommu->sbdf.seg = ivhd_block->pci_segment_group; + iommu->sbdf.bdf = ivhd_block->header.device_id; iommu->cap_offset = ivhd_block->capability_offset; iommu->mmio_base_phys = ivhd_block->base_address; @@ -210,16 +210,16 @@ int __init amd_iommu_detect_one_acpi( /* override IOMMU HT flags */ iommu->ht_flags = ivhd_block->header.flags; - bus = PCI_BUS(iommu->bdf); - dev = PCI_SLOT(iommu->bdf); - func = PCI_FUNC(iommu->bdf); + bus = PCI_BUS(iommu->sbdf.bdf); + dev = PCI_SLOT(iommu->sbdf.bdf); + func = PCI_FUNC(iommu->sbdf.bdf); - rt = get_iommu_capabilities(iommu->seg, bus, dev, func, + rt = get_iommu_capabilities(iommu->sbdf.seg, bus, dev, func, iommu->cap_offset, iommu); if ( rt ) goto out; - rt = get_iommu_msi_capabilities(iommu->seg, bus, dev, func, iommu); + rt = get_iommu_msi_capabilities(iommu->sbdf.seg, bus, dev, func, iommu); if ( rt ) goto out; @@ -228,10 +228,10 @@ int __init amd_iommu_detect_one_acpi( if ( !iommu->domid_map ) goto out; - rt = pci_ro_device(iommu->seg, bus, PCI_DEVFN(dev, func)); + rt = pci_ro_device(iommu->sbdf.seg, bus, PCI_DEVFN(dev, func)); if ( rt ) printk(XENLOG_ERR "Could not mark config space of %pp read-only (%d)\n", - &PCI_SBDF(iommu->seg, iommu->bdf), rt); + &iommu->sbdf, rt); list_add_tail(&iommu->list, &amd_iommu_head); rt = 0; diff --git a/xen/drivers/passthrough/amd/iommu_init.c b/xen/drivers/passthrough/amd/iommu_init.c index bb25b55c85..58d657060a 100644 --- a/xen/drivers/passthrough/amd/iommu_init.c +++ b/xen/drivers/passthrough/amd/iommu_init.c @@ -409,9 +409,7 @@ static void iommu_reset_log(struct amd_iommu *iommu, static void amd_iommu_msi_enable(struct amd_iommu *iommu, int flag) { - pci_sbdf_t sbdf = { .seg = iommu->seg, .bdf = iommu->bdf }; - - __msi_set_enable(sbdf, iommu->msi.msi_attrib.pos, flag); + __msi_set_enable(iommu->sbdf, iommu->msi.msi_attrib.pos, flag); } static void cf_check iommu_msi_unmask(struct irq_desc *desc) @@ -566,7 +564,7 @@ static void cf_check parse_event_log_entry(struct amd_iommu *iommu, u32 entry[]) printk(XENLOG_ERR "AMD-Vi: %s: %pp d%u addr %016"PRIx64 " flags %#x%s%s%s%s%s%s%s%s%s%s\n", - code_str, &PCI_SBDF(iommu->seg, device_id), + code_str, &PCI_SBDF(iommu->sbdf.seg, device_id), domain_id, addr, flags, (flags & 0xe00) ? " ??" : "", (flags & 0x100) ? " TR" : "", @@ -583,8 +581,8 @@ static void cf_check parse_event_log_entry(struct amd_iommu *iommu, u32 entry[]) amd_iommu_print_entries(iommu, device_id, daddr_to_dfn(addr)); for ( bdf = 0; bdf < ivrs_bdf_entries; bdf++ ) - if ( get_dma_requestor_id(iommu->seg, bdf) == device_id ) - pci_check_disable_device(iommu->seg, PCI_BUS(bdf), + if ( get_dma_requestor_id(iommu->sbdf.seg, bdf) == device_id ) + pci_check_disable_device(iommu->sbdf.seg, PCI_BUS(bdf), PCI_DEVFN(bdf)); } else @@ -643,7 +641,7 @@ static void cf_check parse_ppr_log_entry(struct amd_iommu *iommu, u32 entry[]) struct pci_dev *pdev; pcidevs_lock(); - pdev = pci_get_real_pdev(PCI_SBDF(iommu->seg, device_id)); + pdev = pci_get_real_pdev(PCI_SBDF(iommu->sbdf.seg, device_id)); pcidevs_unlock(); if ( pdev ) @@ -752,12 +750,11 @@ static bool __init set_iommu_interrupt_handler(struct amd_iommu *iommu) } pcidevs_lock(); - iommu->msi.dev = pci_get_pdev(NULL, PCI_SBDF(iommu->seg, iommu->bdf)); + iommu->msi.dev = pci_get_pdev(NULL, iommu->sbdf); pcidevs_unlock(); if ( !iommu->msi.dev ) { - AMD_IOMMU_WARN("no pdev for %pp\n", - &PCI_SBDF(iommu->seg, iommu->bdf)); + AMD_IOMMU_WARN("no pdev for %pp\n", &iommu->sbdf); return 0; } @@ -779,7 +776,7 @@ static bool __init set_iommu_interrupt_handler(struct amd_iommu *iommu) hw_irq_controller *handler; u16 control; - control = pci_conf_read16(PCI_SBDF(iommu->seg, iommu->bdf), + control = pci_conf_read16(iommu->sbdf, iommu->msi.msi_attrib.pos + PCI_MSI_FLAGS); iommu->msi.msi.nvec = 1; @@ -843,22 +840,22 @@ static void amd_iommu_erratum_746_workaround(struct amd_iommu *iommu) (boot_cpu_data.x86_model > 0x1f) ) return; - pci_conf_write32(PCI_SBDF(iommu->seg, iommu->bdf), 0xf0, 0x90); - value = pci_conf_read32(PCI_SBDF(iommu->seg, iommu->bdf), 0xf4); + pci_conf_write32(iommu->sbdf, 0xf0, 0x90); + value = pci_conf_read32(iommu->sbdf, 0xf4); if ( value & (1 << 2) ) return; /* Select NB indirect register 0x90 and enable writing */ - pci_conf_write32(PCI_SBDF(iommu->seg, iommu->bdf), 0xf0, 0x90 | (1 << 8)); + pci_conf_write32(iommu->sbdf, 0xf0, 0x90 | (1 << 8)); - pci_conf_write32(PCI_SBDF(iommu->seg, iommu->bdf), 0xf4, value | (1 << 2)); + pci_conf_write32(iommu->sbdf, 0xf4, value | (1 << 2)); printk(XENLOG_INFO "AMD-Vi: Applying erratum 746 workaround for IOMMU at %pp\n", - &PCI_SBDF(iommu->seg, iommu->bdf)); + &iommu->sbdf); /* Clear the enable writing bit */ - pci_conf_write32(PCI_SBDF(iommu->seg, iommu->bdf), 0xf0, 0x90); + pci_conf_write32(iommu->sbdf, 0xf0, 0x90); } static void enable_iommu(struct amd_iommu *iommu) @@ -1095,7 +1092,7 @@ static int __init amd_iommu_init_one(struct amd_iommu *iommu, bool intr) goto error_out; /* Make sure that the device table has been successfully allocated. */ - ivrs_mappings = get_ivrs_mappings(iommu->seg); + ivrs_mappings = get_ivrs_mappings(iommu->sbdf.seg); if ( !IVRS_MAPPINGS_DEVTAB(ivrs_mappings) ) goto error_out; @@ -1363,7 +1360,7 @@ static bool __init amd_sp5100_erratum28(void) static int __init amd_iommu_prepare_one(struct amd_iommu *iommu) { - int rc = alloc_ivrs_mappings(iommu->seg); + int rc = alloc_ivrs_mappings(iommu->sbdf.seg); if ( !rc ) rc = map_iommu_mmio_region(iommu); diff --git a/xen/drivers/passthrough/amd/iommu_intr.c b/xen/drivers/passthrough/amd/iommu_intr.c index 9abdc38053..754e9051dc 100644 --- a/xen/drivers/passthrough/amd/iommu_intr.c +++ b/xen/drivers/passthrough/amd/iommu_intr.c @@ -132,7 +132,7 @@ static int get_intremap_requestor_id(int seg, int bdf) static unsigned int alloc_intremap_entry(const struct amd_iommu *iommu, unsigned int bdf, unsigned int nr) { - const struct ivrs_mappings *ivrs_mappings = get_ivrs_mappings(iommu->seg); + const struct ivrs_mappings *ivrs_mappings = get_ivrs_mappings(iommu->sbdf.seg); unsigned long *inuse = ivrs_mappings[bdf].intremap_inuse; unsigned int nr_ents = intremap_table_entries(ivrs_mappings[bdf].intremap_table, iommu); @@ -167,7 +167,7 @@ static union irte_ptr get_intremap_entry(const struct amd_iommu *iommu, unsigned int bdf, unsigned int index) { union irte_ptr table = { - .ptr = get_ivrs_mappings(iommu->seg)[bdf].intremap_table + .ptr = get_ivrs_mappings(iommu->sbdf.seg)[bdf].intremap_table }; ASSERT(table.ptr && (index < intremap_table_entries(table.ptr, iommu))); @@ -184,7 +184,7 @@ static void free_intremap_entry(const struct amd_iommu *iommu, unsigned int bdf, unsigned int index) { union irte_ptr entry = get_intremap_entry(iommu, bdf, index); - struct ivrs_mappings *ivrs = get_ivrs_mappings(iommu->seg); + struct ivrs_mappings *ivrs = get_ivrs_mappings(iommu->sbdf.seg); if ( iommu->ctrl.ga_en ) { @@ -281,8 +281,8 @@ static int update_intremap_entry_from_ioapic( unsigned int dest, offset; bool fresh = false; - req_id = get_intremap_requestor_id(iommu->seg, bdf); - lock = get_intremap_lock(iommu->seg, req_id); + req_id = get_intremap_requestor_id(iommu->sbdf.seg, bdf); + lock = get_intremap_lock(iommu->sbdf.seg, req_id); delivery_mode = rte->delivery_mode; vector = rte->vector; @@ -419,10 +419,10 @@ static int update_intremap_entry_from_msi_msg( unsigned int dest, offset, i; bool fresh = false; - req_id = get_dma_requestor_id(iommu->seg, bdf); - alias_id = get_intremap_requestor_id(iommu->seg, bdf); + req_id = get_dma_requestor_id(iommu->sbdf.seg, bdf); + alias_id = get_intremap_requestor_id(iommu->sbdf.seg, bdf); - lock = get_intremap_lock(iommu->seg, req_id); + lock = get_intremap_lock(iommu->sbdf.seg, req_id); spin_lock_irqsave(lock, flags); if ( msg == NULL ) @@ -486,10 +486,10 @@ static int update_intremap_entry_from_msi_msg( */ if ( ( req_id != alias_id ) && - get_ivrs_mappings(iommu->seg)[alias_id].intremap_table != NULL ) + get_ivrs_mappings(iommu->sbdf.seg)[alias_id].intremap_table != NULL ) { - BUG_ON(get_ivrs_mappings(iommu->seg)[req_id].intremap_table != - get_ivrs_mappings(iommu->seg)[alias_id].intremap_table); + BUG_ON(get_ivrs_mappings(iommu->sbdf.seg)[req_id].intremap_table != + get_ivrs_mappings(iommu->sbdf.seg)[alias_id].intremap_table); } return fresh; @@ -500,7 +500,7 @@ static struct amd_iommu *_find_iommu_for_device(int seg, int bdf) struct amd_iommu *iommu; for_each_amd_iommu ( iommu ) - if ( iommu->seg == seg && iommu->bdf == bdf ) + if ( iommu->sbdf.seg == seg && iommu->sbdf.bdf == bdf ) return NULL; iommu = find_iommu_for_device(seg, bdf); @@ -730,7 +730,7 @@ static void dump_intremap_table(const struct amd_iommu *iommu, if ( ivrs_mapping ) { printk(" %pp:\n", - &PCI_SBDF(iommu->seg, ivrs_mapping->dte_requestor_id)); + &PCI_SBDF(iommu->sbdf.seg, ivrs_mapping->dte_requestor_id)); ivrs_mapping = NULL; } diff --git a/xen/drivers/passthrough/amd/iommu_map.c b/xen/drivers/passthrough/amd/iommu_map.c index dde393645a..78bde0aeb6 100644 --- a/xen/drivers/passthrough/amd/iommu_map.c +++ b/xen/drivers/passthrough/amd/iommu_map.c @@ -558,14 +558,14 @@ void amd_iommu_print_entries(const struct amd_iommu *iommu, unsigned int dev_id, if ( !dt[dev_id].tv ) { - printk("%pp: no root\n", &PCI_SBDF(iommu->seg, dev_id)); + printk("%pp: no root\n", &PCI_SBDF(iommu->sbdf.seg, dev_id)); return; } pt_mfn = _mfn(dt[dev_id].pt_root); level = dt[dev_id].paging_mode; printk("%pp root @ %"PRI_mfn" (%u levels) dfn=%"PRI_dfn"\n", - &PCI_SBDF(iommu->seg, dev_id), mfn_x(pt_mfn), level, dfn_x(dfn)); + &PCI_SBDF(iommu->sbdf.seg, dev_id), mfn_x(pt_mfn), level, dfn_x(dfn)); while ( level ) { @@ -730,7 +730,7 @@ int cf_check amd_iommu_get_reserved_device_memory( * the same alias ID. */ if ( bdf != req && ivrs_mappings[req].iommu && - func(0, 0, PCI_SBDF(seg, req).sbdf, ctxt) ) + func(0, 0, sbdf.sbdf, ctxt) ) continue; if ( global == pending ) diff --git a/xen/drivers/passthrough/amd/pci_amd_iommu.c b/xen/drivers/passthrough/amd/pci_amd_iommu.c index d00697edb3..55218d9a1d 100644 --- a/xen/drivers/passthrough/amd/pci_amd_iommu.c +++ b/xen/drivers/passthrough/amd/pci_amd_iommu.c @@ -43,7 +43,7 @@ struct amd_iommu *find_iommu_for_device(int seg, int bdf) { unsigned int bd0 = bdf & ~PCI_FUNC(~0); - if ( ivrs_mappings[bd0].iommu && ivrs_mappings[bd0].iommu->bdf != bdf ) + if ( ivrs_mappings[bd0].iommu && ivrs_mappings[bd0].iommu->sbdf.bdf != bdf ) { struct ivrs_mappings tmp = ivrs_mappings[bd0]; @@ -121,7 +121,7 @@ static bool use_ats( { return !ivrs_dev->block_ats && iommu_has_cap(iommu, PCI_CAP_IOTLB_SHIFT) && - pci_ats_device(iommu->seg, pdev->bus, pdev->devfn); + pci_ats_device(iommu->sbdf.seg, pdev->bus, pdev->devfn); } static int __must_check amd_iommu_setup_domain_device( @@ -147,17 +147,17 @@ static int __must_check amd_iommu_setup_domain_device( if ( rc ) return rc; - req_id = get_dma_requestor_id(iommu->seg, pdev->sbdf.bdf); - ivrs_dev = &get_ivrs_mappings(iommu->seg)[req_id]; + req_id = get_dma_requestor_id(iommu->sbdf.seg, pdev->sbdf.bdf); + ivrs_dev = &get_ivrs_mappings(iommu->sbdf.seg)[req_id]; sr_flags = (iommu_hwdom_passthrough && is_hardware_domain(domain) ? 0 : SET_ROOT_VALID) | (ivrs_dev->unity_map ? SET_ROOT_WITH_UNITY_MAP : 0); /* get device-table entry */ - req_id = get_dma_requestor_id(iommu->seg, PCI_BDF(bus, devfn)); + req_id = get_dma_requestor_id(iommu->sbdf.seg, PCI_BDF(bus, devfn)); table = iommu->dev_table.buffer; dte = &table[req_id]; - ivrs_dev = &get_ivrs_mappings(iommu->seg)[req_id]; + ivrs_dev = &get_ivrs_mappings(iommu->sbdf.seg)[req_id]; if ( domain != dom_io ) { @@ -275,7 +275,7 @@ static int __must_check amd_iommu_setup_domain_device( ASSERT(pcidevs_locked()); if ( use_ats(pdev, iommu, ivrs_dev) && - !pci_ats_enabled(iommu->seg, bus, pdev->devfn) ) + !pci_ats_enabled(iommu->sbdf.seg, bus, pdev->devfn) ) { if ( devfn == pdev->devfn ) enable_ats_device(pdev, &iommu->ats_devices); @@ -418,12 +418,12 @@ static void amd_iommu_disable_domain_device(const struct domain *domain, ASSERT(pcidevs_locked()); - if ( pci_ats_device(iommu->seg, bus, pdev->devfn) && - pci_ats_enabled(iommu->seg, bus, pdev->devfn) ) + if ( pci_ats_device(iommu->sbdf.seg, bus, pdev->devfn) && + pci_ats_enabled(iommu->sbdf.seg, bus, pdev->devfn) ) disable_ats_device(pdev); BUG_ON ( iommu->dev_table.buffer == NULL ); - req_id = get_dma_requestor_id(iommu->seg, PCI_BDF(bus, devfn)); + req_id = get_dma_requestor_id(iommu->sbdf.seg, PCI_BDF(bus, devfn)); table = iommu->dev_table.buffer; dte = &table[req_id]; @@ -578,7 +578,7 @@ static int cf_check amd_iommu_add_device(u8 devfn, struct pci_dev *pdev) return -EINVAL; for_each_amd_iommu(iommu) - if ( pdev->seg == iommu->seg && pdev->sbdf.bdf == iommu->bdf ) + if ( pdev->seg == iommu->sbdf.seg && pdev->sbdf.bdf == iommu->sbdf.bdf ) return is_hardware_domain(pdev->domain) ? 0 : -ENODEV; iommu = find_iommu_for_device(pdev->seg, pdev->sbdf.bdf);