Message ID | 2-v2-fd55d00a60b2+c69-arm_no_split_jgg@nvidia.com (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | Remove split on unmap behavior | expand |
On 2024-11-04 5:41 pm, Jason Gunthorpe wrote: > A minority of page table implementations (arm_lpae, armv7) are unique in > how they handle partial unmap of large IOPTEs. > > Other implementations will unmap the large IOPTE and return it's > length. For example if a 2M IOPTE is present and the first 4K is requested > to be unmapped then unmap will remove the whole 2M and report 2M as the > result. > > armv7 instead will break up contiguous entries and replace an entry with a > whole table so it can unmap the requested 4k. > > This seems copied from the arm_lpae implementation, which was analyzed > here: > > https://lore.kernel.org/all/20241024134411.GA6956@nvidia.com/ > > Bring consistency to the implementations and remove this unused > functionality. > > There are no uses outside iommu, this effects the ARM_V7S drivers > msm_iommu, mtk_iommu, and arm-smmmu. > > Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> > --- > drivers/iommu/io-pgtable-arm-v7s.c | 125 +---------------------------- > 1 file changed, 4 insertions(+), 121 deletions(-) Yikes, I'd forgotten quite how much horribleness was devoted to this, despite it being the "simpler" non-recursive one... However, there are also "partial unmap" cases in both sets of selftests, so I think there's still a bit more to remove yet :) Thanks, Robin. > diff --git a/drivers/iommu/io-pgtable-arm-v7s.c b/drivers/iommu/io-pgtable-arm-v7s.c > index 06ffc683b28fee..7e37459cd28332 100644 > --- a/drivers/iommu/io-pgtable-arm-v7s.c > +++ b/drivers/iommu/io-pgtable-arm-v7s.c > @@ -166,7 +166,6 @@ struct arm_v7s_io_pgtable { > > arm_v7s_iopte *pgd; > struct kmem_cache *l2_tables; > - spinlock_t split_lock; > }; > > static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl); > @@ -363,25 +362,6 @@ static arm_v7s_iopte arm_v7s_prot_to_pte(int prot, int lvl, > return pte; > } > > -static int arm_v7s_pte_to_prot(arm_v7s_iopte pte, int lvl) > -{ > - int prot = IOMMU_READ; > - arm_v7s_iopte attr = pte >> ARM_V7S_ATTR_SHIFT(lvl); > - > - if (!(attr & ARM_V7S_PTE_AP_RDONLY)) > - prot |= IOMMU_WRITE; > - if (!(attr & ARM_V7S_PTE_AP_UNPRIV)) > - prot |= IOMMU_PRIV; > - if ((attr & (ARM_V7S_TEX_MASK << ARM_V7S_TEX_SHIFT)) == 0) > - prot |= IOMMU_MMIO; > - else if (pte & ARM_V7S_ATTR_C) > - prot |= IOMMU_CACHE; > - if (pte & ARM_V7S_ATTR_XN(lvl)) > - prot |= IOMMU_NOEXEC; > - > - return prot; > -} > - > static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl) > { > if (lvl == 1) { > @@ -398,23 +378,6 @@ static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl) > return pte; > } > > -static arm_v7s_iopte arm_v7s_cont_to_pte(arm_v7s_iopte pte, int lvl) > -{ > - if (lvl == 1) { > - pte &= ~ARM_V7S_CONT_SECTION; > - } else if (lvl == 2) { > - arm_v7s_iopte xn = pte & BIT(ARM_V7S_CONT_PAGE_XN_SHIFT); > - arm_v7s_iopte tex = pte & (ARM_V7S_CONT_PAGE_TEX_MASK << > - ARM_V7S_CONT_PAGE_TEX_SHIFT); > - > - pte ^= xn | tex | ARM_V7S_PTE_TYPE_CONT_PAGE; > - pte |= (xn >> ARM_V7S_CONT_PAGE_XN_SHIFT) | > - (tex >> ARM_V7S_CONT_PAGE_TEX_SHIFT) | > - ARM_V7S_PTE_TYPE_PAGE; > - } > - return pte; > -} > - > static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl) > { > if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte, lvl)) > @@ -591,77 +554,6 @@ static void arm_v7s_free_pgtable(struct io_pgtable *iop) > kfree(data); > } > > -static arm_v7s_iopte arm_v7s_split_cont(struct arm_v7s_io_pgtable *data, > - unsigned long iova, int idx, int lvl, > - arm_v7s_iopte *ptep) > -{ > - struct io_pgtable *iop = &data->iop; > - arm_v7s_iopte pte; > - size_t size = ARM_V7S_BLOCK_SIZE(lvl); > - int i; > - > - /* Check that we didn't lose a race to get the lock */ > - pte = *ptep; > - if (!arm_v7s_pte_is_cont(pte, lvl)) > - return pte; > - > - ptep -= idx & (ARM_V7S_CONT_PAGES - 1); > - pte = arm_v7s_cont_to_pte(pte, lvl); > - for (i = 0; i < ARM_V7S_CONT_PAGES; i++) > - ptep[i] = pte + i * size; > - > - __arm_v7s_pte_sync(ptep, ARM_V7S_CONT_PAGES, &iop->cfg); > - > - size *= ARM_V7S_CONT_PAGES; > - io_pgtable_tlb_flush_walk(iop, iova, size, size); > - return pte; > -} > - > -static size_t arm_v7s_split_blk_unmap(struct arm_v7s_io_pgtable *data, > - struct iommu_iotlb_gather *gather, > - unsigned long iova, size_t size, > - arm_v7s_iopte blk_pte, > - arm_v7s_iopte *ptep) > -{ > - struct io_pgtable_cfg *cfg = &data->iop.cfg; > - arm_v7s_iopte pte, *tablep; > - int i, unmap_idx, num_entries, num_ptes; > - > - tablep = __arm_v7s_alloc_table(2, GFP_ATOMIC, data); > - if (!tablep) > - return 0; /* Bytes unmapped */ > - > - num_ptes = ARM_V7S_PTES_PER_LVL(2, cfg); > - num_entries = size >> ARM_V7S_LVL_SHIFT(2); > - unmap_idx = ARM_V7S_LVL_IDX(iova, 2, cfg); > - > - pte = arm_v7s_prot_to_pte(arm_v7s_pte_to_prot(blk_pte, 1), 2, cfg); > - if (num_entries > 1) > - pte = arm_v7s_pte_to_cont(pte, 2); > - > - for (i = 0; i < num_ptes; i += num_entries, pte += size) { > - /* Unmap! */ > - if (i == unmap_idx) > - continue; > - > - __arm_v7s_set_pte(&tablep[i], pte, num_entries, cfg); > - } > - > - pte = arm_v7s_install_table(tablep, ptep, blk_pte, cfg); > - if (pte != blk_pte) { > - __arm_v7s_free_table(tablep, 2, data); > - > - if (!ARM_V7S_PTE_IS_TABLE(pte, 1)) > - return 0; > - > - tablep = iopte_deref(pte, 1, data); > - return __arm_v7s_unmap(data, gather, iova, size, 2, tablep); > - } > - > - io_pgtable_tlb_add_page(&data->iop, gather, iova, size); > - return size; > -} > - > static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *data, > struct iommu_iotlb_gather *gather, > unsigned long iova, size_t size, int lvl, > @@ -694,11 +586,8 @@ static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *data, > * case in a lock for the sake of correctness and be done with it. > */ > if (num_entries <= 1 && arm_v7s_pte_is_cont(pte[0], lvl)) { > - unsigned long flags; > - > - spin_lock_irqsave(&data->split_lock, flags); > - pte[0] = arm_v7s_split_cont(data, iova, idx, lvl, ptep); > - spin_unlock_irqrestore(&data->split_lock, flags); > + WARN_ONCE(true, "Unmap of a partial large IOPTE is not allowed"); > + return 0; > } > > /* If the size matches this level, we're in the right place */ > @@ -721,12 +610,8 @@ static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *data, > } > return size; > } else if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte[0], lvl)) { > - /* > - * Insert a table at the next level to map the old region, > - * minus the part we want to unmap > - */ > - return arm_v7s_split_blk_unmap(data, gather, iova, size, pte[0], > - ptep); > + WARN_ONCE(true, "Unmap of a partial large IOPTE is not allowed"); > + return 0; > } > > /* Keep on walkin' */ > @@ -811,8 +696,6 @@ static struct io_pgtable *arm_v7s_alloc_pgtable(struct io_pgtable_cfg *cfg, > if (!data) > return NULL; > > - spin_lock_init(&data->split_lock); > - > /* > * ARM_MTK_TTBR_EXT extend the translation table base support larger > * memory address.
On Mon, Nov 04, 2024 at 07:53:46PM +0000, Robin Murphy wrote: > On 2024-11-04 5:41 pm, Jason Gunthorpe wrote: > > A minority of page table implementations (arm_lpae, armv7) are unique in > > how they handle partial unmap of large IOPTEs. > > > > Other implementations will unmap the large IOPTE and return it's > > length. For example if a 2M IOPTE is present and the first 4K is requested > > to be unmapped then unmap will remove the whole 2M and report 2M as the > > result. > > > > armv7 instead will break up contiguous entries and replace an entry with a > > whole table so it can unmap the requested 4k. > > > > This seems copied from the arm_lpae implementation, which was analyzed > > here: > > > > https://lore.kernel.org/all/20241024134411.GA6956@nvidia.com/ > > > > Bring consistency to the implementations and remove this unused > > functionality. > > > > There are no uses outside iommu, this effects the ARM_V7S drivers > > msm_iommu, mtk_iommu, and arm-smmmu. > > > > Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> > > --- > > drivers/iommu/io-pgtable-arm-v7s.c | 125 +---------------------------- > > 1 file changed, 4 insertions(+), 121 deletions(-) > > Yikes, I'd forgotten quite how much horribleness was devoted to this, > despite it being the "simpler" non-recursive one... Yes, it is the contiguous page support that makes it so complex.. > However, there are also "partial unmap" cases in both sets of selftests, so > I think there's still a bit more to remove yet :) Sneaky, I got it thanks Runs OK now: arm-v7s io-pgtable: self test ok arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 32 Jason --- a/drivers/iommu/io-pgtable-arm-v7s.c +++ b/drivers/iommu/io-pgtable-arm-v7s.c @@ -819,7 +819,7 @@ static int __init arm_v7s_do_selftests(void) .quirks = IO_PGTABLE_QUIRK_ARM_NS, .pgsize_bitmap = SZ_4K | SZ_64K | SZ_1M | SZ_16M, }; - unsigned int iova, size, iova_start; + unsigned int iova, size; unsigned int i, loopnr = 0; size_t mapped; @@ -871,25 +871,6 @@ static int __init arm_v7s_do_selftests(void) loopnr++; } - /* Partial unmap */ - i = 1; - size = 1UL << __ffs(cfg.pgsize_bitmap); - while (i < loopnr) { - iova_start = i * SZ_16M; - if (ops->unmap_pages(ops, iova_start + size, size, 1, NULL) != size) - return __FAIL(ops); - - /* Remap of partial unmap */ - if (ops->map_pages(ops, iova_start + size, size, size, 1, - IOMMU_READ, GFP_KERNEL, &mapped)) - return __FAIL(ops); - - if (ops->iova_to_phys(ops, iova_start + size + 42) - != (size + 42)) - return __FAIL(ops); - i++; - } - /* Full unmap */ iova = 0; for_each_set_bit(i, &cfg.pgsize_bitmap, BITS_PER_LONG) {
On Mon, Nov 04, 2024 at 04:09:51PM -0400, Jason Gunthorpe wrote: > Runs OK now: > > arm-v7s io-pgtable: self test ok > arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 32 > > Jason > > --- a/drivers/iommu/io-pgtable-arm-v7s.c > +++ b/drivers/iommu/io-pgtable-arm-v7s.c > @@ -819,7 +819,7 @@ static int __init arm_v7s_do_selftests(void) > .quirks = IO_PGTABLE_QUIRK_ARM_NS, > .pgsize_bitmap = SZ_4K | SZ_64K | SZ_1M | SZ_16M, > }; > - unsigned int iova, size, iova_start; > + unsigned int iova, size; > unsigned int i, loopnr = 0; > size_t mapped; > > @@ -871,25 +871,6 @@ static int __init arm_v7s_do_selftests(void) > loopnr++; > } > > - /* Partial unmap */ > - i = 1; > - size = 1UL << __ffs(cfg.pgsize_bitmap); > - while (i < loopnr) { > - iova_start = i * SZ_16M; > - if (ops->unmap_pages(ops, iova_start + size, size, 1, NULL) != size) > - return __FAIL(ops); > - > - /* Remap of partial unmap */ > - if (ops->map_pages(ops, iova_start + size, size, size, 1, > - IOMMU_READ, GFP_KERNEL, &mapped)) > - return __FAIL(ops); > - > - if (ops->iova_to_phys(ops, iova_start + size + 42) > - != (size + 42)) > - return __FAIL(ops); > - i++; > - } > - > /* Full unmap */ > iova = 0; > for_each_set_bit(i, &cfg.pgsize_bitmap, BITS_PER_LONG) { Yup, and you can do the same for the other selftest in io-pgtable-arm.c Will
On Tue, Nov 05, 2024 at 04:59:43PM +0000, Will Deacon wrote: > > /* Full unmap */ > > iova = 0; > > for_each_set_bit(i, &cfg.pgsize_bitmap, BITS_PER_LONG) { > > Yup, and you can do the same for the other selftest in io-pgtable-arm.c Ugh, yes, I ran it and thought the log it printed was the success log, it did actually fail too. This seems like the right output: arm-v7s io-pgtable: self test ok arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 32 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 36 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 40 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 42 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 44 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x40201000, IAS 48 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x02004000, IAS 32 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x02004000, IAS 36 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x02004000, IAS 40 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x02004000, IAS 42 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x02004000, IAS 44 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x02004000, IAS 48 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x20010000, IAS 32 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x20010000, IAS 36 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x20010000, IAS 40 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x20010000, IAS 42 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x20010000, IAS 44 arm-lpae io-pgtable: selftest: pgsize_bitmap 0x20010000, IAS 48 arm-lpae io-pgtable: selftest: completed with 18 PASS 0 FAIL diff --git a/drivers/iommu/io-pgtable-arm.c b/drivers/iommu/io-pgtable-arm.c index 9c9ecfdf87be90..abaf323843e3c0 100644 --- a/drivers/iommu/io-pgtable-arm.c +++ b/drivers/iommu/io-pgtable-arm.c @@ -1283,19 +1283,6 @@ static int __init arm_lpae_run_tests(struct io_pgtable_cfg *cfg) iova += SZ_1G; } - /* Partial unmap */ - size = 1UL << __ffs(cfg->pgsize_bitmap); - if (ops->unmap_pages(ops, SZ_1G + size, size, 1, NULL) != size) - return __FAIL(ops, i); - - /* Remap of partial unmap */ - if (ops->map_pages(ops, SZ_1G + size, size, size, 1, - IOMMU_READ, GFP_KERNEL, &mapped)) - return __FAIL(ops, i); - - if (ops->iova_to_phys(ops, SZ_1G + size + 42) != (size + 42)) - return __FAIL(ops, i); - /* Full unmap */ iova = 0; for_each_set_bit(j, &cfg->pgsize_bitmap, BITS_PER_LONG) {
diff --git a/drivers/iommu/io-pgtable-arm-v7s.c b/drivers/iommu/io-pgtable-arm-v7s.c index 06ffc683b28fee..7e37459cd28332 100644 --- a/drivers/iommu/io-pgtable-arm-v7s.c +++ b/drivers/iommu/io-pgtable-arm-v7s.c @@ -166,7 +166,6 @@ struct arm_v7s_io_pgtable { arm_v7s_iopte *pgd; struct kmem_cache *l2_tables; - spinlock_t split_lock; }; static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl); @@ -363,25 +362,6 @@ static arm_v7s_iopte arm_v7s_prot_to_pte(int prot, int lvl, return pte; } -static int arm_v7s_pte_to_prot(arm_v7s_iopte pte, int lvl) -{ - int prot = IOMMU_READ; - arm_v7s_iopte attr = pte >> ARM_V7S_ATTR_SHIFT(lvl); - - if (!(attr & ARM_V7S_PTE_AP_RDONLY)) - prot |= IOMMU_WRITE; - if (!(attr & ARM_V7S_PTE_AP_UNPRIV)) - prot |= IOMMU_PRIV; - if ((attr & (ARM_V7S_TEX_MASK << ARM_V7S_TEX_SHIFT)) == 0) - prot |= IOMMU_MMIO; - else if (pte & ARM_V7S_ATTR_C) - prot |= IOMMU_CACHE; - if (pte & ARM_V7S_ATTR_XN(lvl)) - prot |= IOMMU_NOEXEC; - - return prot; -} - static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl) { if (lvl == 1) { @@ -398,23 +378,6 @@ static arm_v7s_iopte arm_v7s_pte_to_cont(arm_v7s_iopte pte, int lvl) return pte; } -static arm_v7s_iopte arm_v7s_cont_to_pte(arm_v7s_iopte pte, int lvl) -{ - if (lvl == 1) { - pte &= ~ARM_V7S_CONT_SECTION; - } else if (lvl == 2) { - arm_v7s_iopte xn = pte & BIT(ARM_V7S_CONT_PAGE_XN_SHIFT); - arm_v7s_iopte tex = pte & (ARM_V7S_CONT_PAGE_TEX_MASK << - ARM_V7S_CONT_PAGE_TEX_SHIFT); - - pte ^= xn | tex | ARM_V7S_PTE_TYPE_CONT_PAGE; - pte |= (xn >> ARM_V7S_CONT_PAGE_XN_SHIFT) | - (tex >> ARM_V7S_CONT_PAGE_TEX_SHIFT) | - ARM_V7S_PTE_TYPE_PAGE; - } - return pte; -} - static bool arm_v7s_pte_is_cont(arm_v7s_iopte pte, int lvl) { if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte, lvl)) @@ -591,77 +554,6 @@ static void arm_v7s_free_pgtable(struct io_pgtable *iop) kfree(data); } -static arm_v7s_iopte arm_v7s_split_cont(struct arm_v7s_io_pgtable *data, - unsigned long iova, int idx, int lvl, - arm_v7s_iopte *ptep) -{ - struct io_pgtable *iop = &data->iop; - arm_v7s_iopte pte; - size_t size = ARM_V7S_BLOCK_SIZE(lvl); - int i; - - /* Check that we didn't lose a race to get the lock */ - pte = *ptep; - if (!arm_v7s_pte_is_cont(pte, lvl)) - return pte; - - ptep -= idx & (ARM_V7S_CONT_PAGES - 1); - pte = arm_v7s_cont_to_pte(pte, lvl); - for (i = 0; i < ARM_V7S_CONT_PAGES; i++) - ptep[i] = pte + i * size; - - __arm_v7s_pte_sync(ptep, ARM_V7S_CONT_PAGES, &iop->cfg); - - size *= ARM_V7S_CONT_PAGES; - io_pgtable_tlb_flush_walk(iop, iova, size, size); - return pte; -} - -static size_t arm_v7s_split_blk_unmap(struct arm_v7s_io_pgtable *data, - struct iommu_iotlb_gather *gather, - unsigned long iova, size_t size, - arm_v7s_iopte blk_pte, - arm_v7s_iopte *ptep) -{ - struct io_pgtable_cfg *cfg = &data->iop.cfg; - arm_v7s_iopte pte, *tablep; - int i, unmap_idx, num_entries, num_ptes; - - tablep = __arm_v7s_alloc_table(2, GFP_ATOMIC, data); - if (!tablep) - return 0; /* Bytes unmapped */ - - num_ptes = ARM_V7S_PTES_PER_LVL(2, cfg); - num_entries = size >> ARM_V7S_LVL_SHIFT(2); - unmap_idx = ARM_V7S_LVL_IDX(iova, 2, cfg); - - pte = arm_v7s_prot_to_pte(arm_v7s_pte_to_prot(blk_pte, 1), 2, cfg); - if (num_entries > 1) - pte = arm_v7s_pte_to_cont(pte, 2); - - for (i = 0; i < num_ptes; i += num_entries, pte += size) { - /* Unmap! */ - if (i == unmap_idx) - continue; - - __arm_v7s_set_pte(&tablep[i], pte, num_entries, cfg); - } - - pte = arm_v7s_install_table(tablep, ptep, blk_pte, cfg); - if (pte != blk_pte) { - __arm_v7s_free_table(tablep, 2, data); - - if (!ARM_V7S_PTE_IS_TABLE(pte, 1)) - return 0; - - tablep = iopte_deref(pte, 1, data); - return __arm_v7s_unmap(data, gather, iova, size, 2, tablep); - } - - io_pgtable_tlb_add_page(&data->iop, gather, iova, size); - return size; -} - static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *data, struct iommu_iotlb_gather *gather, unsigned long iova, size_t size, int lvl, @@ -694,11 +586,8 @@ static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *data, * case in a lock for the sake of correctness and be done with it. */ if (num_entries <= 1 && arm_v7s_pte_is_cont(pte[0], lvl)) { - unsigned long flags; - - spin_lock_irqsave(&data->split_lock, flags); - pte[0] = arm_v7s_split_cont(data, iova, idx, lvl, ptep); - spin_unlock_irqrestore(&data->split_lock, flags); + WARN_ONCE(true, "Unmap of a partial large IOPTE is not allowed"); + return 0; } /* If the size matches this level, we're in the right place */ @@ -721,12 +610,8 @@ static size_t __arm_v7s_unmap(struct arm_v7s_io_pgtable *data, } return size; } else if (lvl == 1 && !ARM_V7S_PTE_IS_TABLE(pte[0], lvl)) { - /* - * Insert a table at the next level to map the old region, - * minus the part we want to unmap - */ - return arm_v7s_split_blk_unmap(data, gather, iova, size, pte[0], - ptep); + WARN_ONCE(true, "Unmap of a partial large IOPTE is not allowed"); + return 0; } /* Keep on walkin' */ @@ -811,8 +696,6 @@ static struct io_pgtable *arm_v7s_alloc_pgtable(struct io_pgtable_cfg *cfg, if (!data) return NULL; - spin_lock_init(&data->split_lock); - /* * ARM_MTK_TTBR_EXT extend the translation table base support larger * memory address.
A minority of page table implementations (arm_lpae, armv7) are unique in how they handle partial unmap of large IOPTEs. Other implementations will unmap the large IOPTE and return it's length. For example if a 2M IOPTE is present and the first 4K is requested to be unmapped then unmap will remove the whole 2M and report 2M as the result. armv7 instead will break up contiguous entries and replace an entry with a whole table so it can unmap the requested 4k. This seems copied from the arm_lpae implementation, which was analyzed here: https://lore.kernel.org/all/20241024134411.GA6956@nvidia.com/ Bring consistency to the implementations and remove this unused functionality. There are no uses outside iommu, this effects the ARM_V7S drivers msm_iommu, mtk_iommu, and arm-smmmu. Signed-off-by: Jason Gunthorpe <jgg@nvidia.com> --- drivers/iommu/io-pgtable-arm-v7s.c | 125 +---------------------------- 1 file changed, 4 insertions(+), 121 deletions(-)