Message ID | 20210422013058.6305-9-Felix.Kuehling@amd.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Implement multi-GPU DMA mappings for KFD | expand |
Regards,
Oak
On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote:
Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs.
Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com>
---
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2 +
.../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 76 ++++++++++++++++++-
2 files changed, 77 insertions(+), 1 deletion(-)
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
index 63668433f5a6..b706e5a54782 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
@@ -41,6 +41,7 @@ struct amdgpu_device;
enum kfd_mem_attachment_type {
KFD_MEM_ATT_SHARED, /* Share kgd_mem->bo or another attachment's */
KFD_MEM_ATT_USERPTR, /* SG bo to DMA map pages from a userptr bo */
+ KFD_MEM_ATT_DMABUF, /* DMAbuf to DMA map TTM BOs */
};
struct kfd_mem_attachment {
@@ -56,6 +57,7 @@ struct kfd_mem_attachment {
struct kgd_mem {
struct mutex lock;
struct amdgpu_bo *bo;
+ struct dma_buf *dmabuf;
struct list_head attachments;
/* protected by amdkfd_process_info.lock */
struct ttm_validate_buffer validate_list;
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
index 9eeedd0c7920..18a1f9222a59 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
@@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem,
return ret;
}
+static int
+kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment)
+{
+ struct ttm_operation_ctx ctx = {.interruptible = true};
+ struct amdgpu_bo *bo = attachment->bo_va->base.bo;
+
+ amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT);
+ return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation
+}
+
static int
kfd_mem_dmamap_attachment(struct kgd_mem *mem,
struct kfd_mem_attachment *attachment)
@@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem,
return 0;
case KFD_MEM_ATT_USERPTR:
return kfd_mem_dmamap_userptr(mem, attachment);
+ case KFD_MEM_ATT_DMABUF:
+ return kfd_mem_dmamap_dmabuf(attachment);
default:
WARN_ON_ONCE(1);
}
@@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
ttm->sg = NULL;
}
+static void
+kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment)
+{
+ struct ttm_operation_ctx ctx = {.interruptible = true};
+ struct amdgpu_bo *bo = attachment->bo_va->base.bo;
+
+ amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU);
+ ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
+ /* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is
+ * called
+ */
+}
+
static void
kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
struct kfd_mem_attachment *attachment)
@@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
case KFD_MEM_ATT_USERPTR:
kfd_mem_dmaunmap_userptr(mem, attachment);
break;
+ case KFD_MEM_ATT_DMABUF:
+ kfd_mem_dmaunmap_dmabuf(attachment);
+ break;
default:
WARN_ON_ONCE(1);
}
@@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
return 0;
}
+static int
+kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
+ struct amdgpu_bo **bo)
+{
+ struct drm_gem_object *gobj;
+
+ if (!mem->dmabuf) {
+ mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base,
+ mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
+ DRM_RDWR : 0);
+ if (IS_ERR(mem->dmabuf)) {
+ mem->dmabuf = NULL;
+ return PTR_ERR(mem->dmabuf);
+ }
+ }
+
+ gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf);
+ if (IS_ERR(gobj))
+ return PTR_ERR(gobj);
+
+ /* Import takes an extra reference on the dmabuf. Drop it now to
+ * avoid leaking it. We only need the one reference in
+ * kgd_mem->dmabuf.
+ */
+ dma_buf_put(mem->dmabuf);
+
+ *bo = gem_to_amdgpu_bo(gobj);
+ (*bo)->parent = amdgpu_bo_ref(mem->bo);
+
+ return 0;
+}
+
/* kfd_mem_attach - Add a BO to a VM
*
* Everything that needs to bo done only once when a BO is first added
@@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
if (ret)
goto unwind;
+ } else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT &&
+ mem->bo->tbo.type != ttm_bo_type_sg) {
+ /* GTT BOs use DMA-mapping ability of dynamic-attach
+ * DMA bufs. TODO: The same should work for VRAM on
+ * large-BAR GPUs.
+ */
+ attachment[i]->type = KFD_MEM_ATT_DMABUF;
+ ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
+ if (ret)
+ goto unwind;
} else {
- /* FIXME: Need to DMA-map other BO types */
+ /* FIXME: Need to DMA-map other BO types:
+ * large-BAR VRAM, doorbells, MMIO remap
+ */
attachment[i]->type = KFD_MEM_ATT_SHARED;
bo[i] = mem->bo;
drm_gem_object_get(&bo[i]->tbo.base);
@@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(
/* Free the BO*/
drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
+ if (mem->dmabuf)
+ dma_buf_put(mem->dmabuf);
drm_gem_object_put(&mem->bo->tbo.base);
mutex_destroy(&mem->lock);
kfree(mem);
--
2.31.1
_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.freedesktop.org%2Fmailman%2Flistinfo%2Famd-gfx&data=04%7C01%7Coak.zeng%40amd.com%7Ca14e8f1d4ba6450b5f1308d9052e630b%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637546519053906547%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=AD%2FnbE6kQDFHLTr0kbzZ2sl3jKwuOqUKfpPEcPHwwfY%3D&reserved=0
Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak: > Regards, > Oak > > > > On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote: > > Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs. > > Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2 + > .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 76 ++++++++++++++++++- > 2 files changed, 77 insertions(+), 1 deletion(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > index 63668433f5a6..b706e5a54782 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > @@ -41,6 +41,7 @@ struct amdgpu_device; > enum kfd_mem_attachment_type { > KFD_MEM_ATT_SHARED, /* Share kgd_mem->bo or another attachment's */ > KFD_MEM_ATT_USERPTR, /* SG bo to DMA map pages from a userptr bo */ > + KFD_MEM_ATT_DMABUF, /* DMAbuf to DMA map TTM BOs */ > }; > > struct kfd_mem_attachment { > @@ -56,6 +57,7 @@ struct kfd_mem_attachment { > struct kgd_mem { > struct mutex lock; > struct amdgpu_bo *bo; > + struct dma_buf *dmabuf; > struct list_head attachments; > /* protected by amdkfd_process_info.lock */ > struct ttm_validate_buffer validate_list; > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > index 9eeedd0c7920..18a1f9222a59 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem, > return ret; > } > > +static int > +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment) > +{ > + struct ttm_operation_ctx ctx = {.interruptible = true}; > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > + > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT); > + return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation Conceptually, calling ttm_bo_validate ensures that the BO is in the specified domain, in this case GTT. Before calling validate, it can be in the CPU domain, which means it may be swapped to disk so it's not GPU accessible. For a DMABuf attachment, the CPU domain means, that the DMABuf is not attached because the underlying memory object may be on the move or swapped out. The actual implementation of the dmabuf attachment is currently in amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this series fixes that to move the actual dmabuf attachment into amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when a BO is moved between the CPU and GTT domains. Regards, Felix > +} > + > static int > kfd_mem_dmamap_attachment(struct kgd_mem *mem, > struct kfd_mem_attachment *attachment) > @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem, > return 0; > case KFD_MEM_ATT_USERPTR: > return kfd_mem_dmamap_userptr(mem, attachment); > + case KFD_MEM_ATT_DMABUF: > + return kfd_mem_dmamap_dmabuf(attachment); > default: > WARN_ON_ONCE(1); > } > @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem, > ttm->sg = NULL; > } > > +static void > +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment) > +{ > + struct ttm_operation_ctx ctx = {.interruptible = true}; > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > + > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU); > + ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > + /* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is > + * called > + */ > +} > + > static void > kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > struct kfd_mem_attachment *attachment) > @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > case KFD_MEM_ATT_USERPTR: > kfd_mem_dmaunmap_userptr(mem, attachment); > break; > + case KFD_MEM_ATT_DMABUF: > + kfd_mem_dmaunmap_dmabuf(attachment); > + break; > default: > WARN_ON_ONCE(1); > } > @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem, > return 0; > } > > +static int > +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem, > + struct amdgpu_bo **bo) > +{ > + struct drm_gem_object *gobj; > + > + if (!mem->dmabuf) { > + mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base, > + mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? > + DRM_RDWR : 0); > + if (IS_ERR(mem->dmabuf)) { > + mem->dmabuf = NULL; > + return PTR_ERR(mem->dmabuf); > + } > + } > + > + gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf); > + if (IS_ERR(gobj)) > + return PTR_ERR(gobj); > + > + /* Import takes an extra reference on the dmabuf. Drop it now to > + * avoid leaking it. We only need the one reference in > + * kgd_mem->dmabuf. > + */ > + dma_buf_put(mem->dmabuf); > + > + *bo = gem_to_amdgpu_bo(gobj); > + (*bo)->parent = amdgpu_bo_ref(mem->bo); > + > + return 0; > +} > + > /* kfd_mem_attach - Add a BO to a VM > * > * Everything that needs to bo done only once when a BO is first added > @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem, > ret = kfd_mem_attach_userptr(adev, mem, &bo[i]); > if (ret) > goto unwind; > + } else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT && > + mem->bo->tbo.type != ttm_bo_type_sg) { > + /* GTT BOs use DMA-mapping ability of dynamic-attach > + * DMA bufs. TODO: The same should work for VRAM on > + * large-BAR GPUs. > + */ > + attachment[i]->type = KFD_MEM_ATT_DMABUF; > + ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]); > + if (ret) > + goto unwind; > } else { > - /* FIXME: Need to DMA-map other BO types */ > + /* FIXME: Need to DMA-map other BO types: > + * large-BAR VRAM, doorbells, MMIO remap > + */ > attachment[i]->type = KFD_MEM_ATT_SHARED; > bo[i] = mem->bo; > drm_gem_object_get(&bo[i]->tbo.base); > @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu( > > /* Free the BO*/ > drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv); > + if (mem->dmabuf) > + dma_buf_put(mem->dmabuf); > drm_gem_object_put(&mem->bo->tbo.base); > mutex_destroy(&mem->lock); > kfree(mem); > -- > 2.31.1 > > _______________________________________________ > amd-gfx mailing list > amd-gfx@lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/amd-gfx >
Regards, Oak On 2021-04-26, 11:56 PM, "Kuehling, Felix" <Felix.Kuehling@amd.com> wrote: Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak: > Regards, > Oak > > > > On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote: > > Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs. > > Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2 + > .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 76 ++++++++++++++++++- > 2 files changed, 77 insertions(+), 1 deletion(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > index 63668433f5a6..b706e5a54782 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > @@ -41,6 +41,7 @@ struct amdgpu_device; > enum kfd_mem_attachment_type { > KFD_MEM_ATT_SHARED, /* Share kgd_mem->bo or another attachment's */ > KFD_MEM_ATT_USERPTR, /* SG bo to DMA map pages from a userptr bo */ > + KFD_MEM_ATT_DMABUF, /* DMAbuf to DMA map TTM BOs */ > }; > > struct kfd_mem_attachment { > @@ -56,6 +57,7 @@ struct kfd_mem_attachment { > struct kgd_mem { > struct mutex lock; > struct amdgpu_bo *bo; > + struct dma_buf *dmabuf; > struct list_head attachments; > /* protected by amdkfd_process_info.lock */ > struct ttm_validate_buffer validate_list; > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > index 9eeedd0c7920..18a1f9222a59 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem, > return ret; > } > > +static int > +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment) > +{ > + struct ttm_operation_ctx ctx = {.interruptible = true}; > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > + > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT); > + return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation Conceptually, calling ttm_bo_validate ensures that the BO is in the specified domain, in this case GTT. Before calling validate, it can be in the CPU domain, which means it may be swapped to disk so it's not GPU accessible. For a DMABuf attachment, the CPU domain means, that the DMABuf is not attached because the underlying memory object may be on the move or swapped out. The actual implementation of the dmabuf attachment is currently in amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this series fixes that to move the actual dmabuf attachment into amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when a BO is moved between the CPU and GTT domains. Thanks for the explanation. One more thing I don't quite understand: before this series, GTT memory should already has been validated somewhere before GTT memory is mapped to GPU. You added GTT memory validation here - will this validation be duplicated? The function naming kfd_mem_dmamap_dmabuf is still confusing since it seems to me it is only some preparation work before dynamically dma-map a GTT memory. But I understand from this series' perspective, compared to usrptr (where you actually do the dma-mapping in function kfd_mem_dmamap_usrptr), for gtt memory you leveraged the amdgpu ttm function of dynamic dma-mapping. So maybe the naming here makes sense from that perspective. Another thing related but not directly to this series: for GTT memory, it is dma-mapped when it is allocated. See function ttm_populate_and_map_pages calling dma_map_page. The question is, will gtt be first dma-unmapping before it is mapped in amdgpu_ttm_backend_bind? It is existing work, not from your series. Maybe there is not issue but I just want to make sure while we are looking at this area. Regards, Felix > +} > + > static int > kfd_mem_dmamap_attachment(struct kgd_mem *mem, > struct kfd_mem_attachment *attachment) > @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem, > return 0; > case KFD_MEM_ATT_USERPTR: > return kfd_mem_dmamap_userptr(mem, attachment); > + case KFD_MEM_ATT_DMABUF: > + return kfd_mem_dmamap_dmabuf(attachment); > default: > WARN_ON_ONCE(1); > } > @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem, > ttm->sg = NULL; > } > > +static void > +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment) > +{ > + struct ttm_operation_ctx ctx = {.interruptible = true}; > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > + > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU); > + ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > + /* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is > + * called > + */ > +} > + > static void > kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > struct kfd_mem_attachment *attachment) > @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > case KFD_MEM_ATT_USERPTR: > kfd_mem_dmaunmap_userptr(mem, attachment); > break; > + case KFD_MEM_ATT_DMABUF: > + kfd_mem_dmaunmap_dmabuf(attachment); > + break; > default: > WARN_ON_ONCE(1); > } > @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem, > return 0; > } > > +static int > +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem, > + struct amdgpu_bo **bo) > +{ > + struct drm_gem_object *gobj; > + > + if (!mem->dmabuf) { > + mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base, > + mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? > + DRM_RDWR : 0); > + if (IS_ERR(mem->dmabuf)) { > + mem->dmabuf = NULL; > + return PTR_ERR(mem->dmabuf); > + } > + } > + > + gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf); > + if (IS_ERR(gobj)) > + return PTR_ERR(gobj); > + > + /* Import takes an extra reference on the dmabuf. Drop it now to > + * avoid leaking it. We only need the one reference in > + * kgd_mem->dmabuf. > + */ > + dma_buf_put(mem->dmabuf); > + > + *bo = gem_to_amdgpu_bo(gobj); > + (*bo)->parent = amdgpu_bo_ref(mem->bo); > + > + return 0; > +} > + > /* kfd_mem_attach - Add a BO to a VM > * > * Everything that needs to bo done only once when a BO is first added > @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem, > ret = kfd_mem_attach_userptr(adev, mem, &bo[i]); > if (ret) > goto unwind; > + } else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT && > + mem->bo->tbo.type != ttm_bo_type_sg) { > + /* GTT BOs use DMA-mapping ability of dynamic-attach > + * DMA bufs. TODO: The same should work for VRAM on > + * large-BAR GPUs. > + */ > + attachment[i]->type = KFD_MEM_ATT_DMABUF; > + ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]); > + if (ret) > + goto unwind; > } else { > - /* FIXME: Need to DMA-map other BO types */ > + /* FIXME: Need to DMA-map other BO types: > + * large-BAR VRAM, doorbells, MMIO remap > + */ > attachment[i]->type = KFD_MEM_ATT_SHARED; > bo[i] = mem->bo; > drm_gem_object_get(&bo[i]->tbo.base); > @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu( > > /* Free the BO*/ > drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv); > + if (mem->dmabuf) > + dma_buf_put(mem->dmabuf); > drm_gem_object_put(&mem->bo->tbo.base); > mutex_destroy(&mem->lock); > kfree(mem); > -- > 2.31.1 > > _______________________________________________ > amd-gfx mailing list > amd-gfx@lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/amd-gfx >
Am 2021-04-27 um 10:29 a.m. schrieb Zeng, Oak: > Regards, > Oak > > > > On 2021-04-26, 11:56 PM, "Kuehling, Felix" <Felix.Kuehling@amd.com> wrote: > > Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak: > > Regards, > > Oak > > > > > > > > On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote: > > > > Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs. > > > > Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> > > --- > > drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2 + > > .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 76 ++++++++++++++++++- > > 2 files changed, 77 insertions(+), 1 deletion(-) > > > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > > index 63668433f5a6..b706e5a54782 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > > @@ -41,6 +41,7 @@ struct amdgpu_device; > > enum kfd_mem_attachment_type { > > KFD_MEM_ATT_SHARED, /* Share kgd_mem->bo or another attachment's */ > > KFD_MEM_ATT_USERPTR, /* SG bo to DMA map pages from a userptr bo */ > > + KFD_MEM_ATT_DMABUF, /* DMAbuf to DMA map TTM BOs */ > > }; > > > > struct kfd_mem_attachment { > > @@ -56,6 +57,7 @@ struct kfd_mem_attachment { > > struct kgd_mem { > > struct mutex lock; > > struct amdgpu_bo *bo; > > + struct dma_buf *dmabuf; > > struct list_head attachments; > > /* protected by amdkfd_process_info.lock */ > > struct ttm_validate_buffer validate_list; > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > > index 9eeedd0c7920..18a1f9222a59 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > > @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem, > > return ret; > > } > > > > +static int > > +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment) > > +{ > > + struct ttm_operation_ctx ctx = {.interruptible = true}; > > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > > + > > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT); > > + return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > > How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation > > Conceptually, calling ttm_bo_validate ensures that the BO is in the > specified domain, in this case GTT. Before calling validate, it can be > in the CPU domain, which means it may be swapped to disk so it's not GPU > accessible. For a DMABuf attachment, the CPU domain means, that the > DMABuf is not attached because the underlying memory object may be on > the move or swapped out. > > The actual implementation of the dmabuf attachment is currently in > amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this > series fixes that to move the actual dmabuf attachment into > amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when > a BO is moved between the CPU and GTT domains. > > Thanks for the explanation. One more thing I don't quite understand: before this series, GTT memory should already has been validated somewhere before GTT memory is mapped to GPU. You added GTT memory validation here - will this validation be duplicated? When you have N GPUs there are now N BOs involved. Each GPU needs its own BO because it needs its own DMA mapping. There will be one actual GTT BO that allocates physical pages in TTM. The other BOs are dmabuf imports that DMA-map the same physical pages for access by the other GPUs. The validate call here validates one of the dmabuf imports. This does not duplicate the validation of the underlying TTM BO with the actual physical memory allocation. > > The function naming kfd_mem_dmamap_dmabuf is still confusing since it seems to me it is only some preparation work before dynamically dma-map a GTT memory. No, this series is not just preparation. It implements DMA mapping of BOs for multiple GPUs. TTM already handles DMA mapping of the memory for the device where the memory was allocated. (Yes, even GTT memory is associated with a specific GPU even though it's physically in system memory). What this patch series adds, is additional DMA mappings for the other GPUs. Without this patch, we were using the DMA mapping for GPU-1 in the page table of GPU-X, which is incorrect. It works in many cases where the DMA mapping is a direct mapping: * IOMMU disabled * IOMMU in passthrough mode But it breaks when you have multiple GPUs with an IOMMU that's not disabled or in passthrough mode. > But I understand from this series' perspective, compared to usrptr (where you actually do the dma-mapping in function kfd_mem_dmamap_usrptr), for gtt memory you leveraged the amdgpu ttm function of dynamic dma-mapping. So maybe the naming here makes sense from that perspective. Yes. > > Another thing related but not directly to this series: for GTT memory, it is dma-mapped when it is allocated. See function ttm_populate_and_map_pages calling dma_map_page. The question is, will gtt be first dma-unmapping before it is mapped in amdgpu_ttm_backend_bind? It is existing work, not from your series. Maybe there is not issue but I just want to make sure while we are looking at this area. Right. The problem is, that the DMA mappings only work for a specific device. Using the same DMA mapping on multiple devices is broken. The reason we got away with it for a long time is, that we were running with IOMMU disabled or in passthrough mode. Regards, Felix > > Regards, > Felix > > > > +} > > + > > static int > > kfd_mem_dmamap_attachment(struct kgd_mem *mem, > > struct kfd_mem_attachment *attachment) > > @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem, > > return 0; > > case KFD_MEM_ATT_USERPTR: > > return kfd_mem_dmamap_userptr(mem, attachment); > > + case KFD_MEM_ATT_DMABUF: > > + return kfd_mem_dmamap_dmabuf(attachment); > > default: > > WARN_ON_ONCE(1); > > } > > @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem, > > ttm->sg = NULL; > > } > > > > +static void > > +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment) > > +{ > > + struct ttm_operation_ctx ctx = {.interruptible = true}; > > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > > + > > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU); > > + ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > > + /* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is > > + * called > > + */ > > +} > > + > > static void > > kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > > struct kfd_mem_attachment *attachment) > > @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > > case KFD_MEM_ATT_USERPTR: > > kfd_mem_dmaunmap_userptr(mem, attachment); > > break; > > + case KFD_MEM_ATT_DMABUF: > > + kfd_mem_dmaunmap_dmabuf(attachment); > > + break; > > default: > > WARN_ON_ONCE(1); > > } > > @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem, > > return 0; > > } > > > > +static int > > +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem, > > + struct amdgpu_bo **bo) > > +{ > > + struct drm_gem_object *gobj; > > + > > + if (!mem->dmabuf) { > > + mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base, > > + mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? > > + DRM_RDWR : 0); > > + if (IS_ERR(mem->dmabuf)) { > > + mem->dmabuf = NULL; > > + return PTR_ERR(mem->dmabuf); > > + } > > + } > > + > > + gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf); > > + if (IS_ERR(gobj)) > > + return PTR_ERR(gobj); > > + > > + /* Import takes an extra reference on the dmabuf. Drop it now to > > + * avoid leaking it. We only need the one reference in > > + * kgd_mem->dmabuf. > > + */ > > + dma_buf_put(mem->dmabuf); > > + > > + *bo = gem_to_amdgpu_bo(gobj); > > + (*bo)->parent = amdgpu_bo_ref(mem->bo); > > + > > + return 0; > > +} > > + > > /* kfd_mem_attach - Add a BO to a VM > > * > > * Everything that needs to bo done only once when a BO is first added > > @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem, > > ret = kfd_mem_attach_userptr(adev, mem, &bo[i]); > > if (ret) > > goto unwind; > > + } else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT && > > + mem->bo->tbo.type != ttm_bo_type_sg) { > > + /* GTT BOs use DMA-mapping ability of dynamic-attach > > + * DMA bufs. TODO: The same should work for VRAM on > > + * large-BAR GPUs. > > + */ > > + attachment[i]->type = KFD_MEM_ATT_DMABUF; > > + ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]); > > + if (ret) > > + goto unwind; > > } else { > > - /* FIXME: Need to DMA-map other BO types */ > > + /* FIXME: Need to DMA-map other BO types: > > + * large-BAR VRAM, doorbells, MMIO remap > > + */ > > attachment[i]->type = KFD_MEM_ATT_SHARED; > > bo[i] = mem->bo; > > drm_gem_object_get(&bo[i]->tbo.base); > > @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu( > > > > /* Free the BO*/ > > drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv); > > + if (mem->dmabuf) > > + dma_buf_put(mem->dmabuf); > > drm_gem_object_put(&mem->bo->tbo.base); > > mutex_destroy(&mem->lock); > > kfree(mem); > > -- > > 2.31.1 > > > > _______________________________________________ > > amd-gfx mailing list > > amd-gfx@lists.freedesktop.org > > https://lists.freedesktop.org/mailman/listinfo/amd-gfx > > >
[AMD Official Use Only - Internal Distribution Only] Acked-by: Ramesh Errabolu <ramesh.errabolu@amd.com> -----Original Message----- From: amd-gfx <amd-gfx-bounces@lists.freedesktop.org> On Behalf Of Kuehling, Felix Sent: Tuesday, April 27, 2021 10:09 AM To: Zeng, Oak <Oak.Zeng@amd.com>; amd-gfx@lists.freedesktop.org; dri-devel@lists.freedesktop.org Subject: Re: [PATCH v2 08/10] drm/amdgpu: Add DMA mapping of GTT BOs Am 2021-04-27 um 10:29 a.m. schrieb Zeng, Oak: > Regards, > Oak > > > > On 2021-04-26, 11:56 PM, "Kuehling, Felix" <Felix.Kuehling@amd.com> wrote: > > Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak: > > Regards, > > Oak > > > > > > > > On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote: > > > > Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs. > > > > Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> > > --- > > drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2 + > > .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 76 ++++++++++++++++++- > > 2 files changed, 77 insertions(+), 1 deletion(-) > > > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > > index 63668433f5a6..b706e5a54782 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h > > @@ -41,6 +41,7 @@ struct amdgpu_device; > > enum kfd_mem_attachment_type { > > KFD_MEM_ATT_SHARED, /* Share kgd_mem->bo or another attachment's */ > > KFD_MEM_ATT_USERPTR, /* SG bo to DMA map pages from a userptr bo */ > > + KFD_MEM_ATT_DMABUF, /* DMAbuf to DMA map TTM BOs */ > > }; > > > > struct kfd_mem_attachment { > > @@ -56,6 +57,7 @@ struct kfd_mem_attachment { > > struct kgd_mem { > > struct mutex lock; > > struct amdgpu_bo *bo; > > + struct dma_buf *dmabuf; > > struct list_head attachments; > > /* protected by amdkfd_process_info.lock */ > > struct ttm_validate_buffer validate_list; > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > > index 9eeedd0c7920..18a1f9222a59 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c > > @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem, > > return ret; > > } > > > > +static int > > +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment) > > +{ > > + struct ttm_operation_ctx ctx = {.interruptible = true}; > > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > > + > > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT); > > + return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > > How does this work? The function name says this is dma mapping a > buffer but from the implementation, it is just a placement and > validation > > Conceptually, calling ttm_bo_validate ensures that the BO is in the > specified domain, in this case GTT. Before calling validate, it can be > in the CPU domain, which means it may be swapped to disk so it's not GPU > accessible. For a DMABuf attachment, the CPU domain means, that the > DMABuf is not attached because the underlying memory object may be on > the move or swapped out. > > The actual implementation of the dmabuf attachment is currently in > amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this > series fixes that to move the actual dmabuf attachment into > amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when > a BO is moved between the CPU and GTT domains. > > Thanks for the explanation. One more thing I don't quite understand: before this series, GTT memory should already has been validated somewhere before GTT memory is mapped to GPU. You added GTT memory validation here - will this validation be duplicated? When you have N GPUs there are now N BOs involved. Each GPU needs its own BO because it needs its own DMA mapping. There will be one actual GTT BO that allocates physical pages in TTM. The other BOs are dmabuf imports that DMA-map the same physical pages for access by the other GPUs. The validate call here validates one of the dmabuf imports. This does not duplicate the validation of the underlying TTM BO with the actual physical memory allocation. > > The function naming kfd_mem_dmamap_dmabuf is still confusing since it seems to me it is only some preparation work before dynamically dma-map a GTT memory. No, this series is not just preparation. It implements DMA mapping of BOs for multiple GPUs. TTM already handles DMA mapping of the memory for the device where the memory was allocated. (Yes, even GTT memory is associated with a specific GPU even though it's physically in system memory). What this patch series adds, is additional DMA mappings for the other GPUs. Without this patch, we were using the DMA mapping for GPU-1 in the page table of GPU-X, which is incorrect. It works in many cases where the DMA mapping is a direct mapping: * IOMMU disabled * IOMMU in passthrough mode But it breaks when you have multiple GPUs with an IOMMU that's not disabled or in passthrough mode. > But I understand from this series' perspective, compared to usrptr (where you actually do the dma-mapping in function kfd_mem_dmamap_usrptr), for gtt memory you leveraged the amdgpu ttm function of dynamic dma-mapping. So maybe the naming here makes sense from that perspective. Yes. > > Another thing related but not directly to this series: for GTT memory, it is dma-mapped when it is allocated. See function ttm_populate_and_map_pages calling dma_map_page. The question is, will gtt be first dma-unmapping before it is mapped in amdgpu_ttm_backend_bind? It is existing work, not from your series. Maybe there is not issue but I just want to make sure while we are looking at this area. Right. The problem is, that the DMA mappings only work for a specific device. Using the same DMA mapping on multiple devices is broken. The reason we got away with it for a long time is, that we were running with IOMMU disabled or in passthrough mode. Regards, Felix > > Regards, > Felix > > > > +} > > + > > static int > > kfd_mem_dmamap_attachment(struct kgd_mem *mem, > > struct kfd_mem_attachment *attachment) > > @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem, > > return 0; > > case KFD_MEM_ATT_USERPTR: > > return kfd_mem_dmamap_userptr(mem, attachment); > > + case KFD_MEM_ATT_DMABUF: > > + return kfd_mem_dmamap_dmabuf(attachment); > > default: > > WARN_ON_ONCE(1); > > } > > @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem, > > ttm->sg = NULL; > > } > > > > +static void > > +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment) > > +{ > > + struct ttm_operation_ctx ctx = {.interruptible = true}; > > + struct amdgpu_bo *bo = attachment->bo_va->base.bo; > > + > > + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU); > > + ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); > > + /* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is > > + * called > > + */ > > +} > > + > > static void > > kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > > struct kfd_mem_attachment *attachment) > > @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, > > case KFD_MEM_ATT_USERPTR: > > kfd_mem_dmaunmap_userptr(mem, attachment); > > break; > > + case KFD_MEM_ATT_DMABUF: > > + kfd_mem_dmaunmap_dmabuf(attachment); > > + break; > > default: > > WARN_ON_ONCE(1); > > } > > @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem, > > return 0; > > } > > > > +static int > > +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem, > > + struct amdgpu_bo **bo) > > +{ > > + struct drm_gem_object *gobj; > > + > > + if (!mem->dmabuf) { > > + mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base, > > + mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? > > + DRM_RDWR : 0); > > + if (IS_ERR(mem->dmabuf)) { > > + mem->dmabuf = NULL; > > + return PTR_ERR(mem->dmabuf); > > + } > > + } > > + > > + gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf); > > + if (IS_ERR(gobj)) > > + return PTR_ERR(gobj); > > + > > + /* Import takes an extra reference on the dmabuf. Drop it now to > > + * avoid leaking it. We only need the one reference in > > + * kgd_mem->dmabuf. > > + */ > > + dma_buf_put(mem->dmabuf); > > + > > + *bo = gem_to_amdgpu_bo(gobj); > > + (*bo)->parent = amdgpu_bo_ref(mem->bo); > > + > > + return 0; > > +} > > + > > /* kfd_mem_attach - Add a BO to a VM > > * > > * Everything that needs to bo done only once when a BO is first added > > @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem, > > ret = kfd_mem_attach_userptr(adev, mem, &bo[i]); > > if (ret) > > goto unwind; > > + } else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT && > > + mem->bo->tbo.type != ttm_bo_type_sg) { > > + /* GTT BOs use DMA-mapping ability of dynamic-attach > > + * DMA bufs. TODO: The same should work for VRAM on > > + * large-BAR GPUs. > > + */ > > + attachment[i]->type = KFD_MEM_ATT_DMABUF; > > + ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]); > > + if (ret) > > + goto unwind; > > } else { > > - /* FIXME: Need to DMA-map other BO types */ > > + /* FIXME: Need to DMA-map other BO types: > > + * large-BAR VRAM, doorbells, MMIO remap > > + */ > > attachment[i]->type = KFD_MEM_ATT_SHARED; > > bo[i] = mem->bo; > > drm_gem_object_get(&bo[i]->tbo.base); > > @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu( > > > > /* Free the BO*/ > > drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv); > > + if (mem->dmabuf) > > + dma_buf_put(mem->dmabuf); > > drm_gem_object_put(&mem->bo->tbo.base); > > mutex_destroy(&mem->lock); > > kfree(mem); > > -- > > 2.31.1 > > > > _______________________________________________ > > amd-gfx mailing list > > amd-gfx@lists.freedesktop.org > > https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.freedesktop.org%2Fmailman%2Flistinfo%2Famd-gfx&data=04%7C01%7Cphilip.yang%40amd.com%7Cf3c9c824ef6447cbe26808d9098e6606%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637551329471854747%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&sdata=P7SYxlhPoYfyGCNaiiCA%2BaqJS%2BxvGZEMIZkuvqCpCLI%3D&reserved=0 > > >
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h index 63668433f5a6..b706e5a54782 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h @@ -41,6 +41,7 @@ struct amdgpu_device; enum kfd_mem_attachment_type { KFD_MEM_ATT_SHARED, /* Share kgd_mem->bo or another attachment's */ KFD_MEM_ATT_USERPTR, /* SG bo to DMA map pages from a userptr bo */ + KFD_MEM_ATT_DMABUF, /* DMAbuf to DMA map TTM BOs */ }; struct kfd_mem_attachment { @@ -56,6 +57,7 @@ struct kfd_mem_attachment { struct kgd_mem { struct mutex lock; struct amdgpu_bo *bo; + struct dma_buf *dmabuf; struct list_head attachments; /* protected by amdkfd_process_info.lock */ struct ttm_validate_buffer validate_list; diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c index 9eeedd0c7920..18a1f9222a59 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem, return ret; } +static int +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment) +{ + struct ttm_operation_ctx ctx = {.interruptible = true}; + struct amdgpu_bo *bo = attachment->bo_va->base.bo; + + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT); + return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); +} + static int kfd_mem_dmamap_attachment(struct kgd_mem *mem, struct kfd_mem_attachment *attachment) @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem, return 0; case KFD_MEM_ATT_USERPTR: return kfd_mem_dmamap_userptr(mem, attachment); + case KFD_MEM_ATT_DMABUF: + return kfd_mem_dmamap_dmabuf(attachment); default: WARN_ON_ONCE(1); } @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem, ttm->sg = NULL; } +static void +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment) +{ + struct ttm_operation_ctx ctx = {.interruptible = true}; + struct amdgpu_bo *bo = attachment->bo_va->base.bo; + + amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU); + ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); + /* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is + * called + */ +} + static void kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, struct kfd_mem_attachment *attachment) @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem, case KFD_MEM_ATT_USERPTR: kfd_mem_dmaunmap_userptr(mem, attachment); break; + case KFD_MEM_ATT_DMABUF: + kfd_mem_dmaunmap_dmabuf(attachment); + break; default: WARN_ON_ONCE(1); } @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem, return 0; } +static int +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem, + struct amdgpu_bo **bo) +{ + struct drm_gem_object *gobj; + + if (!mem->dmabuf) { + mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base, + mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ? + DRM_RDWR : 0); + if (IS_ERR(mem->dmabuf)) { + mem->dmabuf = NULL; + return PTR_ERR(mem->dmabuf); + } + } + + gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf); + if (IS_ERR(gobj)) + return PTR_ERR(gobj); + + /* Import takes an extra reference on the dmabuf. Drop it now to + * avoid leaking it. We only need the one reference in + * kgd_mem->dmabuf. + */ + dma_buf_put(mem->dmabuf); + + *bo = gem_to_amdgpu_bo(gobj); + (*bo)->parent = amdgpu_bo_ref(mem->bo); + + return 0; +} + /* kfd_mem_attach - Add a BO to a VM * * Everything that needs to bo done only once when a BO is first added @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem, ret = kfd_mem_attach_userptr(adev, mem, &bo[i]); if (ret) goto unwind; + } else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT && + mem->bo->tbo.type != ttm_bo_type_sg) { + /* GTT BOs use DMA-mapping ability of dynamic-attach + * DMA bufs. TODO: The same should work for VRAM on + * large-BAR GPUs. + */ + attachment[i]->type = KFD_MEM_ATT_DMABUF; + ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]); + if (ret) + goto unwind; } else { - /* FIXME: Need to DMA-map other BO types */ + /* FIXME: Need to DMA-map other BO types: + * large-BAR VRAM, doorbells, MMIO remap + */ attachment[i]->type = KFD_MEM_ATT_SHARED; bo[i] = mem->bo; drm_gem_object_get(&bo[i]->tbo.base); @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu( /* Free the BO*/ drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv); + if (mem->dmabuf) + dma_buf_put(mem->dmabuf); drm_gem_object_put(&mem->bo->tbo.base); mutex_destroy(&mem->lock); kfree(mem);
Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs. Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> --- drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2 + .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 76 ++++++++++++++++++- 2 files changed, 77 insertions(+), 1 deletion(-)