diff mbox series

[v2,08/10] drm/amdgpu: Add DMA mapping of GTT BOs

Message ID 20210422013058.6305-9-Felix.Kuehling@amd.com (mailing list archive)
State New, archived
Headers show
Series Implement multi-GPU DMA mappings for KFD | expand

Commit Message

Felix Kuehling April 22, 2021, 1:30 a.m. UTC
Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs.

Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h    |  2 +
 .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c  | 76 ++++++++++++++++++-
 2 files changed, 77 insertions(+), 1 deletion(-)

Comments

Zeng, Oak April 27, 2021, 12:35 a.m. UTC | #1
Regards,
Oak 

 

On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote:

    Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs.

    Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com>
    ---
     drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h    |  2 +
     .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c  | 76 ++++++++++++++++++-
     2 files changed, 77 insertions(+), 1 deletion(-)

    diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
    index 63668433f5a6..b706e5a54782 100644
    --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
    +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
    @@ -41,6 +41,7 @@ struct amdgpu_device;
     enum kfd_mem_attachment_type {
     	KFD_MEM_ATT_SHARED,	/* Share kgd_mem->bo or another attachment's */
     	KFD_MEM_ATT_USERPTR,	/* SG bo to DMA map pages from a userptr bo */
    +	KFD_MEM_ATT_DMABUF,	/* DMAbuf to DMA map TTM BOs */
     };

     struct kfd_mem_attachment {
    @@ -56,6 +57,7 @@ struct kfd_mem_attachment {
     struct kgd_mem {
     	struct mutex lock;
     	struct amdgpu_bo *bo;
    +	struct dma_buf *dmabuf;
     	struct list_head attachments;
     	/* protected by amdkfd_process_info.lock */
     	struct ttm_validate_buffer validate_list;
    diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
    index 9eeedd0c7920..18a1f9222a59 100644
    --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
    +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
    @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem,
     	return ret;
     }

    +static int
    +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment)
    +{
    +	struct ttm_operation_ctx ctx = {.interruptible = true};
    +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
    +
    +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT);
    +	return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation
    +}
    +
     static int
     kfd_mem_dmamap_attachment(struct kgd_mem *mem,
     			  struct kfd_mem_attachment *attachment)
    @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem,
     		return 0;
     	case KFD_MEM_ATT_USERPTR:
     		return kfd_mem_dmamap_userptr(mem, attachment);
    +	case KFD_MEM_ATT_DMABUF:
    +		return kfd_mem_dmamap_dmabuf(attachment);
     	default:
     		WARN_ON_ONCE(1);
     	}
    @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
     	ttm->sg = NULL;
     }

    +static void
    +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment)
    +{
    +	struct ttm_operation_ctx ctx = {.interruptible = true};
    +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
    +
    +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU);
    +	ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
    +	/* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is
    +	 * called
    +	 */
    +}
    +
     static void
     kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
     			    struct kfd_mem_attachment *attachment)
    @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
     	case KFD_MEM_ATT_USERPTR:
     		kfd_mem_dmaunmap_userptr(mem, attachment);
     		break;
    +	case KFD_MEM_ATT_DMABUF:
    +		kfd_mem_dmaunmap_dmabuf(attachment);
    +		break;
     	default:
     		WARN_ON_ONCE(1);
     	}
    @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
     	return 0;
     }

    +static int
    +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
    +		      struct amdgpu_bo **bo)
    +{
    +	struct drm_gem_object *gobj;
    +
    +	if (!mem->dmabuf) {
    +		mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base,
    +			mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
    +				DRM_RDWR : 0);
    +		if (IS_ERR(mem->dmabuf)) {
    +			mem->dmabuf = NULL;
    +			return PTR_ERR(mem->dmabuf);
    +		}
    +	}
    +
    +	gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf);
    +	if (IS_ERR(gobj))
    +		return PTR_ERR(gobj);
    +
    +	/* Import takes an extra reference on the dmabuf. Drop it now to
    +	 * avoid leaking it. We only need the one reference in
    +	 * kgd_mem->dmabuf.
    +	 */
    +	dma_buf_put(mem->dmabuf);
    +
    +	*bo = gem_to_amdgpu_bo(gobj);
    +	(*bo)->parent = amdgpu_bo_ref(mem->bo);
    +
    +	return 0;
    +}
    +
     /* kfd_mem_attach - Add a BO to a VM
      *
      * Everything that needs to bo done only once when a BO is first added
    @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
     			ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
     			if (ret)
     				goto unwind;
    +		} else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT &&
    +			   mem->bo->tbo.type != ttm_bo_type_sg) {
    +			/* GTT BOs use DMA-mapping ability of dynamic-attach
    +			 * DMA bufs. TODO: The same should work for VRAM on
    +			 * large-BAR GPUs.
    +			 */
    +			attachment[i]->type = KFD_MEM_ATT_DMABUF;
    +			ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
    +			if (ret)
    +				goto unwind;
     		} else {
    -			/* FIXME: Need to DMA-map other BO types */
    +			/* FIXME: Need to DMA-map other BO types:
    +			 * large-BAR VRAM, doorbells, MMIO remap
    +			 */
     			attachment[i]->type = KFD_MEM_ATT_SHARED;
     			bo[i] = mem->bo;
     			drm_gem_object_get(&bo[i]->tbo.base);
    @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(

     	/* Free the BO*/
     	drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
    +	if (mem->dmabuf)
    +		dma_buf_put(mem->dmabuf);
     	drm_gem_object_put(&mem->bo->tbo.base);
     	mutex_destroy(&mem->lock);
     	kfree(mem);
    -- 
    2.31.1

    _______________________________________________
    amd-gfx mailing list
    amd-gfx@lists.freedesktop.org
    https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.freedesktop.org%2Fmailman%2Flistinfo%2Famd-gfx&amp;data=04%7C01%7Coak.zeng%40amd.com%7Ca14e8f1d4ba6450b5f1308d9052e630b%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637546519053906547%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&amp;sdata=AD%2FnbE6kQDFHLTr0kbzZ2sl3jKwuOqUKfpPEcPHwwfY%3D&amp;reserved=0
Felix Kuehling April 27, 2021, 3:56 a.m. UTC | #2
Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak:
> Regards,
> Oak 
>
>  
>
> On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote:
>
>     Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs.
>
>     Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com>
>     ---
>      drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h    |  2 +
>      .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c  | 76 ++++++++++++++++++-
>      2 files changed, 77 insertions(+), 1 deletion(-)
>
>     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     index 63668433f5a6..b706e5a54782 100644
>     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     @@ -41,6 +41,7 @@ struct amdgpu_device;
>      enum kfd_mem_attachment_type {
>      	KFD_MEM_ATT_SHARED,	/* Share kgd_mem->bo or another attachment's */
>      	KFD_MEM_ATT_USERPTR,	/* SG bo to DMA map pages from a userptr bo */
>     +	KFD_MEM_ATT_DMABUF,	/* DMAbuf to DMA map TTM BOs */
>      };
>
>      struct kfd_mem_attachment {
>     @@ -56,6 +57,7 @@ struct kfd_mem_attachment {
>      struct kgd_mem {
>      	struct mutex lock;
>      	struct amdgpu_bo *bo;
>     +	struct dma_buf *dmabuf;
>      	struct list_head attachments;
>      	/* protected by amdkfd_process_info.lock */
>      	struct ttm_validate_buffer validate_list;
>     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     index 9eeedd0c7920..18a1f9222a59 100644
>     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem,
>      	return ret;
>      }
>
>     +static int
>     +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment)
>     +{
>     +	struct ttm_operation_ctx ctx = {.interruptible = true};
>     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
>     +
>     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT);
>     +	return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
> How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation

Conceptually, calling ttm_bo_validate ensures that the BO is in the
specified domain, in this case GTT. Before calling validate, it can be
in the CPU domain, which means it may be swapped to disk so it's not GPU
accessible. For a DMABuf attachment, the CPU domain means, that the
DMABuf is not attached because the underlying memory object may be on
the move or swapped out.

The actual implementation of the dmabuf attachment is currently in
amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this
series fixes that to move the actual dmabuf attachment into
amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when
a BO is moved between the CPU and GTT domains.

Regards,
  Felix


>     +}
>     +
>      static int
>      kfd_mem_dmamap_attachment(struct kgd_mem *mem,
>      			  struct kfd_mem_attachment *attachment)
>     @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem,
>      		return 0;
>      	case KFD_MEM_ATT_USERPTR:
>      		return kfd_mem_dmamap_userptr(mem, attachment);
>     +	case KFD_MEM_ATT_DMABUF:
>     +		return kfd_mem_dmamap_dmabuf(attachment);
>      	default:
>      		WARN_ON_ONCE(1);
>      	}
>     @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
>      	ttm->sg = NULL;
>      }
>
>     +static void
>     +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment)
>     +{
>     +	struct ttm_operation_ctx ctx = {.interruptible = true};
>     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
>     +
>     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU);
>     +	ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
>     +	/* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is
>     +	 * called
>     +	 */
>     +}
>     +
>      static void
>      kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
>      			    struct kfd_mem_attachment *attachment)
>     @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
>      	case KFD_MEM_ATT_USERPTR:
>      		kfd_mem_dmaunmap_userptr(mem, attachment);
>      		break;
>     +	case KFD_MEM_ATT_DMABUF:
>     +		kfd_mem_dmaunmap_dmabuf(attachment);
>     +		break;
>      	default:
>      		WARN_ON_ONCE(1);
>      	}
>     @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
>      	return 0;
>      }
>
>     +static int
>     +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
>     +		      struct amdgpu_bo **bo)
>     +{
>     +	struct drm_gem_object *gobj;
>     +
>     +	if (!mem->dmabuf) {
>     +		mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base,
>     +			mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
>     +				DRM_RDWR : 0);
>     +		if (IS_ERR(mem->dmabuf)) {
>     +			mem->dmabuf = NULL;
>     +			return PTR_ERR(mem->dmabuf);
>     +		}
>     +	}
>     +
>     +	gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf);
>     +	if (IS_ERR(gobj))
>     +		return PTR_ERR(gobj);
>     +
>     +	/* Import takes an extra reference on the dmabuf. Drop it now to
>     +	 * avoid leaking it. We only need the one reference in
>     +	 * kgd_mem->dmabuf.
>     +	 */
>     +	dma_buf_put(mem->dmabuf);
>     +
>     +	*bo = gem_to_amdgpu_bo(gobj);
>     +	(*bo)->parent = amdgpu_bo_ref(mem->bo);
>     +
>     +	return 0;
>     +}
>     +
>      /* kfd_mem_attach - Add a BO to a VM
>       *
>       * Everything that needs to bo done only once when a BO is first added
>     @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
>      			ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
>      			if (ret)
>      				goto unwind;
>     +		} else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT &&
>     +			   mem->bo->tbo.type != ttm_bo_type_sg) {
>     +			/* GTT BOs use DMA-mapping ability of dynamic-attach
>     +			 * DMA bufs. TODO: The same should work for VRAM on
>     +			 * large-BAR GPUs.
>     +			 */
>     +			attachment[i]->type = KFD_MEM_ATT_DMABUF;
>     +			ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
>     +			if (ret)
>     +				goto unwind;
>      		} else {
>     -			/* FIXME: Need to DMA-map other BO types */
>     +			/* FIXME: Need to DMA-map other BO types:
>     +			 * large-BAR VRAM, doorbells, MMIO remap
>     +			 */
>      			attachment[i]->type = KFD_MEM_ATT_SHARED;
>      			bo[i] = mem->bo;
>      			drm_gem_object_get(&bo[i]->tbo.base);
>     @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(
>
>      	/* Free the BO*/
>      	drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
>     +	if (mem->dmabuf)
>     +		dma_buf_put(mem->dmabuf);
>      	drm_gem_object_put(&mem->bo->tbo.base);
>      	mutex_destroy(&mem->lock);
>      	kfree(mem);
>     -- 
>     2.31.1
>
>     _______________________________________________
>     amd-gfx mailing list
>     amd-gfx@lists.freedesktop.org
>     https://lists.freedesktop.org/mailman/listinfo/amd-gfx
>
Zeng, Oak April 27, 2021, 2:29 p.m. UTC | #3
Regards,
Oak 

 

On 2021-04-26, 11:56 PM, "Kuehling, Felix" <Felix.Kuehling@amd.com> wrote:

    Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak:
    > Regards,
    > Oak 
    >
    >  
    >
    > On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote:
    >
    >     Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs.
    >
    >     Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com>
    >     ---
    >      drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h    |  2 +
    >      .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c  | 76 ++++++++++++++++++-
    >      2 files changed, 77 insertions(+), 1 deletion(-)
    >
    >     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
    >     index 63668433f5a6..b706e5a54782 100644
    >     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
    >     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
    >     @@ -41,6 +41,7 @@ struct amdgpu_device;
    >      enum kfd_mem_attachment_type {
    >      	KFD_MEM_ATT_SHARED,	/* Share kgd_mem->bo or another attachment's */
    >      	KFD_MEM_ATT_USERPTR,	/* SG bo to DMA map pages from a userptr bo */
    >     +	KFD_MEM_ATT_DMABUF,	/* DMAbuf to DMA map TTM BOs */
    >      };
    >
    >      struct kfd_mem_attachment {
    >     @@ -56,6 +57,7 @@ struct kfd_mem_attachment {
    >      struct kgd_mem {
    >      	struct mutex lock;
    >      	struct amdgpu_bo *bo;
    >     +	struct dma_buf *dmabuf;
    >      	struct list_head attachments;
    >      	/* protected by amdkfd_process_info.lock */
    >      	struct ttm_validate_buffer validate_list;
    >     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
    >     index 9eeedd0c7920..18a1f9222a59 100644
    >     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
    >     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
    >     @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem,
    >      	return ret;
    >      }
    >
    >     +static int
    >     +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment)
    >     +{
    >     +	struct ttm_operation_ctx ctx = {.interruptible = true};
    >     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
    >     +
    >     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT);
    >     +	return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
    > How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation

    Conceptually, calling ttm_bo_validate ensures that the BO is in the
    specified domain, in this case GTT. Before calling validate, it can be
    in the CPU domain, which means it may be swapped to disk so it's not GPU
    accessible. For a DMABuf attachment, the CPU domain means, that the
    DMABuf is not attached because the underlying memory object may be on
    the move or swapped out.

    The actual implementation of the dmabuf attachment is currently in
    amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this
    series fixes that to move the actual dmabuf attachment into
    amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when
    a BO is moved between the CPU and GTT domains.

Thanks for the explanation. One more thing I don't quite understand: before this series, GTT memory should already has been validated somewhere before GTT memory is mapped to GPU. You added GTT memory validation here - will this validation be duplicated?

The function naming kfd_mem_dmamap_dmabuf is still confusing since it seems to me it is only some preparation work before dynamically dma-map a GTT memory. But I understand from this series' perspective, compared to usrptr (where you actually do the dma-mapping in function kfd_mem_dmamap_usrptr), for gtt memory you leveraged the amdgpu ttm function of dynamic dma-mapping. So maybe the naming here makes sense from that perspective.

Another thing related but not directly to this series: for GTT memory, it is dma-mapped when it is allocated. See function ttm_populate_and_map_pages calling dma_map_page. The question is, will gtt be first dma-unmapping before it is mapped in amdgpu_ttm_backend_bind? It is existing work, not from your series. Maybe there is not issue but I just want to make sure while we are looking at this area. 

    Regards,
      Felix


    >     +}
    >     +
    >      static int
    >      kfd_mem_dmamap_attachment(struct kgd_mem *mem,
    >      			  struct kfd_mem_attachment *attachment)
    >     @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem,
    >      		return 0;
    >      	case KFD_MEM_ATT_USERPTR:
    >      		return kfd_mem_dmamap_userptr(mem, attachment);
    >     +	case KFD_MEM_ATT_DMABUF:
    >     +		return kfd_mem_dmamap_dmabuf(attachment);
    >      	default:
    >      		WARN_ON_ONCE(1);
    >      	}
    >     @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
    >      	ttm->sg = NULL;
    >      }
    >
    >     +static void
    >     +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment)
    >     +{
    >     +	struct ttm_operation_ctx ctx = {.interruptible = true};
    >     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
    >     +
    >     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU);
    >     +	ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
    >     +	/* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is
    >     +	 * called
    >     +	 */
    >     +}
    >     +
    >      static void
    >      kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
    >      			    struct kfd_mem_attachment *attachment)
    >     @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
    >      	case KFD_MEM_ATT_USERPTR:
    >      		kfd_mem_dmaunmap_userptr(mem, attachment);
    >      		break;
    >     +	case KFD_MEM_ATT_DMABUF:
    >     +		kfd_mem_dmaunmap_dmabuf(attachment);
    >     +		break;
    >      	default:
    >      		WARN_ON_ONCE(1);
    >      	}
    >     @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
    >      	return 0;
    >      }
    >
    >     +static int
    >     +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
    >     +		      struct amdgpu_bo **bo)
    >     +{
    >     +	struct drm_gem_object *gobj;
    >     +
    >     +	if (!mem->dmabuf) {
    >     +		mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base,
    >     +			mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
    >     +				DRM_RDWR : 0);
    >     +		if (IS_ERR(mem->dmabuf)) {
    >     +			mem->dmabuf = NULL;
    >     +			return PTR_ERR(mem->dmabuf);
    >     +		}
    >     +	}
    >     +
    >     +	gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf);
    >     +	if (IS_ERR(gobj))
    >     +		return PTR_ERR(gobj);
    >     +
    >     +	/* Import takes an extra reference on the dmabuf. Drop it now to
    >     +	 * avoid leaking it. We only need the one reference in
    >     +	 * kgd_mem->dmabuf.
    >     +	 */
    >     +	dma_buf_put(mem->dmabuf);
    >     +
    >     +	*bo = gem_to_amdgpu_bo(gobj);
    >     +	(*bo)->parent = amdgpu_bo_ref(mem->bo);
    >     +
    >     +	return 0;
    >     +}
    >     +
    >      /* kfd_mem_attach - Add a BO to a VM
    >       *
    >       * Everything that needs to bo done only once when a BO is first added
    >     @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
    >      			ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
    >      			if (ret)
    >      				goto unwind;
    >     +		} else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT &&
    >     +			   mem->bo->tbo.type != ttm_bo_type_sg) {
    >     +			/* GTT BOs use DMA-mapping ability of dynamic-attach
    >     +			 * DMA bufs. TODO: The same should work for VRAM on
    >     +			 * large-BAR GPUs.
    >     +			 */
    >     +			attachment[i]->type = KFD_MEM_ATT_DMABUF;
    >     +			ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
    >     +			if (ret)
    >     +				goto unwind;
    >      		} else {
    >     -			/* FIXME: Need to DMA-map other BO types */
    >     +			/* FIXME: Need to DMA-map other BO types:
    >     +			 * large-BAR VRAM, doorbells, MMIO remap
    >     +			 */
    >      			attachment[i]->type = KFD_MEM_ATT_SHARED;
    >      			bo[i] = mem->bo;
    >      			drm_gem_object_get(&bo[i]->tbo.base);
    >     @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(
    >
    >      	/* Free the BO*/
    >      	drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
    >     +	if (mem->dmabuf)
    >     +		dma_buf_put(mem->dmabuf);
    >      	drm_gem_object_put(&mem->bo->tbo.base);
    >      	mutex_destroy(&mem->lock);
    >      	kfree(mem);
    >     -- 
    >     2.31.1
    >
    >     _______________________________________________
    >     amd-gfx mailing list
    >     amd-gfx@lists.freedesktop.org
    >     https://lists.freedesktop.org/mailman/listinfo/amd-gfx
    >
Felix Kuehling April 27, 2021, 3:08 p.m. UTC | #4
Am 2021-04-27 um 10:29 a.m. schrieb Zeng, Oak:
> Regards,
> Oak 
>
>  
>
> On 2021-04-26, 11:56 PM, "Kuehling, Felix" <Felix.Kuehling@amd.com> wrote:
>
>     Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak:
>     > Regards,
>     > Oak 
>     >
>     >  
>     >
>     > On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote:
>     >
>     >     Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs.
>     >
>     >     Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com>
>     >     ---
>     >      drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h    |  2 +
>     >      .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c  | 76 ++++++++++++++++++-
>     >      2 files changed, 77 insertions(+), 1 deletion(-)
>     >
>     >     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     >     index 63668433f5a6..b706e5a54782 100644
>     >     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     >     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     >     @@ -41,6 +41,7 @@ struct amdgpu_device;
>     >      enum kfd_mem_attachment_type {
>     >      	KFD_MEM_ATT_SHARED,	/* Share kgd_mem->bo or another attachment's */
>     >      	KFD_MEM_ATT_USERPTR,	/* SG bo to DMA map pages from a userptr bo */
>     >     +	KFD_MEM_ATT_DMABUF,	/* DMAbuf to DMA map TTM BOs */
>     >      };
>     >
>     >      struct kfd_mem_attachment {
>     >     @@ -56,6 +57,7 @@ struct kfd_mem_attachment {
>     >      struct kgd_mem {
>     >      	struct mutex lock;
>     >      	struct amdgpu_bo *bo;
>     >     +	struct dma_buf *dmabuf;
>     >      	struct list_head attachments;
>     >      	/* protected by amdkfd_process_info.lock */
>     >      	struct ttm_validate_buffer validate_list;
>     >     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     >     index 9eeedd0c7920..18a1f9222a59 100644
>     >     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     >     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     >     @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem,
>     >      	return ret;
>     >      }
>     >
>     >     +static int
>     >     +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment)
>     >     +{
>     >     +	struct ttm_operation_ctx ctx = {.interruptible = true};
>     >     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
>     >     +
>     >     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT);
>     >     +	return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
>     > How does this work? The function name says this is dma mapping a buffer but from the implementation, it is just a placement and validation
>
>     Conceptually, calling ttm_bo_validate ensures that the BO is in the
>     specified domain, in this case GTT. Before calling validate, it can be
>     in the CPU domain, which means it may be swapped to disk so it's not GPU
>     accessible. For a DMABuf attachment, the CPU domain means, that the
>     DMABuf is not attached because the underlying memory object may be on
>     the move or swapped out.
>
>     The actual implementation of the dmabuf attachment is currently in
>     amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this
>     series fixes that to move the actual dmabuf attachment into
>     amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when
>     a BO is moved between the CPU and GTT domains.
>
> Thanks for the explanation. One more thing I don't quite understand: before this series, GTT memory should already has been validated somewhere before GTT memory is mapped to GPU. You added GTT memory validation here - will this validation be duplicated?

When you have N GPUs there are now N BOs involved. Each GPU needs its
own BO because it needs its own DMA mapping. There will be one actual
GTT BO that allocates physical pages in TTM. The other BOs are dmabuf
imports that DMA-map the same physical pages for access by the other GPUs.

The validate call here validates one of the dmabuf imports. This does
not duplicate the validation of the underlying TTM BO with the actual
physical memory allocation.


>
> The function naming kfd_mem_dmamap_dmabuf is still confusing since it seems to me it is only some preparation work before dynamically dma-map a GTT memory.

No, this series is not just preparation. It implements DMA mapping of
BOs for multiple GPUs. TTM already handles DMA mapping of the memory for
the device where the memory was allocated. (Yes, even GTT memory is
associated with a specific GPU even though it's physically in system
memory). What this patch series adds, is additional DMA mappings for the
other GPUs. Without this patch, we were using the DMA mapping for GPU-1
in the page table of GPU-X, which is incorrect. It works in many cases
where the DMA mapping is a direct mapping:

  * IOMMU disabled
  * IOMMU in passthrough mode

But it breaks when you have multiple GPUs with an IOMMU that's not
disabled or in passthrough mode.


>  But I understand from this series' perspective, compared to usrptr (where you actually do the dma-mapping in function kfd_mem_dmamap_usrptr), for gtt memory you leveraged the amdgpu ttm function of dynamic dma-mapping. So maybe the naming here makes sense from that perspective.

Yes.


>
> Another thing related but not directly to this series: for GTT memory, it is dma-mapped when it is allocated. See function ttm_populate_and_map_pages calling dma_map_page. The question is, will gtt be first dma-unmapping before it is mapped in amdgpu_ttm_backend_bind? It is existing work, not from your series. Maybe there is not issue but I just want to make sure while we are looking at this area. 

Right. The problem is, that the DMA mappings only work for a specific
device. Using the same DMA mapping on multiple devices is broken. The
reason we got away with it for a long time is, that we were running with
IOMMU disabled or in passthrough mode.

Regards,
  Felix


>
>     Regards,
>       Felix
>
>
>     >     +}
>     >     +
>     >      static int
>     >      kfd_mem_dmamap_attachment(struct kgd_mem *mem,
>     >      			  struct kfd_mem_attachment *attachment)
>     >     @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem,
>     >      		return 0;
>     >      	case KFD_MEM_ATT_USERPTR:
>     >      		return kfd_mem_dmamap_userptr(mem, attachment);
>     >     +	case KFD_MEM_ATT_DMABUF:
>     >     +		return kfd_mem_dmamap_dmabuf(attachment);
>     >      	default:
>     >      		WARN_ON_ONCE(1);
>     >      	}
>     >     @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
>     >      	ttm->sg = NULL;
>     >      }
>     >
>     >     +static void
>     >     +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment)
>     >     +{
>     >     +	struct ttm_operation_ctx ctx = {.interruptible = true};
>     >     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
>     >     +
>     >     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU);
>     >     +	ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
>     >     +	/* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is
>     >     +	 * called
>     >     +	 */
>     >     +}
>     >     +
>     >      static void
>     >      kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
>     >      			    struct kfd_mem_attachment *attachment)
>     >     @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
>     >      	case KFD_MEM_ATT_USERPTR:
>     >      		kfd_mem_dmaunmap_userptr(mem, attachment);
>     >      		break;
>     >     +	case KFD_MEM_ATT_DMABUF:
>     >     +		kfd_mem_dmaunmap_dmabuf(attachment);
>     >     +		break;
>     >      	default:
>     >      		WARN_ON_ONCE(1);
>     >      	}
>     >     @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
>     >      	return 0;
>     >      }
>     >
>     >     +static int
>     >     +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
>     >     +		      struct amdgpu_bo **bo)
>     >     +{
>     >     +	struct drm_gem_object *gobj;
>     >     +
>     >     +	if (!mem->dmabuf) {
>     >     +		mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base,
>     >     +			mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
>     >     +				DRM_RDWR : 0);
>     >     +		if (IS_ERR(mem->dmabuf)) {
>     >     +			mem->dmabuf = NULL;
>     >     +			return PTR_ERR(mem->dmabuf);
>     >     +		}
>     >     +	}
>     >     +
>     >     +	gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf);
>     >     +	if (IS_ERR(gobj))
>     >     +		return PTR_ERR(gobj);
>     >     +
>     >     +	/* Import takes an extra reference on the dmabuf. Drop it now to
>     >     +	 * avoid leaking it. We only need the one reference in
>     >     +	 * kgd_mem->dmabuf.
>     >     +	 */
>     >     +	dma_buf_put(mem->dmabuf);
>     >     +
>     >     +	*bo = gem_to_amdgpu_bo(gobj);
>     >     +	(*bo)->parent = amdgpu_bo_ref(mem->bo);
>     >     +
>     >     +	return 0;
>     >     +}
>     >     +
>     >      /* kfd_mem_attach - Add a BO to a VM
>     >       *
>     >       * Everything that needs to bo done only once when a BO is first added
>     >     @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
>     >      			ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
>     >      			if (ret)
>     >      				goto unwind;
>     >     +		} else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT &&
>     >     +			   mem->bo->tbo.type != ttm_bo_type_sg) {
>     >     +			/* GTT BOs use DMA-mapping ability of dynamic-attach
>     >     +			 * DMA bufs. TODO: The same should work for VRAM on
>     >     +			 * large-BAR GPUs.
>     >     +			 */
>     >     +			attachment[i]->type = KFD_MEM_ATT_DMABUF;
>     >     +			ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
>     >     +			if (ret)
>     >     +				goto unwind;
>     >      		} else {
>     >     -			/* FIXME: Need to DMA-map other BO types */
>     >     +			/* FIXME: Need to DMA-map other BO types:
>     >     +			 * large-BAR VRAM, doorbells, MMIO remap
>     >     +			 */
>     >      			attachment[i]->type = KFD_MEM_ATT_SHARED;
>     >      			bo[i] = mem->bo;
>     >      			drm_gem_object_get(&bo[i]->tbo.base);
>     >     @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(
>     >
>     >      	/* Free the BO*/
>     >      	drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
>     >     +	if (mem->dmabuf)
>     >     +		dma_buf_put(mem->dmabuf);
>     >      	drm_gem_object_put(&mem->bo->tbo.base);
>     >      	mutex_destroy(&mem->lock);
>     >      	kfree(mem);
>     >     -- 
>     >     2.31.1
>     >
>     >     _______________________________________________
>     >     amd-gfx mailing list
>     >     amd-gfx@lists.freedesktop.org
>     >     https://lists.freedesktop.org/mailman/listinfo/amd-gfx
>     >
>
Ramesh Errabolu May 10, 2021, 10:07 p.m. UTC | #5
[AMD Official Use Only - Internal Distribution Only]

Acked-by: Ramesh Errabolu <ramesh.errabolu@amd.com>

-----Original Message-----
From: amd-gfx <amd-gfx-bounces@lists.freedesktop.org> On Behalf Of Kuehling, Felix
Sent: Tuesday, April 27, 2021 10:09 AM
To: Zeng, Oak <Oak.Zeng@amd.com>; amd-gfx@lists.freedesktop.org; dri-devel@lists.freedesktop.org
Subject: Re: [PATCH v2 08/10] drm/amdgpu: Add DMA mapping of GTT BOs

Am 2021-04-27 um 10:29 a.m. schrieb Zeng, Oak:
> Regards,
> Oak
>
>  
>
> On 2021-04-26, 11:56 PM, "Kuehling, Felix" <Felix.Kuehling@amd.com> wrote:
>
>     Am 2021-04-26 um 8:35 p.m. schrieb Zeng, Oak:
>     > Regards,
>     > Oak 
>     >
>     >  
>     >
>     > On 2021-04-21, 9:31 PM, "amd-gfx on behalf of Felix Kuehling" <amd-gfx-bounces@lists.freedesktop.org on behalf of Felix.Kuehling@amd.com> wrote:
>     >
>     >     Use DMABufs with dynamic attachment to DMA-map GTT BOs on other GPUs.
>     >
>     >     Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com>
>     >     ---
>     >      drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h    |  2 +
>     >      .../gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c  | 76 ++++++++++++++++++-
>     >      2 files changed, 77 insertions(+), 1 deletion(-)
>     >
>     >     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     >     index 63668433f5a6..b706e5a54782 100644
>     >     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     >     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
>     >     @@ -41,6 +41,7 @@ struct amdgpu_device;
>     >      enum kfd_mem_attachment_type {
>     >      	KFD_MEM_ATT_SHARED,	/* Share kgd_mem->bo or another attachment's */
>     >      	KFD_MEM_ATT_USERPTR,	/* SG bo to DMA map pages from a userptr bo */
>     >     +	KFD_MEM_ATT_DMABUF,	/* DMAbuf to DMA map TTM BOs */
>     >      };
>     >
>     >      struct kfd_mem_attachment {
>     >     @@ -56,6 +57,7 @@ struct kfd_mem_attachment {
>     >      struct kgd_mem {
>     >      	struct mutex lock;
>     >      	struct amdgpu_bo *bo;
>     >     +	struct dma_buf *dmabuf;
>     >      	struct list_head attachments;
>     >      	/* protected by amdkfd_process_info.lock */
>     >      	struct ttm_validate_buffer validate_list;
>     >     diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     >     index 9eeedd0c7920..18a1f9222a59 100644
>     >     --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     >     +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
>     >     @@ -524,6 +524,16 @@ kfd_mem_dmamap_userptr(struct kgd_mem *mem,
>     >      	return ret;
>     >      }
>     >
>     >     +static int
>     >     +kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment)
>     >     +{
>     >     +	struct ttm_operation_ctx ctx = {.interruptible = true};
>     >     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
>     >     +
>     >     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT);
>     >     +	return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
>     > How does this work? The function name says this is dma mapping a 
> buffer but from the implementation, it is just a placement and 
> validation
>
>     Conceptually, calling ttm_bo_validate ensures that the BO is in the
>     specified domain, in this case GTT. Before calling validate, it can be
>     in the CPU domain, which means it may be swapped to disk so it's not GPU
>     accessible. For a DMABuf attachment, the CPU domain means, that the
>     DMABuf is not attached because the underlying memory object may be on
>     the move or swapped out.
>
>     The actual implementation of the dmabuf attachment is currently in
>     amdgpu_ttm_populate/unpopulate. This is incorrect. Patch 10 in this
>     series fixes that to move the actual dmabuf attachment into
>     amdgpu_ttm_backend_bind/unbind, which is called from amdgpu_bo_move when
>     a BO is moved between the CPU and GTT domains.
>
> Thanks for the explanation. One more thing I don't quite understand: before this series, GTT memory should already has been validated somewhere before GTT memory is mapped to GPU. You added GTT memory validation here - will this validation be duplicated?

When you have N GPUs there are now N BOs involved. Each GPU needs its own BO because it needs its own DMA mapping. There will be one actual GTT BO that allocates physical pages in TTM. The other BOs are dmabuf imports that DMA-map the same physical pages for access by the other GPUs.

The validate call here validates one of the dmabuf imports. This does not duplicate the validation of the underlying TTM BO with the actual physical memory allocation.


>
> The function naming kfd_mem_dmamap_dmabuf is still confusing since it seems to me it is only some preparation work before dynamically dma-map a GTT memory.

No, this series is not just preparation. It implements DMA mapping of BOs for multiple GPUs. TTM already handles DMA mapping of the memory for the device where the memory was allocated. (Yes, even GTT memory is associated with a specific GPU even though it's physically in system memory). What this patch series adds, is additional DMA mappings for the other GPUs. Without this patch, we were using the DMA mapping for GPU-1 in the page table of GPU-X, which is incorrect. It works in many cases where the DMA mapping is a direct mapping:

  * IOMMU disabled
  * IOMMU in passthrough mode

But it breaks when you have multiple GPUs with an IOMMU that's not disabled or in passthrough mode.


>  But I understand from this series' perspective, compared to usrptr (where you actually do the dma-mapping in function kfd_mem_dmamap_usrptr), for gtt memory you leveraged the amdgpu ttm function of dynamic dma-mapping. So maybe the naming here makes sense from that perspective.

Yes.


>
> Another thing related but not directly to this series: for GTT memory, it is dma-mapped when it is allocated. See function ttm_populate_and_map_pages calling dma_map_page. The question is, will gtt be first dma-unmapping before it is mapped in amdgpu_ttm_backend_bind? It is existing work, not from your series. Maybe there is not issue but I just want to make sure while we are looking at this area. 

Right. The problem is, that the DMA mappings only work for a specific device. Using the same DMA mapping on multiple devices is broken. The reason we got away with it for a long time is, that we were running with IOMMU disabled or in passthrough mode.

Regards,
  Felix


>
>     Regards,
>       Felix
>
>
>     >     +}
>     >     +
>     >      static int
>     >      kfd_mem_dmamap_attachment(struct kgd_mem *mem,
>     >      			  struct kfd_mem_attachment *attachment)
>     >     @@ -533,6 +543,8 @@ kfd_mem_dmamap_attachment(struct kgd_mem *mem,
>     >      		return 0;
>     >      	case KFD_MEM_ATT_USERPTR:
>     >      		return kfd_mem_dmamap_userptr(mem, attachment);
>     >     +	case KFD_MEM_ATT_DMABUF:
>     >     +		return kfd_mem_dmamap_dmabuf(attachment);
>     >      	default:
>     >      		WARN_ON_ONCE(1);
>     >      	}
>     >     @@ -562,6 +574,19 @@ kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
>     >      	ttm->sg = NULL;
>     >      }
>     >
>     >     +static void
>     >     +kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment)
>     >     +{
>     >     +	struct ttm_operation_ctx ctx = {.interruptible = true};
>     >     +	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
>     >     +
>     >     +	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU);
>     >     +	ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
>     >     +	/* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is
>     >     +	 * called
>     >     +	 */
>     >     +}
>     >     +
>     >      static void
>     >      kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
>     >      			    struct kfd_mem_attachment *attachment)
>     >     @@ -572,6 +597,9 @@ kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
>     >      	case KFD_MEM_ATT_USERPTR:
>     >      		kfd_mem_dmaunmap_userptr(mem, attachment);
>     >      		break;
>     >     +	case KFD_MEM_ATT_DMABUF:
>     >     +		kfd_mem_dmaunmap_dmabuf(attachment);
>     >     +		break;
>     >      	default:
>     >      		WARN_ON_ONCE(1);
>     >      	}
>     >     @@ -605,6 +633,38 @@ kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
>     >      	return 0;
>     >      }
>     >
>     >     +static int
>     >     +kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
>     >     +		      struct amdgpu_bo **bo)
>     >     +{
>     >     +	struct drm_gem_object *gobj;
>     >     +
>     >     +	if (!mem->dmabuf) {
>     >     +		mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base,
>     >     +			mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
>     >     +				DRM_RDWR : 0);
>     >     +		if (IS_ERR(mem->dmabuf)) {
>     >     +			mem->dmabuf = NULL;
>     >     +			return PTR_ERR(mem->dmabuf);
>     >     +		}
>     >     +	}
>     >     +
>     >     +	gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf);
>     >     +	if (IS_ERR(gobj))
>     >     +		return PTR_ERR(gobj);
>     >     +
>     >     +	/* Import takes an extra reference on the dmabuf. Drop it now to
>     >     +	 * avoid leaking it. We only need the one reference in
>     >     +	 * kgd_mem->dmabuf.
>     >     +	 */
>     >     +	dma_buf_put(mem->dmabuf);
>     >     +
>     >     +	*bo = gem_to_amdgpu_bo(gobj);
>     >     +	(*bo)->parent = amdgpu_bo_ref(mem->bo);
>     >     +
>     >     +	return 0;
>     >     +}
>     >     +
>     >      /* kfd_mem_attach - Add a BO to a VM
>     >       *
>     >       * Everything that needs to bo done only once when a BO is first added
>     >     @@ -662,8 +722,20 @@ static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
>     >      			ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
>     >      			if (ret)
>     >      				goto unwind;
>     >     +		} else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT &&
>     >     +			   mem->bo->tbo.type != ttm_bo_type_sg) {
>     >     +			/* GTT BOs use DMA-mapping ability of dynamic-attach
>     >     +			 * DMA bufs. TODO: The same should work for VRAM on
>     >     +			 * large-BAR GPUs.
>     >     +			 */
>     >     +			attachment[i]->type = KFD_MEM_ATT_DMABUF;
>     >     +			ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
>     >     +			if (ret)
>     >     +				goto unwind;
>     >      		} else {
>     >     -			/* FIXME: Need to DMA-map other BO types */
>     >     +			/* FIXME: Need to DMA-map other BO types:
>     >     +			 * large-BAR VRAM, doorbells, MMIO remap
>     >     +			 */
>     >      			attachment[i]->type = KFD_MEM_ATT_SHARED;
>     >      			bo[i] = mem->bo;
>     >      			drm_gem_object_get(&bo[i]->tbo.base);
>     >     @@ -1522,6 +1594,8 @@ int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(
>     >
>     >      	/* Free the BO*/
>     >      	drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
>     >     +	if (mem->dmabuf)
>     >     +		dma_buf_put(mem->dmabuf);
>     >      	drm_gem_object_put(&mem->bo->tbo.base);
>     >      	mutex_destroy(&mem->lock);
>     >      	kfree(mem);
>     >     -- 
>     >     2.31.1
>     >
>     >     _______________________________________________
>     >     amd-gfx mailing list
>     >     amd-gfx@lists.freedesktop.org
>     >     https://nam11.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.freedesktop.org%2Fmailman%2Flistinfo%2Famd-gfx&amp;data=04%7C01%7Cphilip.yang%40amd.com%7Cf3c9c824ef6447cbe26808d9098e6606%7C3dd8961fe4884e608e11a82d994e183d%7C0%7C0%7C637551329471854747%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C1000&amp;sdata=P7SYxlhPoYfyGCNaiiCA%2BaqJS%2BxvGZEMIZkuvqCpCLI%3D&amp;reserved=0
>     >
>
diff mbox series

Patch

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
index 63668433f5a6..b706e5a54782 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
@@ -41,6 +41,7 @@  struct amdgpu_device;
 enum kfd_mem_attachment_type {
 	KFD_MEM_ATT_SHARED,	/* Share kgd_mem->bo or another attachment's */
 	KFD_MEM_ATT_USERPTR,	/* SG bo to DMA map pages from a userptr bo */
+	KFD_MEM_ATT_DMABUF,	/* DMAbuf to DMA map TTM BOs */
 };
 
 struct kfd_mem_attachment {
@@ -56,6 +57,7 @@  struct kfd_mem_attachment {
 struct kgd_mem {
 	struct mutex lock;
 	struct amdgpu_bo *bo;
+	struct dma_buf *dmabuf;
 	struct list_head attachments;
 	/* protected by amdkfd_process_info.lock */
 	struct ttm_validate_buffer validate_list;
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
index 9eeedd0c7920..18a1f9222a59 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
@@ -524,6 +524,16 @@  kfd_mem_dmamap_userptr(struct kgd_mem *mem,
 	return ret;
 }
 
+static int
+kfd_mem_dmamap_dmabuf(struct kfd_mem_attachment *attachment)
+{
+	struct ttm_operation_ctx ctx = {.interruptible = true};
+	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
+
+	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_GTT);
+	return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
+}
+
 static int
 kfd_mem_dmamap_attachment(struct kgd_mem *mem,
 			  struct kfd_mem_attachment *attachment)
@@ -533,6 +543,8 @@  kfd_mem_dmamap_attachment(struct kgd_mem *mem,
 		return 0;
 	case KFD_MEM_ATT_USERPTR:
 		return kfd_mem_dmamap_userptr(mem, attachment);
+	case KFD_MEM_ATT_DMABUF:
+		return kfd_mem_dmamap_dmabuf(attachment);
 	default:
 		WARN_ON_ONCE(1);
 	}
@@ -562,6 +574,19 @@  kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
 	ttm->sg = NULL;
 }
 
+static void
+kfd_mem_dmaunmap_dmabuf(struct kfd_mem_attachment *attachment)
+{
+	struct ttm_operation_ctx ctx = {.interruptible = true};
+	struct amdgpu_bo *bo = attachment->bo_va->base.bo;
+
+	amdgpu_bo_placement_from_domain(bo, AMDGPU_GEM_DOMAIN_CPU);
+	ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
+	/* FIXME: This does not guarantee that amdgpu_ttm_tt_unpopulate is
+	 * called
+	 */
+}
+
 static void
 kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
 			    struct kfd_mem_attachment *attachment)
@@ -572,6 +597,9 @@  kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
 	case KFD_MEM_ATT_USERPTR:
 		kfd_mem_dmaunmap_userptr(mem, attachment);
 		break;
+	case KFD_MEM_ATT_DMABUF:
+		kfd_mem_dmaunmap_dmabuf(attachment);
+		break;
 	default:
 		WARN_ON_ONCE(1);
 	}
@@ -605,6 +633,38 @@  kfd_mem_attach_userptr(struct amdgpu_device *adev, struct kgd_mem *mem,
 	return 0;
 }
 
+static int
+kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
+		      struct amdgpu_bo **bo)
+{
+	struct drm_gem_object *gobj;
+
+	if (!mem->dmabuf) {
+		mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base,
+			mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
+				DRM_RDWR : 0);
+		if (IS_ERR(mem->dmabuf)) {
+			mem->dmabuf = NULL;
+			return PTR_ERR(mem->dmabuf);
+		}
+	}
+
+	gobj = amdgpu_gem_prime_import(&adev->ddev, mem->dmabuf);
+	if (IS_ERR(gobj))
+		return PTR_ERR(gobj);
+
+	/* Import takes an extra reference on the dmabuf. Drop it now to
+	 * avoid leaking it. We only need the one reference in
+	 * kgd_mem->dmabuf.
+	 */
+	dma_buf_put(mem->dmabuf);
+
+	*bo = gem_to_amdgpu_bo(gobj);
+	(*bo)->parent = amdgpu_bo_ref(mem->bo);
+
+	return 0;
+}
+
 /* kfd_mem_attach - Add a BO to a VM
  *
  * Everything that needs to bo done only once when a BO is first added
@@ -662,8 +722,20 @@  static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
 			ret = kfd_mem_attach_userptr(adev, mem, &bo[i]);
 			if (ret)
 				goto unwind;
+		} else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT &&
+			   mem->bo->tbo.type != ttm_bo_type_sg) {
+			/* GTT BOs use DMA-mapping ability of dynamic-attach
+			 * DMA bufs. TODO: The same should work for VRAM on
+			 * large-BAR GPUs.
+			 */
+			attachment[i]->type = KFD_MEM_ATT_DMABUF;
+			ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
+			if (ret)
+				goto unwind;
 		} else {
-			/* FIXME: Need to DMA-map other BO types */
+			/* FIXME: Need to DMA-map other BO types:
+			 * large-BAR VRAM, doorbells, MMIO remap
+			 */
 			attachment[i]->type = KFD_MEM_ATT_SHARED;
 			bo[i] = mem->bo;
 			drm_gem_object_get(&bo[i]->tbo.base);
@@ -1522,6 +1594,8 @@  int amdgpu_amdkfd_gpuvm_free_memory_of_gpu(
 
 	/* Free the BO*/
 	drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
+	if (mem->dmabuf)
+		dma_buf_put(mem->dmabuf);
 	drm_gem_object_put(&mem->bo->tbo.base);
 	mutex_destroy(&mem->lock);
 	kfree(mem);