Message ID | 20221125102137.1801-3-christian.koenig@amd.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [1/9] drm/amdgpu: generally allow over-commit during BO allocation | expand |
On 2022-11-25 05:21, Christian König wrote: > Instead of a single worker going over the list of delete BOs in regular > intervals use a per BO worker which blocks for the resv object and > locking of the BO. > > This not only simplifies the handling massively, but also results in > much better response time when cleaning up buffers. > > Signed-off-by: Christian König <christian.koenig@amd.com> Just thinking out loud: If I understand it correctly, this can cause a lot of sleeping worker threads when AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE is used and many BOs are freed at the same time. This happens e.g. when a KFD process terminates or crashes. I guess with a concurrency-managed workqueue this isn't going to be excessive. And since it's on a per device workqueue, it doesn't stall work items on the system work queue or from other devices. I'm trying to understand why you set WQ_MEM_RECLAIM. This work queue is not about freeing ttm_resources but about freeing the BOs. But it affects freeing of ghost_objs that are holding the ttm_resources being freed. If those assumptions all make sense, patches 1-3 are Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- > drivers/gpu/drm/i915/i915_gem.c | 2 +- > drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- > drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- > drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - > drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- > include/drm/ttm/ttm_bo_api.h | 18 +--- > include/drm/ttm/ttm_device.h | 7 +- > 8 files changed, 57 insertions(+), 111 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > index 2b1db37e25c1..74ccbd566777 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device *adev) > amdgpu_fence_driver_hw_fini(adev); > > if (adev->mman.initialized) > - flush_delayed_work(&adev->mman.bdev.wq); > + drain_workqueue(adev->mman.bdev.wq); > > if (adev->pm_sysfs_en) > amdgpu_pm_sysfs_fini(adev); > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > index 8468ca9885fd..c38306f156d6 100644 > --- a/drivers/gpu/drm/i915/i915_gem.c > +++ b/drivers/gpu/drm/i915/i915_gem.c > @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct drm_i915_private *i915) > { > while (atomic_read(&i915->mm.free_count)) { > flush_work(&i915->mm.free_work); > - flush_delayed_work(&i915->bdev.wq); > + drain_workqueue(i915->bdev.wq); > rcu_barrier(); > } > } > diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c b/drivers/gpu/drm/i915/intel_region_ttm.c > index cf89d0c2a2d9..657bbc16a48a 100644 > --- a/drivers/gpu/drm/i915/intel_region_ttm.c > +++ b/drivers/gpu/drm/i915/intel_region_ttm.c > @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct intel_memory_region *mem) > break; > > msleep(20); > - flush_delayed_work(&mem->i915->bdev.wq); > + drain_workqueue(mem->i915->bdev.wq); > } > > /* If we leaked objects, Don't free the region causing use after free */ > diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > index b77262a623e0..4749b65bedc4 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo.c > +++ b/drivers/gpu/drm/ttm/ttm_bo.c > @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object *bo, > ret = 0; > } > > - if (ret || unlikely(list_empty(&bo->ddestroy))) { > + if (ret) { > if (unlock_resv) > dma_resv_unlock(bo->base.resv); > spin_unlock(&bo->bdev->lru_lock); > return ret; > } > > - list_del_init(&bo->ddestroy); > spin_unlock(&bo->bdev->lru_lock); > ttm_bo_cleanup_memtype_use(bo); > > @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object *bo, > } > > /* > - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all > - * encountered buffers. > + * Block for the dma_resv object to become idle, lock the buffer and clean up > + * the resource and tt object. > */ > -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) > +static void ttm_bo_delayed_delete(struct work_struct *work) > { > - struct list_head removed; > - bool empty; > - > - INIT_LIST_HEAD(&removed); > - > - spin_lock(&bdev->lru_lock); > - while (!list_empty(&bdev->ddestroy)) { > - struct ttm_buffer_object *bo; > - > - bo = list_first_entry(&bdev->ddestroy, struct ttm_buffer_object, > - ddestroy); > - list_move_tail(&bo->ddestroy, &removed); > - if (!ttm_bo_get_unless_zero(bo)) > - continue; > - > - if (remove_all || bo->base.resv != &bo->base._resv) { > - spin_unlock(&bdev->lru_lock); > - dma_resv_lock(bo->base.resv, NULL); > - > - spin_lock(&bdev->lru_lock); > - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > - > - } else if (dma_resv_trylock(bo->base.resv)) { > - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > - } else { > - spin_unlock(&bdev->lru_lock); > - } > + struct ttm_buffer_object *bo; > > - ttm_bo_put(bo); > - spin_lock(&bdev->lru_lock); > - } > - list_splice_tail(&removed, &bdev->ddestroy); > - empty = list_empty(&bdev->ddestroy); > - spin_unlock(&bdev->lru_lock); > + bo = container_of(work, typeof(*bo), delayed_delete); > > - return empty; > + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, false, > + MAX_SCHEDULE_TIMEOUT); > + dma_resv_lock(bo->base.resv, NULL); > + ttm_bo_cleanup_memtype_use(bo); > + dma_resv_unlock(bo->base.resv); > + ttm_bo_put(bo); > } > > static void ttm_bo_release(struct kref *kref) > @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) > > drm_vma_offset_remove(bdev->vma_manager, &bo->base.vma_node); > ttm_mem_io_free(bdev, bo->resource); > - } > - > - if (!dma_resv_test_signaled(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP) || > - !dma_resv_trylock(bo->base.resv)) { > - /* The BO is not idle, resurrect it for delayed destroy */ > - ttm_bo_flush_all_fences(bo); > - bo->deleted = true; > > - spin_lock(&bo->bdev->lru_lock); > + if (!dma_resv_test_signaled(bo->base.resv, > + DMA_RESV_USAGE_BOOKKEEP) || > + !dma_resv_trylock(bo->base.resv)) { > + /* The BO is not idle, resurrect it for delayed destroy */ > + ttm_bo_flush_all_fences(bo); > + bo->deleted = true; > > - /* > - * Make pinned bos immediately available to > - * shrinkers, now that they are queued for > - * destruction. > - * > - * FIXME: QXL is triggering this. Can be removed when the > - * driver is fixed. > - */ > - if (bo->pin_count) { > - bo->pin_count = 0; > - ttm_resource_move_to_lru_tail(bo->resource); > - } > + spin_lock(&bo->bdev->lru_lock); > > - kref_init(&bo->kref); > - list_add_tail(&bo->ddestroy, &bdev->ddestroy); > - spin_unlock(&bo->bdev->lru_lock); > + /* > + * Make pinned bos immediately available to > + * shrinkers, now that they are queued for > + * destruction. > + * > + * FIXME: QXL is triggering this. Can be removed when the > + * driver is fixed. > + */ > + if (bo->pin_count) { > + bo->pin_count = 0; > + ttm_resource_move_to_lru_tail(bo->resource); > + } > > - schedule_delayed_work(&bdev->wq, > - ((HZ / 100) < 1) ? 1 : HZ / 100); > - return; > - } > + kref_init(&bo->kref); > + spin_unlock(&bo->bdev->lru_lock); > > - spin_lock(&bo->bdev->lru_lock); > - list_del(&bo->ddestroy); > - spin_unlock(&bo->bdev->lru_lock); > + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); > + queue_work(bdev->wq, &bo->delayed_delete); > + return; > + } > > - ttm_bo_cleanup_memtype_use(bo); > - dma_resv_unlock(bo->base.resv); > + ttm_bo_cleanup_memtype_use(bo); > + dma_resv_unlock(bo->base.resv); > + } > > atomic_dec(&ttm_glob.bo_count); > bo->destroy(bo); > @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, struct ttm_buffer_object *bo, > int ret; > > kref_init(&bo->kref); > - INIT_LIST_HEAD(&bo->ddestroy); > bo->bdev = bdev; > bo->type = type; > bo->page_alignment = alignment; > diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c > index ba3aa0a0fc43..ae4b7922ee1a 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct ttm_buffer_object *bo, > */ > > atomic_inc(&ttm_glob.bo_count); > - INIT_LIST_HEAD(&fbo->base.ddestroy); > drm_vma_node_reset(&fbo->base.base.vma_node); > > kref_init(&fbo->base.kref); > diff --git a/drivers/gpu/drm/ttm/ttm_device.c b/drivers/gpu/drm/ttm/ttm_device.c > index e7147e304637..e9bedca4dfdc 100644 > --- a/drivers/gpu/drm/ttm/ttm_device.c > +++ b/drivers/gpu/drm/ttm/ttm_device.c > @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, struct ttm_operation_ctx *ctx, > } > EXPORT_SYMBOL(ttm_device_swapout); > > -static void ttm_device_delayed_workqueue(struct work_struct *work) > -{ > - struct ttm_device *bdev = > - container_of(work, struct ttm_device, wq.work); > - > - if (!ttm_bo_delayed_delete(bdev, false)) > - schedule_delayed_work(&bdev->wq, > - ((HZ / 100) < 1) ? 1 : HZ / 100); > -} > - > /** > * ttm_device_init > * > @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, struct ttm_device_funcs *funcs, > if (ret) > return ret; > > + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, 16); > + if (!bdev->wq) { > + ttm_global_release(); > + return -ENOMEM; > + } > + > bdev->funcs = funcs; > > ttm_sys_man_init(bdev); > ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); > > bdev->vma_manager = vma_manager; > - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); > spin_lock_init(&bdev->lru_lock); > - INIT_LIST_HEAD(&bdev->ddestroy); > INIT_LIST_HEAD(&bdev->pinned); > bdev->dev_mapping = mapping; > mutex_lock(&ttm_global_mutex); > @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) > list_del(&bdev->device_list); > mutex_unlock(&ttm_global_mutex); > > - cancel_delayed_work_sync(&bdev->wq); > - > - if (ttm_bo_delayed_delete(bdev, true)) > - pr_debug("Delayed destroy list was clean\n"); > + drain_workqueue(bdev->wq); > + destroy_workqueue(bdev->wq); > > spin_lock(&bdev->lru_lock); > for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) > diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h > index 7758347c461c..69e62bbb01e3 100644 > --- a/include/drm/ttm/ttm_bo_api.h > +++ b/include/drm/ttm/ttm_bo_api.h > @@ -92,7 +92,6 @@ struct ttm_tt; > * @ttm: TTM structure holding system pages. > * @evicted: Whether the object was evicted without user-space knowing. > * @deleted: True if the object is only a zombie and already deleted. > - * @ddestroy: List head for the delayed destroy list. > * @swap: List head for swap LRU list. > * @offset: The current GPU offset, which can have different meanings > * depending on the memory type. For SYSTEM type memory, it should be 0. > @@ -135,19 +134,14 @@ struct ttm_buffer_object { > struct ttm_tt *ttm; > bool deleted; > struct ttm_lru_bulk_move *bulk_move; > + unsigned priority; > + unsigned pin_count; > > /** > - * Members protected by the bdev::lru_lock. > - */ > - > - struct list_head ddestroy; > - > - /** > - * Members protected by a bo reservation. > + * @delayed_delete: Work item used when we can't delete the BO > + * immediately > */ > - > - unsigned priority; > - unsigned pin_count; > + struct work_struct delayed_delete; > > /** > * Special members that are protected by the reserve lock > @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); > > int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, > void *buf, int len, int write); > -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); > - > vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); > > #endif > diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h > index 95b3c04b1ab9..4f3e81eac6f3 100644 > --- a/include/drm/ttm/ttm_device.h > +++ b/include/drm/ttm/ttm_device.h > @@ -251,11 +251,6 @@ struct ttm_device { > */ > spinlock_t lru_lock; > > - /** > - * @ddestroy: Destroyed but not yet cleaned up buffer objects. > - */ > - struct list_head ddestroy; > - > /** > * @pinned: Buffer objects which are pinned and so not on any LRU list. > */ > @@ -270,7 +265,7 @@ struct ttm_device { > /** > * @wq: Work queue structure for the delayed delete workqueue. > */ > - struct delayed_work wq; > + struct workqueue_struct *wq; > }; > > int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t gfp_flags);
Am 29.11.22 um 22:14 schrieb Felix Kuehling: > On 2022-11-25 05:21, Christian König wrote: >> Instead of a single worker going over the list of delete BOs in regular >> intervals use a per BO worker which blocks for the resv object and >> locking of the BO. >> >> This not only simplifies the handling massively, but also results in >> much better response time when cleaning up buffers. >> >> Signed-off-by: Christian König <christian.koenig@amd.com> > > Just thinking out loud: If I understand it correctly, this can cause a > lot of sleeping worker threads when > AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE is used and many BOs are freed > at the same time. This happens e.g. when a KFD process terminates or > crashes. I guess with a concurrency-managed workqueue this isn't going > to be excessive. And since it's on a per device workqueue, it doesn't > stall work items on the system work queue or from other devices. Yes, exactly that. The last parameter to alloc_workqueue() limits how many work items can be sleeping. > I'm trying to understand why you set WQ_MEM_RECLAIM. This work queue > is not about freeing ttm_resources but about freeing the BOs. But it > affects freeing of ghost_objs that are holding the ttm_resources being > freed. Well if the BO is idle, but not immediately lockable we delegate freeing the backing pages in the TT object to those workers as well. It might even be a good idea to use a separate wq for this case. > > If those assumptions all make sense, patches 1-3 are > > Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> Thanks, Christian. > > >> --- >> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- >> drivers/gpu/drm/i915/i915_gem.c | 2 +- >> drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- >> drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- >> drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - >> drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- >> include/drm/ttm/ttm_bo_api.h | 18 +--- >> include/drm/ttm/ttm_device.h | 7 +- >> 8 files changed, 57 insertions(+), 111 deletions(-) >> >> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> index 2b1db37e25c1..74ccbd566777 100644 >> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device >> *adev) >> amdgpu_fence_driver_hw_fini(adev); >> if (adev->mman.initialized) >> - flush_delayed_work(&adev->mman.bdev.wq); >> + drain_workqueue(adev->mman.bdev.wq); >> if (adev->pm_sysfs_en) >> amdgpu_pm_sysfs_fini(adev); >> diff --git a/drivers/gpu/drm/i915/i915_gem.c >> b/drivers/gpu/drm/i915/i915_gem.c >> index 8468ca9885fd..c38306f156d6 100644 >> --- a/drivers/gpu/drm/i915/i915_gem.c >> +++ b/drivers/gpu/drm/i915/i915_gem.c >> @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct >> drm_i915_private *i915) >> { >> while (atomic_read(&i915->mm.free_count)) { >> flush_work(&i915->mm.free_work); >> - flush_delayed_work(&i915->bdev.wq); >> + drain_workqueue(i915->bdev.wq); >> rcu_barrier(); >> } >> } >> diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c >> b/drivers/gpu/drm/i915/intel_region_ttm.c >> index cf89d0c2a2d9..657bbc16a48a 100644 >> --- a/drivers/gpu/drm/i915/intel_region_ttm.c >> +++ b/drivers/gpu/drm/i915/intel_region_ttm.c >> @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct >> intel_memory_region *mem) >> break; >> msleep(20); >> - flush_delayed_work(&mem->i915->bdev.wq); >> + drain_workqueue(mem->i915->bdev.wq); >> } >> /* If we leaked objects, Don't free the region causing use >> after free */ >> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c >> index b77262a623e0..4749b65bedc4 100644 >> --- a/drivers/gpu/drm/ttm/ttm_bo.c >> +++ b/drivers/gpu/drm/ttm/ttm_bo.c >> @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct >> ttm_buffer_object *bo, >> ret = 0; >> } >> - if (ret || unlikely(list_empty(&bo->ddestroy))) { >> + if (ret) { >> if (unlock_resv) >> dma_resv_unlock(bo->base.resv); >> spin_unlock(&bo->bdev->lru_lock); >> return ret; >> } >> - list_del_init(&bo->ddestroy); >> spin_unlock(&bo->bdev->lru_lock); >> ttm_bo_cleanup_memtype_use(bo); >> @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct >> ttm_buffer_object *bo, >> } >> /* >> - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all >> - * encountered buffers. >> + * Block for the dma_resv object to become idle, lock the buffer and >> clean up >> + * the resource and tt object. >> */ >> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) >> +static void ttm_bo_delayed_delete(struct work_struct *work) >> { >> - struct list_head removed; >> - bool empty; >> - >> - INIT_LIST_HEAD(&removed); >> - >> - spin_lock(&bdev->lru_lock); >> - while (!list_empty(&bdev->ddestroy)) { >> - struct ttm_buffer_object *bo; >> - >> - bo = list_first_entry(&bdev->ddestroy, struct >> ttm_buffer_object, >> - ddestroy); >> - list_move_tail(&bo->ddestroy, &removed); >> - if (!ttm_bo_get_unless_zero(bo)) >> - continue; >> - >> - if (remove_all || bo->base.resv != &bo->base._resv) { >> - spin_unlock(&bdev->lru_lock); >> - dma_resv_lock(bo->base.resv, NULL); >> - >> - spin_lock(&bdev->lru_lock); >> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); >> - >> - } else if (dma_resv_trylock(bo->base.resv)) { >> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); >> - } else { >> - spin_unlock(&bdev->lru_lock); >> - } >> + struct ttm_buffer_object *bo; >> - ttm_bo_put(bo); >> - spin_lock(&bdev->lru_lock); >> - } >> - list_splice_tail(&removed, &bdev->ddestroy); >> - empty = list_empty(&bdev->ddestroy); >> - spin_unlock(&bdev->lru_lock); >> + bo = container_of(work, typeof(*bo), delayed_delete); >> - return empty; >> + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, >> false, >> + MAX_SCHEDULE_TIMEOUT); >> + dma_resv_lock(bo->base.resv, NULL); >> + ttm_bo_cleanup_memtype_use(bo); >> + dma_resv_unlock(bo->base.resv); >> + ttm_bo_put(bo); >> } >> static void ttm_bo_release(struct kref *kref) >> @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) >> drm_vma_offset_remove(bdev->vma_manager, >> &bo->base.vma_node); >> ttm_mem_io_free(bdev, bo->resource); >> - } >> - >> - if (!dma_resv_test_signaled(bo->base.resv, >> DMA_RESV_USAGE_BOOKKEEP) || >> - !dma_resv_trylock(bo->base.resv)) { >> - /* The BO is not idle, resurrect it for delayed destroy */ >> - ttm_bo_flush_all_fences(bo); >> - bo->deleted = true; >> - spin_lock(&bo->bdev->lru_lock); >> + if (!dma_resv_test_signaled(bo->base.resv, >> + DMA_RESV_USAGE_BOOKKEEP) || >> + !dma_resv_trylock(bo->base.resv)) { >> + /* The BO is not idle, resurrect it for delayed destroy */ >> + ttm_bo_flush_all_fences(bo); >> + bo->deleted = true; >> - /* >> - * Make pinned bos immediately available to >> - * shrinkers, now that they are queued for >> - * destruction. >> - * >> - * FIXME: QXL is triggering this. Can be removed when the >> - * driver is fixed. >> - */ >> - if (bo->pin_count) { >> - bo->pin_count = 0; >> - ttm_resource_move_to_lru_tail(bo->resource); >> - } >> + spin_lock(&bo->bdev->lru_lock); >> - kref_init(&bo->kref); >> - list_add_tail(&bo->ddestroy, &bdev->ddestroy); >> - spin_unlock(&bo->bdev->lru_lock); >> + /* >> + * Make pinned bos immediately available to >> + * shrinkers, now that they are queued for >> + * destruction. >> + * >> + * FIXME: QXL is triggering this. Can be removed when the >> + * driver is fixed. >> + */ >> + if (bo->pin_count) { >> + bo->pin_count = 0; >> + ttm_resource_move_to_lru_tail(bo->resource); >> + } >> - schedule_delayed_work(&bdev->wq, >> - ((HZ / 100) < 1) ? 1 : HZ / 100); >> - return; >> - } >> + kref_init(&bo->kref); >> + spin_unlock(&bo->bdev->lru_lock); >> - spin_lock(&bo->bdev->lru_lock); >> - list_del(&bo->ddestroy); >> - spin_unlock(&bo->bdev->lru_lock); >> + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); >> + queue_work(bdev->wq, &bo->delayed_delete); >> + return; >> + } >> - ttm_bo_cleanup_memtype_use(bo); >> - dma_resv_unlock(bo->base.resv); >> + ttm_bo_cleanup_memtype_use(bo); >> + dma_resv_unlock(bo->base.resv); >> + } >> atomic_dec(&ttm_glob.bo_count); >> bo->destroy(bo); >> @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, >> struct ttm_buffer_object *bo, >> int ret; >> kref_init(&bo->kref); >> - INIT_LIST_HEAD(&bo->ddestroy); >> bo->bdev = bdev; >> bo->type = type; >> bo->page_alignment = alignment; >> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c >> b/drivers/gpu/drm/ttm/ttm_bo_util.c >> index ba3aa0a0fc43..ae4b7922ee1a 100644 >> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c >> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c >> @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct >> ttm_buffer_object *bo, >> */ >> atomic_inc(&ttm_glob.bo_count); >> - INIT_LIST_HEAD(&fbo->base.ddestroy); >> drm_vma_node_reset(&fbo->base.base.vma_node); >> kref_init(&fbo->base.kref); >> diff --git a/drivers/gpu/drm/ttm/ttm_device.c >> b/drivers/gpu/drm/ttm/ttm_device.c >> index e7147e304637..e9bedca4dfdc 100644 >> --- a/drivers/gpu/drm/ttm/ttm_device.c >> +++ b/drivers/gpu/drm/ttm/ttm_device.c >> @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, >> struct ttm_operation_ctx *ctx, >> } >> EXPORT_SYMBOL(ttm_device_swapout); >> -static void ttm_device_delayed_workqueue(struct work_struct *work) >> -{ >> - struct ttm_device *bdev = >> - container_of(work, struct ttm_device, wq.work); >> - >> - if (!ttm_bo_delayed_delete(bdev, false)) >> - schedule_delayed_work(&bdev->wq, >> - ((HZ / 100) < 1) ? 1 : HZ / 100); >> -} >> - >> /** >> * ttm_device_init >> * >> @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, >> struct ttm_device_funcs *funcs, >> if (ret) >> return ret; >> + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, >> 16); >> + if (!bdev->wq) { >> + ttm_global_release(); >> + return -ENOMEM; >> + } >> + >> bdev->funcs = funcs; >> ttm_sys_man_init(bdev); >> ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); >> bdev->vma_manager = vma_manager; >> - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); >> spin_lock_init(&bdev->lru_lock); >> - INIT_LIST_HEAD(&bdev->ddestroy); >> INIT_LIST_HEAD(&bdev->pinned); >> bdev->dev_mapping = mapping; >> mutex_lock(&ttm_global_mutex); >> @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) >> list_del(&bdev->device_list); >> mutex_unlock(&ttm_global_mutex); >> - cancel_delayed_work_sync(&bdev->wq); >> - >> - if (ttm_bo_delayed_delete(bdev, true)) >> - pr_debug("Delayed destroy list was clean\n"); >> + drain_workqueue(bdev->wq); >> + destroy_workqueue(bdev->wq); >> spin_lock(&bdev->lru_lock); >> for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) >> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h >> index 7758347c461c..69e62bbb01e3 100644 >> --- a/include/drm/ttm/ttm_bo_api.h >> +++ b/include/drm/ttm/ttm_bo_api.h >> @@ -92,7 +92,6 @@ struct ttm_tt; >> * @ttm: TTM structure holding system pages. >> * @evicted: Whether the object was evicted without user-space >> knowing. >> * @deleted: True if the object is only a zombie and already deleted. >> - * @ddestroy: List head for the delayed destroy list. >> * @swap: List head for swap LRU list. >> * @offset: The current GPU offset, which can have different meanings >> * depending on the memory type. For SYSTEM type memory, it should >> be 0. >> @@ -135,19 +134,14 @@ struct ttm_buffer_object { >> struct ttm_tt *ttm; >> bool deleted; >> struct ttm_lru_bulk_move *bulk_move; >> + unsigned priority; >> + unsigned pin_count; >> /** >> - * Members protected by the bdev::lru_lock. >> - */ >> - >> - struct list_head ddestroy; >> - >> - /** >> - * Members protected by a bo reservation. >> + * @delayed_delete: Work item used when we can't delete the BO >> + * immediately >> */ >> - >> - unsigned priority; >> - unsigned pin_count; >> + struct work_struct delayed_delete; >> /** >> * Special members that are protected by the reserve lock >> @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); >> int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, >> void *buf, int len, int write); >> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); >> - >> vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); >> #endif >> diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h >> index 95b3c04b1ab9..4f3e81eac6f3 100644 >> --- a/include/drm/ttm/ttm_device.h >> +++ b/include/drm/ttm/ttm_device.h >> @@ -251,11 +251,6 @@ struct ttm_device { >> */ >> spinlock_t lru_lock; >> - /** >> - * @ddestroy: Destroyed but not yet cleaned up buffer objects. >> - */ >> - struct list_head ddestroy; >> - >> /** >> * @pinned: Buffer objects which are pinned and so not on any >> LRU list. >> */ >> @@ -270,7 +265,7 @@ struct ttm_device { >> /** >> * @wq: Work queue structure for the delayed delete workqueue. >> */ >> - struct delayed_work wq; >> + struct workqueue_struct *wq; >> }; >> int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t >> gfp_flags);
On Mon, Dec 5, 2022 at 2:40 PM Christian König <christian.koenig@amd.com> wrote: > > Am 29.11.22 um 22:14 schrieb Felix Kuehling: > > On 2022-11-25 05:21, Christian König wrote: > >> Instead of a single worker going over the list of delete BOs in regular > >> intervals use a per BO worker which blocks for the resv object and > >> locking of the BO. > >> > >> This not only simplifies the handling massively, but also results in > >> much better response time when cleaning up buffers. > >> > >> Signed-off-by: Christian König <christian.koenig@amd.com> > > > > Just thinking out loud: If I understand it correctly, this can cause a > > lot of sleeping worker threads when > > AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE is used and many BOs are freed > > at the same time. This happens e.g. when a KFD process terminates or > > crashes. I guess with a concurrency-managed workqueue this isn't going > > to be excessive. And since it's on a per device workqueue, it doesn't > > stall work items on the system work queue or from other devices. > > Yes, exactly that. The last parameter to alloc_workqueue() limits how > many work items can be sleeping. > > > I'm trying to understand why you set WQ_MEM_RECLAIM. This work queue > > is not about freeing ttm_resources but about freeing the BOs. But it > > affects freeing of ghost_objs that are holding the ttm_resources being > > freed. > > Well if the BO is idle, but not immediately lockable we delegate freeing > the backing pages in the TT object to those workers as well. It might > even be a good idea to use a separate wq for this case. > > > > > If those assumptions all make sense, patches 1-3 are > > > > Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> > > Thanks, > Christian. > This patch causes a heap use-after-free when using nouveau with the potential of trashing filesystems, is there a way to revert it until we figure out a proper solution to the problem? Bug: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213 example trace on affected systems: [ 4102.946946] general protection fault, probably for non-canonical address 0x5f775ce3bd949b45: 0000 [#3] PREEMPT SMP NOPTI [ 4102.957794] CPU: 12 PID: 89561 Comm: glcts Tainted: G D 6.3.5-200.fc38.x86_64 #1 [ 4102.966556] Hardware name: ASUS System Product Name/PRIME B660-PLUS D4, BIOS 0418 10/13/2021 [ 4102.974972] RIP: 0010:__kmem_cache_alloc_node+0x1ba/0x320 [ 4102.980362] Code: 2b 14 25 28 00 00 00 0f 85 74 01 00 00 48 83 c4 18 5b 5d 41 5c 41 5d 41 5e 41 5f c3 cc cc cc cc 41 8b 47 28 4d 8b 07 48 01 f8 <48> 8b 18 48 89 c1 49 33 9f b8 00 00 00 48 0f c9 48 31 cb 41 f6 c0 [ 4102.999073] RSP: 0018:ffff9764e0057b40 EFLAGS: 00010202 [ 4103.004291] RAX: 5f775ce3bd949b45 RBX: 0000000000000dc0 RCX: 0000000000000046 [ 4103.011408] RDX: 00000002cf87600c RSI: 0000000000000dc0 RDI: 5f775ce3bd949b15 [ 4103.018528] RBP: 0000000000000dc0 R08: 00000000000390c0 R09: 0000000030302d6d [ 4103.025649] R10: 00000000756c7473 R11: 0000000020090298 R12: 0000000000000000 [ 4103.032767] R13: 00000000ffffffff R14: 0000000000000046 R15: ffff8bda80042600 [ 4103.039887] FS: 00007f386a85ef00(0000) GS:ffff8be1df700000(0000) knlGS:0000000000000000 [ 4103.047958] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 4103.053692] CR2: 000000000493b868 CR3: 000000014c3ba000 CR4: 0000000000f50ee0 [ 4103.060812] PKRU: 55555554 [ 4103.063520] Call Trace: [ 4103.065970] <TASK> [ 4103.068071] ? die_addr+0x36/0x90 [ 4103.071384] ? exc_general_protection+0x1be/0x420 [ 4103.076081] ? asm_exc_general_protection+0x26/0x30 [ 4103.080952] ? __kmem_cache_alloc_node+0x1ba/0x320 [ 4103.085734] ? ext4_htree_store_dirent+0x42/0x180 [ 4103.090431] ? ext4_htree_store_dirent+0x42/0x180 [ 4103.095132] __kmalloc+0x4d/0x150 [ 4103.098444] ext4_htree_store_dirent+0x42/0x180 [ 4103.102970] htree_dirblock_to_tree+0x1ed/0x370 [ 4103.107494] ext4_htree_fill_tree+0x109/0x3d0 [ 4103.111846] ext4_readdir+0x6d4/0xa80 [ 4103.115505] iterate_dir+0x178/0x1c0 [ 4103.119076] __x64_sys_getdents64+0x88/0x130 [ 4103.123341] ? __pfx_filldir64+0x10/0x10 [ 4103.127260] do_syscall_64+0x5d/0x90 [ 4103.130835] ? handle_mm_fault+0x11e/0x310 [ 4103.134927] ? do_user_addr_fault+0x1e0/0x720 [ 4103.139278] ? exc_page_fault+0x7c/0x180 [ 4103.143195] entry_SYSCALL_64_after_hwframe+0x72/0xdc [ 4103.148240] RIP: 0033:0x7f386a418047 [ 4103.151828] Code: 24 fb ff 4c 89 e0 5b 41 5c 5d c3 0f 1f 84 00 00 00 00 00 f3 0f 1e fa b8 ff ff ff 7f 48 39 c2 48 0f 47 d0 b8 d9 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 01 c3 48 8b 15 91 cd 0f 00 f7 d8 64 89 02 48 [ 4103.170543] RSP: 002b:00007ffd4793ff38 EFLAGS: 00000293 ORIG_RAX: 00000000000000d9 [ 4103.178095] RAX: ffffffffffffffda RBX: 0000000004933830 RCX: 00007f386a418047 [ 4103.185214] RDX: 0000000000008000 RSI: 0000000004933860 RDI: 0000000000000006 [ 4103.192335] RBP: 00007ffd4793ff70 R08: 0000000000000000 R09: 0000000000000001 [ 4103.199454] R10: 0000000000000004 R11: 0000000000000293 R12: 0000000004933834 [ 4103.206573] R13: 0000000004933860 R14: ffffffffffffff60 R15: 0000000000000000 [ 4103.213695] </TASK> [ 4103.215883] Modules linked in: snd_seq_dummy snd_hrtimer nf_conntrack_netbios_ns nf_conntrack_broadcast nft_fib_inet nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 nf_reject_ipv6 nft_reject nft_ct nft_chain_nat ip6table_nat ip6table_mangle ip6table_raw ip6table [ 4103.215911] kvm_intel snd_hwdep snd_seq eeepc_wmi kvm snd_seq_device asus_wmi iTCO_wdt mei_pxp mei_hdcp ledtrig_audio irqbypass snd_pcm ee1004 intel_pmc_bxt sparse_keymap rapl snd_timer pmt_telemetry mei_me iTCO_vendor_support platform_profile joydev intel_cstate pmt_class snde [ 4103.366194] ---[ end trace 0000000000000000 ]--- > > > > > >> --- > >> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- > >> drivers/gpu/drm/i915/i915_gem.c | 2 +- > >> drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- > >> drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- > >> drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - > >> drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- > >> include/drm/ttm/ttm_bo_api.h | 18 +--- > >> include/drm/ttm/ttm_device.h | 7 +- > >> 8 files changed, 57 insertions(+), 111 deletions(-) > >> > >> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >> index 2b1db37e25c1..74ccbd566777 100644 > >> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >> @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device > >> *adev) > >> amdgpu_fence_driver_hw_fini(adev); > >> if (adev->mman.initialized) > >> - flush_delayed_work(&adev->mman.bdev.wq); > >> + drain_workqueue(adev->mman.bdev.wq); > >> if (adev->pm_sysfs_en) > >> amdgpu_pm_sysfs_fini(adev); > >> diff --git a/drivers/gpu/drm/i915/i915_gem.c > >> b/drivers/gpu/drm/i915/i915_gem.c > >> index 8468ca9885fd..c38306f156d6 100644 > >> --- a/drivers/gpu/drm/i915/i915_gem.c > >> +++ b/drivers/gpu/drm/i915/i915_gem.c > >> @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct > >> drm_i915_private *i915) > >> { > >> while (atomic_read(&i915->mm.free_count)) { > >> flush_work(&i915->mm.free_work); > >> - flush_delayed_work(&i915->bdev.wq); > >> + drain_workqueue(i915->bdev.wq); > >> rcu_barrier(); > >> } > >> } > >> diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c > >> b/drivers/gpu/drm/i915/intel_region_ttm.c > >> index cf89d0c2a2d9..657bbc16a48a 100644 > >> --- a/drivers/gpu/drm/i915/intel_region_ttm.c > >> +++ b/drivers/gpu/drm/i915/intel_region_ttm.c > >> @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct > >> intel_memory_region *mem) > >> break; > >> msleep(20); > >> - flush_delayed_work(&mem->i915->bdev.wq); > >> + drain_workqueue(mem->i915->bdev.wq); > >> } > >> /* If we leaked objects, Don't free the region causing use > >> after free */ > >> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > >> index b77262a623e0..4749b65bedc4 100644 > >> --- a/drivers/gpu/drm/ttm/ttm_bo.c > >> +++ b/drivers/gpu/drm/ttm/ttm_bo.c > >> @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct > >> ttm_buffer_object *bo, > >> ret = 0; > >> } > >> - if (ret || unlikely(list_empty(&bo->ddestroy))) { > >> + if (ret) { > >> if (unlock_resv) > >> dma_resv_unlock(bo->base.resv); > >> spin_unlock(&bo->bdev->lru_lock); > >> return ret; > >> } > >> - list_del_init(&bo->ddestroy); > >> spin_unlock(&bo->bdev->lru_lock); > >> ttm_bo_cleanup_memtype_use(bo); > >> @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct > >> ttm_buffer_object *bo, > >> } > >> /* > >> - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all > >> - * encountered buffers. > >> + * Block for the dma_resv object to become idle, lock the buffer and > >> clean up > >> + * the resource and tt object. > >> */ > >> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) > >> +static void ttm_bo_delayed_delete(struct work_struct *work) > >> { > >> - struct list_head removed; > >> - bool empty; > >> - > >> - INIT_LIST_HEAD(&removed); > >> - > >> - spin_lock(&bdev->lru_lock); > >> - while (!list_empty(&bdev->ddestroy)) { > >> - struct ttm_buffer_object *bo; > >> - > >> - bo = list_first_entry(&bdev->ddestroy, struct > >> ttm_buffer_object, > >> - ddestroy); > >> - list_move_tail(&bo->ddestroy, &removed); > >> - if (!ttm_bo_get_unless_zero(bo)) > >> - continue; > >> - > >> - if (remove_all || bo->base.resv != &bo->base._resv) { > >> - spin_unlock(&bdev->lru_lock); > >> - dma_resv_lock(bo->base.resv, NULL); > >> - > >> - spin_lock(&bdev->lru_lock); > >> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > >> - > >> - } else if (dma_resv_trylock(bo->base.resv)) { > >> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > >> - } else { > >> - spin_unlock(&bdev->lru_lock); > >> - } > >> + struct ttm_buffer_object *bo; > >> - ttm_bo_put(bo); > >> - spin_lock(&bdev->lru_lock); > >> - } > >> - list_splice_tail(&removed, &bdev->ddestroy); > >> - empty = list_empty(&bdev->ddestroy); > >> - spin_unlock(&bdev->lru_lock); > >> + bo = container_of(work, typeof(*bo), delayed_delete); > >> - return empty; > >> + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, > >> false, > >> + MAX_SCHEDULE_TIMEOUT); > >> + dma_resv_lock(bo->base.resv, NULL); > >> + ttm_bo_cleanup_memtype_use(bo); > >> + dma_resv_unlock(bo->base.resv); > >> + ttm_bo_put(bo); > >> } > >> static void ttm_bo_release(struct kref *kref) > >> @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) > >> drm_vma_offset_remove(bdev->vma_manager, > >> &bo->base.vma_node); > >> ttm_mem_io_free(bdev, bo->resource); > >> - } > >> - > >> - if (!dma_resv_test_signaled(bo->base.resv, > >> DMA_RESV_USAGE_BOOKKEEP) || > >> - !dma_resv_trylock(bo->base.resv)) { > >> - /* The BO is not idle, resurrect it for delayed destroy */ > >> - ttm_bo_flush_all_fences(bo); > >> - bo->deleted = true; > >> - spin_lock(&bo->bdev->lru_lock); > >> + if (!dma_resv_test_signaled(bo->base.resv, > >> + DMA_RESV_USAGE_BOOKKEEP) || > >> + !dma_resv_trylock(bo->base.resv)) { > >> + /* The BO is not idle, resurrect it for delayed destroy */ > >> + ttm_bo_flush_all_fences(bo); > >> + bo->deleted = true; > >> - /* > >> - * Make pinned bos immediately available to > >> - * shrinkers, now that they are queued for > >> - * destruction. > >> - * > >> - * FIXME: QXL is triggering this. Can be removed when the > >> - * driver is fixed. > >> - */ > >> - if (bo->pin_count) { > >> - bo->pin_count = 0; > >> - ttm_resource_move_to_lru_tail(bo->resource); > >> - } > >> + spin_lock(&bo->bdev->lru_lock); > >> - kref_init(&bo->kref); > >> - list_add_tail(&bo->ddestroy, &bdev->ddestroy); > >> - spin_unlock(&bo->bdev->lru_lock); > >> + /* > >> + * Make pinned bos immediately available to > >> + * shrinkers, now that they are queued for > >> + * destruction. > >> + * > >> + * FIXME: QXL is triggering this. Can be removed when the > >> + * driver is fixed. > >> + */ > >> + if (bo->pin_count) { > >> + bo->pin_count = 0; > >> + ttm_resource_move_to_lru_tail(bo->resource); > >> + } > >> - schedule_delayed_work(&bdev->wq, > >> - ((HZ / 100) < 1) ? 1 : HZ / 100); > >> - return; > >> - } > >> + kref_init(&bo->kref); > >> + spin_unlock(&bo->bdev->lru_lock); > >> - spin_lock(&bo->bdev->lru_lock); > >> - list_del(&bo->ddestroy); > >> - spin_unlock(&bo->bdev->lru_lock); > >> + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); > >> + queue_work(bdev->wq, &bo->delayed_delete); > >> + return; > >> + } > >> - ttm_bo_cleanup_memtype_use(bo); > >> - dma_resv_unlock(bo->base.resv); > >> + ttm_bo_cleanup_memtype_use(bo); > >> + dma_resv_unlock(bo->base.resv); > >> + } > >> atomic_dec(&ttm_glob.bo_count); > >> bo->destroy(bo); > >> @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, > >> struct ttm_buffer_object *bo, > >> int ret; > >> kref_init(&bo->kref); > >> - INIT_LIST_HEAD(&bo->ddestroy); > >> bo->bdev = bdev; > >> bo->type = type; > >> bo->page_alignment = alignment; > >> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c > >> b/drivers/gpu/drm/ttm/ttm_bo_util.c > >> index ba3aa0a0fc43..ae4b7922ee1a 100644 > >> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > >> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > >> @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct > >> ttm_buffer_object *bo, > >> */ > >> atomic_inc(&ttm_glob.bo_count); > >> - INIT_LIST_HEAD(&fbo->base.ddestroy); > >> drm_vma_node_reset(&fbo->base.base.vma_node); > >> kref_init(&fbo->base.kref); > >> diff --git a/drivers/gpu/drm/ttm/ttm_device.c > >> b/drivers/gpu/drm/ttm/ttm_device.c > >> index e7147e304637..e9bedca4dfdc 100644 > >> --- a/drivers/gpu/drm/ttm/ttm_device.c > >> +++ b/drivers/gpu/drm/ttm/ttm_device.c > >> @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, > >> struct ttm_operation_ctx *ctx, > >> } > >> EXPORT_SYMBOL(ttm_device_swapout); > >> -static void ttm_device_delayed_workqueue(struct work_struct *work) > >> -{ > >> - struct ttm_device *bdev = > >> - container_of(work, struct ttm_device, wq.work); > >> - > >> - if (!ttm_bo_delayed_delete(bdev, false)) > >> - schedule_delayed_work(&bdev->wq, > >> - ((HZ / 100) < 1) ? 1 : HZ / 100); > >> -} > >> - > >> /** > >> * ttm_device_init > >> * > >> @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, > >> struct ttm_device_funcs *funcs, > >> if (ret) > >> return ret; > >> + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, > >> 16); > >> + if (!bdev->wq) { > >> + ttm_global_release(); > >> + return -ENOMEM; > >> + } > >> + > >> bdev->funcs = funcs; > >> ttm_sys_man_init(bdev); > >> ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); > >> bdev->vma_manager = vma_manager; > >> - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); > >> spin_lock_init(&bdev->lru_lock); > >> - INIT_LIST_HEAD(&bdev->ddestroy); > >> INIT_LIST_HEAD(&bdev->pinned); > >> bdev->dev_mapping = mapping; > >> mutex_lock(&ttm_global_mutex); > >> @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) > >> list_del(&bdev->device_list); > >> mutex_unlock(&ttm_global_mutex); > >> - cancel_delayed_work_sync(&bdev->wq); > >> - > >> - if (ttm_bo_delayed_delete(bdev, true)) > >> - pr_debug("Delayed destroy list was clean\n"); > >> + drain_workqueue(bdev->wq); > >> + destroy_workqueue(bdev->wq); > >> spin_lock(&bdev->lru_lock); > >> for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) > >> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h > >> index 7758347c461c..69e62bbb01e3 100644 > >> --- a/include/drm/ttm/ttm_bo_api.h > >> +++ b/include/drm/ttm/ttm_bo_api.h > >> @@ -92,7 +92,6 @@ struct ttm_tt; > >> * @ttm: TTM structure holding system pages. > >> * @evicted: Whether the object was evicted without user-space > >> knowing. > >> * @deleted: True if the object is only a zombie and already deleted. > >> - * @ddestroy: List head for the delayed destroy list. > >> * @swap: List head for swap LRU list. > >> * @offset: The current GPU offset, which can have different meanings > >> * depending on the memory type. For SYSTEM type memory, it should > >> be 0. > >> @@ -135,19 +134,14 @@ struct ttm_buffer_object { > >> struct ttm_tt *ttm; > >> bool deleted; > >> struct ttm_lru_bulk_move *bulk_move; > >> + unsigned priority; > >> + unsigned pin_count; > >> /** > >> - * Members protected by the bdev::lru_lock. > >> - */ > >> - > >> - struct list_head ddestroy; > >> - > >> - /** > >> - * Members protected by a bo reservation. > >> + * @delayed_delete: Work item used when we can't delete the BO > >> + * immediately > >> */ > >> - > >> - unsigned priority; > >> - unsigned pin_count; > >> + struct work_struct delayed_delete; > >> /** > >> * Special members that are protected by the reserve lock > >> @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); > >> int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, > >> void *buf, int len, int write); > >> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); > >> - > >> vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); > >> #endif > >> diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h > >> index 95b3c04b1ab9..4f3e81eac6f3 100644 > >> --- a/include/drm/ttm/ttm_device.h > >> +++ b/include/drm/ttm/ttm_device.h > >> @@ -251,11 +251,6 @@ struct ttm_device { > >> */ > >> spinlock_t lru_lock; > >> - /** > >> - * @ddestroy: Destroyed but not yet cleaned up buffer objects. > >> - */ > >> - struct list_head ddestroy; > >> - > >> /** > >> * @pinned: Buffer objects which are pinned and so not on any > >> LRU list. > >> */ > >> @@ -270,7 +265,7 @@ struct ttm_device { > >> /** > >> * @wq: Work queue structure for the delayed delete workqueue. > >> */ > >> - struct delayed_work wq; > >> + struct workqueue_struct *wq; > >> }; > >> int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t > >> gfp_flags); >
Am 13.06.23 um 15:05 schrieb Karol Herbst: > On Mon, Dec 5, 2022 at 2:40 PM Christian König <christian.koenig@amd.com> wrote: >> Am 29.11.22 um 22:14 schrieb Felix Kuehling: >>> On 2022-11-25 05:21, Christian König wrote: >>>> Instead of a single worker going over the list of delete BOs in regular >>>> intervals use a per BO worker which blocks for the resv object and >>>> locking of the BO. >>>> >>>> This not only simplifies the handling massively, but also results in >>>> much better response time when cleaning up buffers. >>>> >>>> Signed-off-by: Christian König <christian.koenig@amd.com> >>> Just thinking out loud: If I understand it correctly, this can cause a >>> lot of sleeping worker threads when >>> AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE is used and many BOs are freed >>> at the same time. This happens e.g. when a KFD process terminates or >>> crashes. I guess with a concurrency-managed workqueue this isn't going >>> to be excessive. And since it's on a per device workqueue, it doesn't >>> stall work items on the system work queue or from other devices. >> Yes, exactly that. The last parameter to alloc_workqueue() limits how >> many work items can be sleeping. >> >>> I'm trying to understand why you set WQ_MEM_RECLAIM. This work queue >>> is not about freeing ttm_resources but about freeing the BOs. But it >>> affects freeing of ghost_objs that are holding the ttm_resources being >>> freed. >> Well if the BO is idle, but not immediately lockable we delegate freeing >> the backing pages in the TT object to those workers as well. It might >> even be a good idea to use a separate wq for this case. >> >>> If those assumptions all make sense, patches 1-3 are >>> >>> Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> >> Thanks, >> Christian. >> > This patch causes a heap use-after-free when using nouveau with the > potential of trashing filesystems, is there a way to revert it until > we figure out a proper solution to the problem? Uff I don't think so, we have quite some work based on top of this. But let me double check. On the other hand have you tried running this with KASAN to catch use after free errors? Since we now block for work to finish and not check every few milliseconds to garbage collect memory will now be reclaimed much faster after freeing it. Regards, Christian. > > Bug: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213 > > example trace on affected systems: > > [ 4102.946946] general protection fault, probably for non-canonical > address 0x5f775ce3bd949b45: 0000 [#3] PREEMPT SMP NOPTI > [ 4102.957794] CPU: 12 PID: 89561 Comm: glcts Tainted: G D > 6.3.5-200.fc38.x86_64 #1 > [ 4102.966556] Hardware name: ASUS System Product Name/PRIME B660-PLUS > D4, BIOS 0418 10/13/2021 > [ 4102.974972] RIP: 0010:__kmem_cache_alloc_node+0x1ba/0x320 > [ 4102.980362] Code: 2b 14 25 28 00 00 00 0f 85 74 01 00 00 48 83 c4 > 18 5b 5d 41 5c 41 5d 41 5e 41 5f c3 cc cc cc cc 41 8b 47 28 4d 8b 07 > 48 01 f8 <48> 8b 18 48 89 c1 49 33 9f b8 00 00 00 48 0f c9 48 31 cb 41 > f6 c0 > [ 4102.999073] RSP: 0018:ffff9764e0057b40 EFLAGS: 00010202 > [ 4103.004291] RAX: 5f775ce3bd949b45 RBX: 0000000000000dc0 RCX: 0000000000000046 > [ 4103.011408] RDX: 00000002cf87600c RSI: 0000000000000dc0 RDI: 5f775ce3bd949b15 > [ 4103.018528] RBP: 0000000000000dc0 R08: 00000000000390c0 R09: 0000000030302d6d > [ 4103.025649] R10: 00000000756c7473 R11: 0000000020090298 R12: 0000000000000000 > [ 4103.032767] R13: 00000000ffffffff R14: 0000000000000046 R15: ffff8bda80042600 > [ 4103.039887] FS: 00007f386a85ef00(0000) GS:ffff8be1df700000(0000) > knlGS:0000000000000000 > [ 4103.047958] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > [ 4103.053692] CR2: 000000000493b868 CR3: 000000014c3ba000 CR4: 0000000000f50ee0 > [ 4103.060812] PKRU: 55555554 > [ 4103.063520] Call Trace: > [ 4103.065970] <TASK> > [ 4103.068071] ? die_addr+0x36/0x90 > [ 4103.071384] ? exc_general_protection+0x1be/0x420 > [ 4103.076081] ? asm_exc_general_protection+0x26/0x30 > [ 4103.080952] ? __kmem_cache_alloc_node+0x1ba/0x320 > [ 4103.085734] ? ext4_htree_store_dirent+0x42/0x180 > [ 4103.090431] ? ext4_htree_store_dirent+0x42/0x180 > [ 4103.095132] __kmalloc+0x4d/0x150 > [ 4103.098444] ext4_htree_store_dirent+0x42/0x180 > [ 4103.102970] htree_dirblock_to_tree+0x1ed/0x370 > [ 4103.107494] ext4_htree_fill_tree+0x109/0x3d0 > [ 4103.111846] ext4_readdir+0x6d4/0xa80 > [ 4103.115505] iterate_dir+0x178/0x1c0 > [ 4103.119076] __x64_sys_getdents64+0x88/0x130 > [ 4103.123341] ? __pfx_filldir64+0x10/0x10 > [ 4103.127260] do_syscall_64+0x5d/0x90 > [ 4103.130835] ? handle_mm_fault+0x11e/0x310 > [ 4103.134927] ? do_user_addr_fault+0x1e0/0x720 > [ 4103.139278] ? exc_page_fault+0x7c/0x180 > [ 4103.143195] entry_SYSCALL_64_after_hwframe+0x72/0xdc > [ 4103.148240] RIP: 0033:0x7f386a418047 > [ 4103.151828] Code: 24 fb ff 4c 89 e0 5b 41 5c 5d c3 0f 1f 84 00 00 > 00 00 00 f3 0f 1e fa b8 ff ff ff 7f 48 39 c2 48 0f 47 d0 b8 d9 00 00 > 00 0f 05 <48> 3d 00 f0 ff ff 77 01 c3 48 8b 15 91 cd 0f 00 f7 d8 64 89 > 02 48 > [ 4103.170543] RSP: 002b:00007ffd4793ff38 EFLAGS: 00000293 ORIG_RAX: > 00000000000000d9 > [ 4103.178095] RAX: ffffffffffffffda RBX: 0000000004933830 RCX: 00007f386a418047 > [ 4103.185214] RDX: 0000000000008000 RSI: 0000000004933860 RDI: 0000000000000006 > [ 4103.192335] RBP: 00007ffd4793ff70 R08: 0000000000000000 R09: 0000000000000001 > [ 4103.199454] R10: 0000000000000004 R11: 0000000000000293 R12: 0000000004933834 > [ 4103.206573] R13: 0000000004933860 R14: ffffffffffffff60 R15: 0000000000000000 > [ 4103.213695] </TASK> > [ 4103.215883] Modules linked in: snd_seq_dummy snd_hrtimer > nf_conntrack_netbios_ns nf_conntrack_broadcast nft_fib_inet > nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 > nf_reject_ipv6 nft_reject nft_ct nft_chain_nat ip6table_nat > ip6table_mangle ip6table_raw ip6table > [ 4103.215911] kvm_intel snd_hwdep snd_seq eeepc_wmi kvm > snd_seq_device asus_wmi iTCO_wdt mei_pxp mei_hdcp ledtrig_audio > irqbypass snd_pcm ee1004 intel_pmc_bxt sparse_keymap rapl snd_timer > pmt_telemetry mei_me iTCO_vendor_support platform_profile joydev > intel_cstate pmt_class snde > [ 4103.366194] ---[ end trace 0000000000000000 ]--- > >>> >>>> --- >>>> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- >>>> drivers/gpu/drm/i915/i915_gem.c | 2 +- >>>> drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- >>>> drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- >>>> drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - >>>> drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- >>>> include/drm/ttm/ttm_bo_api.h | 18 +--- >>>> include/drm/ttm/ttm_device.h | 7 +- >>>> 8 files changed, 57 insertions(+), 111 deletions(-) >>>> >>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>> index 2b1db37e25c1..74ccbd566777 100644 >>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>> @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device >>>> *adev) >>>> amdgpu_fence_driver_hw_fini(adev); >>>> if (adev->mman.initialized) >>>> - flush_delayed_work(&adev->mman.bdev.wq); >>>> + drain_workqueue(adev->mman.bdev.wq); >>>> if (adev->pm_sysfs_en) >>>> amdgpu_pm_sysfs_fini(adev); >>>> diff --git a/drivers/gpu/drm/i915/i915_gem.c >>>> b/drivers/gpu/drm/i915/i915_gem.c >>>> index 8468ca9885fd..c38306f156d6 100644 >>>> --- a/drivers/gpu/drm/i915/i915_gem.c >>>> +++ b/drivers/gpu/drm/i915/i915_gem.c >>>> @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct >>>> drm_i915_private *i915) >>>> { >>>> while (atomic_read(&i915->mm.free_count)) { >>>> flush_work(&i915->mm.free_work); >>>> - flush_delayed_work(&i915->bdev.wq); >>>> + drain_workqueue(i915->bdev.wq); >>>> rcu_barrier(); >>>> } >>>> } >>>> diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c >>>> b/drivers/gpu/drm/i915/intel_region_ttm.c >>>> index cf89d0c2a2d9..657bbc16a48a 100644 >>>> --- a/drivers/gpu/drm/i915/intel_region_ttm.c >>>> +++ b/drivers/gpu/drm/i915/intel_region_ttm.c >>>> @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct >>>> intel_memory_region *mem) >>>> break; >>>> msleep(20); >>>> - flush_delayed_work(&mem->i915->bdev.wq); >>>> + drain_workqueue(mem->i915->bdev.wq); >>>> } >>>> /* If we leaked objects, Don't free the region causing use >>>> after free */ >>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c >>>> index b77262a623e0..4749b65bedc4 100644 >>>> --- a/drivers/gpu/drm/ttm/ttm_bo.c >>>> +++ b/drivers/gpu/drm/ttm/ttm_bo.c >>>> @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct >>>> ttm_buffer_object *bo, >>>> ret = 0; >>>> } >>>> - if (ret || unlikely(list_empty(&bo->ddestroy))) { >>>> + if (ret) { >>>> if (unlock_resv) >>>> dma_resv_unlock(bo->base.resv); >>>> spin_unlock(&bo->bdev->lru_lock); >>>> return ret; >>>> } >>>> - list_del_init(&bo->ddestroy); >>>> spin_unlock(&bo->bdev->lru_lock); >>>> ttm_bo_cleanup_memtype_use(bo); >>>> @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct >>>> ttm_buffer_object *bo, >>>> } >>>> /* >>>> - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all >>>> - * encountered buffers. >>>> + * Block for the dma_resv object to become idle, lock the buffer and >>>> clean up >>>> + * the resource and tt object. >>>> */ >>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) >>>> +static void ttm_bo_delayed_delete(struct work_struct *work) >>>> { >>>> - struct list_head removed; >>>> - bool empty; >>>> - >>>> - INIT_LIST_HEAD(&removed); >>>> - >>>> - spin_lock(&bdev->lru_lock); >>>> - while (!list_empty(&bdev->ddestroy)) { >>>> - struct ttm_buffer_object *bo; >>>> - >>>> - bo = list_first_entry(&bdev->ddestroy, struct >>>> ttm_buffer_object, >>>> - ddestroy); >>>> - list_move_tail(&bo->ddestroy, &removed); >>>> - if (!ttm_bo_get_unless_zero(bo)) >>>> - continue; >>>> - >>>> - if (remove_all || bo->base.resv != &bo->base._resv) { >>>> - spin_unlock(&bdev->lru_lock); >>>> - dma_resv_lock(bo->base.resv, NULL); >>>> - >>>> - spin_lock(&bdev->lru_lock); >>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); >>>> - >>>> - } else if (dma_resv_trylock(bo->base.resv)) { >>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); >>>> - } else { >>>> - spin_unlock(&bdev->lru_lock); >>>> - } >>>> + struct ttm_buffer_object *bo; >>>> - ttm_bo_put(bo); >>>> - spin_lock(&bdev->lru_lock); >>>> - } >>>> - list_splice_tail(&removed, &bdev->ddestroy); >>>> - empty = list_empty(&bdev->ddestroy); >>>> - spin_unlock(&bdev->lru_lock); >>>> + bo = container_of(work, typeof(*bo), delayed_delete); >>>> - return empty; >>>> + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, >>>> false, >>>> + MAX_SCHEDULE_TIMEOUT); >>>> + dma_resv_lock(bo->base.resv, NULL); >>>> + ttm_bo_cleanup_memtype_use(bo); >>>> + dma_resv_unlock(bo->base.resv); >>>> + ttm_bo_put(bo); >>>> } >>>> static void ttm_bo_release(struct kref *kref) >>>> @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) >>>> drm_vma_offset_remove(bdev->vma_manager, >>>> &bo->base.vma_node); >>>> ttm_mem_io_free(bdev, bo->resource); >>>> - } >>>> - >>>> - if (!dma_resv_test_signaled(bo->base.resv, >>>> DMA_RESV_USAGE_BOOKKEEP) || >>>> - !dma_resv_trylock(bo->base.resv)) { >>>> - /* The BO is not idle, resurrect it for delayed destroy */ >>>> - ttm_bo_flush_all_fences(bo); >>>> - bo->deleted = true; >>>> - spin_lock(&bo->bdev->lru_lock); >>>> + if (!dma_resv_test_signaled(bo->base.resv, >>>> + DMA_RESV_USAGE_BOOKKEEP) || >>>> + !dma_resv_trylock(bo->base.resv)) { >>>> + /* The BO is not idle, resurrect it for delayed destroy */ >>>> + ttm_bo_flush_all_fences(bo); >>>> + bo->deleted = true; >>>> - /* >>>> - * Make pinned bos immediately available to >>>> - * shrinkers, now that they are queued for >>>> - * destruction. >>>> - * >>>> - * FIXME: QXL is triggering this. Can be removed when the >>>> - * driver is fixed. >>>> - */ >>>> - if (bo->pin_count) { >>>> - bo->pin_count = 0; >>>> - ttm_resource_move_to_lru_tail(bo->resource); >>>> - } >>>> + spin_lock(&bo->bdev->lru_lock); >>>> - kref_init(&bo->kref); >>>> - list_add_tail(&bo->ddestroy, &bdev->ddestroy); >>>> - spin_unlock(&bo->bdev->lru_lock); >>>> + /* >>>> + * Make pinned bos immediately available to >>>> + * shrinkers, now that they are queued for >>>> + * destruction. >>>> + * >>>> + * FIXME: QXL is triggering this. Can be removed when the >>>> + * driver is fixed. >>>> + */ >>>> + if (bo->pin_count) { >>>> + bo->pin_count = 0; >>>> + ttm_resource_move_to_lru_tail(bo->resource); >>>> + } >>>> - schedule_delayed_work(&bdev->wq, >>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); >>>> - return; >>>> - } >>>> + kref_init(&bo->kref); >>>> + spin_unlock(&bo->bdev->lru_lock); >>>> - spin_lock(&bo->bdev->lru_lock); >>>> - list_del(&bo->ddestroy); >>>> - spin_unlock(&bo->bdev->lru_lock); >>>> + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); >>>> + queue_work(bdev->wq, &bo->delayed_delete); >>>> + return; >>>> + } >>>> - ttm_bo_cleanup_memtype_use(bo); >>>> - dma_resv_unlock(bo->base.resv); >>>> + ttm_bo_cleanup_memtype_use(bo); >>>> + dma_resv_unlock(bo->base.resv); >>>> + } >>>> atomic_dec(&ttm_glob.bo_count); >>>> bo->destroy(bo); >>>> @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, >>>> struct ttm_buffer_object *bo, >>>> int ret; >>>> kref_init(&bo->kref); >>>> - INIT_LIST_HEAD(&bo->ddestroy); >>>> bo->bdev = bdev; >>>> bo->type = type; >>>> bo->page_alignment = alignment; >>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c >>>> b/drivers/gpu/drm/ttm/ttm_bo_util.c >>>> index ba3aa0a0fc43..ae4b7922ee1a 100644 >>>> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c >>>> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c >>>> @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct >>>> ttm_buffer_object *bo, >>>> */ >>>> atomic_inc(&ttm_glob.bo_count); >>>> - INIT_LIST_HEAD(&fbo->base.ddestroy); >>>> drm_vma_node_reset(&fbo->base.base.vma_node); >>>> kref_init(&fbo->base.kref); >>>> diff --git a/drivers/gpu/drm/ttm/ttm_device.c >>>> b/drivers/gpu/drm/ttm/ttm_device.c >>>> index e7147e304637..e9bedca4dfdc 100644 >>>> --- a/drivers/gpu/drm/ttm/ttm_device.c >>>> +++ b/drivers/gpu/drm/ttm/ttm_device.c >>>> @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, >>>> struct ttm_operation_ctx *ctx, >>>> } >>>> EXPORT_SYMBOL(ttm_device_swapout); >>>> -static void ttm_device_delayed_workqueue(struct work_struct *work) >>>> -{ >>>> - struct ttm_device *bdev = >>>> - container_of(work, struct ttm_device, wq.work); >>>> - >>>> - if (!ttm_bo_delayed_delete(bdev, false)) >>>> - schedule_delayed_work(&bdev->wq, >>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); >>>> -} >>>> - >>>> /** >>>> * ttm_device_init >>>> * >>>> @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, >>>> struct ttm_device_funcs *funcs, >>>> if (ret) >>>> return ret; >>>> + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, >>>> 16); >>>> + if (!bdev->wq) { >>>> + ttm_global_release(); >>>> + return -ENOMEM; >>>> + } >>>> + >>>> bdev->funcs = funcs; >>>> ttm_sys_man_init(bdev); >>>> ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); >>>> bdev->vma_manager = vma_manager; >>>> - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); >>>> spin_lock_init(&bdev->lru_lock); >>>> - INIT_LIST_HEAD(&bdev->ddestroy); >>>> INIT_LIST_HEAD(&bdev->pinned); >>>> bdev->dev_mapping = mapping; >>>> mutex_lock(&ttm_global_mutex); >>>> @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) >>>> list_del(&bdev->device_list); >>>> mutex_unlock(&ttm_global_mutex); >>>> - cancel_delayed_work_sync(&bdev->wq); >>>> - >>>> - if (ttm_bo_delayed_delete(bdev, true)) >>>> - pr_debug("Delayed destroy list was clean\n"); >>>> + drain_workqueue(bdev->wq); >>>> + destroy_workqueue(bdev->wq); >>>> spin_lock(&bdev->lru_lock); >>>> for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) >>>> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h >>>> index 7758347c461c..69e62bbb01e3 100644 >>>> --- a/include/drm/ttm/ttm_bo_api.h >>>> +++ b/include/drm/ttm/ttm_bo_api.h >>>> @@ -92,7 +92,6 @@ struct ttm_tt; >>>> * @ttm: TTM structure holding system pages. >>>> * @evicted: Whether the object was evicted without user-space >>>> knowing. >>>> * @deleted: True if the object is only a zombie and already deleted. >>>> - * @ddestroy: List head for the delayed destroy list. >>>> * @swap: List head for swap LRU list. >>>> * @offset: The current GPU offset, which can have different meanings >>>> * depending on the memory type. For SYSTEM type memory, it should >>>> be 0. >>>> @@ -135,19 +134,14 @@ struct ttm_buffer_object { >>>> struct ttm_tt *ttm; >>>> bool deleted; >>>> struct ttm_lru_bulk_move *bulk_move; >>>> + unsigned priority; >>>> + unsigned pin_count; >>>> /** >>>> - * Members protected by the bdev::lru_lock. >>>> - */ >>>> - >>>> - struct list_head ddestroy; >>>> - >>>> - /** >>>> - * Members protected by a bo reservation. >>>> + * @delayed_delete: Work item used when we can't delete the BO >>>> + * immediately >>>> */ >>>> - >>>> - unsigned priority; >>>> - unsigned pin_count; >>>> + struct work_struct delayed_delete; >>>> /** >>>> * Special members that are protected by the reserve lock >>>> @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); >>>> int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, >>>> void *buf, int len, int write); >>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); >>>> - >>>> vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); >>>> #endif >>>> diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h >>>> index 95b3c04b1ab9..4f3e81eac6f3 100644 >>>> --- a/include/drm/ttm/ttm_device.h >>>> +++ b/include/drm/ttm/ttm_device.h >>>> @@ -251,11 +251,6 @@ struct ttm_device { >>>> */ >>>> spinlock_t lru_lock; >>>> - /** >>>> - * @ddestroy: Destroyed but not yet cleaned up buffer objects. >>>> - */ >>>> - struct list_head ddestroy; >>>> - >>>> /** >>>> * @pinned: Buffer objects which are pinned and so not on any >>>> LRU list. >>>> */ >>>> @@ -270,7 +265,7 @@ struct ttm_device { >>>> /** >>>> * @wq: Work queue structure for the delayed delete workqueue. >>>> */ >>>> - struct delayed_work wq; >>>> + struct workqueue_struct *wq; >>>> }; >>>> int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t >>>> gfp_flags);
On Tue, Jun 13, 2023 at 3:59 PM Christian König <christian.koenig@amd.com> wrote: > > Am 13.06.23 um 15:05 schrieb Karol Herbst: > > On Mon, Dec 5, 2022 at 2:40 PM Christian König <christian.koenig@amd.com> wrote: > >> Am 29.11.22 um 22:14 schrieb Felix Kuehling: > >>> On 2022-11-25 05:21, Christian König wrote: > >>>> Instead of a single worker going over the list of delete BOs in regular > >>>> intervals use a per BO worker which blocks for the resv object and > >>>> locking of the BO. > >>>> > >>>> This not only simplifies the handling massively, but also results in > >>>> much better response time when cleaning up buffers. > >>>> > >>>> Signed-off-by: Christian König <christian.koenig@amd.com> > >>> Just thinking out loud: If I understand it correctly, this can cause a > >>> lot of sleeping worker threads when > >>> AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE is used and many BOs are freed > >>> at the same time. This happens e.g. when a KFD process terminates or > >>> crashes. I guess with a concurrency-managed workqueue this isn't going > >>> to be excessive. And since it's on a per device workqueue, it doesn't > >>> stall work items on the system work queue or from other devices. > >> Yes, exactly that. The last parameter to alloc_workqueue() limits how > >> many work items can be sleeping. > >> > >>> I'm trying to understand why you set WQ_MEM_RECLAIM. This work queue > >>> is not about freeing ttm_resources but about freeing the BOs. But it > >>> affects freeing of ghost_objs that are holding the ttm_resources being > >>> freed. > >> Well if the BO is idle, but not immediately lockable we delegate freeing > >> the backing pages in the TT object to those workers as well. It might > >> even be a good idea to use a separate wq for this case. > >> > >>> If those assumptions all make sense, patches 1-3 are > >>> > >>> Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> > >> Thanks, > >> Christian. > >> > > This patch causes a heap use-after-free when using nouveau with the > > potential of trashing filesystems, is there a way to revert it until > > we figure out a proper solution to the problem? > > Uff I don't think so, we have quite some work based on top of this. But > let me double check. > yeah.. I already talked with Dave about fixing this issue as Dave has more knowledge on this part of the driver (I hope), so we might have a fix soonish, but the concerning part is, that it's already out to users, so might be better to be able to revert it if the fix takes a while to emerge. > On the other hand have you tried running this with KASAN to catch use > after free errors? yes: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213#note_1942777 > > Since we now block for work to finish and not check every few > milliseconds to garbage collect memory will now be reclaimed much faster > after freeing it. yeah, that kinda makes sense. This entire issue feels like a race happening as I need to run the OpenGL CTS in parallel with 8+ threads to trigger it reliably. > > Regards, > Christian. > > > > > Bug: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213 > > > > example trace on affected systems: > > > > [ 4102.946946] general protection fault, probably for non-canonical > > address 0x5f775ce3bd949b45: 0000 [#3] PREEMPT SMP NOPTI > > [ 4102.957794] CPU: 12 PID: 89561 Comm: glcts Tainted: G D > > 6.3.5-200.fc38.x86_64 #1 > > [ 4102.966556] Hardware name: ASUS System Product Name/PRIME B660-PLUS > > D4, BIOS 0418 10/13/2021 > > [ 4102.974972] RIP: 0010:__kmem_cache_alloc_node+0x1ba/0x320 > > [ 4102.980362] Code: 2b 14 25 28 00 00 00 0f 85 74 01 00 00 48 83 c4 > > 18 5b 5d 41 5c 41 5d 41 5e 41 5f c3 cc cc cc cc 41 8b 47 28 4d 8b 07 > > 48 01 f8 <48> 8b 18 48 89 c1 49 33 9f b8 00 00 00 48 0f c9 48 31 cb 41 > > f6 c0 > > [ 4102.999073] RSP: 0018:ffff9764e0057b40 EFLAGS: 00010202 > > [ 4103.004291] RAX: 5f775ce3bd949b45 RBX: 0000000000000dc0 RCX: 0000000000000046 > > [ 4103.011408] RDX: 00000002cf87600c RSI: 0000000000000dc0 RDI: 5f775ce3bd949b15 > > [ 4103.018528] RBP: 0000000000000dc0 R08: 00000000000390c0 R09: 0000000030302d6d > > [ 4103.025649] R10: 00000000756c7473 R11: 0000000020090298 R12: 0000000000000000 > > [ 4103.032767] R13: 00000000ffffffff R14: 0000000000000046 R15: ffff8bda80042600 > > [ 4103.039887] FS: 00007f386a85ef00(0000) GS:ffff8be1df700000(0000) > > knlGS:0000000000000000 > > [ 4103.047958] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > [ 4103.053692] CR2: 000000000493b868 CR3: 000000014c3ba000 CR4: 0000000000f50ee0 > > [ 4103.060812] PKRU: 55555554 > > [ 4103.063520] Call Trace: > > [ 4103.065970] <TASK> > > [ 4103.068071] ? die_addr+0x36/0x90 > > [ 4103.071384] ? exc_general_protection+0x1be/0x420 > > [ 4103.076081] ? asm_exc_general_protection+0x26/0x30 > > [ 4103.080952] ? __kmem_cache_alloc_node+0x1ba/0x320 > > [ 4103.085734] ? ext4_htree_store_dirent+0x42/0x180 > > [ 4103.090431] ? ext4_htree_store_dirent+0x42/0x180 > > [ 4103.095132] __kmalloc+0x4d/0x150 > > [ 4103.098444] ext4_htree_store_dirent+0x42/0x180 > > [ 4103.102970] htree_dirblock_to_tree+0x1ed/0x370 > > [ 4103.107494] ext4_htree_fill_tree+0x109/0x3d0 > > [ 4103.111846] ext4_readdir+0x6d4/0xa80 > > [ 4103.115505] iterate_dir+0x178/0x1c0 > > [ 4103.119076] __x64_sys_getdents64+0x88/0x130 > > [ 4103.123341] ? __pfx_filldir64+0x10/0x10 > > [ 4103.127260] do_syscall_64+0x5d/0x90 > > [ 4103.130835] ? handle_mm_fault+0x11e/0x310 > > [ 4103.134927] ? do_user_addr_fault+0x1e0/0x720 > > [ 4103.139278] ? exc_page_fault+0x7c/0x180 > > [ 4103.143195] entry_SYSCALL_64_after_hwframe+0x72/0xdc > > [ 4103.148240] RIP: 0033:0x7f386a418047 > > [ 4103.151828] Code: 24 fb ff 4c 89 e0 5b 41 5c 5d c3 0f 1f 84 00 00 > > 00 00 00 f3 0f 1e fa b8 ff ff ff 7f 48 39 c2 48 0f 47 d0 b8 d9 00 00 > > 00 0f 05 <48> 3d 00 f0 ff ff 77 01 c3 48 8b 15 91 cd 0f 00 f7 d8 64 89 > > 02 48 > > [ 4103.170543] RSP: 002b:00007ffd4793ff38 EFLAGS: 00000293 ORIG_RAX: > > 00000000000000d9 > > [ 4103.178095] RAX: ffffffffffffffda RBX: 0000000004933830 RCX: 00007f386a418047 > > [ 4103.185214] RDX: 0000000000008000 RSI: 0000000004933860 RDI: 0000000000000006 > > [ 4103.192335] RBP: 00007ffd4793ff70 R08: 0000000000000000 R09: 0000000000000001 > > [ 4103.199454] R10: 0000000000000004 R11: 0000000000000293 R12: 0000000004933834 > > [ 4103.206573] R13: 0000000004933860 R14: ffffffffffffff60 R15: 0000000000000000 > > [ 4103.213695] </TASK> > > [ 4103.215883] Modules linked in: snd_seq_dummy snd_hrtimer > > nf_conntrack_netbios_ns nf_conntrack_broadcast nft_fib_inet > > nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 > > nf_reject_ipv6 nft_reject nft_ct nft_chain_nat ip6table_nat > > ip6table_mangle ip6table_raw ip6table > > [ 4103.215911] kvm_intel snd_hwdep snd_seq eeepc_wmi kvm > > snd_seq_device asus_wmi iTCO_wdt mei_pxp mei_hdcp ledtrig_audio > > irqbypass snd_pcm ee1004 intel_pmc_bxt sparse_keymap rapl snd_timer > > pmt_telemetry mei_me iTCO_vendor_support platform_profile joydev > > intel_cstate pmt_class snde > > [ 4103.366194] ---[ end trace 0000000000000000 ]--- > > > >>> > >>>> --- > >>>> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- > >>>> drivers/gpu/drm/i915/i915_gem.c | 2 +- > >>>> drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- > >>>> drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- > >>>> drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - > >>>> drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- > >>>> include/drm/ttm/ttm_bo_api.h | 18 +--- > >>>> include/drm/ttm/ttm_device.h | 7 +- > >>>> 8 files changed, 57 insertions(+), 111 deletions(-) > >>>> > >>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>> index 2b1db37e25c1..74ccbd566777 100644 > >>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>> @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device > >>>> *adev) > >>>> amdgpu_fence_driver_hw_fini(adev); > >>>> if (adev->mman.initialized) > >>>> - flush_delayed_work(&adev->mman.bdev.wq); > >>>> + drain_workqueue(adev->mman.bdev.wq); > >>>> if (adev->pm_sysfs_en) > >>>> amdgpu_pm_sysfs_fini(adev); > >>>> diff --git a/drivers/gpu/drm/i915/i915_gem.c > >>>> b/drivers/gpu/drm/i915/i915_gem.c > >>>> index 8468ca9885fd..c38306f156d6 100644 > >>>> --- a/drivers/gpu/drm/i915/i915_gem.c > >>>> +++ b/drivers/gpu/drm/i915/i915_gem.c > >>>> @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct > >>>> drm_i915_private *i915) > >>>> { > >>>> while (atomic_read(&i915->mm.free_count)) { > >>>> flush_work(&i915->mm.free_work); > >>>> - flush_delayed_work(&i915->bdev.wq); > >>>> + drain_workqueue(i915->bdev.wq); > >>>> rcu_barrier(); > >>>> } > >>>> } > >>>> diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c > >>>> b/drivers/gpu/drm/i915/intel_region_ttm.c > >>>> index cf89d0c2a2d9..657bbc16a48a 100644 > >>>> --- a/drivers/gpu/drm/i915/intel_region_ttm.c > >>>> +++ b/drivers/gpu/drm/i915/intel_region_ttm.c > >>>> @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct > >>>> intel_memory_region *mem) > >>>> break; > >>>> msleep(20); > >>>> - flush_delayed_work(&mem->i915->bdev.wq); > >>>> + drain_workqueue(mem->i915->bdev.wq); > >>>> } > >>>> /* If we leaked objects, Don't free the region causing use > >>>> after free */ > >>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > >>>> index b77262a623e0..4749b65bedc4 100644 > >>>> --- a/drivers/gpu/drm/ttm/ttm_bo.c > >>>> +++ b/drivers/gpu/drm/ttm/ttm_bo.c > >>>> @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct > >>>> ttm_buffer_object *bo, > >>>> ret = 0; > >>>> } > >>>> - if (ret || unlikely(list_empty(&bo->ddestroy))) { > >>>> + if (ret) { > >>>> if (unlock_resv) > >>>> dma_resv_unlock(bo->base.resv); > >>>> spin_unlock(&bo->bdev->lru_lock); > >>>> return ret; > >>>> } > >>>> - list_del_init(&bo->ddestroy); > >>>> spin_unlock(&bo->bdev->lru_lock); > >>>> ttm_bo_cleanup_memtype_use(bo); > >>>> @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct > >>>> ttm_buffer_object *bo, > >>>> } > >>>> /* > >>>> - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all > >>>> - * encountered buffers. > >>>> + * Block for the dma_resv object to become idle, lock the buffer and > >>>> clean up > >>>> + * the resource and tt object. > >>>> */ > >>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) > >>>> +static void ttm_bo_delayed_delete(struct work_struct *work) > >>>> { > >>>> - struct list_head removed; > >>>> - bool empty; > >>>> - > >>>> - INIT_LIST_HEAD(&removed); > >>>> - > >>>> - spin_lock(&bdev->lru_lock); > >>>> - while (!list_empty(&bdev->ddestroy)) { > >>>> - struct ttm_buffer_object *bo; > >>>> - > >>>> - bo = list_first_entry(&bdev->ddestroy, struct > >>>> ttm_buffer_object, > >>>> - ddestroy); > >>>> - list_move_tail(&bo->ddestroy, &removed); > >>>> - if (!ttm_bo_get_unless_zero(bo)) > >>>> - continue; > >>>> - > >>>> - if (remove_all || bo->base.resv != &bo->base._resv) { > >>>> - spin_unlock(&bdev->lru_lock); > >>>> - dma_resv_lock(bo->base.resv, NULL); > >>>> - > >>>> - spin_lock(&bdev->lru_lock); > >>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > >>>> - > >>>> - } else if (dma_resv_trylock(bo->base.resv)) { > >>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > >>>> - } else { > >>>> - spin_unlock(&bdev->lru_lock); > >>>> - } > >>>> + struct ttm_buffer_object *bo; > >>>> - ttm_bo_put(bo); > >>>> - spin_lock(&bdev->lru_lock); > >>>> - } > >>>> - list_splice_tail(&removed, &bdev->ddestroy); > >>>> - empty = list_empty(&bdev->ddestroy); > >>>> - spin_unlock(&bdev->lru_lock); > >>>> + bo = container_of(work, typeof(*bo), delayed_delete); > >>>> - return empty; > >>>> + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, > >>>> false, > >>>> + MAX_SCHEDULE_TIMEOUT); > >>>> + dma_resv_lock(bo->base.resv, NULL); > >>>> + ttm_bo_cleanup_memtype_use(bo); > >>>> + dma_resv_unlock(bo->base.resv); > >>>> + ttm_bo_put(bo); > >>>> } > >>>> static void ttm_bo_release(struct kref *kref) > >>>> @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) > >>>> drm_vma_offset_remove(bdev->vma_manager, > >>>> &bo->base.vma_node); > >>>> ttm_mem_io_free(bdev, bo->resource); > >>>> - } > >>>> - > >>>> - if (!dma_resv_test_signaled(bo->base.resv, > >>>> DMA_RESV_USAGE_BOOKKEEP) || > >>>> - !dma_resv_trylock(bo->base.resv)) { > >>>> - /* The BO is not idle, resurrect it for delayed destroy */ > >>>> - ttm_bo_flush_all_fences(bo); > >>>> - bo->deleted = true; > >>>> - spin_lock(&bo->bdev->lru_lock); > >>>> + if (!dma_resv_test_signaled(bo->base.resv, > >>>> + DMA_RESV_USAGE_BOOKKEEP) || > >>>> + !dma_resv_trylock(bo->base.resv)) { > >>>> + /* The BO is not idle, resurrect it for delayed destroy */ > >>>> + ttm_bo_flush_all_fences(bo); > >>>> + bo->deleted = true; > >>>> - /* > >>>> - * Make pinned bos immediately available to > >>>> - * shrinkers, now that they are queued for > >>>> - * destruction. > >>>> - * > >>>> - * FIXME: QXL is triggering this. Can be removed when the > >>>> - * driver is fixed. > >>>> - */ > >>>> - if (bo->pin_count) { > >>>> - bo->pin_count = 0; > >>>> - ttm_resource_move_to_lru_tail(bo->resource); > >>>> - } > >>>> + spin_lock(&bo->bdev->lru_lock); > >>>> - kref_init(&bo->kref); > >>>> - list_add_tail(&bo->ddestroy, &bdev->ddestroy); > >>>> - spin_unlock(&bo->bdev->lru_lock); > >>>> + /* > >>>> + * Make pinned bos immediately available to > >>>> + * shrinkers, now that they are queued for > >>>> + * destruction. > >>>> + * > >>>> + * FIXME: QXL is triggering this. Can be removed when the > >>>> + * driver is fixed. > >>>> + */ > >>>> + if (bo->pin_count) { > >>>> + bo->pin_count = 0; > >>>> + ttm_resource_move_to_lru_tail(bo->resource); > >>>> + } > >>>> - schedule_delayed_work(&bdev->wq, > >>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); > >>>> - return; > >>>> - } > >>>> + kref_init(&bo->kref); > >>>> + spin_unlock(&bo->bdev->lru_lock); > >>>> - spin_lock(&bo->bdev->lru_lock); > >>>> - list_del(&bo->ddestroy); > >>>> - spin_unlock(&bo->bdev->lru_lock); > >>>> + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); > >>>> + queue_work(bdev->wq, &bo->delayed_delete); > >>>> + return; > >>>> + } > >>>> - ttm_bo_cleanup_memtype_use(bo); > >>>> - dma_resv_unlock(bo->base.resv); > >>>> + ttm_bo_cleanup_memtype_use(bo); > >>>> + dma_resv_unlock(bo->base.resv); > >>>> + } > >>>> atomic_dec(&ttm_glob.bo_count); > >>>> bo->destroy(bo); > >>>> @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, > >>>> struct ttm_buffer_object *bo, > >>>> int ret; > >>>> kref_init(&bo->kref); > >>>> - INIT_LIST_HEAD(&bo->ddestroy); > >>>> bo->bdev = bdev; > >>>> bo->type = type; > >>>> bo->page_alignment = alignment; > >>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>> b/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>> index ba3aa0a0fc43..ae4b7922ee1a 100644 > >>>> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>> @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct > >>>> ttm_buffer_object *bo, > >>>> */ > >>>> atomic_inc(&ttm_glob.bo_count); > >>>> - INIT_LIST_HEAD(&fbo->base.ddestroy); > >>>> drm_vma_node_reset(&fbo->base.base.vma_node); > >>>> kref_init(&fbo->base.kref); > >>>> diff --git a/drivers/gpu/drm/ttm/ttm_device.c > >>>> b/drivers/gpu/drm/ttm/ttm_device.c > >>>> index e7147e304637..e9bedca4dfdc 100644 > >>>> --- a/drivers/gpu/drm/ttm/ttm_device.c > >>>> +++ b/drivers/gpu/drm/ttm/ttm_device.c > >>>> @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, > >>>> struct ttm_operation_ctx *ctx, > >>>> } > >>>> EXPORT_SYMBOL(ttm_device_swapout); > >>>> -static void ttm_device_delayed_workqueue(struct work_struct *work) > >>>> -{ > >>>> - struct ttm_device *bdev = > >>>> - container_of(work, struct ttm_device, wq.work); > >>>> - > >>>> - if (!ttm_bo_delayed_delete(bdev, false)) > >>>> - schedule_delayed_work(&bdev->wq, > >>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); > >>>> -} > >>>> - > >>>> /** > >>>> * ttm_device_init > >>>> * > >>>> @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, > >>>> struct ttm_device_funcs *funcs, > >>>> if (ret) > >>>> return ret; > >>>> + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, > >>>> 16); > >>>> + if (!bdev->wq) { > >>>> + ttm_global_release(); > >>>> + return -ENOMEM; > >>>> + } > >>>> + > >>>> bdev->funcs = funcs; > >>>> ttm_sys_man_init(bdev); > >>>> ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); > >>>> bdev->vma_manager = vma_manager; > >>>> - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); > >>>> spin_lock_init(&bdev->lru_lock); > >>>> - INIT_LIST_HEAD(&bdev->ddestroy); > >>>> INIT_LIST_HEAD(&bdev->pinned); > >>>> bdev->dev_mapping = mapping; > >>>> mutex_lock(&ttm_global_mutex); > >>>> @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) > >>>> list_del(&bdev->device_list); > >>>> mutex_unlock(&ttm_global_mutex); > >>>> - cancel_delayed_work_sync(&bdev->wq); > >>>> - > >>>> - if (ttm_bo_delayed_delete(bdev, true)) > >>>> - pr_debug("Delayed destroy list was clean\n"); > >>>> + drain_workqueue(bdev->wq); > >>>> + destroy_workqueue(bdev->wq); > >>>> spin_lock(&bdev->lru_lock); > >>>> for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) > >>>> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h > >>>> index 7758347c461c..69e62bbb01e3 100644 > >>>> --- a/include/drm/ttm/ttm_bo_api.h > >>>> +++ b/include/drm/ttm/ttm_bo_api.h > >>>> @@ -92,7 +92,6 @@ struct ttm_tt; > >>>> * @ttm: TTM structure holding system pages. > >>>> * @evicted: Whether the object was evicted without user-space > >>>> knowing. > >>>> * @deleted: True if the object is only a zombie and already deleted. > >>>> - * @ddestroy: List head for the delayed destroy list. > >>>> * @swap: List head for swap LRU list. > >>>> * @offset: The current GPU offset, which can have different meanings > >>>> * depending on the memory type. For SYSTEM type memory, it should > >>>> be 0. > >>>> @@ -135,19 +134,14 @@ struct ttm_buffer_object { > >>>> struct ttm_tt *ttm; > >>>> bool deleted; > >>>> struct ttm_lru_bulk_move *bulk_move; > >>>> + unsigned priority; > >>>> + unsigned pin_count; > >>>> /** > >>>> - * Members protected by the bdev::lru_lock. > >>>> - */ > >>>> - > >>>> - struct list_head ddestroy; > >>>> - > >>>> - /** > >>>> - * Members protected by a bo reservation. > >>>> + * @delayed_delete: Work item used when we can't delete the BO > >>>> + * immediately > >>>> */ > >>>> - > >>>> - unsigned priority; > >>>> - unsigned pin_count; > >>>> + struct work_struct delayed_delete; > >>>> /** > >>>> * Special members that are protected by the reserve lock > >>>> @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); > >>>> int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, > >>>> void *buf, int len, int write); > >>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); > >>>> - > >>>> vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); > >>>> #endif > >>>> diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h > >>>> index 95b3c04b1ab9..4f3e81eac6f3 100644 > >>>> --- a/include/drm/ttm/ttm_device.h > >>>> +++ b/include/drm/ttm/ttm_device.h > >>>> @@ -251,11 +251,6 @@ struct ttm_device { > >>>> */ > >>>> spinlock_t lru_lock; > >>>> - /** > >>>> - * @ddestroy: Destroyed but not yet cleaned up buffer objects. > >>>> - */ > >>>> - struct list_head ddestroy; > >>>> - > >>>> /** > >>>> * @pinned: Buffer objects which are pinned and so not on any > >>>> LRU list. > >>>> */ > >>>> @@ -270,7 +265,7 @@ struct ttm_device { > >>>> /** > >>>> * @wq: Work queue structure for the delayed delete workqueue. > >>>> */ > >>>> - struct delayed_work wq; > >>>> + struct workqueue_struct *wq; > >>>> }; > >>>> int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t > >>>> gfp_flags); >
Am 13.06.23 um 16:18 schrieb Karol Herbst: > On Tue, Jun 13, 2023 at 3:59 PM Christian König > <christian.koenig@amd.com> wrote: >> Am 13.06.23 um 15:05 schrieb Karol Herbst: >>> On Mon, Dec 5, 2022 at 2:40 PM Christian König <christian.koenig@amd.com> wrote: >>>> Am 29.11.22 um 22:14 schrieb Felix Kuehling: >>>>> On 2022-11-25 05:21, Christian König wrote: >>>>>> Instead of a single worker going over the list of delete BOs in regular >>>>>> intervals use a per BO worker which blocks for the resv object and >>>>>> locking of the BO. >>>>>> >>>>>> This not only simplifies the handling massively, but also results in >>>>>> much better response time when cleaning up buffers. >>>>>> >>>>>> Signed-off-by: Christian König <christian.koenig@amd.com> >>>>> Just thinking out loud: If I understand it correctly, this can cause a >>>>> lot of sleeping worker threads when >>>>> AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE is used and many BOs are freed >>>>> at the same time. This happens e.g. when a KFD process terminates or >>>>> crashes. I guess with a concurrency-managed workqueue this isn't going >>>>> to be excessive. And since it's on a per device workqueue, it doesn't >>>>> stall work items on the system work queue or from other devices. >>>> Yes, exactly that. The last parameter to alloc_workqueue() limits how >>>> many work items can be sleeping. >>>> >>>>> I'm trying to understand why you set WQ_MEM_RECLAIM. This work queue >>>>> is not about freeing ttm_resources but about freeing the BOs. But it >>>>> affects freeing of ghost_objs that are holding the ttm_resources being >>>>> freed. >>>> Well if the BO is idle, but not immediately lockable we delegate freeing >>>> the backing pages in the TT object to those workers as well. It might >>>> even be a good idea to use a separate wq for this case. >>>> >>>>> If those assumptions all make sense, patches 1-3 are >>>>> >>>>> Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> >>>> Thanks, >>>> Christian. >>>> >>> This patch causes a heap use-after-free when using nouveau with the >>> potential of trashing filesystems, is there a way to revert it until >>> we figure out a proper solution to the problem? >> Uff I don't think so, we have quite some work based on top of this. But >> let me double check. >> > yeah.. I already talked with Dave about fixing this issue as Dave has > more knowledge on this part of the driver (I hope), so we might have a > fix soonish, but the concerning part is, that it's already out to > users, so might be better to be able to revert it if the fix takes a > while to emerge. We at least can't revert easily. This was fixing issues where we have seen OOM situations because TTM wasn't releasing resources fast enough. >> On the other hand have you tried running this with KASAN to catch use >> after free errors? > yes: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213#note_1942777 > >> Since we now block for work to finish and not check every few >> milliseconds to garbage collect memory will now be reclaimed much faster >> after freeing it. > yeah, that kinda makes sense. This entire issue feels like a race > happening as I need to run the OpenGL CTS in parallel with 8+ threads > to trigger it reliably. Yeah, from the bug report that's a classic use after free because of a race. Easiest would be to make sure everybody has a reference to the fence. Regards, Christian. > >> Regards, >> Christian. >> >>> Bug: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213 >>> >>> example trace on affected systems: >>> >>> [ 4102.946946] general protection fault, probably for non-canonical >>> address 0x5f775ce3bd949b45: 0000 [#3] PREEMPT SMP NOPTI >>> [ 4102.957794] CPU: 12 PID: 89561 Comm: glcts Tainted: G D >>> 6.3.5-200.fc38.x86_64 #1 >>> [ 4102.966556] Hardware name: ASUS System Product Name/PRIME B660-PLUS >>> D4, BIOS 0418 10/13/2021 >>> [ 4102.974972] RIP: 0010:__kmem_cache_alloc_node+0x1ba/0x320 >>> [ 4102.980362] Code: 2b 14 25 28 00 00 00 0f 85 74 01 00 00 48 83 c4 >>> 18 5b 5d 41 5c 41 5d 41 5e 41 5f c3 cc cc cc cc 41 8b 47 28 4d 8b 07 >>> 48 01 f8 <48> 8b 18 48 89 c1 49 33 9f b8 00 00 00 48 0f c9 48 31 cb 41 >>> f6 c0 >>> [ 4102.999073] RSP: 0018:ffff9764e0057b40 EFLAGS: 00010202 >>> [ 4103.004291] RAX: 5f775ce3bd949b45 RBX: 0000000000000dc0 RCX: 0000000000000046 >>> [ 4103.011408] RDX: 00000002cf87600c RSI: 0000000000000dc0 RDI: 5f775ce3bd949b15 >>> [ 4103.018528] RBP: 0000000000000dc0 R08: 00000000000390c0 R09: 0000000030302d6d >>> [ 4103.025649] R10: 00000000756c7473 R11: 0000000020090298 R12: 0000000000000000 >>> [ 4103.032767] R13: 00000000ffffffff R14: 0000000000000046 R15: ffff8bda80042600 >>> [ 4103.039887] FS: 00007f386a85ef00(0000) GS:ffff8be1df700000(0000) >>> knlGS:0000000000000000 >>> [ 4103.047958] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >>> [ 4103.053692] CR2: 000000000493b868 CR3: 000000014c3ba000 CR4: 0000000000f50ee0 >>> [ 4103.060812] PKRU: 55555554 >>> [ 4103.063520] Call Trace: >>> [ 4103.065970] <TASK> >>> [ 4103.068071] ? die_addr+0x36/0x90 >>> [ 4103.071384] ? exc_general_protection+0x1be/0x420 >>> [ 4103.076081] ? asm_exc_general_protection+0x26/0x30 >>> [ 4103.080952] ? __kmem_cache_alloc_node+0x1ba/0x320 >>> [ 4103.085734] ? ext4_htree_store_dirent+0x42/0x180 >>> [ 4103.090431] ? ext4_htree_store_dirent+0x42/0x180 >>> [ 4103.095132] __kmalloc+0x4d/0x150 >>> [ 4103.098444] ext4_htree_store_dirent+0x42/0x180 >>> [ 4103.102970] htree_dirblock_to_tree+0x1ed/0x370 >>> [ 4103.107494] ext4_htree_fill_tree+0x109/0x3d0 >>> [ 4103.111846] ext4_readdir+0x6d4/0xa80 >>> [ 4103.115505] iterate_dir+0x178/0x1c0 >>> [ 4103.119076] __x64_sys_getdents64+0x88/0x130 >>> [ 4103.123341] ? __pfx_filldir64+0x10/0x10 >>> [ 4103.127260] do_syscall_64+0x5d/0x90 >>> [ 4103.130835] ? handle_mm_fault+0x11e/0x310 >>> [ 4103.134927] ? do_user_addr_fault+0x1e0/0x720 >>> [ 4103.139278] ? exc_page_fault+0x7c/0x180 >>> [ 4103.143195] entry_SYSCALL_64_after_hwframe+0x72/0xdc >>> [ 4103.148240] RIP: 0033:0x7f386a418047 >>> [ 4103.151828] Code: 24 fb ff 4c 89 e0 5b 41 5c 5d c3 0f 1f 84 00 00 >>> 00 00 00 f3 0f 1e fa b8 ff ff ff 7f 48 39 c2 48 0f 47 d0 b8 d9 00 00 >>> 00 0f 05 <48> 3d 00 f0 ff ff 77 01 c3 48 8b 15 91 cd 0f 00 f7 d8 64 89 >>> 02 48 >>> [ 4103.170543] RSP: 002b:00007ffd4793ff38 EFLAGS: 00000293 ORIG_RAX: >>> 00000000000000d9 >>> [ 4103.178095] RAX: ffffffffffffffda RBX: 0000000004933830 RCX: 00007f386a418047 >>> [ 4103.185214] RDX: 0000000000008000 RSI: 0000000004933860 RDI: 0000000000000006 >>> [ 4103.192335] RBP: 00007ffd4793ff70 R08: 0000000000000000 R09: 0000000000000001 >>> [ 4103.199454] R10: 0000000000000004 R11: 0000000000000293 R12: 0000000004933834 >>> [ 4103.206573] R13: 0000000004933860 R14: ffffffffffffff60 R15: 0000000000000000 >>> [ 4103.213695] </TASK> >>> [ 4103.215883] Modules linked in: snd_seq_dummy snd_hrtimer >>> nf_conntrack_netbios_ns nf_conntrack_broadcast nft_fib_inet >>> nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 >>> nf_reject_ipv6 nft_reject nft_ct nft_chain_nat ip6table_nat >>> ip6table_mangle ip6table_raw ip6table >>> [ 4103.215911] kvm_intel snd_hwdep snd_seq eeepc_wmi kvm >>> snd_seq_device asus_wmi iTCO_wdt mei_pxp mei_hdcp ledtrig_audio >>> irqbypass snd_pcm ee1004 intel_pmc_bxt sparse_keymap rapl snd_timer >>> pmt_telemetry mei_me iTCO_vendor_support platform_profile joydev >>> intel_cstate pmt_class snde >>> [ 4103.366194] ---[ end trace 0000000000000000 ]--- >>> >>>>>> --- >>>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- >>>>>> drivers/gpu/drm/i915/i915_gem.c | 2 +- >>>>>> drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- >>>>>> drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- >>>>>> drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - >>>>>> drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- >>>>>> include/drm/ttm/ttm_bo_api.h | 18 +--- >>>>>> include/drm/ttm/ttm_device.h | 7 +- >>>>>> 8 files changed, 57 insertions(+), 111 deletions(-) >>>>>> >>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>>> index 2b1db37e25c1..74ccbd566777 100644 >>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>>> @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device >>>>>> *adev) >>>>>> amdgpu_fence_driver_hw_fini(adev); >>>>>> if (adev->mman.initialized) >>>>>> - flush_delayed_work(&adev->mman.bdev.wq); >>>>>> + drain_workqueue(adev->mman.bdev.wq); >>>>>> if (adev->pm_sysfs_en) >>>>>> amdgpu_pm_sysfs_fini(adev); >>>>>> diff --git a/drivers/gpu/drm/i915/i915_gem.c >>>>>> b/drivers/gpu/drm/i915/i915_gem.c >>>>>> index 8468ca9885fd..c38306f156d6 100644 >>>>>> --- a/drivers/gpu/drm/i915/i915_gem.c >>>>>> +++ b/drivers/gpu/drm/i915/i915_gem.c >>>>>> @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct >>>>>> drm_i915_private *i915) >>>>>> { >>>>>> while (atomic_read(&i915->mm.free_count)) { >>>>>> flush_work(&i915->mm.free_work); >>>>>> - flush_delayed_work(&i915->bdev.wq); >>>>>> + drain_workqueue(i915->bdev.wq); >>>>>> rcu_barrier(); >>>>>> } >>>>>> } >>>>>> diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c >>>>>> b/drivers/gpu/drm/i915/intel_region_ttm.c >>>>>> index cf89d0c2a2d9..657bbc16a48a 100644 >>>>>> --- a/drivers/gpu/drm/i915/intel_region_ttm.c >>>>>> +++ b/drivers/gpu/drm/i915/intel_region_ttm.c >>>>>> @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct >>>>>> intel_memory_region *mem) >>>>>> break; >>>>>> msleep(20); >>>>>> - flush_delayed_work(&mem->i915->bdev.wq); >>>>>> + drain_workqueue(mem->i915->bdev.wq); >>>>>> } >>>>>> /* If we leaked objects, Don't free the region causing use >>>>>> after free */ >>>>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c >>>>>> index b77262a623e0..4749b65bedc4 100644 >>>>>> --- a/drivers/gpu/drm/ttm/ttm_bo.c >>>>>> +++ b/drivers/gpu/drm/ttm/ttm_bo.c >>>>>> @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct >>>>>> ttm_buffer_object *bo, >>>>>> ret = 0; >>>>>> } >>>>>> - if (ret || unlikely(list_empty(&bo->ddestroy))) { >>>>>> + if (ret) { >>>>>> if (unlock_resv) >>>>>> dma_resv_unlock(bo->base.resv); >>>>>> spin_unlock(&bo->bdev->lru_lock); >>>>>> return ret; >>>>>> } >>>>>> - list_del_init(&bo->ddestroy); >>>>>> spin_unlock(&bo->bdev->lru_lock); >>>>>> ttm_bo_cleanup_memtype_use(bo); >>>>>> @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct >>>>>> ttm_buffer_object *bo, >>>>>> } >>>>>> /* >>>>>> - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all >>>>>> - * encountered buffers. >>>>>> + * Block for the dma_resv object to become idle, lock the buffer and >>>>>> clean up >>>>>> + * the resource and tt object. >>>>>> */ >>>>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) >>>>>> +static void ttm_bo_delayed_delete(struct work_struct *work) >>>>>> { >>>>>> - struct list_head removed; >>>>>> - bool empty; >>>>>> - >>>>>> - INIT_LIST_HEAD(&removed); >>>>>> - >>>>>> - spin_lock(&bdev->lru_lock); >>>>>> - while (!list_empty(&bdev->ddestroy)) { >>>>>> - struct ttm_buffer_object *bo; >>>>>> - >>>>>> - bo = list_first_entry(&bdev->ddestroy, struct >>>>>> ttm_buffer_object, >>>>>> - ddestroy); >>>>>> - list_move_tail(&bo->ddestroy, &removed); >>>>>> - if (!ttm_bo_get_unless_zero(bo)) >>>>>> - continue; >>>>>> - >>>>>> - if (remove_all || bo->base.resv != &bo->base._resv) { >>>>>> - spin_unlock(&bdev->lru_lock); >>>>>> - dma_resv_lock(bo->base.resv, NULL); >>>>>> - >>>>>> - spin_lock(&bdev->lru_lock); >>>>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); >>>>>> - >>>>>> - } else if (dma_resv_trylock(bo->base.resv)) { >>>>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); >>>>>> - } else { >>>>>> - spin_unlock(&bdev->lru_lock); >>>>>> - } >>>>>> + struct ttm_buffer_object *bo; >>>>>> - ttm_bo_put(bo); >>>>>> - spin_lock(&bdev->lru_lock); >>>>>> - } >>>>>> - list_splice_tail(&removed, &bdev->ddestroy); >>>>>> - empty = list_empty(&bdev->ddestroy); >>>>>> - spin_unlock(&bdev->lru_lock); >>>>>> + bo = container_of(work, typeof(*bo), delayed_delete); >>>>>> - return empty; >>>>>> + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, >>>>>> false, >>>>>> + MAX_SCHEDULE_TIMEOUT); >>>>>> + dma_resv_lock(bo->base.resv, NULL); >>>>>> + ttm_bo_cleanup_memtype_use(bo); >>>>>> + dma_resv_unlock(bo->base.resv); >>>>>> + ttm_bo_put(bo); >>>>>> } >>>>>> static void ttm_bo_release(struct kref *kref) >>>>>> @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) >>>>>> drm_vma_offset_remove(bdev->vma_manager, >>>>>> &bo->base.vma_node); >>>>>> ttm_mem_io_free(bdev, bo->resource); >>>>>> - } >>>>>> - >>>>>> - if (!dma_resv_test_signaled(bo->base.resv, >>>>>> DMA_RESV_USAGE_BOOKKEEP) || >>>>>> - !dma_resv_trylock(bo->base.resv)) { >>>>>> - /* The BO is not idle, resurrect it for delayed destroy */ >>>>>> - ttm_bo_flush_all_fences(bo); >>>>>> - bo->deleted = true; >>>>>> - spin_lock(&bo->bdev->lru_lock); >>>>>> + if (!dma_resv_test_signaled(bo->base.resv, >>>>>> + DMA_RESV_USAGE_BOOKKEEP) || >>>>>> + !dma_resv_trylock(bo->base.resv)) { >>>>>> + /* The BO is not idle, resurrect it for delayed destroy */ >>>>>> + ttm_bo_flush_all_fences(bo); >>>>>> + bo->deleted = true; >>>>>> - /* >>>>>> - * Make pinned bos immediately available to >>>>>> - * shrinkers, now that they are queued for >>>>>> - * destruction. >>>>>> - * >>>>>> - * FIXME: QXL is triggering this. Can be removed when the >>>>>> - * driver is fixed. >>>>>> - */ >>>>>> - if (bo->pin_count) { >>>>>> - bo->pin_count = 0; >>>>>> - ttm_resource_move_to_lru_tail(bo->resource); >>>>>> - } >>>>>> + spin_lock(&bo->bdev->lru_lock); >>>>>> - kref_init(&bo->kref); >>>>>> - list_add_tail(&bo->ddestroy, &bdev->ddestroy); >>>>>> - spin_unlock(&bo->bdev->lru_lock); >>>>>> + /* >>>>>> + * Make pinned bos immediately available to >>>>>> + * shrinkers, now that they are queued for >>>>>> + * destruction. >>>>>> + * >>>>>> + * FIXME: QXL is triggering this. Can be removed when the >>>>>> + * driver is fixed. >>>>>> + */ >>>>>> + if (bo->pin_count) { >>>>>> + bo->pin_count = 0; >>>>>> + ttm_resource_move_to_lru_tail(bo->resource); >>>>>> + } >>>>>> - schedule_delayed_work(&bdev->wq, >>>>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); >>>>>> - return; >>>>>> - } >>>>>> + kref_init(&bo->kref); >>>>>> + spin_unlock(&bo->bdev->lru_lock); >>>>>> - spin_lock(&bo->bdev->lru_lock); >>>>>> - list_del(&bo->ddestroy); >>>>>> - spin_unlock(&bo->bdev->lru_lock); >>>>>> + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); >>>>>> + queue_work(bdev->wq, &bo->delayed_delete); >>>>>> + return; >>>>>> + } >>>>>> - ttm_bo_cleanup_memtype_use(bo); >>>>>> - dma_resv_unlock(bo->base.resv); >>>>>> + ttm_bo_cleanup_memtype_use(bo); >>>>>> + dma_resv_unlock(bo->base.resv); >>>>>> + } >>>>>> atomic_dec(&ttm_glob.bo_count); >>>>>> bo->destroy(bo); >>>>>> @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, >>>>>> struct ttm_buffer_object *bo, >>>>>> int ret; >>>>>> kref_init(&bo->kref); >>>>>> - INIT_LIST_HEAD(&bo->ddestroy); >>>>>> bo->bdev = bdev; >>>>>> bo->type = type; >>>>>> bo->page_alignment = alignment; >>>>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c >>>>>> b/drivers/gpu/drm/ttm/ttm_bo_util.c >>>>>> index ba3aa0a0fc43..ae4b7922ee1a 100644 >>>>>> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c >>>>>> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c >>>>>> @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct >>>>>> ttm_buffer_object *bo, >>>>>> */ >>>>>> atomic_inc(&ttm_glob.bo_count); >>>>>> - INIT_LIST_HEAD(&fbo->base.ddestroy); >>>>>> drm_vma_node_reset(&fbo->base.base.vma_node); >>>>>> kref_init(&fbo->base.kref); >>>>>> diff --git a/drivers/gpu/drm/ttm/ttm_device.c >>>>>> b/drivers/gpu/drm/ttm/ttm_device.c >>>>>> index e7147e304637..e9bedca4dfdc 100644 >>>>>> --- a/drivers/gpu/drm/ttm/ttm_device.c >>>>>> +++ b/drivers/gpu/drm/ttm/ttm_device.c >>>>>> @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, >>>>>> struct ttm_operation_ctx *ctx, >>>>>> } >>>>>> EXPORT_SYMBOL(ttm_device_swapout); >>>>>> -static void ttm_device_delayed_workqueue(struct work_struct *work) >>>>>> -{ >>>>>> - struct ttm_device *bdev = >>>>>> - container_of(work, struct ttm_device, wq.work); >>>>>> - >>>>>> - if (!ttm_bo_delayed_delete(bdev, false)) >>>>>> - schedule_delayed_work(&bdev->wq, >>>>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); >>>>>> -} >>>>>> - >>>>>> /** >>>>>> * ttm_device_init >>>>>> * >>>>>> @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, >>>>>> struct ttm_device_funcs *funcs, >>>>>> if (ret) >>>>>> return ret; >>>>>> + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, >>>>>> 16); >>>>>> + if (!bdev->wq) { >>>>>> + ttm_global_release(); >>>>>> + return -ENOMEM; >>>>>> + } >>>>>> + >>>>>> bdev->funcs = funcs; >>>>>> ttm_sys_man_init(bdev); >>>>>> ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); >>>>>> bdev->vma_manager = vma_manager; >>>>>> - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); >>>>>> spin_lock_init(&bdev->lru_lock); >>>>>> - INIT_LIST_HEAD(&bdev->ddestroy); >>>>>> INIT_LIST_HEAD(&bdev->pinned); >>>>>> bdev->dev_mapping = mapping; >>>>>> mutex_lock(&ttm_global_mutex); >>>>>> @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) >>>>>> list_del(&bdev->device_list); >>>>>> mutex_unlock(&ttm_global_mutex); >>>>>> - cancel_delayed_work_sync(&bdev->wq); >>>>>> - >>>>>> - if (ttm_bo_delayed_delete(bdev, true)) >>>>>> - pr_debug("Delayed destroy list was clean\n"); >>>>>> + drain_workqueue(bdev->wq); >>>>>> + destroy_workqueue(bdev->wq); >>>>>> spin_lock(&bdev->lru_lock); >>>>>> for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) >>>>>> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h >>>>>> index 7758347c461c..69e62bbb01e3 100644 >>>>>> --- a/include/drm/ttm/ttm_bo_api.h >>>>>> +++ b/include/drm/ttm/ttm_bo_api.h >>>>>> @@ -92,7 +92,6 @@ struct ttm_tt; >>>>>> * @ttm: TTM structure holding system pages. >>>>>> * @evicted: Whether the object was evicted without user-space >>>>>> knowing. >>>>>> * @deleted: True if the object is only a zombie and already deleted. >>>>>> - * @ddestroy: List head for the delayed destroy list. >>>>>> * @swap: List head for swap LRU list. >>>>>> * @offset: The current GPU offset, which can have different meanings >>>>>> * depending on the memory type. For SYSTEM type memory, it should >>>>>> be 0. >>>>>> @@ -135,19 +134,14 @@ struct ttm_buffer_object { >>>>>> struct ttm_tt *ttm; >>>>>> bool deleted; >>>>>> struct ttm_lru_bulk_move *bulk_move; >>>>>> + unsigned priority; >>>>>> + unsigned pin_count; >>>>>> /** >>>>>> - * Members protected by the bdev::lru_lock. >>>>>> - */ >>>>>> - >>>>>> - struct list_head ddestroy; >>>>>> - >>>>>> - /** >>>>>> - * Members protected by a bo reservation. >>>>>> + * @delayed_delete: Work item used when we can't delete the BO >>>>>> + * immediately >>>>>> */ >>>>>> - >>>>>> - unsigned priority; >>>>>> - unsigned pin_count; >>>>>> + struct work_struct delayed_delete; >>>>>> /** >>>>>> * Special members that are protected by the reserve lock >>>>>> @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); >>>>>> int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, >>>>>> void *buf, int len, int write); >>>>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); >>>>>> - >>>>>> vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); >>>>>> #endif >>>>>> diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h >>>>>> index 95b3c04b1ab9..4f3e81eac6f3 100644 >>>>>> --- a/include/drm/ttm/ttm_device.h >>>>>> +++ b/include/drm/ttm/ttm_device.h >>>>>> @@ -251,11 +251,6 @@ struct ttm_device { >>>>>> */ >>>>>> spinlock_t lru_lock; >>>>>> - /** >>>>>> - * @ddestroy: Destroyed but not yet cleaned up buffer objects. >>>>>> - */ >>>>>> - struct list_head ddestroy; >>>>>> - >>>>>> /** >>>>>> * @pinned: Buffer objects which are pinned and so not on any >>>>>> LRU list. >>>>>> */ >>>>>> @@ -270,7 +265,7 @@ struct ttm_device { >>>>>> /** >>>>>> * @wq: Work queue structure for the delayed delete workqueue. >>>>>> */ >>>>>> - struct delayed_work wq; >>>>>> + struct workqueue_struct *wq; >>>>>> }; >>>>>> int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t >>>>>> gfp_flags);
On Thu, Jun 15, 2023 at 1:19 PM Christian König <ckoenig.leichtzumerken@gmail.com> wrote: > > Am 13.06.23 um 16:18 schrieb Karol Herbst: > > On Tue, Jun 13, 2023 at 3:59 PM Christian König > > <christian.koenig@amd.com> wrote: > >> Am 13.06.23 um 15:05 schrieb Karol Herbst: > >>> On Mon, Dec 5, 2022 at 2:40 PM Christian König <christian.koenig@amd.com> wrote: > >>>> Am 29.11.22 um 22:14 schrieb Felix Kuehling: > >>>>> On 2022-11-25 05:21, Christian König wrote: > >>>>>> Instead of a single worker going over the list of delete BOs in regular > >>>>>> intervals use a per BO worker which blocks for the resv object and > >>>>>> locking of the BO. > >>>>>> > >>>>>> This not only simplifies the handling massively, but also results in > >>>>>> much better response time when cleaning up buffers. > >>>>>> > >>>>>> Signed-off-by: Christian König <christian.koenig@amd.com> > >>>>> Just thinking out loud: If I understand it correctly, this can cause a > >>>>> lot of sleeping worker threads when > >>>>> AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE is used and many BOs are freed > >>>>> at the same time. This happens e.g. when a KFD process terminates or > >>>>> crashes. I guess with a concurrency-managed workqueue this isn't going > >>>>> to be excessive. And since it's on a per device workqueue, it doesn't > >>>>> stall work items on the system work queue or from other devices. > >>>> Yes, exactly that. The last parameter to alloc_workqueue() limits how > >>>> many work items can be sleeping. > >>>> > >>>>> I'm trying to understand why you set WQ_MEM_RECLAIM. This work queue > >>>>> is not about freeing ttm_resources but about freeing the BOs. But it > >>>>> affects freeing of ghost_objs that are holding the ttm_resources being > >>>>> freed. > >>>> Well if the BO is idle, but not immediately lockable we delegate freeing > >>>> the backing pages in the TT object to those workers as well. It might > >>>> even be a good idea to use a separate wq for this case. > >>>> > >>>>> If those assumptions all make sense, patches 1-3 are > >>>>> > >>>>> Reviewed-by: Felix Kuehling <Felix.Kuehling@amd.com> > >>>> Thanks, > >>>> Christian. > >>>> > >>> This patch causes a heap use-after-free when using nouveau with the > >>> potential of trashing filesystems, is there a way to revert it until > >>> we figure out a proper solution to the problem? > >> Uff I don't think so, we have quite some work based on top of this. But > >> let me double check. > >> > > yeah.. I already talked with Dave about fixing this issue as Dave has > > more knowledge on this part of the driver (I hope), so we might have a > > fix soonish, but the concerning part is, that it's already out to > > users, so might be better to be able to revert it if the fix takes a > > while to emerge. > > We at least can't revert easily. This was fixing issues where we have > seen OOM situations because TTM wasn't releasing resources fast enough. > > >> On the other hand have you tried running this with KASAN to catch use > >> after free errors? > > yes: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213#note_1942777 > > > >> Since we now block for work to finish and not check every few > >> milliseconds to garbage collect memory will now be reclaimed much faster > >> after freeing it. > > yeah, that kinda makes sense. This entire issue feels like a race > > happening as I need to run the OpenGL CTS in parallel with 8+ threads > > to trigger it reliably. > > Yeah, from the bug report that's a classic use after free because of a race. > > Easiest would be to make sure everybody has a reference to the fence. > turns out nouveau was using `dma_fence_is_signaled_locked` without taking the lock. Fixing that seems to improve the situation by a lot, so I think we have a fix to resolve this problem. > Regards, > Christian. > > > > >> Regards, > >> Christian. > >> > >>> Bug: https://gitlab.freedesktop.org/drm/nouveau/-/issues/213 > >>> > >>> example trace on affected systems: > >>> > >>> [ 4102.946946] general protection fault, probably for non-canonical > >>> address 0x5f775ce3bd949b45: 0000 [#3] PREEMPT SMP NOPTI > >>> [ 4102.957794] CPU: 12 PID: 89561 Comm: glcts Tainted: G D > >>> 6.3.5-200.fc38.x86_64 #1 > >>> [ 4102.966556] Hardware name: ASUS System Product Name/PRIME B660-PLUS > >>> D4, BIOS 0418 10/13/2021 > >>> [ 4102.974972] RIP: 0010:__kmem_cache_alloc_node+0x1ba/0x320 > >>> [ 4102.980362] Code: 2b 14 25 28 00 00 00 0f 85 74 01 00 00 48 83 c4 > >>> 18 5b 5d 41 5c 41 5d 41 5e 41 5f c3 cc cc cc cc 41 8b 47 28 4d 8b 07 > >>> 48 01 f8 <48> 8b 18 48 89 c1 49 33 9f b8 00 00 00 48 0f c9 48 31 cb 41 > >>> f6 c0 > >>> [ 4102.999073] RSP: 0018:ffff9764e0057b40 EFLAGS: 00010202 > >>> [ 4103.004291] RAX: 5f775ce3bd949b45 RBX: 0000000000000dc0 RCX: 0000000000000046 > >>> [ 4103.011408] RDX: 00000002cf87600c RSI: 0000000000000dc0 RDI: 5f775ce3bd949b15 > >>> [ 4103.018528] RBP: 0000000000000dc0 R08: 00000000000390c0 R09: 0000000030302d6d > >>> [ 4103.025649] R10: 00000000756c7473 R11: 0000000020090298 R12: 0000000000000000 > >>> [ 4103.032767] R13: 00000000ffffffff R14: 0000000000000046 R15: ffff8bda80042600 > >>> [ 4103.039887] FS: 00007f386a85ef00(0000) GS:ffff8be1df700000(0000) > >>> knlGS:0000000000000000 > >>> [ 4103.047958] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > >>> [ 4103.053692] CR2: 000000000493b868 CR3: 000000014c3ba000 CR4: 0000000000f50ee0 > >>> [ 4103.060812] PKRU: 55555554 > >>> [ 4103.063520] Call Trace: > >>> [ 4103.065970] <TASK> > >>> [ 4103.068071] ? die_addr+0x36/0x90 > >>> [ 4103.071384] ? exc_general_protection+0x1be/0x420 > >>> [ 4103.076081] ? asm_exc_general_protection+0x26/0x30 > >>> [ 4103.080952] ? __kmem_cache_alloc_node+0x1ba/0x320 > >>> [ 4103.085734] ? ext4_htree_store_dirent+0x42/0x180 > >>> [ 4103.090431] ? ext4_htree_store_dirent+0x42/0x180 > >>> [ 4103.095132] __kmalloc+0x4d/0x150 > >>> [ 4103.098444] ext4_htree_store_dirent+0x42/0x180 > >>> [ 4103.102970] htree_dirblock_to_tree+0x1ed/0x370 > >>> [ 4103.107494] ext4_htree_fill_tree+0x109/0x3d0 > >>> [ 4103.111846] ext4_readdir+0x6d4/0xa80 > >>> [ 4103.115505] iterate_dir+0x178/0x1c0 > >>> [ 4103.119076] __x64_sys_getdents64+0x88/0x130 > >>> [ 4103.123341] ? __pfx_filldir64+0x10/0x10 > >>> [ 4103.127260] do_syscall_64+0x5d/0x90 > >>> [ 4103.130835] ? handle_mm_fault+0x11e/0x310 > >>> [ 4103.134927] ? do_user_addr_fault+0x1e0/0x720 > >>> [ 4103.139278] ? exc_page_fault+0x7c/0x180 > >>> [ 4103.143195] entry_SYSCALL_64_after_hwframe+0x72/0xdc > >>> [ 4103.148240] RIP: 0033:0x7f386a418047 > >>> [ 4103.151828] Code: 24 fb ff 4c 89 e0 5b 41 5c 5d c3 0f 1f 84 00 00 > >>> 00 00 00 f3 0f 1e fa b8 ff ff ff 7f 48 39 c2 48 0f 47 d0 b8 d9 00 00 > >>> 00 0f 05 <48> 3d 00 f0 ff ff 77 01 c3 48 8b 15 91 cd 0f 00 f7 d8 64 89 > >>> 02 48 > >>> [ 4103.170543] RSP: 002b:00007ffd4793ff38 EFLAGS: 00000293 ORIG_RAX: > >>> 00000000000000d9 > >>> [ 4103.178095] RAX: ffffffffffffffda RBX: 0000000004933830 RCX: 00007f386a418047 > >>> [ 4103.185214] RDX: 0000000000008000 RSI: 0000000004933860 RDI: 0000000000000006 > >>> [ 4103.192335] RBP: 00007ffd4793ff70 R08: 0000000000000000 R09: 0000000000000001 > >>> [ 4103.199454] R10: 0000000000000004 R11: 0000000000000293 R12: 0000000004933834 > >>> [ 4103.206573] R13: 0000000004933860 R14: ffffffffffffff60 R15: 0000000000000000 > >>> [ 4103.213695] </TASK> > >>> [ 4103.215883] Modules linked in: snd_seq_dummy snd_hrtimer > >>> nf_conntrack_netbios_ns nf_conntrack_broadcast nft_fib_inet > >>> nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 > >>> nf_reject_ipv6 nft_reject nft_ct nft_chain_nat ip6table_nat > >>> ip6table_mangle ip6table_raw ip6table > >>> [ 4103.215911] kvm_intel snd_hwdep snd_seq eeepc_wmi kvm > >>> snd_seq_device asus_wmi iTCO_wdt mei_pxp mei_hdcp ledtrig_audio > >>> irqbypass snd_pcm ee1004 intel_pmc_bxt sparse_keymap rapl snd_timer > >>> pmt_telemetry mei_me iTCO_vendor_support platform_profile joydev > >>> intel_cstate pmt_class snde > >>> [ 4103.366194] ---[ end trace 0000000000000000 ]--- > >>> > >>>>>> --- > >>>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- > >>>>>> drivers/gpu/drm/i915/i915_gem.c | 2 +- > >>>>>> drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- > >>>>>> drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- > >>>>>> drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - > >>>>>> drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- > >>>>>> include/drm/ttm/ttm_bo_api.h | 18 +--- > >>>>>> include/drm/ttm/ttm_device.h | 7 +- > >>>>>> 8 files changed, 57 insertions(+), 111 deletions(-) > >>>>>> > >>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>>>> index 2b1db37e25c1..74ccbd566777 100644 > >>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > >>>>>> @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device > >>>>>> *adev) > >>>>>> amdgpu_fence_driver_hw_fini(adev); > >>>>>> if (adev->mman.initialized) > >>>>>> - flush_delayed_work(&adev->mman.bdev.wq); > >>>>>> + drain_workqueue(adev->mman.bdev.wq); > >>>>>> if (adev->pm_sysfs_en) > >>>>>> amdgpu_pm_sysfs_fini(adev); > >>>>>> diff --git a/drivers/gpu/drm/i915/i915_gem.c > >>>>>> b/drivers/gpu/drm/i915/i915_gem.c > >>>>>> index 8468ca9885fd..c38306f156d6 100644 > >>>>>> --- a/drivers/gpu/drm/i915/i915_gem.c > >>>>>> +++ b/drivers/gpu/drm/i915/i915_gem.c > >>>>>> @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct > >>>>>> drm_i915_private *i915) > >>>>>> { > >>>>>> while (atomic_read(&i915->mm.free_count)) { > >>>>>> flush_work(&i915->mm.free_work); > >>>>>> - flush_delayed_work(&i915->bdev.wq); > >>>>>> + drain_workqueue(i915->bdev.wq); > >>>>>> rcu_barrier(); > >>>>>> } > >>>>>> } > >>>>>> diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c > >>>>>> b/drivers/gpu/drm/i915/intel_region_ttm.c > >>>>>> index cf89d0c2a2d9..657bbc16a48a 100644 > >>>>>> --- a/drivers/gpu/drm/i915/intel_region_ttm.c > >>>>>> +++ b/drivers/gpu/drm/i915/intel_region_ttm.c > >>>>>> @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct > >>>>>> intel_memory_region *mem) > >>>>>> break; > >>>>>> msleep(20); > >>>>>> - flush_delayed_work(&mem->i915->bdev.wq); > >>>>>> + drain_workqueue(mem->i915->bdev.wq); > >>>>>> } > >>>>>> /* If we leaked objects, Don't free the region causing use > >>>>>> after free */ > >>>>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > >>>>>> index b77262a623e0..4749b65bedc4 100644 > >>>>>> --- a/drivers/gpu/drm/ttm/ttm_bo.c > >>>>>> +++ b/drivers/gpu/drm/ttm/ttm_bo.c > >>>>>> @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct > >>>>>> ttm_buffer_object *bo, > >>>>>> ret = 0; > >>>>>> } > >>>>>> - if (ret || unlikely(list_empty(&bo->ddestroy))) { > >>>>>> + if (ret) { > >>>>>> if (unlock_resv) > >>>>>> dma_resv_unlock(bo->base.resv); > >>>>>> spin_unlock(&bo->bdev->lru_lock); > >>>>>> return ret; > >>>>>> } > >>>>>> - list_del_init(&bo->ddestroy); > >>>>>> spin_unlock(&bo->bdev->lru_lock); > >>>>>> ttm_bo_cleanup_memtype_use(bo); > >>>>>> @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct > >>>>>> ttm_buffer_object *bo, > >>>>>> } > >>>>>> /* > >>>>>> - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all > >>>>>> - * encountered buffers. > >>>>>> + * Block for the dma_resv object to become idle, lock the buffer and > >>>>>> clean up > >>>>>> + * the resource and tt object. > >>>>>> */ > >>>>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) > >>>>>> +static void ttm_bo_delayed_delete(struct work_struct *work) > >>>>>> { > >>>>>> - struct list_head removed; > >>>>>> - bool empty; > >>>>>> - > >>>>>> - INIT_LIST_HEAD(&removed); > >>>>>> - > >>>>>> - spin_lock(&bdev->lru_lock); > >>>>>> - while (!list_empty(&bdev->ddestroy)) { > >>>>>> - struct ttm_buffer_object *bo; > >>>>>> - > >>>>>> - bo = list_first_entry(&bdev->ddestroy, struct > >>>>>> ttm_buffer_object, > >>>>>> - ddestroy); > >>>>>> - list_move_tail(&bo->ddestroy, &removed); > >>>>>> - if (!ttm_bo_get_unless_zero(bo)) > >>>>>> - continue; > >>>>>> - > >>>>>> - if (remove_all || bo->base.resv != &bo->base._resv) { > >>>>>> - spin_unlock(&bdev->lru_lock); > >>>>>> - dma_resv_lock(bo->base.resv, NULL); > >>>>>> - > >>>>>> - spin_lock(&bdev->lru_lock); > >>>>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > >>>>>> - > >>>>>> - } else if (dma_resv_trylock(bo->base.resv)) { > >>>>>> - ttm_bo_cleanup_refs(bo, false, !remove_all, true); > >>>>>> - } else { > >>>>>> - spin_unlock(&bdev->lru_lock); > >>>>>> - } > >>>>>> + struct ttm_buffer_object *bo; > >>>>>> - ttm_bo_put(bo); > >>>>>> - spin_lock(&bdev->lru_lock); > >>>>>> - } > >>>>>> - list_splice_tail(&removed, &bdev->ddestroy); > >>>>>> - empty = list_empty(&bdev->ddestroy); > >>>>>> - spin_unlock(&bdev->lru_lock); > >>>>>> + bo = container_of(work, typeof(*bo), delayed_delete); > >>>>>> - return empty; > >>>>>> + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, > >>>>>> false, > >>>>>> + MAX_SCHEDULE_TIMEOUT); > >>>>>> + dma_resv_lock(bo->base.resv, NULL); > >>>>>> + ttm_bo_cleanup_memtype_use(bo); > >>>>>> + dma_resv_unlock(bo->base.resv); > >>>>>> + ttm_bo_put(bo); > >>>>>> } > >>>>>> static void ttm_bo_release(struct kref *kref) > >>>>>> @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) > >>>>>> drm_vma_offset_remove(bdev->vma_manager, > >>>>>> &bo->base.vma_node); > >>>>>> ttm_mem_io_free(bdev, bo->resource); > >>>>>> - } > >>>>>> - > >>>>>> - if (!dma_resv_test_signaled(bo->base.resv, > >>>>>> DMA_RESV_USAGE_BOOKKEEP) || > >>>>>> - !dma_resv_trylock(bo->base.resv)) { > >>>>>> - /* The BO is not idle, resurrect it for delayed destroy */ > >>>>>> - ttm_bo_flush_all_fences(bo); > >>>>>> - bo->deleted = true; > >>>>>> - spin_lock(&bo->bdev->lru_lock); > >>>>>> + if (!dma_resv_test_signaled(bo->base.resv, > >>>>>> + DMA_RESV_USAGE_BOOKKEEP) || > >>>>>> + !dma_resv_trylock(bo->base.resv)) { > >>>>>> + /* The BO is not idle, resurrect it for delayed destroy */ > >>>>>> + ttm_bo_flush_all_fences(bo); > >>>>>> + bo->deleted = true; > >>>>>> - /* > >>>>>> - * Make pinned bos immediately available to > >>>>>> - * shrinkers, now that they are queued for > >>>>>> - * destruction. > >>>>>> - * > >>>>>> - * FIXME: QXL is triggering this. Can be removed when the > >>>>>> - * driver is fixed. > >>>>>> - */ > >>>>>> - if (bo->pin_count) { > >>>>>> - bo->pin_count = 0; > >>>>>> - ttm_resource_move_to_lru_tail(bo->resource); > >>>>>> - } > >>>>>> + spin_lock(&bo->bdev->lru_lock); > >>>>>> - kref_init(&bo->kref); > >>>>>> - list_add_tail(&bo->ddestroy, &bdev->ddestroy); > >>>>>> - spin_unlock(&bo->bdev->lru_lock); > >>>>>> + /* > >>>>>> + * Make pinned bos immediately available to > >>>>>> + * shrinkers, now that they are queued for > >>>>>> + * destruction. > >>>>>> + * > >>>>>> + * FIXME: QXL is triggering this. Can be removed when the > >>>>>> + * driver is fixed. > >>>>>> + */ > >>>>>> + if (bo->pin_count) { > >>>>>> + bo->pin_count = 0; > >>>>>> + ttm_resource_move_to_lru_tail(bo->resource); > >>>>>> + } > >>>>>> - schedule_delayed_work(&bdev->wq, > >>>>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); > >>>>>> - return; > >>>>>> - } > >>>>>> + kref_init(&bo->kref); > >>>>>> + spin_unlock(&bo->bdev->lru_lock); > >>>>>> - spin_lock(&bo->bdev->lru_lock); > >>>>>> - list_del(&bo->ddestroy); > >>>>>> - spin_unlock(&bo->bdev->lru_lock); > >>>>>> + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); > >>>>>> + queue_work(bdev->wq, &bo->delayed_delete); > >>>>>> + return; > >>>>>> + } > >>>>>> - ttm_bo_cleanup_memtype_use(bo); > >>>>>> - dma_resv_unlock(bo->base.resv); > >>>>>> + ttm_bo_cleanup_memtype_use(bo); > >>>>>> + dma_resv_unlock(bo->base.resv); > >>>>>> + } > >>>>>> atomic_dec(&ttm_glob.bo_count); > >>>>>> bo->destroy(bo); > >>>>>> @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, > >>>>>> struct ttm_buffer_object *bo, > >>>>>> int ret; > >>>>>> kref_init(&bo->kref); > >>>>>> - INIT_LIST_HEAD(&bo->ddestroy); > >>>>>> bo->bdev = bdev; > >>>>>> bo->type = type; > >>>>>> bo->page_alignment = alignment; > >>>>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>>>> b/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>>>> index ba3aa0a0fc43..ae4b7922ee1a 100644 > >>>>>> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>>>> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > >>>>>> @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct > >>>>>> ttm_buffer_object *bo, > >>>>>> */ > >>>>>> atomic_inc(&ttm_glob.bo_count); > >>>>>> - INIT_LIST_HEAD(&fbo->base.ddestroy); > >>>>>> drm_vma_node_reset(&fbo->base.base.vma_node); > >>>>>> kref_init(&fbo->base.kref); > >>>>>> diff --git a/drivers/gpu/drm/ttm/ttm_device.c > >>>>>> b/drivers/gpu/drm/ttm/ttm_device.c > >>>>>> index e7147e304637..e9bedca4dfdc 100644 > >>>>>> --- a/drivers/gpu/drm/ttm/ttm_device.c > >>>>>> +++ b/drivers/gpu/drm/ttm/ttm_device.c > >>>>>> @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, > >>>>>> struct ttm_operation_ctx *ctx, > >>>>>> } > >>>>>> EXPORT_SYMBOL(ttm_device_swapout); > >>>>>> -static void ttm_device_delayed_workqueue(struct work_struct *work) > >>>>>> -{ > >>>>>> - struct ttm_device *bdev = > >>>>>> - container_of(work, struct ttm_device, wq.work); > >>>>>> - > >>>>>> - if (!ttm_bo_delayed_delete(bdev, false)) > >>>>>> - schedule_delayed_work(&bdev->wq, > >>>>>> - ((HZ / 100) < 1) ? 1 : HZ / 100); > >>>>>> -} > >>>>>> - > >>>>>> /** > >>>>>> * ttm_device_init > >>>>>> * > >>>>>> @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, > >>>>>> struct ttm_device_funcs *funcs, > >>>>>> if (ret) > >>>>>> return ret; > >>>>>> + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, > >>>>>> 16); > >>>>>> + if (!bdev->wq) { > >>>>>> + ttm_global_release(); > >>>>>> + return -ENOMEM; > >>>>>> + } > >>>>>> + > >>>>>> bdev->funcs = funcs; > >>>>>> ttm_sys_man_init(bdev); > >>>>>> ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); > >>>>>> bdev->vma_manager = vma_manager; > >>>>>> - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); > >>>>>> spin_lock_init(&bdev->lru_lock); > >>>>>> - INIT_LIST_HEAD(&bdev->ddestroy); > >>>>>> INIT_LIST_HEAD(&bdev->pinned); > >>>>>> bdev->dev_mapping = mapping; > >>>>>> mutex_lock(&ttm_global_mutex); > >>>>>> @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) > >>>>>> list_del(&bdev->device_list); > >>>>>> mutex_unlock(&ttm_global_mutex); > >>>>>> - cancel_delayed_work_sync(&bdev->wq); > >>>>>> - > >>>>>> - if (ttm_bo_delayed_delete(bdev, true)) > >>>>>> - pr_debug("Delayed destroy list was clean\n"); > >>>>>> + drain_workqueue(bdev->wq); > >>>>>> + destroy_workqueue(bdev->wq); > >>>>>> spin_lock(&bdev->lru_lock); > >>>>>> for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) > >>>>>> diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h > >>>>>> index 7758347c461c..69e62bbb01e3 100644 > >>>>>> --- a/include/drm/ttm/ttm_bo_api.h > >>>>>> +++ b/include/drm/ttm/ttm_bo_api.h > >>>>>> @@ -92,7 +92,6 @@ struct ttm_tt; > >>>>>> * @ttm: TTM structure holding system pages. > >>>>>> * @evicted: Whether the object was evicted without user-space > >>>>>> knowing. > >>>>>> * @deleted: True if the object is only a zombie and already deleted. > >>>>>> - * @ddestroy: List head for the delayed destroy list. > >>>>>> * @swap: List head for swap LRU list. > >>>>>> * @offset: The current GPU offset, which can have different meanings > >>>>>> * depending on the memory type. For SYSTEM type memory, it should > >>>>>> be 0. > >>>>>> @@ -135,19 +134,14 @@ struct ttm_buffer_object { > >>>>>> struct ttm_tt *ttm; > >>>>>> bool deleted; > >>>>>> struct ttm_lru_bulk_move *bulk_move; > >>>>>> + unsigned priority; > >>>>>> + unsigned pin_count; > >>>>>> /** > >>>>>> - * Members protected by the bdev::lru_lock. > >>>>>> - */ > >>>>>> - > >>>>>> - struct list_head ddestroy; > >>>>>> - > >>>>>> - /** > >>>>>> - * Members protected by a bo reservation. > >>>>>> + * @delayed_delete: Work item used when we can't delete the BO > >>>>>> + * immediately > >>>>>> */ > >>>>>> - > >>>>>> - unsigned priority; > >>>>>> - unsigned pin_count; > >>>>>> + struct work_struct delayed_delete; > >>>>>> /** > >>>>>> * Special members that are protected by the reserve lock > >>>>>> @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); > >>>>>> int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, > >>>>>> void *buf, int len, int write); > >>>>>> -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); > >>>>>> - > >>>>>> vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); > >>>>>> #endif > >>>>>> diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h > >>>>>> index 95b3c04b1ab9..4f3e81eac6f3 100644 > >>>>>> --- a/include/drm/ttm/ttm_device.h > >>>>>> +++ b/include/drm/ttm/ttm_device.h > >>>>>> @@ -251,11 +251,6 @@ struct ttm_device { > >>>>>> */ > >>>>>> spinlock_t lru_lock; > >>>>>> - /** > >>>>>> - * @ddestroy: Destroyed but not yet cleaned up buffer objects. > >>>>>> - */ > >>>>>> - struct list_head ddestroy; > >>>>>> - > >>>>>> /** > >>>>>> * @pinned: Buffer objects which are pinned and so not on any > >>>>>> LRU list. > >>>>>> */ > >>>>>> @@ -270,7 +265,7 @@ struct ttm_device { > >>>>>> /** > >>>>>> * @wq: Work queue structure for the delayed delete workqueue. > >>>>>> */ > >>>>>> - struct delayed_work wq; > >>>>>> + struct workqueue_struct *wq; > >>>>>> }; > >>>>>> int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t > >>>>>> gfp_flags); >
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c index 2b1db37e25c1..74ccbd566777 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c @@ -3984,7 +3984,7 @@ void amdgpu_device_fini_hw(struct amdgpu_device *adev) amdgpu_fence_driver_hw_fini(adev); if (adev->mman.initialized) - flush_delayed_work(&adev->mman.bdev.wq); + drain_workqueue(adev->mman.bdev.wq); if (adev->pm_sysfs_en) amdgpu_pm_sysfs_fini(adev); diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 8468ca9885fd..c38306f156d6 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c @@ -1099,7 +1099,7 @@ void i915_gem_drain_freed_objects(struct drm_i915_private *i915) { while (atomic_read(&i915->mm.free_count)) { flush_work(&i915->mm.free_work); - flush_delayed_work(&i915->bdev.wq); + drain_workqueue(i915->bdev.wq); rcu_barrier(); } } diff --git a/drivers/gpu/drm/i915/intel_region_ttm.c b/drivers/gpu/drm/i915/intel_region_ttm.c index cf89d0c2a2d9..657bbc16a48a 100644 --- a/drivers/gpu/drm/i915/intel_region_ttm.c +++ b/drivers/gpu/drm/i915/intel_region_ttm.c @@ -132,7 +132,7 @@ int intel_region_ttm_fini(struct intel_memory_region *mem) break; msleep(20); - flush_delayed_work(&mem->i915->bdev.wq); + drain_workqueue(mem->i915->bdev.wq); } /* If we leaked objects, Don't free the region causing use after free */ diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c index b77262a623e0..4749b65bedc4 100644 --- a/drivers/gpu/drm/ttm/ttm_bo.c +++ b/drivers/gpu/drm/ttm/ttm_bo.c @@ -280,14 +280,13 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object *bo, ret = 0; } - if (ret || unlikely(list_empty(&bo->ddestroy))) { + if (ret) { if (unlock_resv) dma_resv_unlock(bo->base.resv); spin_unlock(&bo->bdev->lru_lock); return ret; } - list_del_init(&bo->ddestroy); spin_unlock(&bo->bdev->lru_lock); ttm_bo_cleanup_memtype_use(bo); @@ -300,47 +299,21 @@ static int ttm_bo_cleanup_refs(struct ttm_buffer_object *bo, } /* - * Traverse the delayed list, and call ttm_bo_cleanup_refs on all - * encountered buffers. + * Block for the dma_resv object to become idle, lock the buffer and clean up + * the resource and tt object. */ -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all) +static void ttm_bo_delayed_delete(struct work_struct *work) { - struct list_head removed; - bool empty; - - INIT_LIST_HEAD(&removed); - - spin_lock(&bdev->lru_lock); - while (!list_empty(&bdev->ddestroy)) { - struct ttm_buffer_object *bo; - - bo = list_first_entry(&bdev->ddestroy, struct ttm_buffer_object, - ddestroy); - list_move_tail(&bo->ddestroy, &removed); - if (!ttm_bo_get_unless_zero(bo)) - continue; - - if (remove_all || bo->base.resv != &bo->base._resv) { - spin_unlock(&bdev->lru_lock); - dma_resv_lock(bo->base.resv, NULL); - - spin_lock(&bdev->lru_lock); - ttm_bo_cleanup_refs(bo, false, !remove_all, true); - - } else if (dma_resv_trylock(bo->base.resv)) { - ttm_bo_cleanup_refs(bo, false, !remove_all, true); - } else { - spin_unlock(&bdev->lru_lock); - } + struct ttm_buffer_object *bo; - ttm_bo_put(bo); - spin_lock(&bdev->lru_lock); - } - list_splice_tail(&removed, &bdev->ddestroy); - empty = list_empty(&bdev->ddestroy); - spin_unlock(&bdev->lru_lock); + bo = container_of(work, typeof(*bo), delayed_delete); - return empty; + dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP, false, + MAX_SCHEDULE_TIMEOUT); + dma_resv_lock(bo->base.resv, NULL); + ttm_bo_cleanup_memtype_use(bo); + dma_resv_unlock(bo->base.resv); + ttm_bo_put(bo); } static void ttm_bo_release(struct kref *kref) @@ -369,44 +342,40 @@ static void ttm_bo_release(struct kref *kref) drm_vma_offset_remove(bdev->vma_manager, &bo->base.vma_node); ttm_mem_io_free(bdev, bo->resource); - } - - if (!dma_resv_test_signaled(bo->base.resv, DMA_RESV_USAGE_BOOKKEEP) || - !dma_resv_trylock(bo->base.resv)) { - /* The BO is not idle, resurrect it for delayed destroy */ - ttm_bo_flush_all_fences(bo); - bo->deleted = true; - spin_lock(&bo->bdev->lru_lock); + if (!dma_resv_test_signaled(bo->base.resv, + DMA_RESV_USAGE_BOOKKEEP) || + !dma_resv_trylock(bo->base.resv)) { + /* The BO is not idle, resurrect it for delayed destroy */ + ttm_bo_flush_all_fences(bo); + bo->deleted = true; - /* - * Make pinned bos immediately available to - * shrinkers, now that they are queued for - * destruction. - * - * FIXME: QXL is triggering this. Can be removed when the - * driver is fixed. - */ - if (bo->pin_count) { - bo->pin_count = 0; - ttm_resource_move_to_lru_tail(bo->resource); - } + spin_lock(&bo->bdev->lru_lock); - kref_init(&bo->kref); - list_add_tail(&bo->ddestroy, &bdev->ddestroy); - spin_unlock(&bo->bdev->lru_lock); + /* + * Make pinned bos immediately available to + * shrinkers, now that they are queued for + * destruction. + * + * FIXME: QXL is triggering this. Can be removed when the + * driver is fixed. + */ + if (bo->pin_count) { + bo->pin_count = 0; + ttm_resource_move_to_lru_tail(bo->resource); + } - schedule_delayed_work(&bdev->wq, - ((HZ / 100) < 1) ? 1 : HZ / 100); - return; - } + kref_init(&bo->kref); + spin_unlock(&bo->bdev->lru_lock); - spin_lock(&bo->bdev->lru_lock); - list_del(&bo->ddestroy); - spin_unlock(&bo->bdev->lru_lock); + INIT_WORK(&bo->delayed_delete, ttm_bo_delayed_delete); + queue_work(bdev->wq, &bo->delayed_delete); + return; + } - ttm_bo_cleanup_memtype_use(bo); - dma_resv_unlock(bo->base.resv); + ttm_bo_cleanup_memtype_use(bo); + dma_resv_unlock(bo->base.resv); + } atomic_dec(&ttm_glob.bo_count); bo->destroy(bo); @@ -946,7 +915,6 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, struct ttm_buffer_object *bo, int ret; kref_init(&bo->kref); - INIT_LIST_HEAD(&bo->ddestroy); bo->bdev = bdev; bo->type = type; bo->page_alignment = alignment; diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c index ba3aa0a0fc43..ae4b7922ee1a 100644 --- a/drivers/gpu/drm/ttm/ttm_bo_util.c +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c @@ -230,7 +230,6 @@ static int ttm_buffer_object_transfer(struct ttm_buffer_object *bo, */ atomic_inc(&ttm_glob.bo_count); - INIT_LIST_HEAD(&fbo->base.ddestroy); drm_vma_node_reset(&fbo->base.base.vma_node); kref_init(&fbo->base.kref); diff --git a/drivers/gpu/drm/ttm/ttm_device.c b/drivers/gpu/drm/ttm/ttm_device.c index e7147e304637..e9bedca4dfdc 100644 --- a/drivers/gpu/drm/ttm/ttm_device.c +++ b/drivers/gpu/drm/ttm/ttm_device.c @@ -175,16 +175,6 @@ int ttm_device_swapout(struct ttm_device *bdev, struct ttm_operation_ctx *ctx, } EXPORT_SYMBOL(ttm_device_swapout); -static void ttm_device_delayed_workqueue(struct work_struct *work) -{ - struct ttm_device *bdev = - container_of(work, struct ttm_device, wq.work); - - if (!ttm_bo_delayed_delete(bdev, false)) - schedule_delayed_work(&bdev->wq, - ((HZ / 100) < 1) ? 1 : HZ / 100); -} - /** * ttm_device_init * @@ -215,15 +205,19 @@ int ttm_device_init(struct ttm_device *bdev, struct ttm_device_funcs *funcs, if (ret) return ret; + bdev->wq = alloc_workqueue("ttm", WQ_MEM_RECLAIM | WQ_HIGHPRI, 16); + if (!bdev->wq) { + ttm_global_release(); + return -ENOMEM; + } + bdev->funcs = funcs; ttm_sys_man_init(bdev); ttm_pool_init(&bdev->pool, dev, use_dma_alloc, use_dma32); bdev->vma_manager = vma_manager; - INIT_DELAYED_WORK(&bdev->wq, ttm_device_delayed_workqueue); spin_lock_init(&bdev->lru_lock); - INIT_LIST_HEAD(&bdev->ddestroy); INIT_LIST_HEAD(&bdev->pinned); bdev->dev_mapping = mapping; mutex_lock(&ttm_global_mutex); @@ -247,10 +241,8 @@ void ttm_device_fini(struct ttm_device *bdev) list_del(&bdev->device_list); mutex_unlock(&ttm_global_mutex); - cancel_delayed_work_sync(&bdev->wq); - - if (ttm_bo_delayed_delete(bdev, true)) - pr_debug("Delayed destroy list was clean\n"); + drain_workqueue(bdev->wq); + destroy_workqueue(bdev->wq); spin_lock(&bdev->lru_lock); for (i = 0; i < TTM_MAX_BO_PRIORITY; ++i) diff --git a/include/drm/ttm/ttm_bo_api.h b/include/drm/ttm/ttm_bo_api.h index 7758347c461c..69e62bbb01e3 100644 --- a/include/drm/ttm/ttm_bo_api.h +++ b/include/drm/ttm/ttm_bo_api.h @@ -92,7 +92,6 @@ struct ttm_tt; * @ttm: TTM structure holding system pages. * @evicted: Whether the object was evicted without user-space knowing. * @deleted: True if the object is only a zombie and already deleted. - * @ddestroy: List head for the delayed destroy list. * @swap: List head for swap LRU list. * @offset: The current GPU offset, which can have different meanings * depending on the memory type. For SYSTEM type memory, it should be 0. @@ -135,19 +134,14 @@ struct ttm_buffer_object { struct ttm_tt *ttm; bool deleted; struct ttm_lru_bulk_move *bulk_move; + unsigned priority; + unsigned pin_count; /** - * Members protected by the bdev::lru_lock. - */ - - struct list_head ddestroy; - - /** - * Members protected by a bo reservation. + * @delayed_delete: Work item used when we can't delete the BO + * immediately */ - - unsigned priority; - unsigned pin_count; + struct work_struct delayed_delete; /** * Special members that are protected by the reserve lock @@ -448,8 +442,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma); int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write); -bool ttm_bo_delayed_delete(struct ttm_device *bdev, bool remove_all); - vm_fault_t ttm_bo_vm_dummy_page(struct vm_fault *vmf, pgprot_t prot); #endif diff --git a/include/drm/ttm/ttm_device.h b/include/drm/ttm/ttm_device.h index 95b3c04b1ab9..4f3e81eac6f3 100644 --- a/include/drm/ttm/ttm_device.h +++ b/include/drm/ttm/ttm_device.h @@ -251,11 +251,6 @@ struct ttm_device { */ spinlock_t lru_lock; - /** - * @ddestroy: Destroyed but not yet cleaned up buffer objects. - */ - struct list_head ddestroy; - /** * @pinned: Buffer objects which are pinned and so not on any LRU list. */ @@ -270,7 +265,7 @@ struct ttm_device { /** * @wq: Work queue structure for the delayed delete workqueue. */ - struct delayed_work wq; + struct workqueue_struct *wq; }; int ttm_global_swapout(struct ttm_operation_ctx *ctx, gfp_t gfp_flags);
Instead of a single worker going over the list of delete BOs in regular intervals use a per BO worker which blocks for the resv object and locking of the BO. This not only simplifies the handling massively, but also results in much better response time when cleaning up buffers. Signed-off-by: Christian König <christian.koenig@amd.com> --- drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- drivers/gpu/drm/i915/i915_gem.c | 2 +- drivers/gpu/drm/i915/intel_region_ttm.c | 2 +- drivers/gpu/drm/ttm/ttm_bo.c | 112 ++++++++------------- drivers/gpu/drm/ttm/ttm_bo_util.c | 1 - drivers/gpu/drm/ttm/ttm_device.c | 24 ++--- include/drm/ttm/ttm_bo_api.h | 18 +--- include/drm/ttm/ttm_device.h | 7 +- 8 files changed, 57 insertions(+), 111 deletions(-)