Message ID | 20221026153557.63541-1-christian.koenig@amd.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [1/5] drm/amd/amdgpu revert "implement tdr advanced mode" | expand |
Hi, Christian A later bad compute job can block a good gfx job, so the original TDR design find a wrong guilty job(good gfx job). Advanced TDR re-submits jobs in order to find the real guilty job(bad compute job). After reverting this commit, how does the new gang-submit promise the isolation between compute jobs and gfx jobs? On 10/26/2022 11:35 PM, Christian König wrote: > This reverts commit e6c6338f393b74ac0b303d567bb918b44ae7ad75. > > This feature basically re-submits one job after another to > figure out which one was the one causing a hang. > > This is obviously incompatible with gang-submit which requires > that multiple jobs run at the same time. It's also absolutely > not helpful to crash the hardware multiple times if a clean > recovery is desired. > > For testing and debugging environments we should rather disable > recovery alltogether to be able to inspect the state with a hw > debugger. > > Additional to that the sw implementation is clearly buggy and causes > reference count issues for the hardware fence. > > Signed-off-by: Christian König <christian.koenig@amd.com> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 103 --------------------- > drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c | 2 +- > drivers/gpu/drm/scheduler/sched_main.c | 58 ++---------- > include/drm/gpu_scheduler.h | 3 - > 4 files changed, 10 insertions(+), 156 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > index 6f958603c8cc..d4584e577b51 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > @@ -5070,94 +5070,6 @@ static int amdgpu_device_suspend_display_audio(struct amdgpu_device *adev) > return 0; > } > > -static void amdgpu_device_recheck_guilty_jobs( > - struct amdgpu_device *adev, struct list_head *device_list_handle, > - struct amdgpu_reset_context *reset_context) > -{ > - int i, r = 0; > - > - for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { > - struct amdgpu_ring *ring = adev->rings[i]; > - int ret = 0; > - struct drm_sched_job *s_job; > - > - if (!ring || !ring->sched.thread) > - continue; > - > - s_job = list_first_entry_or_null(&ring->sched.pending_list, > - struct drm_sched_job, list); > - if (s_job == NULL) > - continue; > - > - /* clear job's guilty and depend the folowing step to decide the real one */ > - drm_sched_reset_karma(s_job); > - drm_sched_resubmit_jobs_ext(&ring->sched, 1); > - > - if (!s_job->s_fence->parent) { > - DRM_WARN("Failed to get a HW fence for job!"); > - continue; > - } > - > - ret = dma_fence_wait_timeout(s_job->s_fence->parent, false, ring->sched.timeout); > - if (ret == 0) { /* timeout */ > - DRM_ERROR("Found the real bad job! ring:%s, job_id:%llx\n", > - ring->sched.name, s_job->id); > - > - > - amdgpu_fence_driver_isr_toggle(adev, true); > - > - /* Clear this failed job from fence array */ > - amdgpu_fence_driver_clear_job_fences(ring); > - > - amdgpu_fence_driver_isr_toggle(adev, false); > - > - /* Since the job won't signal and we go for > - * another resubmit drop this parent pointer > - */ > - dma_fence_put(s_job->s_fence->parent); > - s_job->s_fence->parent = NULL; > - > - /* set guilty */ > - drm_sched_increase_karma(s_job); > - amdgpu_reset_prepare_hwcontext(adev, reset_context); > -retry: > - /* do hw reset */ > - if (amdgpu_sriov_vf(adev)) { > - amdgpu_virt_fini_data_exchange(adev); > - r = amdgpu_device_reset_sriov(adev, false); > - if (r) > - adev->asic_reset_res = r; > - } else { > - clear_bit(AMDGPU_SKIP_HW_RESET, > - &reset_context->flags); > - r = amdgpu_do_asic_reset(device_list_handle, > - reset_context); > - if (r && r == -EAGAIN) > - goto retry; > - } > - > - /* > - * add reset counter so that the following > - * resubmitted job could flush vmid > - */ > - atomic_inc(&adev->gpu_reset_counter); > - continue; > - } > - > - /* got the hw fence, signal finished fence */ > - atomic_dec(ring->sched.score); > - dma_fence_get(&s_job->s_fence->finished); > - dma_fence_signal(&s_job->s_fence->finished); > - dma_fence_put(&s_job->s_fence->finished); > - > - /* remove node from list and free the job */ > - spin_lock(&ring->sched.job_list_lock); > - list_del_init(&s_job->list); > - spin_unlock(&ring->sched.job_list_lock); > - ring->sched.ops->free_job(s_job); > - } > -} > - > static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) > { > struct amdgpu_ras *con = amdgpu_ras_get_context(adev); > @@ -5178,7 +5090,6 @@ static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) > > } > > - > /** > * amdgpu_device_gpu_recover - reset the asic and recover scheduler > * > @@ -5201,7 +5112,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, > int i, r = 0; > bool need_emergency_restart = false; > bool audio_suspended = false; > - int tmp_vram_lost_counter; > bool gpu_reset_for_dev_remove = false; > > gpu_reset_for_dev_remove = > @@ -5347,7 +5257,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, > amdgpu_device_stop_pending_resets(tmp_adev); > } > > - tmp_vram_lost_counter = atomic_read(&((adev)->vram_lost_counter)); > /* Actual ASIC resets if needed.*/ > /* Host driver will handle XGMI hive reset for SRIOV */ > if (amdgpu_sriov_vf(adev)) { > @@ -5372,18 +5281,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, > /* Post ASIC reset for all devs .*/ > list_for_each_entry(tmp_adev, device_list_handle, reset_list) { > > - /* > - * Sometimes a later bad compute job can block a good gfx job as gfx > - * and compute ring share internal GC HW mutually. We add an additional > - * guilty jobs recheck step to find the real guilty job, it synchronously > - * submits and pends for the first job being signaled. If it gets timeout, > - * we identify it as a real guilty job. > - */ > - if (amdgpu_gpu_recovery == 2 && > - !(tmp_vram_lost_counter < atomic_read(&adev->vram_lost_counter))) > - amdgpu_device_recheck_guilty_jobs( > - tmp_adev, device_list_handle, reset_context); > - > for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { > struct amdgpu_ring *ring = tmp_adev->rings[i]; > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c > index 16f6a313335e..f177d8e5b665 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c > @@ -519,7 +519,7 @@ module_param_named(compute_multipipe, amdgpu_compute_multipipe, int, 0444); > * DOC: gpu_recovery (int) > * Set to enable GPU recovery mechanism (1 = enable, 0 = disable). The default is -1 (auto, disabled except SRIOV). > */ > -MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (2 = advanced tdr mode, 1 = enable, 0 = disable, -1 = auto)"); > +MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (1 = enable, 0 = disable, -1 = auto)"); > module_param_named(gpu_recovery, amdgpu_gpu_recovery, int, 0444); > > /** > diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c > index e0ab14e0fb6b..bb28f31bff6f 100644 > --- a/drivers/gpu/drm/scheduler/sched_main.c > +++ b/drivers/gpu/drm/scheduler/sched_main.c > @@ -355,27 +355,6 @@ static void drm_sched_job_timedout(struct work_struct *work) > } > } > > - /** > - * drm_sched_increase_karma - Update sched_entity guilty flag > - * > - * @bad: The job guilty of time out > - * > - * Increment on every hang caused by the 'bad' job. If this exceeds the hang > - * limit of the scheduler then the respective sched entity is marked guilty and > - * jobs from it will not be scheduled further > - */ > -void drm_sched_increase_karma(struct drm_sched_job *bad) > -{ > - drm_sched_increase_karma_ext(bad, 1); > -} > -EXPORT_SYMBOL(drm_sched_increase_karma); > - > -void drm_sched_reset_karma(struct drm_sched_job *bad) > -{ > - drm_sched_increase_karma_ext(bad, 0); > -} > -EXPORT_SYMBOL(drm_sched_reset_karma); > - > /** > * drm_sched_stop - stop the scheduler > * > @@ -516,32 +495,15 @@ EXPORT_SYMBOL(drm_sched_start); > * > */ > void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched) > -{ > - drm_sched_resubmit_jobs_ext(sched, INT_MAX); > -} > -EXPORT_SYMBOL(drm_sched_resubmit_jobs); > - > -/** > - * drm_sched_resubmit_jobs_ext - helper to relunch certain number of jobs from mirror ring list > - * > - * @sched: scheduler instance > - * @max: job numbers to relaunch > - * > - */ > -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) > { > struct drm_sched_job *s_job, *tmp; > uint64_t guilty_context; > bool found_guilty = false; > struct dma_fence *fence; > - int i = 0; > > list_for_each_entry_safe(s_job, tmp, &sched->pending_list, list) { > struct drm_sched_fence *s_fence = s_job->s_fence; > > - if (i >= max) > - break; > - > if (!found_guilty && atomic_read(&s_job->karma) > sched->hang_limit) { > found_guilty = true; > guilty_context = s_job->s_fence->scheduled.context; > @@ -551,7 +513,6 @@ void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) > dma_fence_set_error(&s_fence->finished, -ECANCELED); > > fence = sched->ops->run_job(s_job); > - i++; > > if (IS_ERR_OR_NULL(fence)) { > if (IS_ERR(fence)) > @@ -567,7 +528,7 @@ void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) > } > } > } > -EXPORT_SYMBOL(drm_sched_resubmit_jobs_ext); > +EXPORT_SYMBOL(drm_sched_resubmit_jobs); > > /** > * drm_sched_job_init - init a scheduler job > @@ -1081,13 +1042,15 @@ void drm_sched_fini(struct drm_gpu_scheduler *sched) > EXPORT_SYMBOL(drm_sched_fini); > > /** > - * drm_sched_increase_karma_ext - Update sched_entity guilty flag > + * drm_sched_increase_karma - Update sched_entity guilty flag > * > * @bad: The job guilty of time out > - * @type: type for increase/reset karma > * > + * Increment on every hang caused by the 'bad' job. If this exceeds the hang > + * limit of the scheduler then the respective sched entity is marked guilty and > + * jobs from it will not be scheduled further > */ > -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > +void drm_sched_increase_karma(struct drm_sched_job *bad) > { > int i; > struct drm_sched_entity *tmp; > @@ -1099,10 +1062,7 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > * corrupt but keep in mind that kernel jobs always considered good. > */ > if (bad->s_priority != DRM_SCHED_PRIORITY_KERNEL) { > - if (type == 0) > - atomic_set(&bad->karma, 0); > - else if (type == 1) > - atomic_inc(&bad->karma); > + atomic_inc(&bad->karma); > > for (i = DRM_SCHED_PRIORITY_MIN; i < DRM_SCHED_PRIORITY_KERNEL; > i++) { > @@ -1113,7 +1073,7 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > if (bad->s_fence->scheduled.context == > entity->fence_context) { > if (entity->guilty) > - atomic_set(entity->guilty, type); > + atomic_set(entity->guilty, 1); > break; > } > } > @@ -1123,4 +1083,4 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > } > } > } > -EXPORT_SYMBOL(drm_sched_increase_karma_ext); > +EXPORT_SYMBOL(drm_sched_increase_karma); > diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h > index 0fca8f38bee4..c564be29c5ae 100644 > --- a/include/drm/gpu_scheduler.h > +++ b/include/drm/gpu_scheduler.h > @@ -488,10 +488,7 @@ void drm_sched_wakeup(struct drm_gpu_scheduler *sched); > void drm_sched_stop(struct drm_gpu_scheduler *sched, struct drm_sched_job *bad); > void drm_sched_start(struct drm_gpu_scheduler *sched, bool full_recovery); > void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched); > -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max); > void drm_sched_increase_karma(struct drm_sched_job *bad); > -void drm_sched_reset_karma(struct drm_sched_job *bad); > -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type); > bool drm_sched_dependency_optimized(struct dma_fence* fence, > struct drm_sched_entity *entity); > void drm_sched_fault(struct drm_gpu_scheduler *sched);
The series is, Acked-by: Luben Tuikov <luben.tuikov@amd.com> We don't want the kernel to be in the business of retrying client's requests. Instead we want the kernel to provide a conduit for such requests to be sent, executed by the GPU, and a result returned. If the kernel cannot process requests for any reason, e.g. GPU is being reset, or in recovery, or OOM, or there's a problem with the request, etc., the driver should return the request back to the client (user space), and let the client decide how to proceed. Regards, Luben On 2022-10-26 11:35, Christian König wrote: > This reverts commit e6c6338f393b74ac0b303d567bb918b44ae7ad75. > > This feature basically re-submits one job after another to > figure out which one was the one causing a hang. > > This is obviously incompatible with gang-submit which requires > that multiple jobs run at the same time. It's also absolutely > not helpful to crash the hardware multiple times if a clean > recovery is desired. > > For testing and debugging environments we should rather disable > recovery alltogether to be able to inspect the state with a hw > debugger. > > Additional to that the sw implementation is clearly buggy and causes > reference count issues for the hardware fence. > > Signed-off-by: Christian König <christian.koenig@amd.com> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 103 --------------------- > drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c | 2 +- > drivers/gpu/drm/scheduler/sched_main.c | 58 ++---------- > include/drm/gpu_scheduler.h | 3 - > 4 files changed, 10 insertions(+), 156 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > index 6f958603c8cc..d4584e577b51 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > @@ -5070,94 +5070,6 @@ static int amdgpu_device_suspend_display_audio(struct amdgpu_device *adev) > return 0; > } > > -static void amdgpu_device_recheck_guilty_jobs( > - struct amdgpu_device *adev, struct list_head *device_list_handle, > - struct amdgpu_reset_context *reset_context) > -{ > - int i, r = 0; > - > - for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { > - struct amdgpu_ring *ring = adev->rings[i]; > - int ret = 0; > - struct drm_sched_job *s_job; > - > - if (!ring || !ring->sched.thread) > - continue; > - > - s_job = list_first_entry_or_null(&ring->sched.pending_list, > - struct drm_sched_job, list); > - if (s_job == NULL) > - continue; > - > - /* clear job's guilty and depend the folowing step to decide the real one */ > - drm_sched_reset_karma(s_job); > - drm_sched_resubmit_jobs_ext(&ring->sched, 1); > - > - if (!s_job->s_fence->parent) { > - DRM_WARN("Failed to get a HW fence for job!"); > - continue; > - } > - > - ret = dma_fence_wait_timeout(s_job->s_fence->parent, false, ring->sched.timeout); > - if (ret == 0) { /* timeout */ > - DRM_ERROR("Found the real bad job! ring:%s, job_id:%llx\n", > - ring->sched.name, s_job->id); > - > - > - amdgpu_fence_driver_isr_toggle(adev, true); > - > - /* Clear this failed job from fence array */ > - amdgpu_fence_driver_clear_job_fences(ring); > - > - amdgpu_fence_driver_isr_toggle(adev, false); > - > - /* Since the job won't signal and we go for > - * another resubmit drop this parent pointer > - */ > - dma_fence_put(s_job->s_fence->parent); > - s_job->s_fence->parent = NULL; > - > - /* set guilty */ > - drm_sched_increase_karma(s_job); > - amdgpu_reset_prepare_hwcontext(adev, reset_context); > -retry: > - /* do hw reset */ > - if (amdgpu_sriov_vf(adev)) { > - amdgpu_virt_fini_data_exchange(adev); > - r = amdgpu_device_reset_sriov(adev, false); > - if (r) > - adev->asic_reset_res = r; > - } else { > - clear_bit(AMDGPU_SKIP_HW_RESET, > - &reset_context->flags); > - r = amdgpu_do_asic_reset(device_list_handle, > - reset_context); > - if (r && r == -EAGAIN) > - goto retry; > - } > - > - /* > - * add reset counter so that the following > - * resubmitted job could flush vmid > - */ > - atomic_inc(&adev->gpu_reset_counter); > - continue; > - } > - > - /* got the hw fence, signal finished fence */ > - atomic_dec(ring->sched.score); > - dma_fence_get(&s_job->s_fence->finished); > - dma_fence_signal(&s_job->s_fence->finished); > - dma_fence_put(&s_job->s_fence->finished); > - > - /* remove node from list and free the job */ > - spin_lock(&ring->sched.job_list_lock); > - list_del_init(&s_job->list); > - spin_unlock(&ring->sched.job_list_lock); > - ring->sched.ops->free_job(s_job); > - } > -} > - > static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) > { > struct amdgpu_ras *con = amdgpu_ras_get_context(adev); > @@ -5178,7 +5090,6 @@ static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) > > } > > - > /** > * amdgpu_device_gpu_recover - reset the asic and recover scheduler > * > @@ -5201,7 +5112,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, > int i, r = 0; > bool need_emergency_restart = false; > bool audio_suspended = false; > - int tmp_vram_lost_counter; > bool gpu_reset_for_dev_remove = false; > > gpu_reset_for_dev_remove = > @@ -5347,7 +5257,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, > amdgpu_device_stop_pending_resets(tmp_adev); > } > > - tmp_vram_lost_counter = atomic_read(&((adev)->vram_lost_counter)); > /* Actual ASIC resets if needed.*/ > /* Host driver will handle XGMI hive reset for SRIOV */ > if (amdgpu_sriov_vf(adev)) { > @@ -5372,18 +5281,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, > /* Post ASIC reset for all devs .*/ > list_for_each_entry(tmp_adev, device_list_handle, reset_list) { > > - /* > - * Sometimes a later bad compute job can block a good gfx job as gfx > - * and compute ring share internal GC HW mutually. We add an additional > - * guilty jobs recheck step to find the real guilty job, it synchronously > - * submits and pends for the first job being signaled. If it gets timeout, > - * we identify it as a real guilty job. > - */ > - if (amdgpu_gpu_recovery == 2 && > - !(tmp_vram_lost_counter < atomic_read(&adev->vram_lost_counter))) > - amdgpu_device_recheck_guilty_jobs( > - tmp_adev, device_list_handle, reset_context); > - > for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { > struct amdgpu_ring *ring = tmp_adev->rings[i]; > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c > index 16f6a313335e..f177d8e5b665 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c > @@ -519,7 +519,7 @@ module_param_named(compute_multipipe, amdgpu_compute_multipipe, int, 0444); > * DOC: gpu_recovery (int) > * Set to enable GPU recovery mechanism (1 = enable, 0 = disable). The default is -1 (auto, disabled except SRIOV). > */ > -MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (2 = advanced tdr mode, 1 = enable, 0 = disable, -1 = auto)"); > +MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (1 = enable, 0 = disable, -1 = auto)"); > module_param_named(gpu_recovery, amdgpu_gpu_recovery, int, 0444); > > /** > diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c > index e0ab14e0fb6b..bb28f31bff6f 100644 > --- a/drivers/gpu/drm/scheduler/sched_main.c > +++ b/drivers/gpu/drm/scheduler/sched_main.c > @@ -355,27 +355,6 @@ static void drm_sched_job_timedout(struct work_struct *work) > } > } > > - /** > - * drm_sched_increase_karma - Update sched_entity guilty flag > - * > - * @bad: The job guilty of time out > - * > - * Increment on every hang caused by the 'bad' job. If this exceeds the hang > - * limit of the scheduler then the respective sched entity is marked guilty and > - * jobs from it will not be scheduled further > - */ > -void drm_sched_increase_karma(struct drm_sched_job *bad) > -{ > - drm_sched_increase_karma_ext(bad, 1); > -} > -EXPORT_SYMBOL(drm_sched_increase_karma); > - > -void drm_sched_reset_karma(struct drm_sched_job *bad) > -{ > - drm_sched_increase_karma_ext(bad, 0); > -} > -EXPORT_SYMBOL(drm_sched_reset_karma); > - > /** > * drm_sched_stop - stop the scheduler > * > @@ -516,32 +495,15 @@ EXPORT_SYMBOL(drm_sched_start); > * > */ > void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched) > -{ > - drm_sched_resubmit_jobs_ext(sched, INT_MAX); > -} > -EXPORT_SYMBOL(drm_sched_resubmit_jobs); > - > -/** > - * drm_sched_resubmit_jobs_ext - helper to relunch certain number of jobs from mirror ring list > - * > - * @sched: scheduler instance > - * @max: job numbers to relaunch > - * > - */ > -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) > { > struct drm_sched_job *s_job, *tmp; > uint64_t guilty_context; > bool found_guilty = false; > struct dma_fence *fence; > - int i = 0; > > list_for_each_entry_safe(s_job, tmp, &sched->pending_list, list) { > struct drm_sched_fence *s_fence = s_job->s_fence; > > - if (i >= max) > - break; > - > if (!found_guilty && atomic_read(&s_job->karma) > sched->hang_limit) { > found_guilty = true; > guilty_context = s_job->s_fence->scheduled.context; > @@ -551,7 +513,6 @@ void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) > dma_fence_set_error(&s_fence->finished, -ECANCELED); > > fence = sched->ops->run_job(s_job); > - i++; > > if (IS_ERR_OR_NULL(fence)) { > if (IS_ERR(fence)) > @@ -567,7 +528,7 @@ void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) > } > } > } > -EXPORT_SYMBOL(drm_sched_resubmit_jobs_ext); > +EXPORT_SYMBOL(drm_sched_resubmit_jobs); > > /** > * drm_sched_job_init - init a scheduler job > @@ -1081,13 +1042,15 @@ void drm_sched_fini(struct drm_gpu_scheduler *sched) > EXPORT_SYMBOL(drm_sched_fini); > > /** > - * drm_sched_increase_karma_ext - Update sched_entity guilty flag > + * drm_sched_increase_karma - Update sched_entity guilty flag > * > * @bad: The job guilty of time out > - * @type: type for increase/reset karma > * > + * Increment on every hang caused by the 'bad' job. If this exceeds the hang > + * limit of the scheduler then the respective sched entity is marked guilty and > + * jobs from it will not be scheduled further > */ > -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > +void drm_sched_increase_karma(struct drm_sched_job *bad) > { > int i; > struct drm_sched_entity *tmp; > @@ -1099,10 +1062,7 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > * corrupt but keep in mind that kernel jobs always considered good. > */ > if (bad->s_priority != DRM_SCHED_PRIORITY_KERNEL) { > - if (type == 0) > - atomic_set(&bad->karma, 0); > - else if (type == 1) > - atomic_inc(&bad->karma); > + atomic_inc(&bad->karma); > > for (i = DRM_SCHED_PRIORITY_MIN; i < DRM_SCHED_PRIORITY_KERNEL; > i++) { > @@ -1113,7 +1073,7 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > if (bad->s_fence->scheduled.context == > entity->fence_context) { > if (entity->guilty) > - atomic_set(entity->guilty, type); > + atomic_set(entity->guilty, 1); > break; > } > } > @@ -1123,4 +1083,4 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) > } > } > } > -EXPORT_SYMBOL(drm_sched_increase_karma_ext); > +EXPORT_SYMBOL(drm_sched_increase_karma); > diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h > index 0fca8f38bee4..c564be29c5ae 100644 > --- a/include/drm/gpu_scheduler.h > +++ b/include/drm/gpu_scheduler.h > @@ -488,10 +488,7 @@ void drm_sched_wakeup(struct drm_gpu_scheduler *sched); > void drm_sched_stop(struct drm_gpu_scheduler *sched, struct drm_sched_job *bad); > void drm_sched_start(struct drm_gpu_scheduler *sched, bool full_recovery); > void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched); > -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max); > void drm_sched_increase_karma(struct drm_sched_job *bad); > -void drm_sched_reset_karma(struct drm_sched_job *bad); > -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type); > bool drm_sched_dependency_optimized(struct dma_fence* fence, > struct drm_sched_entity *entity); > void drm_sched_fault(struct drm_gpu_scheduler *sched);
> A later bad compute job can block a good gfx job The gang submit/barrier approach makes sure that only one application at a time can use the gfx/compute block. So when application B makes a compute submission while a GFX submission of application A is still running we will wait for that GFX submission to complete before B is allowed to access the hardware. Two submissions to the same engine will wait in the hardware for the first submission to fully complete before the second is started. This way neither concurrent submissions to different engines (compute/gfx) nor subsequent submissions to the same engine let different applications access the hardware at the same time. It's essentially the same requirement we have for gang submit and that is now well tested and in production for a couple of month. So we already know that the approach works. Regards, Christian. Am 28.01.23 um 09:28 schrieb Yin, ZhenGuo (Chris): > Hi, Christian > > A later bad compute job can block a good gfx job, so the original TDR > design find a wrong guilty job(good gfx job). > > Advanced TDR re-submits jobs in order to find the real guilty job(bad > compute job). > > After reverting this commit, how does the new gang-submit promise the > isolation between compute jobs and gfx jobs? > > On 10/26/2022 11:35 PM, Christian König wrote: >> This reverts commit e6c6338f393b74ac0b303d567bb918b44ae7ad75. >> >> This feature basically re-submits one job after another to >> figure out which one was the one causing a hang. >> >> This is obviously incompatible with gang-submit which requires >> that multiple jobs run at the same time. It's also absolutely >> not helpful to crash the hardware multiple times if a clean >> recovery is desired. >> >> For testing and debugging environments we should rather disable >> recovery alltogether to be able to inspect the state with a hw >> debugger. >> >> Additional to that the sw implementation is clearly buggy and causes >> reference count issues for the hardware fence. >> >> Signed-off-by: Christian König <christian.koenig@amd.com> >> --- >> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 103 --------------------- >> drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c | 2 +- >> drivers/gpu/drm/scheduler/sched_main.c | 58 ++---------- >> include/drm/gpu_scheduler.h | 3 - >> 4 files changed, 10 insertions(+), 156 deletions(-) >> >> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> index 6f958603c8cc..d4584e577b51 100644 >> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >> @@ -5070,94 +5070,6 @@ static int >> amdgpu_device_suspend_display_audio(struct amdgpu_device *adev) >> return 0; >> } >> -static void amdgpu_device_recheck_guilty_jobs( >> - struct amdgpu_device *adev, struct list_head *device_list_handle, >> - struct amdgpu_reset_context *reset_context) >> -{ >> - int i, r = 0; >> - >> - for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { >> - struct amdgpu_ring *ring = adev->rings[i]; >> - int ret = 0; >> - struct drm_sched_job *s_job; >> - >> - if (!ring || !ring->sched.thread) >> - continue; >> - >> - s_job = list_first_entry_or_null(&ring->sched.pending_list, >> - struct drm_sched_job, list); >> - if (s_job == NULL) >> - continue; >> - >> - /* clear job's guilty and depend the folowing step to decide >> the real one */ >> - drm_sched_reset_karma(s_job); >> - drm_sched_resubmit_jobs_ext(&ring->sched, 1); >> - >> - if (!s_job->s_fence->parent) { >> - DRM_WARN("Failed to get a HW fence for job!"); >> - continue; >> - } >> - >> - ret = dma_fence_wait_timeout(s_job->s_fence->parent, false, >> ring->sched.timeout); >> - if (ret == 0) { /* timeout */ >> - DRM_ERROR("Found the real bad job! ring:%s, job_id:%llx\n", >> - ring->sched.name, s_job->id); >> - >> - >> - amdgpu_fence_driver_isr_toggle(adev, true); >> - >> - /* Clear this failed job from fence array */ >> - amdgpu_fence_driver_clear_job_fences(ring); >> - >> - amdgpu_fence_driver_isr_toggle(adev, false); >> - >> - /* Since the job won't signal and we go for >> - * another resubmit drop this parent pointer >> - */ >> - dma_fence_put(s_job->s_fence->parent); >> - s_job->s_fence->parent = NULL; >> - >> - /* set guilty */ >> - drm_sched_increase_karma(s_job); >> - amdgpu_reset_prepare_hwcontext(adev, reset_context); >> -retry: >> - /* do hw reset */ >> - if (amdgpu_sriov_vf(adev)) { >> - amdgpu_virt_fini_data_exchange(adev); >> - r = amdgpu_device_reset_sriov(adev, false); >> - if (r) >> - adev->asic_reset_res = r; >> - } else { >> - clear_bit(AMDGPU_SKIP_HW_RESET, >> - &reset_context->flags); >> - r = amdgpu_do_asic_reset(device_list_handle, >> - reset_context); >> - if (r && r == -EAGAIN) >> - goto retry; >> - } >> - >> - /* >> - * add reset counter so that the following >> - * resubmitted job could flush vmid >> - */ >> - atomic_inc(&adev->gpu_reset_counter); >> - continue; >> - } >> - >> - /* got the hw fence, signal finished fence */ >> - atomic_dec(ring->sched.score); >> - dma_fence_get(&s_job->s_fence->finished); >> - dma_fence_signal(&s_job->s_fence->finished); >> - dma_fence_put(&s_job->s_fence->finished); >> - >> - /* remove node from list and free the job */ >> - spin_lock(&ring->sched.job_list_lock); >> - list_del_init(&s_job->list); >> - spin_unlock(&ring->sched.job_list_lock); >> - ring->sched.ops->free_job(s_job); >> - } >> -} >> - >> static inline void amdgpu_device_stop_pending_resets(struct >> amdgpu_device *adev) >> { >> struct amdgpu_ras *con = amdgpu_ras_get_context(adev); >> @@ -5178,7 +5090,6 @@ static inline void >> amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) >> } >> - >> /** >> * amdgpu_device_gpu_recover - reset the asic and recover scheduler >> * >> @@ -5201,7 +5112,6 @@ int amdgpu_device_gpu_recover(struct >> amdgpu_device *adev, >> int i, r = 0; >> bool need_emergency_restart = false; >> bool audio_suspended = false; >> - int tmp_vram_lost_counter; >> bool gpu_reset_for_dev_remove = false; >> gpu_reset_for_dev_remove = >> @@ -5347,7 +5257,6 @@ int amdgpu_device_gpu_recover(struct >> amdgpu_device *adev, >> amdgpu_device_stop_pending_resets(tmp_adev); >> } >> - tmp_vram_lost_counter = >> atomic_read(&((adev)->vram_lost_counter)); >> /* Actual ASIC resets if needed.*/ >> /* Host driver will handle XGMI hive reset for SRIOV */ >> if (amdgpu_sriov_vf(adev)) { >> @@ -5372,18 +5281,6 @@ int amdgpu_device_gpu_recover(struct >> amdgpu_device *adev, >> /* Post ASIC reset for all devs .*/ >> list_for_each_entry(tmp_adev, device_list_handle, reset_list) { >> - /* >> - * Sometimes a later bad compute job can block a good gfx >> job as gfx >> - * and compute ring share internal GC HW mutually. We add an >> additional >> - * guilty jobs recheck step to find the real guilty job, it >> synchronously >> - * submits and pends for the first job being signaled. If it >> gets timeout, >> - * we identify it as a real guilty job. >> - */ >> - if (amdgpu_gpu_recovery == 2 && >> - !(tmp_vram_lost_counter < >> atomic_read(&adev->vram_lost_counter))) >> - amdgpu_device_recheck_guilty_jobs( >> - tmp_adev, device_list_handle, reset_context); >> - >> for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { >> struct amdgpu_ring *ring = tmp_adev->rings[i]; >> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c >> b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c >> index 16f6a313335e..f177d8e5b665 100644 >> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c >> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c >> @@ -519,7 +519,7 @@ module_param_named(compute_multipipe, >> amdgpu_compute_multipipe, int, 0444); >> * DOC: gpu_recovery (int) >> * Set to enable GPU recovery mechanism (1 = enable, 0 = disable). >> The default is -1 (auto, disabled except SRIOV). >> */ >> -MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (2 = >> advanced tdr mode, 1 = enable, 0 = disable, -1 = auto)"); >> +MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (1 = >> enable, 0 = disable, -1 = auto)"); >> module_param_named(gpu_recovery, amdgpu_gpu_recovery, int, 0444); >> /** >> diff --git a/drivers/gpu/drm/scheduler/sched_main.c >> b/drivers/gpu/drm/scheduler/sched_main.c >> index e0ab14e0fb6b..bb28f31bff6f 100644 >> --- a/drivers/gpu/drm/scheduler/sched_main.c >> +++ b/drivers/gpu/drm/scheduler/sched_main.c >> @@ -355,27 +355,6 @@ static void drm_sched_job_timedout(struct >> work_struct *work) >> } >> } >> - /** >> - * drm_sched_increase_karma - Update sched_entity guilty flag >> - * >> - * @bad: The job guilty of time out >> - * >> - * Increment on every hang caused by the 'bad' job. If this exceeds >> the hang >> - * limit of the scheduler then the respective sched entity is >> marked guilty and >> - * jobs from it will not be scheduled further >> - */ >> -void drm_sched_increase_karma(struct drm_sched_job *bad) >> -{ >> - drm_sched_increase_karma_ext(bad, 1); >> -} >> -EXPORT_SYMBOL(drm_sched_increase_karma); >> - >> -void drm_sched_reset_karma(struct drm_sched_job *bad) >> -{ >> - drm_sched_increase_karma_ext(bad, 0); >> -} >> -EXPORT_SYMBOL(drm_sched_reset_karma); >> - >> /** >> * drm_sched_stop - stop the scheduler >> * >> @@ -516,32 +495,15 @@ EXPORT_SYMBOL(drm_sched_start); >> * >> */ >> void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched) >> -{ >> - drm_sched_resubmit_jobs_ext(sched, INT_MAX); >> -} >> -EXPORT_SYMBOL(drm_sched_resubmit_jobs); >> - >> -/** >> - * drm_sched_resubmit_jobs_ext - helper to relunch certain number of >> jobs from mirror ring list >> - * >> - * @sched: scheduler instance >> - * @max: job numbers to relaunch >> - * >> - */ >> -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, >> int max) >> { >> struct drm_sched_job *s_job, *tmp; >> uint64_t guilty_context; >> bool found_guilty = false; >> struct dma_fence *fence; >> - int i = 0; >> list_for_each_entry_safe(s_job, tmp, &sched->pending_list, >> list) { >> struct drm_sched_fence *s_fence = s_job->s_fence; >> - if (i >= max) >> - break; >> - >> if (!found_guilty && atomic_read(&s_job->karma) > >> sched->hang_limit) { >> found_guilty = true; >> guilty_context = s_job->s_fence->scheduled.context; >> @@ -551,7 +513,6 @@ void drm_sched_resubmit_jobs_ext(struct >> drm_gpu_scheduler *sched, int max) >> dma_fence_set_error(&s_fence->finished, -ECANCELED); >> fence = sched->ops->run_job(s_job); >> - i++; >> if (IS_ERR_OR_NULL(fence)) { >> if (IS_ERR(fence)) >> @@ -567,7 +528,7 @@ void drm_sched_resubmit_jobs_ext(struct >> drm_gpu_scheduler *sched, int max) >> } >> } >> } >> -EXPORT_SYMBOL(drm_sched_resubmit_jobs_ext); >> +EXPORT_SYMBOL(drm_sched_resubmit_jobs); >> /** >> * drm_sched_job_init - init a scheduler job >> @@ -1081,13 +1042,15 @@ void drm_sched_fini(struct drm_gpu_scheduler >> *sched) >> EXPORT_SYMBOL(drm_sched_fini); >> /** >> - * drm_sched_increase_karma_ext - Update sched_entity guilty flag >> + * drm_sched_increase_karma - Update sched_entity guilty flag >> * >> * @bad: The job guilty of time out >> - * @type: type for increase/reset karma >> * >> + * Increment on every hang caused by the 'bad' job. If this exceeds >> the hang >> + * limit of the scheduler then the respective sched entity is marked >> guilty and >> + * jobs from it will not be scheduled further >> */ >> -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) >> +void drm_sched_increase_karma(struct drm_sched_job *bad) >> { >> int i; >> struct drm_sched_entity *tmp; >> @@ -1099,10 +1062,7 @@ void drm_sched_increase_karma_ext(struct >> drm_sched_job *bad, int type) >> * corrupt but keep in mind that kernel jobs always considered >> good. >> */ >> if (bad->s_priority != DRM_SCHED_PRIORITY_KERNEL) { >> - if (type == 0) >> - atomic_set(&bad->karma, 0); >> - else if (type == 1) >> - atomic_inc(&bad->karma); >> + atomic_inc(&bad->karma); >> for (i = DRM_SCHED_PRIORITY_MIN; i < >> DRM_SCHED_PRIORITY_KERNEL; >> i++) { >> @@ -1113,7 +1073,7 @@ void drm_sched_increase_karma_ext(struct >> drm_sched_job *bad, int type) >> if (bad->s_fence->scheduled.context == >> entity->fence_context) { >> if (entity->guilty) >> - atomic_set(entity->guilty, type); >> + atomic_set(entity->guilty, 1); >> break; >> } >> } >> @@ -1123,4 +1083,4 @@ void drm_sched_increase_karma_ext(struct >> drm_sched_job *bad, int type) >> } >> } >> } >> -EXPORT_SYMBOL(drm_sched_increase_karma_ext); >> +EXPORT_SYMBOL(drm_sched_increase_karma); >> diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h >> index 0fca8f38bee4..c564be29c5ae 100644 >> --- a/include/drm/gpu_scheduler.h >> +++ b/include/drm/gpu_scheduler.h >> @@ -488,10 +488,7 @@ void drm_sched_wakeup(struct drm_gpu_scheduler >> *sched); >> void drm_sched_stop(struct drm_gpu_scheduler *sched, struct >> drm_sched_job *bad); >> void drm_sched_start(struct drm_gpu_scheduler *sched, bool >> full_recovery); >> void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched); >> -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, >> int max); >> void drm_sched_increase_karma(struct drm_sched_job *bad); >> -void drm_sched_reset_karma(struct drm_sched_job *bad); >> -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type); >> bool drm_sched_dependency_optimized(struct dma_fence* fence, >> struct drm_sched_entity *entity); >> void drm_sched_fault(struct drm_gpu_scheduler *sched);
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c index 6f958603c8cc..d4584e577b51 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c @@ -5070,94 +5070,6 @@ static int amdgpu_device_suspend_display_audio(struct amdgpu_device *adev) return 0; } -static void amdgpu_device_recheck_guilty_jobs( - struct amdgpu_device *adev, struct list_head *device_list_handle, - struct amdgpu_reset_context *reset_context) -{ - int i, r = 0; - - for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { - struct amdgpu_ring *ring = adev->rings[i]; - int ret = 0; - struct drm_sched_job *s_job; - - if (!ring || !ring->sched.thread) - continue; - - s_job = list_first_entry_or_null(&ring->sched.pending_list, - struct drm_sched_job, list); - if (s_job == NULL) - continue; - - /* clear job's guilty and depend the folowing step to decide the real one */ - drm_sched_reset_karma(s_job); - drm_sched_resubmit_jobs_ext(&ring->sched, 1); - - if (!s_job->s_fence->parent) { - DRM_WARN("Failed to get a HW fence for job!"); - continue; - } - - ret = dma_fence_wait_timeout(s_job->s_fence->parent, false, ring->sched.timeout); - if (ret == 0) { /* timeout */ - DRM_ERROR("Found the real bad job! ring:%s, job_id:%llx\n", - ring->sched.name, s_job->id); - - - amdgpu_fence_driver_isr_toggle(adev, true); - - /* Clear this failed job from fence array */ - amdgpu_fence_driver_clear_job_fences(ring); - - amdgpu_fence_driver_isr_toggle(adev, false); - - /* Since the job won't signal and we go for - * another resubmit drop this parent pointer - */ - dma_fence_put(s_job->s_fence->parent); - s_job->s_fence->parent = NULL; - - /* set guilty */ - drm_sched_increase_karma(s_job); - amdgpu_reset_prepare_hwcontext(adev, reset_context); -retry: - /* do hw reset */ - if (amdgpu_sriov_vf(adev)) { - amdgpu_virt_fini_data_exchange(adev); - r = amdgpu_device_reset_sriov(adev, false); - if (r) - adev->asic_reset_res = r; - } else { - clear_bit(AMDGPU_SKIP_HW_RESET, - &reset_context->flags); - r = amdgpu_do_asic_reset(device_list_handle, - reset_context); - if (r && r == -EAGAIN) - goto retry; - } - - /* - * add reset counter so that the following - * resubmitted job could flush vmid - */ - atomic_inc(&adev->gpu_reset_counter); - continue; - } - - /* got the hw fence, signal finished fence */ - atomic_dec(ring->sched.score); - dma_fence_get(&s_job->s_fence->finished); - dma_fence_signal(&s_job->s_fence->finished); - dma_fence_put(&s_job->s_fence->finished); - - /* remove node from list and free the job */ - spin_lock(&ring->sched.job_list_lock); - list_del_init(&s_job->list); - spin_unlock(&ring->sched.job_list_lock); - ring->sched.ops->free_job(s_job); - } -} - static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) { struct amdgpu_ras *con = amdgpu_ras_get_context(adev); @@ -5178,7 +5090,6 @@ static inline void amdgpu_device_stop_pending_resets(struct amdgpu_device *adev) } - /** * amdgpu_device_gpu_recover - reset the asic and recover scheduler * @@ -5201,7 +5112,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, int i, r = 0; bool need_emergency_restart = false; bool audio_suspended = false; - int tmp_vram_lost_counter; bool gpu_reset_for_dev_remove = false; gpu_reset_for_dev_remove = @@ -5347,7 +5257,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, amdgpu_device_stop_pending_resets(tmp_adev); } - tmp_vram_lost_counter = atomic_read(&((adev)->vram_lost_counter)); /* Actual ASIC resets if needed.*/ /* Host driver will handle XGMI hive reset for SRIOV */ if (amdgpu_sriov_vf(adev)) { @@ -5372,18 +5281,6 @@ int amdgpu_device_gpu_recover(struct amdgpu_device *adev, /* Post ASIC reset for all devs .*/ list_for_each_entry(tmp_adev, device_list_handle, reset_list) { - /* - * Sometimes a later bad compute job can block a good gfx job as gfx - * and compute ring share internal GC HW mutually. We add an additional - * guilty jobs recheck step to find the real guilty job, it synchronously - * submits and pends for the first job being signaled. If it gets timeout, - * we identify it as a real guilty job. - */ - if (amdgpu_gpu_recovery == 2 && - !(tmp_vram_lost_counter < atomic_read(&adev->vram_lost_counter))) - amdgpu_device_recheck_guilty_jobs( - tmp_adev, device_list_handle, reset_context); - for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { struct amdgpu_ring *ring = tmp_adev->rings[i]; diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c index 16f6a313335e..f177d8e5b665 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c @@ -519,7 +519,7 @@ module_param_named(compute_multipipe, amdgpu_compute_multipipe, int, 0444); * DOC: gpu_recovery (int) * Set to enable GPU recovery mechanism (1 = enable, 0 = disable). The default is -1 (auto, disabled except SRIOV). */ -MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (2 = advanced tdr mode, 1 = enable, 0 = disable, -1 = auto)"); +MODULE_PARM_DESC(gpu_recovery, "Enable GPU recovery mechanism, (1 = enable, 0 = disable, -1 = auto)"); module_param_named(gpu_recovery, amdgpu_gpu_recovery, int, 0444); /** diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c index e0ab14e0fb6b..bb28f31bff6f 100644 --- a/drivers/gpu/drm/scheduler/sched_main.c +++ b/drivers/gpu/drm/scheduler/sched_main.c @@ -355,27 +355,6 @@ static void drm_sched_job_timedout(struct work_struct *work) } } - /** - * drm_sched_increase_karma - Update sched_entity guilty flag - * - * @bad: The job guilty of time out - * - * Increment on every hang caused by the 'bad' job. If this exceeds the hang - * limit of the scheduler then the respective sched entity is marked guilty and - * jobs from it will not be scheduled further - */ -void drm_sched_increase_karma(struct drm_sched_job *bad) -{ - drm_sched_increase_karma_ext(bad, 1); -} -EXPORT_SYMBOL(drm_sched_increase_karma); - -void drm_sched_reset_karma(struct drm_sched_job *bad) -{ - drm_sched_increase_karma_ext(bad, 0); -} -EXPORT_SYMBOL(drm_sched_reset_karma); - /** * drm_sched_stop - stop the scheduler * @@ -516,32 +495,15 @@ EXPORT_SYMBOL(drm_sched_start); * */ void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched) -{ - drm_sched_resubmit_jobs_ext(sched, INT_MAX); -} -EXPORT_SYMBOL(drm_sched_resubmit_jobs); - -/** - * drm_sched_resubmit_jobs_ext - helper to relunch certain number of jobs from mirror ring list - * - * @sched: scheduler instance - * @max: job numbers to relaunch - * - */ -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) { struct drm_sched_job *s_job, *tmp; uint64_t guilty_context; bool found_guilty = false; struct dma_fence *fence; - int i = 0; list_for_each_entry_safe(s_job, tmp, &sched->pending_list, list) { struct drm_sched_fence *s_fence = s_job->s_fence; - if (i >= max) - break; - if (!found_guilty && atomic_read(&s_job->karma) > sched->hang_limit) { found_guilty = true; guilty_context = s_job->s_fence->scheduled.context; @@ -551,7 +513,6 @@ void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) dma_fence_set_error(&s_fence->finished, -ECANCELED); fence = sched->ops->run_job(s_job); - i++; if (IS_ERR_OR_NULL(fence)) { if (IS_ERR(fence)) @@ -567,7 +528,7 @@ void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max) } } } -EXPORT_SYMBOL(drm_sched_resubmit_jobs_ext); +EXPORT_SYMBOL(drm_sched_resubmit_jobs); /** * drm_sched_job_init - init a scheduler job @@ -1081,13 +1042,15 @@ void drm_sched_fini(struct drm_gpu_scheduler *sched) EXPORT_SYMBOL(drm_sched_fini); /** - * drm_sched_increase_karma_ext - Update sched_entity guilty flag + * drm_sched_increase_karma - Update sched_entity guilty flag * * @bad: The job guilty of time out - * @type: type for increase/reset karma * + * Increment on every hang caused by the 'bad' job. If this exceeds the hang + * limit of the scheduler then the respective sched entity is marked guilty and + * jobs from it will not be scheduled further */ -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) +void drm_sched_increase_karma(struct drm_sched_job *bad) { int i; struct drm_sched_entity *tmp; @@ -1099,10 +1062,7 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) * corrupt but keep in mind that kernel jobs always considered good. */ if (bad->s_priority != DRM_SCHED_PRIORITY_KERNEL) { - if (type == 0) - atomic_set(&bad->karma, 0); - else if (type == 1) - atomic_inc(&bad->karma); + atomic_inc(&bad->karma); for (i = DRM_SCHED_PRIORITY_MIN; i < DRM_SCHED_PRIORITY_KERNEL; i++) { @@ -1113,7 +1073,7 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) if (bad->s_fence->scheduled.context == entity->fence_context) { if (entity->guilty) - atomic_set(entity->guilty, type); + atomic_set(entity->guilty, 1); break; } } @@ -1123,4 +1083,4 @@ void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type) } } } -EXPORT_SYMBOL(drm_sched_increase_karma_ext); +EXPORT_SYMBOL(drm_sched_increase_karma); diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h index 0fca8f38bee4..c564be29c5ae 100644 --- a/include/drm/gpu_scheduler.h +++ b/include/drm/gpu_scheduler.h @@ -488,10 +488,7 @@ void drm_sched_wakeup(struct drm_gpu_scheduler *sched); void drm_sched_stop(struct drm_gpu_scheduler *sched, struct drm_sched_job *bad); void drm_sched_start(struct drm_gpu_scheduler *sched, bool full_recovery); void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched); -void drm_sched_resubmit_jobs_ext(struct drm_gpu_scheduler *sched, int max); void drm_sched_increase_karma(struct drm_sched_job *bad); -void drm_sched_reset_karma(struct drm_sched_job *bad); -void drm_sched_increase_karma_ext(struct drm_sched_job *bad, int type); bool drm_sched_dependency_optimized(struct dma_fence* fence, struct drm_sched_entity *entity); void drm_sched_fault(struct drm_gpu_scheduler *sched);
This reverts commit e6c6338f393b74ac0b303d567bb918b44ae7ad75. This feature basically re-submits one job after another to figure out which one was the one causing a hang. This is obviously incompatible with gang-submit which requires that multiple jobs run at the same time. It's also absolutely not helpful to crash the hardware multiple times if a clean recovery is desired. For testing and debugging environments we should rather disable recovery alltogether to be able to inspect the state with a hw debugger. Additional to that the sw implementation is clearly buggy and causes reference count issues for the hardware fence. Signed-off-by: Christian König <christian.koenig@amd.com> --- drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 103 --------------------- drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c | 2 +- drivers/gpu/drm/scheduler/sched_main.c | 58 ++---------- include/drm/gpu_scheduler.h | 3 - 4 files changed, 10 insertions(+), 156 deletions(-)