Feel free to merge 1+2 since they don't really depend on any other work 
in the series and they were previously reviewed.

Nicholas Kazlauskas

On 4/23/19 8:32 AM, Koenig, Christian wrote:
> Well you at least have to give me time till after the holidays to get
> going again :)
> 
> Not sure exactly jet why we need patch number 5.
> 
> And we should probably commit patch #1 and #2.
> 
> Christian.
> 
> Am 22.04.19 um 13:54 schrieb Grodzovsky, Andrey:
>> Ping for patches 3, new patch 5 and patch 6.
>>
>> Andrey
>>
>> On 4/18/19 11:00 AM, Andrey Grodzovsky wrote:
>>> Also reject TDRs if another one already running.
>>>
>>> v2:
>>> Stop all schedulers across device and entire XGMI hive before
>>> force signaling HW fences.
>>> Avoid passing job_signaled to helper fnctions to keep all the decision
>>> making about skipping HW reset in one place.
>>>
>>> v3:
>>> Fix SW sched. hang after non HW reset. sched.hw_rq_count has to be balanced
>>> against it's decrement in drm_sched_stop in non HW reset case.
>>> v4: rebase
>>> v5: Revert v3 as we do it now in sceduler code.
>>>
>>> Signed-off-by: Andrey Grodzovsky <andrey.grodzov...@amd.com>
>>> ---
>>>     drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 143 
>>> +++++++++++++++++++----------
>>>     1 file changed, 95 insertions(+), 48 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 
>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
>>> index a0e165c..85f8792 100644
>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
>>> @@ -3334,8 +3334,6 @@ static int amdgpu_device_pre_asic_reset(struct 
>>> amdgpu_device *adev,
>>>                     if (!ring || !ring->sched.thread)
>>>                             continue;
>>>     
>>> -           drm_sched_stop(&ring->sched, &job->base);
>>> -
>>>                     /* after all hw jobs are reset, hw fence is 
>>> meaningless, so force_completion */
>>>                     amdgpu_fence_driver_force_completion(ring);
>>>             }
>>> @@ -3343,6 +3341,7 @@ static int amdgpu_device_pre_asic_reset(struct 
>>> amdgpu_device *adev,
>>>             if(job)
>>>                     drm_sched_increase_karma(&job->base);
>>>     
>>> +   /* Don't suspend on bare metal if we are not going to HW reset the ASIC 
>>> */
>>>             if (!amdgpu_sriov_vf(adev)) {
>>>     
>>>                     if (!need_full_reset)
>>> @@ -3480,37 +3479,21 @@ static int amdgpu_do_asic_reset(struct 
>>> amdgpu_hive_info *hive,
>>>             return r;
>>>     }
>>>     
>>> -static void amdgpu_device_post_asic_reset(struct amdgpu_device *adev)
>>> +static bool amdgpu_device_lock_adev(struct amdgpu_device *adev, bool 
>>> trylock)
>>>     {
>>> -   int i;
>>> -
>>> -   for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>>> -           struct amdgpu_ring *ring = adev->rings[i];
>>> -
>>> -           if (!ring || !ring->sched.thread)
>>> -                   continue;
>>> -
>>> -           if (!adev->asic_reset_res)
>>> -                   drm_sched_resubmit_jobs(&ring->sched);
>>> +   if (trylock) {
>>> +           if (!mutex_trylock(&adev->lock_reset))
>>> +                   return false;
>>> +   } else
>>> +           mutex_lock(&adev->lock_reset);
>>>     
>>> -           drm_sched_start(&ring->sched, !adev->asic_reset_res);
>>> -   }
>>> -
>>> -   if (!amdgpu_device_has_dc_support(adev)) {
>>> -           drm_helper_resume_force_mode(adev->ddev);
>>> -   }
>>> -
>>> -   adev->asic_reset_res = 0;
>>> -}
>>> -
>>> -static void amdgpu_device_lock_adev(struct amdgpu_device *adev)
>>> -{
>>> -   mutex_lock(&adev->lock_reset);
>>>             atomic_inc(&adev->gpu_reset_counter);
>>>             adev->in_gpu_reset = 1;
>>>             /* Block kfd: SRIOV would do it separately */
>>>             if (!amdgpu_sriov_vf(adev))
>>>                     amdgpu_amdkfd_pre_reset(adev);
>>> +
>>> +   return true;
>>>     }
>>>     
>>>     static void amdgpu_device_unlock_adev(struct amdgpu_device *adev)
>>> @@ -3538,40 +3521,42 @@ static void amdgpu_device_unlock_adev(struct 
>>> amdgpu_device *adev)
>>>     int amdgpu_device_gpu_recover(struct amdgpu_device *adev,
>>>                                   struct amdgpu_job *job)
>>>     {
>>> -   int r;
>>> +   struct list_head device_list, *device_list_handle =  NULL;
>>> +   bool need_full_reset, job_signaled;
>>>             struct amdgpu_hive_info *hive = NULL;
>>> -   bool need_full_reset = false;
>>>             struct amdgpu_device *tmp_adev = NULL;
>>> -   struct list_head device_list, *device_list_handle =  NULL;
>>> +   int i, r = 0;
>>>     
>>> +   need_full_reset = job_signaled = false;
>>>             INIT_LIST_HEAD(&device_list);
>>>     
>>>             dev_info(adev->dev, "GPU reset begin!\n");
>>>     
>>> +   hive = amdgpu_get_xgmi_hive(adev, false);
>>> +
>>>             /*
>>> -    * In case of XGMI hive disallow concurrent resets to be triggered
>>> -    * by different nodes. No point also since the one node already 
>>> executing
>>> -    * reset will also reset all the other nodes in the hive.
>>> +    * Here we trylock to avoid chain of resets executing from
>>> +    * either trigger by jobs on different adevs in XGMI hive or jobs on
>>> +    * different schedulers for same device while this TO handler is 
>>> running.
>>> +    * We always reset all schedulers for device and all devices for XGMI
>>> +    * hive so that should take care of them too.
>>>              */
>>> -   hive = amdgpu_get_xgmi_hive(adev, 0);
>>> -   if (hive && adev->gmc.xgmi.num_physical_nodes > 1 &&
>>> -       !mutex_trylock(&hive->reset_lock))
>>> +
>>> +   if (hive && !mutex_trylock(&hive->reset_lock)) {
>>> +           DRM_INFO("Bailing on TDR for s_job:%llx, hive: %llx as another 
>>> already in progress",
>>> +                    job->base.id, hive->hive_id);
>>>                     return 0;
>>> +   }
>>>     
>>>             /* Start with adev pre asic reset first for soft reset check.*/
>>> -   amdgpu_device_lock_adev(adev);
>>> -   r = amdgpu_device_pre_asic_reset(adev,
>>> -                                    job,
>>> -                                    &need_full_reset);
>>> -   if (r) {
>>> -           /*TODO Should we stop ?*/
>>> -           DRM_ERROR("GPU pre asic reset failed with err, %d for drm dev, 
>>> %s ",
>>> -                     r, adev->ddev->unique);
>>> -           adev->asic_reset_res = r;
>>> +   if (!amdgpu_device_lock_adev(adev, !hive)) {
>>> +           DRM_INFO("Bailing on TDR for s_job:%llx, as another already in 
>>> progress",
>>> +                                    job->base.id);
>>> +           return 0;
>>>             }
>>>     
>>>             /* Build list of devices to reset */
>>> -   if  (need_full_reset && adev->gmc.xgmi.num_physical_nodes > 1) {
>>> +   if  (adev->gmc.xgmi.num_physical_nodes > 1) {
>>>                     if (!hive) {
>>>                             amdgpu_device_unlock_adev(adev);
>>>                             return -ENODEV;
>>> @@ -3588,13 +3573,56 @@ int amdgpu_device_gpu_recover(struct amdgpu_device 
>>> *adev,
>>>                     device_list_handle = &device_list;
>>>             }
>>>     
>>> +   /* block all schedulers and reset given job's ring */
>>> +   list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) {
>>> +           for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>>> +                   struct amdgpu_ring *ring = tmp_adev->rings[i];
>>> +
>>> +                   if (!ring || !ring->sched.thread)
>>> +                           continue;
>>> +
>>> +                   drm_sched_stop(&ring->sched, &job->base);
>>> +           }
>>> +   }
>>> +
>>> +
>>> +   /*
>>> +    * Must check guilty signal here since after this point all old
>>> +    * HW fences are force signaled.
>>> +    *
>>> +    * job->base holds a reference to parent fence
>>> +    */
>>> +   if (job && job->base.s_fence->parent &&
>>> +       dma_fence_is_signaled(job->base.s_fence->parent))
>>> +           job_signaled = true;
>>> +
>>> +   if (!amdgpu_device_ip_need_full_reset(adev))
>>> +           device_list_handle = &device_list;
>>> +
>>> +   if (job_signaled) {
>>> +           dev_info(adev->dev, "Guilty job already signaled, skipping HW 
>>> reset");
>>> +           goto skip_hw_reset;
>>> +   }
>>> +
>>> +
>>> +   /* Guilty job will be freed after this*/
>>> +   r = amdgpu_device_pre_asic_reset(adev,
>>> +                                    job,
>>> +                                    &need_full_reset);
>>> +   if (r) {
>>> +           /*TODO Should we stop ?*/
>>> +           DRM_ERROR("GPU pre asic reset failed with err, %d for drm dev, 
>>> %s ",
>>> +                     r, adev->ddev->unique);
>>> +           adev->asic_reset_res = r;
>>> +   }
>>> +
>>>     retry:  /* Rest of adevs pre asic reset from XGMI hive. */
>>>             list_for_each_entry(tmp_adev, device_list_handle, 
>>> gmc.xgmi.head) {
>>>     
>>>                     if (tmp_adev == adev)
>>>                             continue;
>>>     
>>> -           amdgpu_device_lock_adev(tmp_adev);
>>> +           amdgpu_device_lock_adev(tmp_adev, false);
>>>                     r = amdgpu_device_pre_asic_reset(tmp_adev,
>>>                                                      NULL,
>>>                                                      &need_full_reset);
>>> @@ -3618,9 +3646,28 @@ int amdgpu_device_gpu_recover(struct amdgpu_device 
>>> *adev,
>>>                             goto retry;
>>>             }
>>>     
>>> +skip_hw_reset:
>>> +
>>>             /* Post ASIC reset for all devs .*/
>>>             list_for_each_entry(tmp_adev, device_list_handle, 
>>> gmc.xgmi.head) {
>>> -           amdgpu_device_post_asic_reset(tmp_adev);
>>> +           for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
>>> +                   struct amdgpu_ring *ring = tmp_adev->rings[i];
>>> +
>>> +                   if (!ring || !ring->sched.thread)
>>> +                           continue;
>>> +
>>> +                   /* No point to resubmit jobs if we didn't HW reset*/
>>> +                   if (!tmp_adev->asic_reset_res && !job_signaled)
>>> +                           drm_sched_resubmit_jobs(&ring->sched);
>>> +
>>> +                   drm_sched_start(&ring->sched, 
>>> !tmp_adev->asic_reset_res);
>>> +           }
>>> +
>>> +           if (!amdgpu_device_has_dc_support(tmp_adev) && !job_signaled) {
>>> +                   drm_helper_resume_force_mode(tmp_adev->ddev);
>>> +           }
>>> +
>>> +           tmp_adev->asic_reset_res = 0;
>>>     
>>>                     if (r) {
>>>                             /* bad news, how to tell it to userspace ? */
>>> @@ -3633,7 +3680,7 @@ int amdgpu_device_gpu_recover(struct amdgpu_device 
>>> *adev,
>>>                     amdgpu_device_unlock_adev(tmp_adev);
>>>             }
>>>     
>>> -   if (hive && adev->gmc.xgmi.num_physical_nodes > 1)
>>> +   if (hive)
>>>                     mutex_unlock(&hive->reset_lock);
>>>     
>>>             if (r)
> 
> _______________________________________________
> dri-devel mailing list
> dri-de...@lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/dri-devel
> 

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

Reply via email to