From: Jiadong Zhu <jiadong....@amd.com>

Kiq command unmap_queues only does the dequeueing action.
We have to map the queue back with clean mqd.

Signed-off-by: Jiadong Zhu <jiadong....@amd.com>
Reviewed-by: Alex Deucher <alexander.deuc...@amd.com>
Signed-off-by: Alex Deucher <alexander.deuc...@amd.com>
---
 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c | 36 ++++++++++++++++++++-------
 1 file changed, 27 insertions(+), 9 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 
b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
index 78495df29b5c..3a819c6923c6 100644
--- a/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
@@ -3742,7 +3742,7 @@ static int gfx_v9_0_kiq_init_queue(struct amdgpu_ring 
*ring)
        return 0;
 }
 
-static int gfx_v9_0_kcq_init_queue(struct amdgpu_ring *ring)
+static int gfx_v9_0_kcq_init_queue(struct amdgpu_ring *ring, bool restore)
 {
        struct amdgpu_device *adev = ring->adev;
        struct v9_mqd *mqd = ring->mqd_ptr;
@@ -3754,8 +3754,8 @@ static int gfx_v9_0_kcq_init_queue(struct amdgpu_ring 
*ring)
         */
        tmp_mqd = (struct v9_mqd *)adev->gfx.mec.mqd_backup[mqd_idx];
 
-       if (!tmp_mqd->cp_hqd_pq_control ||
-           (!amdgpu_in_reset(adev) && !adev->in_suspend)) {
+       if (!restore && (!tmp_mqd->cp_hqd_pq_control ||
+           (!amdgpu_in_reset(adev) && !adev->in_suspend))) {
                memset((void *)mqd, 0, sizeof(struct v9_mqd_allocation));
                ((struct v9_mqd_allocation *)mqd)->dynamic_cu_mask = 0xFFFFFFFF;
                ((struct v9_mqd_allocation *)mqd)->dynamic_rb_mask = 0xFFFFFFFF;
@@ -3819,7 +3819,7 @@ static int gfx_v9_0_kcq_resume(struct amdgpu_device *adev)
                        goto done;
                r = amdgpu_bo_kmap(ring->mqd_obj, (void **)&ring->mqd_ptr);
                if (!r) {
-                       r = gfx_v9_0_kcq_init_queue(ring);
+                       r = gfx_v9_0_kcq_init_queue(ring, false);
                        amdgpu_bo_kunmap(ring->mqd_obj);
                        ring->mqd_ptr = NULL;
                }
@@ -7129,11 +7129,29 @@ static int gfx_v9_0_reset_kcq(struct amdgpu_ring *ring,
        if (r)
                return r;
 
-       /* reset the ring */
-       ring->wptr = 0;
-       *ring->wptr_cpu_addr = 0;
-       amdgpu_ring_clear_ring(ring);
-
+       r = amdgpu_bo_reserve(ring->mqd_obj, false);
+       if (unlikely(r != 0)){
+               DRM_ERROR("fail to resv mqd_obj\n");
+               return r;
+       }
+       r = amdgpu_bo_kmap(ring->mqd_obj, (void **)&ring->mqd_ptr);
+       if (!r) {
+               r = gfx_v9_0_kcq_init_queue(ring, true);
+               amdgpu_bo_kunmap(ring->mqd_obj);
+               ring->mqd_ptr = NULL;
+       }
+       amdgpu_bo_unreserve(ring->mqd_obj);
+       if (r){
+               DRM_ERROR("fail to unresv mqd_obj\n");
+               return r;
+       }
+       r = amdgpu_ring_alloc(kiq_ring, kiq->pmf->map_queues_size);
+       kiq->pmf->kiq_map_queues(kiq_ring, ring);
+       r = amdgpu_ring_test_ring(kiq_ring);
+       if (r){
+               DRM_ERROR("fail to remap queue\n");
+               return r;
+       }
        return amdgpu_ring_test_ring(ring);
 }
 
-- 
2.45.2

Reply via email to