This patch adds suspend support for gfx userqueues. It typically does
the following:
- adds an enable_signaling function for the eviction fence, so that it
  can trigger the userqueue suspend,
- adds a delayed function for suspending the userqueues, to suspend all
  the queues under this userq manager and signals the eviction fence,
- adds reference of userq manager in the eviction fence container so
  that it can be used in the suspend function.

V2: Addressed Christian's review comments:
    - schedule suspend work immediately

V4: Addressed Christian's review comments:
    - wait for pending uq fences before starting suspend, added
      queue->last_fence for the same
    - accommodate ev_fence_mgr into existing code
    - some bug fixes and NULL checks

V5: Addressed Christian's review comments (gitlab)
    - Wait for eviction fence to get signaled in destroy, dont signal it
    - Wait for eviction fence to get signaled in replace fence, dont signal it

Cc: Alex Deucher <alexander.deuc...@amd.com>
Cc: Christian Koenig <christian.koe...@amd.com>
Signed-off-by: Shashank Sharma <shashank.sha...@amd.com>
Signed-off-by: Arvind Yadav <arvind.ya...@amd.com>
Change-Id: Ib60a7feda5544e3badc87bd1a991931ee726ee82
---
 .../drm/amd/amdgpu/amdgpu_eviction_fence.c    | 149 ++++++++++++++++++
 .../drm/amd/amdgpu/amdgpu_eviction_fence.h    |   2 +
 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c       |   2 +
 .../gpu/drm/amd/amdgpu/amdgpu_userq_fence.c   |  10 +-
 drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c | 100 ++++++++++++
 .../gpu/drm/amd/include/amdgpu_userqueue.h    |  10 ++
 6 files changed, 272 insertions(+), 1 deletion(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c
index 2d474cb11cf9..3d4fc704adb1 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c
@@ -22,8 +22,12 @@
  *
  */
 #include <linux/sched.h>
+#include <drm/drm_exec.h>
 #include "amdgpu.h"
 
+#define work_to_evf_mgr(w, name) container_of(w, struct 
amdgpu_eviction_fence_mgr, name)
+#define evf_mgr_to_fpriv(e) container_of(e, struct amdgpu_fpriv, evf_mgr)
+
 static const char *
 amdgpu_eviction_fence_get_driver_name(struct dma_fence *fence)
 {
@@ -39,10 +43,150 @@ amdgpu_eviction_fence_get_timeline_name(struct dma_fence 
*f)
        return ef->timeline_name;
 }
 
+static void
+amdgpu_eviction_fence_update_fence(struct amdgpu_eviction_fence_mgr *evf_mgr,
+                                  struct amdgpu_eviction_fence *new_ef)
+{
+       struct dma_fence *old_ef = &evf_mgr->ev_fence->base;
+
+       spin_lock(&evf_mgr->ev_fence_lock);
+       dma_fence_put(old_ef);
+       evf_mgr->ev_fence = new_ef;
+       spin_unlock(&evf_mgr->ev_fence_lock);
+}
+
+int
+amdgpu_eviction_fence_replace_fence(struct amdgpu_fpriv *fpriv)
+{
+       struct amdgpu_eviction_fence_mgr *evf_mgr = &fpriv->evf_mgr;
+       struct amdgpu_vm *vm = &fpriv->vm;
+       struct amdgpu_eviction_fence *old_ef, *new_ef;
+       struct amdgpu_bo_va *bo_va, *tmp;
+       int ret;
+
+       old_ef = evf_mgr->ev_fence;
+       if (old_ef && !dma_fence_is_signaled(&old_ef->base)) {
+               DRM_DEBUG_DRIVER("Old EF not signaled yet\n");
+               dma_fence_wait(&old_ef->base, true);
+       }
+
+       new_ef = amdgpu_eviction_fence_create(evf_mgr);
+       if (!new_ef) {
+               DRM_ERROR("Failed to create new eviction fence\n");
+               return ret;
+       }
+
+       /* Replace fences and free old one */
+       amdgpu_eviction_fence_update_fence(evf_mgr, new_ef);
+
+       /* Attach new eviction fence to BOs */
+       list_for_each_entry_safe(bo_va, tmp, &vm->done, base.vm_status) {
+               struct amdgpu_bo *bo = bo_va->base.bo;
+
+               if (!bo)
+                       continue;
+
+               /* Skip pinned BOs */
+               if (bo->tbo.pin_count)
+                       continue;
+
+               ret = amdgpu_eviction_fence_attach(evf_mgr, bo);
+               if (ret) {
+                       DRM_ERROR("Failed to attch new eviction fence\n");
+                       goto free_err;
+               }
+       }
+
+       return 0;
+
+free_err:
+       kfree(new_ef);
+       return ret;
+}
+
+static void
+amdgpu_eviction_fence_suspend_worker(struct work_struct *work)
+{
+       struct amdgpu_eviction_fence_mgr *evf_mgr = work_to_evf_mgr(work, 
suspend_work.work);
+       struct amdgpu_fpriv *fpriv = evf_mgr_to_fpriv(evf_mgr);
+       struct amdgpu_vm *vm = &fpriv->vm;
+       struct amdgpu_bo_va *bo_va, *tmp;
+       struct drm_exec exec;
+       struct amdgpu_bo *bo;
+       int ret;
+
+       /* Signal old eviction fence */
+       ret = amdgpu_eviction_fence_signal(evf_mgr);
+       if (ret) {
+               DRM_ERROR("Failed to signal eviction fence err=%d\n", ret);
+               return;
+       }
+
+       /* Cleanup old eviction fence entry */
+       amdgpu_eviction_fence_destroy(evf_mgr);
+
+       /* Do not replace eviction fence is fd is getting closed */
+       if (evf_mgr->eviction_allowed)
+               return;
+
+       drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES, 0);
+       drm_exec_until_all_locked(&exec) {
+               ret = amdgpu_vm_lock_pd(vm, &exec, 2);
+               drm_exec_retry_on_contention(&exec);
+               if (unlikely(ret)) {
+                       DRM_ERROR("Failed to lock PD\n");
+                       goto unlock_drm;
+               }
+
+               /* Lock the done list */
+               list_for_each_entry_safe(bo_va, tmp, &vm->done, base.vm_status) 
{
+                       bo = bo_va->base.bo;
+                       if (!bo) continue;
+
+                       ret = drm_exec_lock_obj(&exec, &bo->tbo.base);
+                       drm_exec_retry_on_contention(&exec);
+                       if (unlikely(ret))
+                               goto unlock_drm;
+               }
+       }
+       /* Replace old eviction fence with new one */
+       ret = amdgpu_eviction_fence_replace_fence(fpriv);
+       if (ret)
+               DRM_ERROR("Failed to replace eviction fence\n");
+unlock_drm:
+       drm_exec_fini(&exec);
+}
+
+static bool amdgpu_eviction_fence_enable_signaling(struct dma_fence *f)
+{
+       struct amdgpu_eviction_fence_mgr *evf_mgr;
+       struct amdgpu_eviction_fence *ev_fence;
+       struct amdgpu_userq_mgr *uq_mgr;
+       struct amdgpu_fpriv *fpriv;
+
+       if (!f)
+               return true;
+
+       ev_fence = to_ev_fence(f);
+       uq_mgr = ev_fence->uq_mgr;
+       fpriv = uq_mgr_to_fpriv(uq_mgr);
+       evf_mgr = &fpriv->evf_mgr;
+
+       if (uq_mgr->num_userqs)
+               /* If userqueues are active, suspend userqueues */
+               schedule_delayed_work(&uq_mgr->suspend_work, 0);
+       else
+               /* Else just signal and replace eviction fence */
+               schedule_delayed_work(&evf_mgr->suspend_work, 0);
+
+       return true;
+}
+
 static const struct dma_fence_ops amdgpu_eviction_fence_ops = {
        .use_64bit_seqno = true,
        .get_driver_name = amdgpu_eviction_fence_get_driver_name,
        .get_timeline_name = amdgpu_eviction_fence_get_timeline_name,
+       .enable_signaling = amdgpu_eviction_fence_enable_signaling,
 };
 
 int amdgpu_eviction_fence_signal(struct amdgpu_eviction_fence_mgr *evf_mgr)
@@ -59,11 +203,14 @@ struct amdgpu_eviction_fence *
 amdgpu_eviction_fence_create(struct amdgpu_eviction_fence_mgr *evf_mgr)
 {
        struct amdgpu_eviction_fence *ev_fence;
+       struct amdgpu_fpriv *fpriv = evf_mgr_to_fpriv(evf_mgr);
+       struct amdgpu_userq_mgr *uq_mgr = &fpriv->userq_mgr;
 
        ev_fence = kzalloc(sizeof(*ev_fence), GFP_KERNEL);
        if (!ev_fence)
                return NULL;
 
+       ev_fence->uq_mgr = uq_mgr;
        get_task_comm(ev_fence->timeline_name, current);
        spin_lock_init(&ev_fence->lock);
        dma_fence_init(&ev_fence->base, &amdgpu_eviction_fence_ops,
@@ -143,6 +290,8 @@ void amdgpu_eviction_fence_init(struct 
amdgpu_eviction_fence_mgr *evf_mgr)
                goto unlock;
        }
 
+       INIT_DELAYED_WORK(&evf_mgr->suspend_work, 
amdgpu_eviction_fence_suspend_worker);
+
 unlock:
        spin_unlock(&evf_mgr->ev_fence_lock);
 }
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.h 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.h
index b47ab1307ec5..712fabf09fc1 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.h
@@ -37,6 +37,8 @@ struct amdgpu_eviction_fence_mgr {
        atomic_t                ev_fence_seq;
        spinlock_t              ev_fence_lock;
        struct amdgpu_eviction_fence *ev_fence;
+       struct delayed_work     suspend_work;
+       bool eviction_allowed;
 };
 
 /* Eviction fence helper functions */
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
index e7fb13e20197..88f3a885b1dc 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
@@ -1434,6 +1434,7 @@ void amdgpu_driver_postclose_kms(struct drm_device *dev,
 {
        struct amdgpu_device *adev = drm_to_adev(dev);
        struct amdgpu_fpriv *fpriv = file_priv->driver_priv;
+       struct amdgpu_eviction_fence_mgr *evf_mgr = &fpriv->evf_mgr;
        struct amdgpu_bo_list *list;
        struct amdgpu_bo *pd;
        u32 pasid;
@@ -1466,6 +1467,7 @@ void amdgpu_driver_postclose_kms(struct drm_device *dev,
                amdgpu_bo_unreserve(pd);
        }
 
+       evf_mgr->eviction_allowed = true;
        amdgpu_eviction_fence_destroy(&fpriv->evf_mgr);
        amdgpu_ctx_mgr_fini(&fpriv->ctx_mgr);
        amdgpu_vm_fini(adev, &fpriv->vm);
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
index 614953b0fc19..4cf65aba9a9b 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
@@ -455,10 +455,18 @@ int amdgpu_userq_signal_ioctl(struct drm_device *dev, 
void *data,
        if (r)
                goto exec_fini;
 
-       for (i = 0; i < num_bo_handles; i++)
+       /* Save the fence to wait for during suspend */
+       dma_fence_put(queue->last_fence);
+       queue->last_fence = dma_fence_get(fence);
+
+       for (i = 0; i < num_bo_handles; i++) {
+               if (!gobj || !gobj[i]->resv)
+                       continue;
+
                dma_resv_add_fence(gobj[i]->resv, fence,
                                   dma_resv_usage_rw(args->bo_flags &
                                   AMDGPU_USERQ_BO_WRITE));
+       }
 
        /* Add the created fence to syncobj/BO's */
        for (i = 0; i < num_syncobj_handles; i++)
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c
index ba986d55ceeb..979174f80993 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c
@@ -22,6 +22,7 @@
  *
  */
 #include <drm/drm_syncobj.h>
+#include <drm/drm_exec.h>
 #include "amdgpu.h"
 #include "amdgpu_vm.h"
 #include "amdgpu_userqueue.h"
@@ -282,6 +283,7 @@ amdgpu_userqueue_destroy(struct drm_file *filp, int 
queue_id)
        amdgpu_bo_unpin(queue->db_obj.obj);
        amdgpu_bo_unref(&queue->db_obj.obj);
        amdgpu_userqueue_cleanup(uq_mgr, queue, queue_id);
+       uq_mgr->num_userqs--;
        mutex_unlock(&uq_mgr->userq_mutex);
        return 0;
 }
@@ -369,6 +371,7 @@ amdgpu_userqueue_create(struct drm_file *filp, union 
drm_amdgpu_userq *args)
                goto unlock;
        }
        args->out.queue_id = qid;
+       uq_mgr->num_userqs++;
 
 unlock:
        mutex_unlock(&uq_mgr->userq_mutex);
@@ -402,12 +405,109 @@ int amdgpu_userq_ioctl(struct drm_device *dev, void 
*data,
        return r;
 }
 
+static int
+amdgpu_userqueue_suspend_all(struct amdgpu_userq_mgr *uq_mgr)
+{
+       struct amdgpu_device *adev = uq_mgr->adev;
+       const struct amdgpu_userq_funcs *userq_funcs;
+       struct amdgpu_usermode_queue *queue;
+       int queue_id, ret;
+
+       userq_funcs = adev->userq_funcs[AMDGPU_HW_IP_GFX];
+
+       /* Suspend all the queues for this process */
+       idr_for_each_entry(&uq_mgr->userq_idr, queue, queue_id) {
+               ret = userq_funcs->suspend(uq_mgr, queue);
+               if (ret)
+                       DRM_ERROR("Failed to suspend queue\n");
+       }
+
+       return ret;
+}
+
+static int
+amdgpu_userqueue_wait_for_signal(struct amdgpu_userq_mgr *uq_mgr)
+{
+       struct amdgpu_usermode_queue *queue;
+       int queue_id, ret;
+
+       idr_for_each_entry(&uq_mgr->userq_idr, queue, queue_id) {
+               struct dma_fence *f;
+               struct drm_exec exec;
+
+               drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT, 0);
+               drm_exec_until_all_locked(&exec) {
+                       f = queue->last_fence;
+                       drm_exec_retry_on_contention(&exec);
+               }
+               drm_exec_fini(&exec);
+
+               if (!f || dma_fence_is_signaled(f))
+                       continue;
+               ret = dma_fence_wait_timeout(f, true, msecs_to_jiffies(100));
+               if ( ret <= 0) {
+                       DRM_ERROR("Timed out waiting for fence f=%p\n", f);
+                       return -ETIMEDOUT;
+               }
+       }
+
+       return 0;
+}
+
+static void
+amdgpu_userqueue_suspend_worker(struct work_struct *work)
+{
+       int ret;
+       struct amdgpu_userq_mgr *uq_mgr = work_to_uq_mgr(work, 
suspend_work.work);
+       struct amdgpu_fpriv *fpriv = uq_mgr_to_fpriv(uq_mgr);
+       struct amdgpu_eviction_fence_mgr *evf_mgr = &fpriv->evf_mgr;
+
+       /* Wait for any pending userqueue fence to signal */
+       ret = amdgpu_userqueue_wait_for_signal(uq_mgr);
+       if (ret) {
+               DRM_ERROR("Not suspending userqueue, timeout waiting for 
work\n");
+               return;
+       }
+
+       mutex_lock(&uq_mgr->userq_mutex);
+       ret = amdgpu_userqueue_suspend_all(uq_mgr);
+       if (ret) {
+               DRM_ERROR("Failed to evict userqueue\n");
+               goto unlock;
+       }
+
+       /* Signal current eviction fence */
+       ret = amdgpu_eviction_fence_signal(evf_mgr);
+       if (ret) {
+               DRM_ERROR("Can't signal eviction fence to suspend\n");
+               goto unlock;
+       }
+
+       /* Cleanup old eviction fence entry */
+       amdgpu_eviction_fence_destroy(evf_mgr);
+
+unlock:
+       mutex_unlock(&uq_mgr->userq_mutex);
+}
+
 int amdgpu_userq_mgr_init(struct amdgpu_userq_mgr *userq_mgr, struct 
amdgpu_device *adev)
 {
+       struct amdgpu_fpriv *fpriv;
+
        mutex_init(&userq_mgr->userq_mutex);
        idr_init_base(&userq_mgr->userq_idr, 1);
        userq_mgr->adev = adev;
+       userq_mgr->num_userqs = 0;
+
+       fpriv = uq_mgr_to_fpriv(userq_mgr);
+       if (!fpriv->evf_mgr.ev_fence) {
+               DRM_ERROR("Eviction fence not initialized yet\n");
+               return -EINVAL;
+       }
 
+       /* This reference is required for suspend work */
+       fpriv->evf_mgr.ev_fence->uq_mgr = userq_mgr;
+       INIT_DELAYED_WORK(&userq_mgr->suspend_work, 
amdgpu_userqueue_suspend_worker);
        return 0;
 }
 
diff --git a/drivers/gpu/drm/amd/include/amdgpu_userqueue.h 
b/drivers/gpu/drm/amd/include/amdgpu_userqueue.h
index 37be29048f42..8b3b50fa8b5b 100644
--- a/drivers/gpu/drm/amd/include/amdgpu_userqueue.h
+++ b/drivers/gpu/drm/amd/include/amdgpu_userqueue.h
@@ -27,6 +27,10 @@
 
 #define AMDGPU_MAX_USERQ_COUNT 512
 
+#define to_ev_fence(f) container_of(f, struct amdgpu_eviction_fence, base)
+#define work_to_uq_mgr(w, name) container_of(w, struct amdgpu_userq_mgr, name)
+#define uq_mgr_to_fpriv(u) container_of(u, struct amdgpu_fpriv, userq_mgr)
+
 struct amdgpu_mqd_prop;
 
 struct amdgpu_userq_obj {
@@ -50,6 +54,7 @@ struct amdgpu_usermode_queue {
        struct amdgpu_userq_obj wptr_obj;
        struct xarray           uq_fence_drv_xa;
        struct amdgpu_userq_fence_driver *fence_drv;
+       struct dma_fence        *last_fence;
 };
 
 struct amdgpu_userq_funcs {
@@ -69,6 +74,9 @@ struct amdgpu_userq_mgr {
        struct idr                      userq_idr;
        struct mutex                    userq_mutex;
        struct amdgpu_device            *adev;
+
+       struct delayed_work             suspend_work;
+       int num_userqs;
 };
 
 int amdgpu_userq_ioctl(struct drm_device *dev, void *data, struct drm_file 
*filp);
@@ -86,4 +94,6 @@ void amdgpu_userqueue_destroy_object(struct amdgpu_userq_mgr 
*uq_mgr,
 int amdgpu_userqueue_update_bo_mapping(struct drm_file *filp, struct 
amdgpu_bo_va *bo_va,
                                       uint32_t operation, uint32_t 
syncobj_handle,
                                       uint64_t point);
+
+int amdgpu_userqueue_enable_signaling(struct dma_fence *f);
 #endif
-- 
2.45.1

Reply via email to