From: Christian König <christian.koe...@amd.com>

Remove the mostly unused device parameter, for the few cases
where we really need it keep a copy in the context structure.

Signed-off-by: Christian König <christian.koenig at amd.com>
Reviewed-by: Jammy Zhou <Jammy.Zhou at amd.com>
Reviewed-by: Alex Deucher <alexander.deucher at amd.com>
---
 amdgpu/amdgpu.h            |  24 +++-------
 amdgpu/amdgpu_cs.c         | 115 +++++++++++++++++++--------------------------
 amdgpu/amdgpu_internal.h   |   2 +
 tests/amdgpu/basic_tests.c |  33 +++++++------
 tests/amdgpu/cs_tests.c    |  14 +++---
 5 files changed, 78 insertions(+), 110 deletions(-)

diff --git a/amdgpu/amdgpu.h b/amdgpu/amdgpu.h
index 90dc33c..7a85982 100644
--- a/amdgpu/amdgpu.h
+++ b/amdgpu/amdgpu.h
@@ -882,7 +882,6 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
  *
  * Destroy GPU execution context when not needed any more
  *
- * \param   dev            - \c [in] Device handle. See 
#amdgpu_device_initialize()
  * \param   context - \c [in] GPU Context handle
  *
  * \return   0 on success\n
@@ -892,13 +891,11 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
  * \sa amdgpu_cs_ctx_create()
  *
 */
-int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
-                      amdgpu_context_handle context);
+int amdgpu_cs_ctx_free(amdgpu_context_handle context);

 /**
  * Query reset state for the specific GPU Context
  *
- * \param   dev            - \c [in] Device handle. See 
#amdgpu_device_initialize()
  * \param   context - \c [in]  GPU Context handle
  * \param   state   - \c [out] Reset state status
  *
@@ -909,8 +906,7 @@ int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
  * \sa amdgpu_cs_ctx_create()
  *
 */
-int amdgpu_cs_query_reset_state(amdgpu_device_handle dev,
-                               amdgpu_context_handle context,
+int amdgpu_cs_query_reset_state(amdgpu_context_handle context,
                                enum amdgpu_cs_ctx_reset_state *state);


@@ -924,7 +920,6 @@ int amdgpu_cs_query_reset_state(amdgpu_device_handle dev,
  * Allocate memory to be filled with PM4 packets and be served as the first
  * entry point of execution (a.k.a. Indirect Buffer)
  *
- * \param   dev            - \c [in] Device handle. See 
#amdgpu_device_initialize()
  * \param   context - \c [in]  GPU Context which will use IB
  * \param   ib_size - \c [in]  Size of allocation
  * \param   output  - \c [out] Pointer to structure to get information about
@@ -937,8 +932,7 @@ int amdgpu_cs_query_reset_state(amdgpu_device_handle dev,
  * \sa amdgpu_cs_free_ib()
  *
 */
-int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
-                      amdgpu_context_handle context,
+int amdgpu_cs_alloc_ib(amdgpu_context_handle context,
                       enum amdgpu_cs_ib_size ib_size,
                       struct amdgpu_cs_ib_alloc_result *output);

@@ -946,8 +940,6 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
  * If UMD has allocates IBs which doesn’t need any more than those IBs must
  * be explicitly freed
  *
- * \param   dev            - \c [in] Device handle. See 
#amdgpu_device_initialize()
- * \param   context - \c [in] GPU Context containing IB
  * \param   handle  - \c [in] IB handle
  *
  * \return   0 on success\n
@@ -960,9 +952,7 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
  * \sa amdgpu_cs_alloc_ib()
  *
 */
-int amdgpu_cs_free_ib(amdgpu_device_handle dev,
-                     amdgpu_context_handle context,
-                     amdgpu_ib_handle handle);
+int amdgpu_cs_free_ib(amdgpu_ib_handle handle);

 /**
  * Send request to submit command buffers to hardware.
@@ -1011,8 +1001,7 @@ int amdgpu_cs_free_ib(amdgpu_device_handle dev,
  *     amdgpu_cs_query_fence_status()
  *
 */
-int amdgpu_cs_submit(amdgpu_device_handle  dev,
-                    amdgpu_context_handle context,
+int amdgpu_cs_submit(amdgpu_context_handle context,
                     uint64_t flags,
                     struct amdgpu_cs_request *ibs_request,
                     uint32_t number_of_requests,
@@ -1038,8 +1027,7 @@ int amdgpu_cs_submit(amdgpu_device_handle  dev,
  *
  * \sa amdgpu_cs_submit()
 */
-int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,
-                                struct amdgpu_cs_query_fence *fence,
+int amdgpu_cs_query_fence_status(struct amdgpu_cs_query_fence *fence,
                                 uint32_t *expired);


diff --git a/amdgpu/amdgpu_cs.c b/amdgpu/amdgpu_cs.c
index 614904d..d6b4b2d 100644
--- a/amdgpu/amdgpu_cs.c
+++ b/amdgpu/amdgpu_cs.c
@@ -42,8 +42,7 @@
  *
  * \return  0 on success otherwise POSIX Error code
 */
-static int amdgpu_cs_create_ib(amdgpu_device_handle dev,
-                              amdgpu_context_handle context,
+static int amdgpu_cs_create_ib(amdgpu_context_handle context,
                               enum amdgpu_cs_ib_size ib_size,
                               amdgpu_ib_handle *ib)
 {
@@ -79,7 +78,7 @@ static int amdgpu_cs_create_ib(amdgpu_device_handle dev,

        alloc_buffer.preferred_heap = AMDGPU_GEM_DOMAIN_GTT;

-       r = amdgpu_bo_alloc(dev,
+       r = amdgpu_bo_alloc(context->dev,
                            &alloc_buffer,
                            &info);
        if (r)
@@ -98,6 +97,7 @@ static int amdgpu_cs_create_ib(amdgpu_device_handle dev,
                return -ENOMEM;
        }

+       new_ib->context = context;
        new_ib->buf_handle = info.buf_handle;
        new_ib->cpu = cpu;
        new_ib->virtual_mc_base_address = info.virtual_mc_base_address;
@@ -114,10 +114,10 @@ static int amdgpu_cs_create_ib(amdgpu_device_handle dev,
  *
  * \return  0 on success otherwise POSIX Error code
 */
-static int amdgpu_cs_destroy_ib(amdgpu_device_handle dev,
-                               amdgpu_ib_handle ib)
+static int amdgpu_cs_destroy_ib(amdgpu_ib_handle ib)
 {
        int r;
+
        r = amdgpu_bo_cpu_unmap(ib->buf_handle);
        if (r)
                return r;
@@ -162,8 +162,7 @@ static int amdgpu_cs_init_ib_pool(amdgpu_context_handle 
context)
  *
  * \return  0 on success otherwise POSIX Error code
 */
-static int amdgpu_cs_alloc_from_ib_pool(amdgpu_device_handle dev,
-                                       amdgpu_context_handle context,
+static int amdgpu_cs_alloc_from_ib_pool(amdgpu_context_handle context,
                                        enum amdgpu_cs_ib_size ib_size,
                                        amdgpu_ib_handle *ib)
 {
@@ -210,21 +209,19 @@ static void 
amdgpu_cs_free_to_ib_pool(amdgpu_context_handle context,
  *
  * \return  0 on success otherwise POSIX Error code
 */
-static int amdgpu_cs_destroy_ib_pool(amdgpu_device_handle dev,
-                                    amdgpu_context_handle context)
+static int amdgpu_cs_destroy_ib_pool(amdgpu_context_handle context)
 {
-       int i;
-       int r;
        struct list_head *head;
        struct amdgpu_ib *next;
        struct amdgpu_ib *storage;
+       int i, r;

        r = 0;
        pthread_mutex_lock(&context->pool_mutex);
        for (i = 0; i < AMDGPU_CS_IB_SIZE_NUM; i++) {
                head = &context->ib_pools[i];
                LIST_FOR_EACH_ENTRY_SAFE(next, storage, head, list_node) {
-                       r = amdgpu_cs_destroy_ib(dev, next);
+                       r = amdgpu_cs_destroy_ib(next);
                        if (r)
                                break;
                }
@@ -268,8 +265,7 @@ static int amdgpu_cs_init_pendings(amdgpu_context_handle 
context)
  *
  * \return  0 on success otherwise POSIX Error code
 */
-static int amdgpu_cs_destroy_pendings(amdgpu_device_handle dev,
-                                     amdgpu_context_handle context)
+static int amdgpu_cs_destroy_pendings(amdgpu_context_handle context)
 {
        int ip, inst;
        uint32_t ring;
@@ -285,7 +281,7 @@ static int amdgpu_cs_destroy_pendings(amdgpu_device_handle 
dev,
                        for (ring = 0; ring < AMDGPU_CS_MAX_RINGS; ring++) {
                                head = &context->pendings[ip][inst][ring];
                                LIST_FOR_EACH_ENTRY_SAFE(next, s, head, 
list_node) {
-                                       r = amdgpu_cs_destroy_ib(dev, next);
+                                       r = amdgpu_cs_destroy_ib(next);
                                        if (r)
                                                break;
                                }
@@ -293,7 +289,7 @@ static int amdgpu_cs_destroy_pendings(amdgpu_device_handle 
dev,

        head = &context->freed;
        LIST_FOR_EACH_ENTRY_SAFE(next, s, head, list_node) {
-               r = amdgpu_cs_destroy_ib(dev, next);
+               r = amdgpu_cs_destroy_ib(next);
                if (r)
                        break;
        }
@@ -441,39 +437,35 @@ static void 
amdgpu_cs_all_pending_gc(amdgpu_context_handle context)
  *
  * \return  0 on success otherwise POSIX Error code
 */
-static int amdgpu_cs_alloc_ib_local(amdgpu_device_handle dev,
-                                   amdgpu_context_handle context,
+static int amdgpu_cs_alloc_ib_local(amdgpu_context_handle context,
                                    enum amdgpu_cs_ib_size ib_size,
                                    amdgpu_ib_handle *ib)
 {
        int r;

-       r = amdgpu_cs_alloc_from_ib_pool(dev, context, ib_size, ib);
+       r = amdgpu_cs_alloc_from_ib_pool(context, ib_size, ib);
        if (!r)
                return r;

        amdgpu_cs_all_pending_gc(context);

        /* Retry to allocate from free IB pools after garbage collector. */
-       r = amdgpu_cs_alloc_from_ib_pool(dev, context, ib_size, ib);
+       r = amdgpu_cs_alloc_from_ib_pool(context, ib_size, ib);
        if (!r)
                return r;

        /* There is no suitable IB in free pools. Create one. */
-       r = amdgpu_cs_create_ib(dev, context, ib_size, ib);
+       r = amdgpu_cs_create_ib(context, ib_size, ib);
        return r;
 }

-int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
-                      amdgpu_context_handle context,
+int amdgpu_cs_alloc_ib(amdgpu_context_handle context,
                       enum amdgpu_cs_ib_size ib_size,
                       struct amdgpu_cs_ib_alloc_result *output)
 {
        int r;
        amdgpu_ib_handle ib;

-       if (NULL == dev)
-               return -EINVAL;
        if (NULL == context)
                return -EINVAL;
        if (NULL == output)
@@ -481,7 +473,7 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
        if (ib_size >= AMDGPU_CS_IB_SIZE_NUM)
                return -EINVAL;

-       r = amdgpu_cs_alloc_ib_local(dev, context, ib_size, &ib);
+       r = amdgpu_cs_alloc_ib_local(context, ib_size, &ib);
        if (!r) {
                output->handle = ib;
                output->cpu = ib->cpu;
@@ -491,17 +483,14 @@ int amdgpu_cs_alloc_ib(amdgpu_device_handle dev,
        return r;
 }

-int amdgpu_cs_free_ib(amdgpu_device_handle dev,
-                     amdgpu_context_handle context,
-                     amdgpu_ib_handle handle)
+int amdgpu_cs_free_ib(amdgpu_ib_handle handle)
 {
-       if (NULL == dev)
-               return -EINVAL;
-       if (NULL == context)
-               return -EINVAL;
+       amdgpu_context_handle context;
+
        if (NULL == handle)
                return -EINVAL;

+       context = handle->context;
        pthread_mutex_lock(&context->pendings_mutex);
        LIST_ADD(&handle->list_node, &context->freed);
        pthread_mutex_unlock(&context->pendings_mutex);
@@ -532,6 +521,8 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
        if (NULL == gpu_context)
                return -ENOMEM;

+       gpu_context->dev = dev;
+
        r = pthread_mutex_init(&gpu_context->sequence_mutex, NULL);
        if (r)
                goto error_mutex;
@@ -544,7 +535,7 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
        if (r)
                goto error_pendings;

-       r = amdgpu_cs_alloc_ib_local(dev, gpu_context, amdgpu_cs_ib_size_4K,
+       r = amdgpu_cs_alloc_ib_local(gpu_context, amdgpu_cs_ib_size_4K,
                                     &gpu_context->fence_ib);
        if (r)
                goto error_fence_ib;
@@ -562,13 +553,13 @@ int amdgpu_cs_ctx_create(amdgpu_device_handle dev,
        return 0;

 error_kernel:
-       amdgpu_cs_free_ib(dev, gpu_context, gpu_context->fence_ib);
+       amdgpu_cs_free_ib(gpu_context->fence_ib);

 error_fence_ib:
-       amdgpu_cs_destroy_pendings(dev, gpu_context);
+       amdgpu_cs_destroy_pendings(gpu_context);

 error_pendings:
-       amdgpu_cs_destroy_ib_pool(dev, gpu_context);
+       amdgpu_cs_destroy_ib_pool(gpu_context);

 error_pool:
        pthread_mutex_destroy(&gpu_context->sequence_mutex);
@@ -586,26 +577,23 @@ error_mutex:
  *
  * \return  0 on success otherwise POSIX Error code
 */
-int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
-                      amdgpu_context_handle context)
+int amdgpu_cs_ctx_free(amdgpu_context_handle context)
 {
-       int r;
        union drm_amdgpu_ctx args;
+       int r;

-       if (NULL == dev)
-               return -EINVAL;
        if (NULL == context)
                return -EINVAL;

-       r = amdgpu_cs_free_ib(dev, context, context->fence_ib);
+       r = amdgpu_cs_free_ib(context->fence_ib);
        if (r)
                return r;

-       r = amdgpu_cs_destroy_pendings(dev, context);
+       r = amdgpu_cs_destroy_pendings(context);
        if (r)
                return r;

-       r = amdgpu_cs_destroy_ib_pool(dev, context);
+       r = amdgpu_cs_destroy_ib_pool(context);
        if (r)
                return r;

@@ -615,15 +603,15 @@ int amdgpu_cs_ctx_free(amdgpu_device_handle dev,
        memset(&args, 0, sizeof(args));
        args.in.op = AMDGPU_CTX_OP_FREE_CTX;
        args.in.ctx_id = context->id;
-       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_CTX, &args, sizeof(args));
+       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_CTX,
+                               &args, sizeof(args));

        free(context);

        return r;
 }

-static int amdgpu_cs_create_bo_list(amdgpu_device_handle dev,
-                                   amdgpu_context_handle context,
+static int amdgpu_cs_create_bo_list(amdgpu_context_handle context,
                                    struct amdgpu_cs_request *request,
                                    amdgpu_ib_handle fence_ib,
                                    uint32_t *handle)
@@ -657,7 +645,7 @@ static int amdgpu_cs_create_bo_list(amdgpu_device_handle 
dev,
        if (fence_ib)
                list[i].bo_handle = fence_ib->buf_handle->handle;

-       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
+       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_BO_LIST,
                                &args, sizeof(args));
        if (r)
                return r;
@@ -666,7 +654,7 @@ static int amdgpu_cs_create_bo_list(amdgpu_device_handle 
dev,
        return 0;
 }

-static int amdgpu_cs_free_bo_list(amdgpu_device_handle dev, uint32_t handle)
+static int amdgpu_cs_free_bo_list(amdgpu_context_handle context, uint32_t 
handle)
 {
        union drm_amdgpu_bo_list args;
        int r;
@@ -675,7 +663,7 @@ static int amdgpu_cs_free_bo_list(amdgpu_device_handle dev, 
uint32_t handle)
        args.in.operation = AMDGPU_BO_LIST_OP_DESTROY;
        args.in.list_handle = handle;

-       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
+       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_BO_LIST,
                                &args, sizeof(args));

        return r;
@@ -696,8 +684,7 @@ static uint32_t amdgpu_cs_fence_index(unsigned ip, unsigned 
ring)
  * \return  0 on success otherwise POSIX Error code
  * \sa amdgpu_cs_submit()
 */
-static int amdgpu_cs_submit_one(amdgpu_device_handle dev,
-                               amdgpu_context_handle context,
+static int amdgpu_cs_submit_one(amdgpu_context_handle context,
                                struct amdgpu_cs_request *ibs_request,
                                uint64_t *fence)
 {
@@ -753,7 +740,7 @@ static int amdgpu_cs_submit_one(amdgpu_device_handle dev,
                        chunk_data[i].ib_data.flags = AMDGPU_IB_FLAG_CE;
        }

-       r = amdgpu_cs_create_bo_list(dev, context, ibs_request, NULL,
+       r = amdgpu_cs_create_bo_list(context, ibs_request, NULL,
                                     &cs.in.bo_list_handle);
        if (r)
                goto error_unlock;
@@ -778,7 +765,7 @@ static int amdgpu_cs_submit_one(amdgpu_device_handle dev,
                chunk_data[i].fence_data.offset *= sizeof(uint64_t);
        }

-       r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_CS,
+       r = drmCommandWriteRead(context->dev->fd, DRM_AMDGPU_CS,
                                &cs, sizeof(cs));
        if (r)
                goto error_unlock;
@@ -805,7 +792,7 @@ static int amdgpu_cs_submit_one(amdgpu_device_handle dev,

        pthread_mutex_unlock(&context->sequence_mutex);

-       r = amdgpu_cs_free_bo_list(dev, cs.in.bo_list_handle);
+       r = amdgpu_cs_free_bo_list(context, cs.in.bo_list_handle);
        if (r)
                goto error_free;

@@ -820,18 +807,15 @@ error_free:
        return r;
 }

-int amdgpu_cs_submit(amdgpu_device_handle  dev,
-                    amdgpu_context_handle context,
+int amdgpu_cs_submit(amdgpu_context_handle context,
                     uint64_t flags,
                     struct amdgpu_cs_request *ibs_request,
                     uint32_t number_of_requests,
                     uint64_t *fences)
 {
-       int r;
        uint32_t i;
+       int r;

-       if (NULL == dev)
-               return -EINVAL;
        if (NULL == context)
                return -EINVAL;
        if (NULL == ibs_request)
@@ -841,7 +825,7 @@ int amdgpu_cs_submit(amdgpu_device_handle  dev,

        r = 0;
        for (i = 0; i < number_of_requests; i++) {
-               r = amdgpu_cs_submit_one(dev, context, ibs_request, fences);
+               r = amdgpu_cs_submit_one(context, ibs_request, fences);
                if (r)
                        break;
                fences++;
@@ -904,8 +888,7 @@ static int amdgpu_ioctl_wait_cs(amdgpu_device_handle dev,
        return 0;
 }

-int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,
-                                struct amdgpu_cs_query_fence *fence,
+int amdgpu_cs_query_fence_status(struct amdgpu_cs_query_fence *fence,
                                 uint32_t *expired)
 {
        amdgpu_context_handle context;
@@ -916,8 +899,6 @@ int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,
        bool busy = true;
        int r;

-       if (NULL == dev)
-               return -EINVAL;
        if (NULL == fence)
                return -EINVAL;
        if (NULL == expired)
@@ -958,7 +939,7 @@ int amdgpu_cs_query_fence_status(amdgpu_device_handle dev,

        pthread_mutex_unlock(&context->sequence_mutex);

-       r = amdgpu_ioctl_wait_cs(dev, ip_type, ip_instance, ring,
+       r = amdgpu_ioctl_wait_cs(context->dev, ip_type, ip_instance, ring,
                                 fence->fence, fence->timeout_ns, &busy);
        if (!r && !busy) {
                *expired = true;
diff --git a/amdgpu/amdgpu_internal.h b/amdgpu/amdgpu_internal.h
index b27756d..55db645 100644
--- a/amdgpu/amdgpu_internal.h
+++ b/amdgpu/amdgpu_internal.h
@@ -93,6 +93,7 @@ struct amdgpu_bo {
  * sequence_mutex -> pendings_mutex -> pool_mutex.
 */
 struct amdgpu_context {
+       struct amdgpu_device *dev;
        /** Mutex for accessing fences and to maintain command submissions
            and pending lists in good sequence. */
        pthread_mutex_t sequence_mutex;
@@ -115,6 +116,7 @@ struct amdgpu_context {
 };

 struct amdgpu_ib {
+       amdgpu_context_handle context;
        struct list_head list_node;
        amdgpu_bo_handle buf_handle;
        void *cpu;
diff --git a/tests/amdgpu/basic_tests.c b/tests/amdgpu/basic_tests.c
index c53f6a0..c01a7f4 100644
--- a/tests/amdgpu/basic_tests.c
+++ b/tests/amdgpu/basic_tests.c
@@ -168,11 +168,11 @@ static void amdgpu_command_submission_gfx(void)
        r = amdgpu_cs_ctx_create(device_handle, &context_handle);
        CU_ASSERT_EQUAL(r, 0);

-       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
+       r = amdgpu_cs_alloc_ib(context_handle,
                               amdgpu_cs_ib_size_4K, &ib_result);
        CU_ASSERT_EQUAL(r, 0);

-       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
+       r = amdgpu_cs_alloc_ib(context_handle,
                               amdgpu_cs_ib_size_4K, &ib_result_ce);
        CU_ASSERT_EQUAL(r, 0);

@@ -199,7 +199,7 @@ static void amdgpu_command_submission_gfx(void)
        ibs_request.number_of_ibs = 2;
        ibs_request.ibs = ib_info;

-       r = amdgpu_cs_submit(device_handle, context_handle, 0,
+       r = amdgpu_cs_submit(context_handle, 0,
                             &ibs_request, 1, &fence_status.fence);
        CU_ASSERT_EQUAL(r, 0);

@@ -207,10 +207,10 @@ static void amdgpu_command_submission_gfx(void)
        fence_status.timeout_ns = AMDGPU_TIMEOUT_INFINITE;
        fence_status.ip_type = AMDGPU_HW_IP_GFX;

-       r = amdgpu_cs_query_fence_status(device_handle, &fence_status, 
&expired);
+       r = amdgpu_cs_query_fence_status(&fence_status, &expired);
        CU_ASSERT_EQUAL(r, 0);

-       r = amdgpu_cs_ctx_free(device_handle, context_handle);
+       r = amdgpu_cs_ctx_free(context_handle);
        CU_ASSERT_EQUAL(r, 0);
 }

@@ -230,7 +230,7 @@ static void amdgpu_command_submission_compute(void)

        for (instance = 0; instance < 8; instance++) {
                memset(&ib_result, 0, sizeof(struct amdgpu_cs_ib_alloc_result));
-               r = amdgpu_cs_alloc_ib(device_handle, context_handle,
+               r = amdgpu_cs_alloc_ib(context_handle,
                                       amdgpu_cs_ib_size_4K, &ib_result);
                CU_ASSERT_EQUAL(r, 0);

@@ -249,7 +249,7 @@ static void amdgpu_command_submission_compute(void)
                ibs_request.ibs = &ib_info;

                memset(&fence_status, 0, sizeof(struct amdgpu_cs_query_fence));
-               r = amdgpu_cs_submit(device_handle, context_handle, 0,
+               r = amdgpu_cs_submit(context_handle, 0,
                                     &ibs_request, 1, &fence_status.fence);
                CU_ASSERT_EQUAL(r, 0);

@@ -258,11 +258,11 @@ static void amdgpu_command_submission_compute(void)
                fence_status.ip_type = AMDGPU_HW_IP_COMPUTE;
                fence_status.ring = instance;

-               r = amdgpu_cs_query_fence_status(device_handle, &fence_status, 
&expired);
+               r = amdgpu_cs_query_fence_status(&fence_status, &expired);
                CU_ASSERT_EQUAL(r, 0);
        }

-       r = amdgpu_cs_ctx_free(device_handle, context_handle);
+       r = amdgpu_cs_ctx_free(context_handle);
        CU_ASSERT_EQUAL(r, 0);
 }

@@ -291,7 +291,7 @@ static void amdgpu_sdma_test_exec_cs(amdgpu_context_handle 
context_handle,
        CU_ASSERT_TRUE(pm4_dw <= 1024);

        /* allocate IB */
-       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
+       r = amdgpu_cs_alloc_ib(context_handle,
                           amdgpu_cs_ib_size_4K, &ib_result);
        CU_ASSERT_EQUAL(r, 0);

@@ -313,7 +313,7 @@ static void amdgpu_sdma_test_exec_cs(amdgpu_context_handle 
context_handle,
        CU_ASSERT_NOT_EQUAL(ibs_request, NULL);

        /* submit CS */
-       r = amdgpu_cs_submit(device_handle, context_handle, 0,
+       r = amdgpu_cs_submit(context_handle, 0,
                         ibs_request, 1, &fence_status.fence);
        CU_ASSERT_EQUAL(r, 0);

@@ -323,8 +323,7 @@ static void amdgpu_sdma_test_exec_cs(amdgpu_context_handle 
context_handle,
        fence_status.timeout_ns = AMDGPU_TIMEOUT_INFINITE;

        /* wait for IB accomplished */
-       r = amdgpu_cs_query_fence_status(device_handle, &fence_status,
-                                       &expired);
+       r = amdgpu_cs_query_fence_status(&fence_status, &expired);
        CU_ASSERT_EQUAL(r, 0);
        CU_ASSERT_EQUAL(expired, true);
 }
@@ -410,7 +409,7 @@ static void 
amdgpu_command_submission_sdma_write_linear(void)
        free(pm4);

        /* end of test */
-       r = amdgpu_cs_ctx_free(device_handle, context_handle);
+       r = amdgpu_cs_ctx_free(context_handle);
        CU_ASSERT_EQUAL(r, 0);
 }

@@ -493,7 +492,7 @@ static void amdgpu_command_submission_sdma_const_fill(void)
        free(pm4);

        /* end of test */
-       r = amdgpu_cs_ctx_free(device_handle, context_handle);
+       r = amdgpu_cs_ctx_free(context_handle);
        CU_ASSERT_EQUAL(r, 0);
 }

@@ -596,7 +595,7 @@ static void amdgpu_command_submission_sdma_copy_linear(void)
        free(pm4);

        /* end of test */
-       r = amdgpu_cs_ctx_free(device_handle, context_handle);
+       r = amdgpu_cs_ctx_free(context_handle);
        CU_ASSERT_EQUAL(r, 0);
 }

@@ -671,6 +670,6 @@ static void amdgpu_userptr_test(void)
        CU_ASSERT_EQUAL(r, 0);
        free(ptr);

-       r = amdgpu_cs_ctx_free(device_handle, context_handle);
+       r = amdgpu_cs_ctx_free(context_handle);
        CU_ASSERT_EQUAL(r, 0);
 }
diff --git a/tests/amdgpu/cs_tests.c b/tests/amdgpu/cs_tests.c
index f30d923..c0d0c0d 100644
--- a/tests/amdgpu/cs_tests.c
+++ b/tests/amdgpu/cs_tests.c
@@ -73,8 +73,7 @@ int suite_cs_tests_init(void)
        if (r)
                return CUE_SINIT_FAILED;

-        r = amdgpu_cs_alloc_ib(device_handle, context_handle,
-                               IB_SIZE, &ib_result);
+        r = amdgpu_cs_alloc_ib(context_handle, IB_SIZE, &ib_result);
        if (r)
                return CUE_SINIT_FAILED;

@@ -88,11 +87,11 @@ int suite_cs_tests_clean(void)
 {
        int r;

-       r = amdgpu_cs_free_ib(device_handle, context_handle, ib_handle);
+       r = amdgpu_cs_free_ib(ib_handle);
        if (r)
                return CUE_SCLEAN_FAILED;

-       r = amdgpu_cs_ctx_free(device_handle, context_handle);
+       r = amdgpu_cs_ctx_free(context_handle);
        if (r)
                return CUE_SCLEAN_FAILED;

@@ -121,13 +120,12 @@ static int submit(unsigned ndw, unsigned ip)
        ibs_request.number_of_ibs = 1;
        ibs_request.ibs = &ib_info;

-       r = amdgpu_cs_submit(device_handle, context_handle, 0,
+       r = amdgpu_cs_submit(context_handle, 0,
                             &ibs_request, 1, &fence_status.fence);
        if (r)
                return r;

-       r = amdgpu_cs_alloc_ib(device_handle, context_handle,
-                              IB_SIZE, &ib_result);
+       r = amdgpu_cs_alloc_ib(context_handle, IB_SIZE, &ib_result);
        if (r)
                return r;

@@ -138,7 +136,7 @@ static int submit(unsigned ndw, unsigned ip)
        fence_status.timeout_ns = AMDGPU_TIMEOUT_INFINITE;
        fence_status.ip_type = ip;

-       r = amdgpu_cs_query_fence_status(device_handle, &fence_status, 
&expired);
+       r = amdgpu_cs_query_fence_status(&fence_status, &expired);
        if (r)
                return r;

-- 
1.9.1

Reply via email to