change the DRM_ERROR to drm_err which gives the drm device information too which is useful in case of multiple GPU's.
Signed-off-by: Sunil Khatri <sunil.kha...@amd.com> --- drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c | 75 ++++++++++--------- 1 file changed, 41 insertions(+), 34 deletions(-) diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c index eb96c12e02e7..a1046808ab4b 100644 --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_userqueue.c @@ -47,7 +47,7 @@ amdgpu_userqueue_cleanup(struct amdgpu_userq_mgr *uq_mgr, if (ret <= 0) { file = uq_mgr->file; drm_process_info(file, proc_log, sizeof(proc_log)); - DRM_ERROR("Timed out waiting for fence f=%p for %s\n", + drm_err(adev_to_drm(adev), "Timed out waiting for fence f=%p for %s\n", f, proc_log); return; } @@ -128,25 +128,25 @@ int amdgpu_userqueue_create_object(struct amdgpu_userq_mgr *uq_mgr, r = amdgpu_bo_create(adev, &bp, &userq_obj->obj); if (r) { - DRM_ERROR("Failed to allocate BO for userqueue (%d)", r); + drm_err(adev_to_drm(adev), "Failed to allocate BO for userqueue (%d)", r); return r; } r = amdgpu_bo_reserve(userq_obj->obj, true); if (r) { - DRM_ERROR("Failed to reserve BO to map (%d)", r); + drm_err(adev_to_drm(adev), "Failed to reserve BO to map (%d)", r); goto free_obj; } r = amdgpu_ttm_alloc_gart(&(userq_obj->obj)->tbo); if (r) { - DRM_ERROR("Failed to alloc GART for userqueue object (%d)", r); + drm_err(adev_to_drm(adev), "Failed to alloc GART for userqueue object (%d)", r); goto unresv; } r = amdgpu_bo_kmap(userq_obj->obj, &userq_obj->cpu_ptr); if (r) { - DRM_ERROR("Failed to map BO for userqueue (%d)", r); + drm_err(adev_to_drm(adev), "Failed to map BO for userqueue (%d)", r); goto unresv; } @@ -182,7 +182,7 @@ amdgpu_userqueue_get_doorbell_index(struct amdgpu_userq_mgr *uq_mgr, gobj = drm_gem_object_lookup(filp, db_info->doorbell_handle); if (gobj == NULL) { - DRM_ERROR("Can't find GEM object for doorbell\n"); + drm_err(adev_to_drm(uq_mgr->adev), "Can't find GEM object for doorbell\n"); return -EINVAL; } @@ -192,13 +192,15 @@ amdgpu_userqueue_get_doorbell_index(struct amdgpu_userq_mgr *uq_mgr, /* Pin the BO before generating the index, unpin in queue destroy */ r = amdgpu_bo_pin(db_obj->obj, AMDGPU_GEM_DOMAIN_DOORBELL); if (r) { - DRM_ERROR("[Usermode queues] Failed to pin doorbell object\n"); + drm_err(adev_to_drm(uq_mgr->adev), + "[Usermode queues] Failed to pin doorbell object\n"); goto unref_bo; } r = amdgpu_bo_reserve(db_obj->obj, true); if (r) { - DRM_ERROR("[Usermode queues] Failed to pin doorbell object\n"); + drm_err(adev_to_drm(uq_mgr->adev), + "[Usermode queues] Failed to pin doorbell object\n"); goto unpin_bo; } @@ -220,14 +222,16 @@ amdgpu_userqueue_get_doorbell_index(struct amdgpu_userq_mgr *uq_mgr, break; default: - DRM_ERROR("[Usermode queues] IP %d not support\n", db_info->queue_type); + drm_err(adev_to_drm(uq_mgr->adev), + "[Usermode queues] IP %d not support\n", db_info->queue_type); r = -EINVAL; goto unpin_bo; } index = amdgpu_doorbell_index_on_bar(uq_mgr->adev, db_obj->obj, db_info->doorbell_offset, db_size); - DRM_DEBUG_DRIVER("[Usermode queues] doorbell index=%lld\n", index); + drm_dbg_driver(adev_to_drm(uq_mgr->adev), + "[Usermode queues] doorbell index=%lld\n", index); amdgpu_bo_unreserve(db_obj->obj); return index; @@ -254,7 +258,7 @@ amdgpu_userqueue_destroy(struct drm_file *filp, int queue_id) queue = amdgpu_userqueue_find(uq_mgr, queue_id); if (!queue) { - DRM_DEBUG_DRIVER("Invalid queue id to destroy\n"); + drm_dbg_driver(adev_to_drm(uq_mgr->adev), "Invalid queue id to destroy\n"); mutex_unlock(&uq_mgr->userq_mutex); return -EINVAL; } @@ -287,7 +291,8 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args) if (args->in.ip_type != AMDGPU_HW_IP_GFX && args->in.ip_type != AMDGPU_HW_IP_DMA && args->in.ip_type != AMDGPU_HW_IP_COMPUTE) { - DRM_ERROR("Usermode queue doesn't support IP type %u\n", args->in.ip_type); + drm_err(adev_to_drm(adev), + "Usermode queue doesn't support IP type %u\n", args->in.ip_type); return -EINVAL; } @@ -309,14 +314,16 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args) uq_funcs = adev->userq_funcs[args->in.ip_type]; if (!uq_funcs) { - DRM_ERROR("Usermode queue is not supported for this IP (%u)\n", args->in.ip_type); + drm_err(adev_to_drm(adev), + "Usermode queue is not supported for this IP (%u)\n", + args->in.ip_type); r = -EINVAL; goto unlock; } queue = kzalloc(sizeof(struct amdgpu_usermode_queue), GFP_KERNEL); if (!queue) { - DRM_ERROR("Failed to allocate memory for queue\n"); + drm_err(adev_to_drm(adev), "Failed to allocate memory for queue\n"); r = -ENOMEM; goto unlock; } @@ -332,7 +339,7 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args) /* Convert relative doorbell offset into absolute doorbell index */ index = amdgpu_userqueue_get_doorbell_index(uq_mgr, &db_info, filp); if (index == (uint64_t)-EINVAL) { - DRM_ERROR("Failed to get doorbell for queue\n"); + drm_err(adev_to_drm(adev), "Failed to get doorbell for queue\n"); kfree(queue); goto unlock; } @@ -341,13 +348,13 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args) xa_init_flags(&queue->fence_drv_xa, XA_FLAGS_ALLOC); r = amdgpu_userq_fence_driver_alloc(adev, queue); if (r) { - DRM_ERROR("Failed to alloc fence driver\n"); + drm_err(adev_to_drm(adev), "Failed to alloc fence driver\n"); goto unlock; } r = uq_funcs->mqd_create(uq_mgr, &args->in, queue); if (r) { - DRM_ERROR("Failed to create Queue\n"); + drm_err(adev_to_drm(adev), "Failed to create Queue\n"); amdgpu_userq_fence_driver_free(queue); kfree(queue); goto unlock; @@ -355,7 +362,7 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args) qid = idr_alloc(&uq_mgr->userq_idr, queue, 1, AMDGPU_MAX_USERQ_COUNT, GFP_KERNEL); if (qid < 0) { - DRM_ERROR("Failed to allocate a queue id\n"); + drm_err(adev_to_drm(adev), "Failed to allocate a queue id\n"); amdgpu_userq_fence_driver_free(queue); uq_funcs->mqd_destroy(uq_mgr, queue); kfree(queue); @@ -365,7 +372,7 @@ amdgpu_userqueue_create(struct drm_file *filp, union drm_amdgpu_userq *args) r = uq_funcs->map(uq_mgr, queue); if (r) { - DRM_ERROR("Failed to map Queue\n"); + drm_err(adev_to_drm(adev), "Failed to map Queue\n"); idr_remove(&uq_mgr->userq_idr, qid); amdgpu_userq_fence_driver_free(queue); uq_funcs->mqd_destroy(uq_mgr, queue); @@ -393,7 +400,7 @@ int amdgpu_userq_ioctl(struct drm_device *dev, void *data, return -EINVAL; r = amdgpu_userqueue_create(filp, args); if (r) - DRM_ERROR("Failed to create usermode queue\n"); + drm_err(dev, "Failed to create usermode queue\n"); break; case AMDGPU_USERQ_OP_FREE: @@ -411,11 +418,11 @@ int amdgpu_userq_ioctl(struct drm_device *dev, void *data, return -EINVAL; r = amdgpu_userqueue_destroy(filp, args->in.queue_id); if (r) - DRM_ERROR("Failed to destroy usermode queue\n"); + drm_err(dev, "Failed to destroy usermode queue\n"); break; default: - DRM_DEBUG_DRIVER("Invalid user queue op specified: %d\n", args->in.op); + drm_dbg_driver(dev, "Invalid user queue op specified: %d\n", args->in.op); return -EINVAL; } @@ -449,7 +456,7 @@ amdgpu_userqueue_resume_all(struct amdgpu_userq_mgr *uq_mgr) if (ret) { file = uq_mgr->file; drm_process_info(file, proc_log, sizeof(proc_log)); - DRM_ERROR("Failed to map all the queue for %s\n", + drm_err(adev_to_drm(uq_mgr->adev), "Failed to map all the queue for %s\n", proc_log); } @@ -490,7 +497,7 @@ amdgpu_userqueue_validate_bos(struct amdgpu_userq_mgr *uq_mgr) ret = amdgpu_vm_lock_pd(vm, &exec, 2); drm_exec_retry_on_contention(&exec); if (unlikely(ret)) { - DRM_ERROR("Failed to lock PD\n"); + drm_err(adev_to_drm(adev), "Failed to lock PD\n"); goto unlock_all; } @@ -530,7 +537,7 @@ amdgpu_userqueue_validate_bos(struct amdgpu_userq_mgr *uq_mgr) bo = bo_va->base.bo; ret = amdgpu_userqueue_validate_vm_bo(NULL, bo); if (ret) { - DRM_ERROR("Failed to validate BO\n"); + drm_err(adev_to_drm(adev), "Failed to validate BO\n"); goto unlock_all; } @@ -561,7 +568,7 @@ amdgpu_userqueue_validate_bos(struct amdgpu_userq_mgr *uq_mgr) ret = amdgpu_eviction_fence_replace_fence(&fpriv->evf_mgr, &exec); if (ret) - DRM_ERROR("Failed to replace eviction fence\n"); + drm_err(adev_to_drm(adev), "Failed to replace eviction fence\n"); unlock_all: drm_exec_fini(&exec); @@ -580,13 +587,13 @@ static void amdgpu_userqueue_resume_worker(struct work_struct *work) ret = amdgpu_userqueue_validate_bos(uq_mgr); if (ret) { - DRM_ERROR("Failed to validate BOs to restore\n"); + drm_err(adev_to_drm(uq_mgr->adev), "Failed to validate BOs to restore\n"); goto unlock; } ret = amdgpu_userqueue_resume_all(uq_mgr); if (ret) { - DRM_ERROR("Failed to resume all queues\n"); + drm_err(adev_to_drm(uq_mgr->adev), "Failed to resume all queues\n"); goto unlock; } @@ -614,7 +621,7 @@ amdgpu_userqueue_suspend_all(struct amdgpu_userq_mgr *uq_mgr) if (ret) { file = uq_mgr->file; drm_process_info(file, proc_log, sizeof(proc_log)); - DRM_ERROR("Couldn't unmap all the queues for %s\n", + drm_err(adev_to_drm(adev), "Couldn't unmap all the queues for %s\n", proc_log); } @@ -638,8 +645,8 @@ amdgpu_userqueue_wait_for_signal(struct amdgpu_userq_mgr *uq_mgr) if (ret <= 0) { file = uq_mgr->file; drm_process_info(file, proc_log, sizeof(proc_log)); - DRM_ERROR("Timed out waiting for fence f=%p for %s\n", - f, proc_log); + drm_err(adev_to_drm(uq_mgr->adev), + "Timed out waiting for fence f=%p for %s\n", f, proc_log); return -ETIMEDOUT; } } @@ -662,8 +669,8 @@ amdgpu_userqueue_suspend(struct amdgpu_userq_mgr *uq_mgr, if (ret) { file = uq_mgr->file; drm_process_info(file, proc_log, sizeof(proc_log)); - DRM_ERROR("Not suspending userqueue, timeout waiting for %s\n", - proc_log); + drm_err(adev_to_drm(uq_mgr->adev), + "Not suspending userqueue, timeout waiting for %s\n", proc_log); return; } @@ -671,7 +678,7 @@ amdgpu_userqueue_suspend(struct amdgpu_userq_mgr *uq_mgr, if (ret) { file = uq_mgr->file; drm_process_info(file, proc_log, sizeof(proc_log)); - DRM_ERROR("Failed to evict userqueue for %s\n", proc_log); + drm_err(adev_to_drm(uq_mgr->adev), "Failed to evict userqueue for %s\n", proc_log); return; } -- 2.34.1