[Public] > From: Koenig, Christian <christian.koe...@amd.com> > Sent: Monday, June 2, 2025 7:06 PM > To: Liang, Prike <prike.li...@amd.com>; amd-gfx@lists.freedesktop.org > Cc: Deucher, Alexander <alexander.deuc...@amd.com>; Lazar, Lijo > <lijo.la...@amd.com> > Subject: Re: [PATCH 5/9] drm/amdgpu: add userq object va track helpers > > On 5/30/25 09:55, Prike Liang wrote: > > Add the userq object virtual address get(),mapped() and put() helpers > > for tracking the userq obj va address usage. > > > > Signed-off-by: Prike Liang <prike.li...@amd.com> > > --- > > drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c | 120 > > +++++++++++++++++++++- drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h | > > 11 ++ > > 2 files changed, 130 insertions(+), 1 deletion(-) > > > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c > > b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c > > index 32e88064bdea..3854b1696d4d 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c > > @@ -79,6 +79,122 @@ int amdgpu_userq_input_va_validate(struct amdgpu_vm > *vm, u64 addr, > > return -EINVAL; > > } > > > > +int amdgpu_userq_buffer_va_get(struct amdgpu_vm *vm, u64 addr) { > > + struct amdgpu_bo_va_mapping *mapping; > > + u64 user_addr; > > + int r; > > + > > + addr &= AMDGPU_GMC_HOLE_MASK; > > + user_addr = ALIGN_DOWN(addr >> AMDGPU_GPU_PAGE_SHIFT, 0x100); > > + > > + r = amdgpu_bo_reserve(vm->root.bo, false); > > + if (r) > > + return r; > > + > > + mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr); > > + if (!mapping) > > + goto out_err; > > + > > + /* > > + * Need to unify the following userq va reference. > > + * mqd bo > > + * rptr bo > > + * wptr bo > > + * eop bo > > + * doorbell bo > > + */ > > + //amdgpu_bo_ref(mapping->bo_va->base.bo); > > Please don't use // in kernel code. Note.
> Regards, > Christian. > > > + mapping->bo_va->queue_refcount++; > > + > > + amdgpu_bo_unreserve(vm->root.bo); > > + return 0; > > + > > +out_err: > > + amdgpu_bo_unreserve(vm->root.bo); > > + return -EINVAL; > > +} > > + > > +bool amdgpu_userq_buffer_va_mapped(struct amdgpu_vm *vm, u64 addr) { > > + struct amdgpu_bo_va_mapping *mapping; > > + u64 user_addr; > > + bool r; > > + > > + addr &= AMDGPU_GMC_HOLE_MASK; > > + user_addr = ALIGN_DOWN(addr >> AMDGPU_GPU_PAGE_SHIFT, 0x100); > > + > > + if (amdgpu_bo_reserve(vm->root.bo, false)) > > + return false; > > + > > + mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr); > > + if (!IS_ERR_OR_NULL(mapping) && mapping->bo_va->queue_refcount > 0) > > + r = true; > > + else > > + r = false; > > + amdgpu_bo_unreserve(vm->root.bo); > > + > > + return r; > > +} > > + > > +bool amdgpu_userq_buffer_vas_mapped(struct amdgpu_vm *vm, > > + struct amdgpu_usermode_queue *queue) { > > + > > + if (amdgpu_userq_buffer_va_mapped(vm, queue->doorbell_handle) || > > + amdgpu_userq_buffer_va_mapped(vm, queue->queue_va) || > > + amdgpu_userq_buffer_va_mapped(vm, queue->rptr_va) || > > + amdgpu_userq_buffer_va_mapped(vm, queue->wptr_va)) > > + return true; > > + else > > + return false; > > +} > > + > > +int amdgpu_userq_buffer_va_put(struct amdgpu_vm *vm, u64 addr) { > > + struct amdgpu_bo_va_mapping *mapping; > > + u64 user_addr; > > + int r; > > + > > + addr &= AMDGPU_GMC_HOLE_MASK; > > + user_addr = ALIGN_DOWN(addr >> AMDGPU_GPU_PAGE_SHIFT, 0x100); > > + > > + r = amdgpu_bo_reserve(vm->root.bo, false); > > + if (r) > > + return r; > > + > > + mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr); > > + if (!mapping) > > + goto out_err; > > + /* > > + * As the GEM userq bo will be unmapped by amdgpu_vm_bo_unmap() > which is > > + * invoked before userq is destroyed. So if we want to > > reference/dereference > > + * the userq vm bo, then ensure the userq vm bo dereferenced before > > + * amdgpu_vm_bo_unmap() or fallback to unmap the userq vm bo at > amdgpu_userq_destroy(). > > + */ > > + //amdgpu_bo_unref(&mapping->bo_va->base.bo); > > + > > + mapping->bo_va->queue_refcount--; > > + > > + amdgpu_bo_unreserve(vm->root.bo); > > + return 0; > > + > > +out_err: > > + amdgpu_bo_unreserve(vm->root.bo); > > + return -EINVAL; > > +} > > + > > +int amdgpu_userq_buffer_vas_put(struct amdgpu_vm *vm, > > + struct amdgpu_usermode_queue *queue) { > > + amdgpu_userq_buffer_va_put(vm, queue->doorbell_handle); > > + amdgpu_userq_buffer_va_put(vm, queue->queue_va); > > + amdgpu_userq_buffer_va_put(vm, queue->rptr_va); > > + amdgpu_userq_buffer_va_put(vm, queue->wptr_va); > > + > > + return 0; > > +} > > + > > static int > > amdgpu_userq_unmap_helper(struct amdgpu_userq_mgr *uq_mgr, > > struct amdgpu_usermode_queue *queue) @@ -451,6 > +567,9 @@ > > amdgpu_userq_create(struct drm_file *filp, union drm_amdgpu_userq *args) > > queue->queue_type = args->in.ip_type; > > queue->vm = &fpriv->vm; > > queue->priority = priority; > > + queue->queue_va = args->in.queue_va; > > + queue->rptr_va = args->in.rptr_va; > > + queue->wptr_va = args->in.wptr_va; > > > > db_info.queue_type = queue->queue_type; > > db_info.doorbell_handle = queue->doorbell_handle; @@ -481,7 +600,6 > > @@ amdgpu_userq_create(struct drm_file *filp, union drm_amdgpu_userq *args) > > goto unlock; > > } > > > > - > > qid = idr_alloc(&uq_mgr->userq_idr, queue, 1, > AMDGPU_MAX_USERQ_COUNT, GFP_KERNEL); > > if (qid < 0) { > > drm_file_err(uq_mgr->file, "Failed to allocate a queue id\n"); > > diff > > --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h > > b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h > > index 375fba639c94..65f9bd91ac73 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h > > @@ -52,6 +52,10 @@ struct amdgpu_usermode_queue { > > enum amdgpu_userq_state state; > > uint64_t doorbell_handle; > > uint64_t doorbell_index; > > + uint64_t queue_va; > > + uint64_t rptr_va; > > + uint64_t wptr_va; > > + > > uint64_t flags; > > struct amdgpu_mqd_prop *userq_prop; > > struct amdgpu_userq_mgr *userq_mgr; > > @@ -134,4 +138,11 @@ int > > amdgpu_userq_start_sched_for_enforce_isolation(struct amdgpu_device > > *adev, > > > > int amdgpu_userq_input_va_validate(struct amdgpu_vm *vm, u64 addr, > > u64 expected_size); > > +int amdgpu_userq_buffer_va_get(struct amdgpu_vm *vm, u64 addr); bool > > +amdgpu_userq_buffer_va_mapped(struct amdgpu_vm *vm, u64 addr); bool > > +amdgpu_userq_buffer_vas_mapped(struct amdgpu_vm *vm, > > + struct amdgpu_usermode_queue *queue); int > > +amdgpu_userq_buffer_va_put(struct amdgpu_vm *vm, u64 addr); int > > +amdgpu_userq_buffer_vas_put(struct amdgpu_vm *vm, > > + struct amdgpu_usermode_queue *queue); > > #endif