- DRM_GPUVM_SM_MAP_NOT_MADVISE: Default sm_map operations for the input range.
- DRM_GPUVM_SKIP_GEM_OBJ_VA_SPLIT_MADVISE: This flag is used by drm_gpuvm_sm_map_ops_create to iterate over GPUVMA's in the user-provided range and split the existing non-GEM object VMA if the start or end of the input range lies within it. The operations can create up to 2 REMAPS and 2 MAPs. The purpose of this operation is to be used by the Xe driver to assign attributes to GPUVMA's within the user-defined range. Unlike drm_gpuvm_sm_map_ops_flags in default mode, the operation with this flag will never have UNMAPs and merges, and can be without any final operations. v2 - use drm_gpuvm_sm_map_ops_create with flags instead of defining new ops_create (Danilo) - Add doc (Danilo) Cc: Danilo Krummrich <d...@redhat.com> Cc: Matthew Brost <matthew.br...@intel.com> Cc: Boris Brezillon <bbrezil...@kernel.org> Cc: <dri-devel@lists.freedesktop.org> Signed-off-by: Himal Prasad Ghimiray<himal.prasad.ghimi...@intel.com> --- RFC Link: https://lore.kernel.org/intel-xe/20250314080226.2059819-1-himal.prasad.ghimi...@intel.com/T/#mb706bd1c55232110e42dc7d5c05de61946982472 --- drivers/gpu/drm/drm_gpuvm.c | 93 ++++++++++++++++++++------ drivers/gpu/drm/nouveau/nouveau_uvmm.c | 1 + drivers/gpu/drm/xe/xe_vm.c | 1 + include/drm/drm_gpuvm.h | 25 ++++++- 4 files changed, 98 insertions(+), 22 deletions(-) diff --git a/drivers/gpu/drm/drm_gpuvm.c b/drivers/gpu/drm/drm_gpuvm.c index f9eb56f24bef..9d09d177b9fa 100644 --- a/drivers/gpu/drm/drm_gpuvm.c +++ b/drivers/gpu/drm/drm_gpuvm.c @@ -2102,10 +2102,13 @@ static int __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, const struct drm_gpuvm_ops *ops, void *priv, u64 req_addr, u64 req_range, + enum drm_gpuvm_sm_map_ops_flags flags, struct drm_gem_object *req_obj, u64 req_offset) { struct drm_gpuva *va, *next; u64 req_end = req_addr + req_range; + bool is_madvise_ops = (flags == DRM_GPUVM_SKIP_GEM_OBJ_VA_SPLIT_MADVISE); + bool needs_map = !is_madvise_ops; int ret; if (unlikely(!drm_gpuvm_range_valid(gpuvm, req_addr, req_range))) @@ -2118,26 +2121,35 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, u64 range = va->va.range; u64 end = addr + range; bool merge = !!va->gem.obj; + bool skip_madvise_ops = is_madvise_ops && merge; + needs_map = false; if (addr == req_addr) { merge &= obj == req_obj && offset == req_offset; if (end == req_end) { - ret = op_unmap_cb(ops, priv, va, merge); - if (ret) - return ret; + if (!is_madvise_ops) { + ret = op_unmap_cb(ops, priv, va, merge); + if (ret) + return ret; + } break; } if (end < req_end) { - ret = op_unmap_cb(ops, priv, va, merge); - if (ret) - return ret; + if (!is_madvise_ops) { + ret = op_unmap_cb(ops, priv, va, merge); + if (ret) + return ret; + } continue; } if (end > req_end) { + if (skip_madvise_ops) + break; + struct drm_gpuva_op_map n = { .va.addr = req_end, .va.range = range - req_range, @@ -2152,6 +2164,9 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, ret = op_remap_cb(ops, priv, NULL, &n, &u); if (ret) return ret; + + if (is_madvise_ops) + needs_map = true; break; } } else if (addr < req_addr) { @@ -2169,20 +2184,42 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, u.keep = merge; if (end == req_end) { + if (skip_madvise_ops) + break; + ret = op_remap_cb(ops, priv, &p, NULL, &u); if (ret) return ret; + + if (is_madvise_ops) + needs_map = true; + break; } if (end < req_end) { + if (skip_madvise_ops) + continue; + ret = op_remap_cb(ops, priv, &p, NULL, &u); if (ret) return ret; + + if (is_madvise_ops) { + ret = op_map_cb(ops, priv, req_addr, + min(end - req_addr, req_end - end), + NULL, req_offset); + if (ret) + return ret; + } + continue; } if (end > req_end) { + if (skip_madvise_ops) + break; + struct drm_gpuva_op_map n = { .va.addr = req_end, .va.range = end - req_end, @@ -2194,6 +2231,9 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, ret = op_remap_cb(ops, priv, &p, &n, &u); if (ret) return ret; + + if (is_madvise_ops) + needs_map = true; break; } } else if (addr > req_addr) { @@ -2202,20 +2242,29 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, (addr - req_addr); if (end == req_end) { - ret = op_unmap_cb(ops, priv, va, merge); - if (ret) - return ret; + if (!is_madvise_ops) { + ret = op_unmap_cb(ops, priv, va, merge); + if (ret) + return ret; + } + break; } if (end < req_end) { - ret = op_unmap_cb(ops, priv, va, merge); - if (ret) - return ret; + if (is_madvise_ops) { + ret = op_unmap_cb(ops, priv, va, merge); + if (ret) + return ret; + } + continue; } if (end > req_end) { + if (skip_madvise_ops) + break; + struct drm_gpuva_op_map n = { .va.addr = req_end, .va.range = end - req_end, @@ -2230,14 +2279,16 @@ __drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, ret = op_remap_cb(ops, priv, NULL, &n, &u); if (ret) return ret; + + if (is_madvise_ops) + return op_map_cb(ops, priv, addr, + (req_end - addr), NULL, req_offset); break; } } } - - return op_map_cb(ops, priv, - req_addr, req_range, - req_obj, req_offset); + return needs_map ? op_map_cb(ops, priv, req_addr, + req_range, req_obj, req_offset) : 0; } static int @@ -2336,15 +2387,15 @@ drm_gpuvm_sm_map(struct drm_gpuvm *gpuvm, void *priv, struct drm_gem_object *req_obj, u64 req_offset) { const struct drm_gpuvm_ops *ops = gpuvm->ops; + enum drm_gpuvm_sm_map_ops_flags flags = DRM_GPUVM_SM_MAP_NOT_MADVISE; if (unlikely(!(ops && ops->sm_step_map && ops->sm_step_remap && ops->sm_step_unmap))) return -EINVAL; - return __drm_gpuvm_sm_map(gpuvm, ops, priv, - req_addr, req_range, - req_obj, req_offset); + return __drm_gpuvm_sm_map(gpuvm, ops, priv, req_addr, req_range, + flags, req_obj, req_offset); } EXPORT_SYMBOL_GPL(drm_gpuvm_sm_map); @@ -2486,6 +2537,7 @@ static const struct drm_gpuvm_ops gpuvm_list_ops = { * @gpuvm: the &drm_gpuvm representing the GPU VA space * @req_addr: the start address of the new mapping * @req_range: the range of the new mapping + * @drm_gpuvm_sm_map_ops_flag: ops flag determining madvise or not * @req_obj: the &drm_gem_object to map * @req_offset: the offset within the &drm_gem_object * @@ -2516,6 +2568,7 @@ static const struct drm_gpuvm_ops gpuvm_list_ops = { struct drm_gpuva_ops * drm_gpuvm_sm_map_ops_create(struct drm_gpuvm *gpuvm, u64 req_addr, u64 req_range, + enum drm_gpuvm_sm_map_ops_flags flags, struct drm_gem_object *req_obj, u64 req_offset) { struct drm_gpuva_ops *ops; @@ -2535,7 +2588,7 @@ drm_gpuvm_sm_map_ops_create(struct drm_gpuvm *gpuvm, args.ops = ops; ret = __drm_gpuvm_sm_map(gpuvm, &gpuvm_list_ops, &args, - req_addr, req_range, + req_addr, req_range, flags, req_obj, req_offset); if (ret) goto err_free_ops; diff --git a/drivers/gpu/drm/nouveau/nouveau_uvmm.c b/drivers/gpu/drm/nouveau/nouveau_uvmm.c index 48f105239f42..26e13fcdbdb8 100644 --- a/drivers/gpu/drm/nouveau/nouveau_uvmm.c +++ b/drivers/gpu/drm/nouveau/nouveau_uvmm.c @@ -1303,6 +1303,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job, op->ops = drm_gpuvm_sm_map_ops_create(&uvmm->base, op->va.addr, op->va.range, + DRM_GPUVM_SM_MAP_NOT_MADVISE, op->gem.obj, op->gem.offset); if (IS_ERR(op->ops)) { diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c index fd98e74485f4..27a8dbe709c2 100644 --- a/drivers/gpu/drm/xe/xe_vm.c +++ b/drivers/gpu/drm/xe/xe_vm.c @@ -2314,6 +2314,7 @@ vm_bind_ioctl_ops_create(struct xe_vm *vm, struct xe_vma_ops *vops, case DRM_XE_VM_BIND_OP_MAP: case DRM_XE_VM_BIND_OP_MAP_USERPTR: ops = drm_gpuvm_sm_map_ops_create(&vm->gpuvm, addr, range, + DRM_GPUVM_SM_MAP_NOT_MADVISE, obj, bo_offset_or_userptr); break; case DRM_XE_VM_BIND_OP_UNMAP: diff --git a/include/drm/drm_gpuvm.h b/include/drm/drm_gpuvm.h index 2a9629377633..6c2452537d4f 100644 --- a/include/drm/drm_gpuvm.h +++ b/include/drm/drm_gpuvm.h @@ -211,6 +211,27 @@ enum drm_gpuvm_flags { DRM_GPUVM_USERBITS = BIT(1), }; +/** + * enum drm_gpuvm_madvise_flags - flags for drm_gpuvm split/merge ops + */ +enum drm_gpuvm_sm_map_ops_flags { + /** + * @DRM_GPUVM_SM_MAP_NOT_MADVISE: DEFAULT sm_map ops + */ + DRM_GPUVM_SM_MAP_NOT_MADVISE = 0, + + /** + * @DRM_GPUVM_SKIP_GEM_OBJ_VA_SPLIT_MADVISE: This flag is used by + * drm_gpuvm_sm_map_ops_create to iterate over GPUVMA's in the + * user-provided range and split the existing non-GEM object VMA if the + * start or end of the input range lies within it. The operations can + * create up to 2 REMAPS and 2 MAPs. Unlike drm_gpuvm_sm_map_ops_flags + * in default mode, the operation with this flag will never have UNMAPs and + * merges, and can be without any final operations. + */ + DRM_GPUVM_SKIP_GEM_OBJ_VA_SPLIT_MADVISE = BIT(0), +}; + /** * struct drm_gpuvm - DRM GPU VA Manager * @@ -1059,8 +1080,8 @@ struct drm_gpuva_ops { #define drm_gpuva_next_op(op) list_next_entry(op, entry) struct drm_gpuva_ops * -drm_gpuvm_sm_map_ops_create(struct drm_gpuvm *gpuvm, - u64 addr, u64 range, +drm_gpuvm_sm_map_ops_create(struct drm_gpuvm *gpuvm, u64 addr, u64 range, + enum drm_gpuvm_sm_map_ops_flags flags, struct drm_gem_object *obj, u64 offset); struct drm_gpuva_ops * drm_gpuvm_sm_unmap_ops_create(struct drm_gpuvm *gpuvm, -- 2.34.1