On Thu, Apr 24, 2025 at 01:18:35PM +0100, Matthew Auld wrote:
> We now use the same notifier lock for SVM and userptr, with that we can
> combine xe_pt_userptr_pre_commit and xe_pt_svm_pre_commit.
> 
> v2: (Matt B)
>   - Re-use xe_svm_notifier_lock/unlock for userptr.
>   - Combine svm/userptr handling further down into op_check_svm_userptr.
> 
> Suggested-by: Matthew Brost <matthew.br...@intel.com>

The Kconfig issues pointed out in the previous patch need to be fixed
but implementation here look correct:

Reviewed-by: Matthew Brost <matthew.br...@intel.com>

> Signed-off-by: Matthew Auld <matthew.a...@intel.com>
> Cc: Himal Prasad Ghimiray <himal.prasad.ghimi...@intel.com>
> Cc: Thomas Hellström <thomas.hellst...@linux.intel.com>
> ---
>  drivers/gpu/drm/xe/xe_pt.c       | 90 ++++++++++----------------------
>  drivers/gpu/drm/xe/xe_pt_types.h |  2 -
>  drivers/gpu/drm/xe/xe_svm.h      | 19 +++----
>  3 files changed, 39 insertions(+), 72 deletions(-)
> 
> diff --git a/drivers/gpu/drm/xe/xe_pt.c b/drivers/gpu/drm/xe/xe_pt.c
> index ff898e518afd..a7db65f9913e 100644
> --- a/drivers/gpu/drm/xe/xe_pt.c
> +++ b/drivers/gpu/drm/xe/xe_pt.c
> @@ -1393,8 +1393,8 @@ static int vma_check_userptr(struct xe_vm *vm, struct 
> xe_vma *vma,
>       return 0;
>  }
>  
> -static int op_check_userptr(struct xe_vm *vm, struct xe_vma_op *op,
> -                         struct xe_vm_pgtable_update_ops *pt_update)
> +static int op_check_svm_userptr(struct xe_vm *vm, struct xe_vma_op *op,
> +                             struct xe_vm_pgtable_update_ops *pt_update)
>  {
>       int err = 0;
>  
> @@ -1419,6 +1419,24 @@ static int op_check_userptr(struct xe_vm *vm, struct 
> xe_vma_op *op,
>               err = vma_check_userptr(vm, gpuva_to_vma(op->base.prefetch.va),
>                                       pt_update);
>               break;
> +#if IS_ENABLED(CONFIG_DRM_XE_GPUSVM)
> +     case DRM_GPUVA_OP_DRIVER:
> +             if (op->subop == XE_VMA_SUBOP_MAP_RANGE) {
> +                     struct xe_svm_range *range = op->map_range.range;
> +
> +                     xe_svm_range_debug(range, "PRE-COMMIT");
> +
> +                     xe_assert(vm->xe,
> +                               xe_vma_is_cpu_addr_mirror(op->map_range.vma));
> +                     xe_assert(vm->xe, op->subop == XE_VMA_SUBOP_MAP_RANGE);
> +
> +                     if (!xe_svm_range_pages_valid(range)) {
> +                             xe_svm_range_debug(range, "PRE-COMMIT - RETRY");
> +                             err = -EAGAIN;
> +                     }
> +             }
> +             break;
> +#endif
>       default:
>               drm_warn(&vm->xe->drm, "NOT POSSIBLE");
>       }
> @@ -1426,7 +1444,7 @@ static int op_check_userptr(struct xe_vm *vm, struct 
> xe_vma_op *op,
>       return err;
>  }
>  
> -static int xe_pt_userptr_pre_commit(struct xe_migrate_pt_update *pt_update)
> +static int xe_pt_svm_userptr_pre_commit(struct xe_migrate_pt_update 
> *pt_update)
>  {
>       struct xe_vm *vm = pt_update->vops->vm;
>       struct xe_vma_ops *vops = pt_update->vops;
> @@ -1439,12 +1457,12 @@ static int xe_pt_userptr_pre_commit(struct 
> xe_migrate_pt_update *pt_update)
>       if (err)
>               return err;
>  
> -     down_read(&vm->svm.gpusvm.notifier_lock);
> +     xe_svm_notifier_lock(vm);
>  
>       list_for_each_entry(op, &vops->list, link) {
> -             err = op_check_userptr(vm, op, pt_update_ops);
> +             err = op_check_svm_userptr(vm, op, pt_update_ops);
>               if (err) {
> -                     up_read(&vm->svm.gpusvm.notifier_lock);
> +                     xe_svm_notifier_unlock(vm);
>                       break;
>               }
>       }
> @@ -1452,42 +1470,6 @@ static int xe_pt_userptr_pre_commit(struct 
> xe_migrate_pt_update *pt_update)
>       return err;
>  }
>  
> -#if IS_ENABLED(CONFIG_DRM_XE_GPUSVM)
> -static int xe_pt_svm_pre_commit(struct xe_migrate_pt_update *pt_update)
> -{
> -     struct xe_vm *vm = pt_update->vops->vm;
> -     struct xe_vma_ops *vops = pt_update->vops;
> -     struct xe_vma_op *op;
> -     int err;
> -
> -     err = xe_pt_pre_commit(pt_update);
> -     if (err)
> -             return err;
> -
> -     xe_svm_notifier_lock(vm);
> -
> -     list_for_each_entry(op, &vops->list, link) {
> -             struct xe_svm_range *range = op->map_range.range;
> -
> -             if (op->subop == XE_VMA_SUBOP_UNMAP_RANGE)
> -                     continue;
> -
> -             xe_svm_range_debug(range, "PRE-COMMIT");
> -
> -             xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(op->map_range.vma));
> -             xe_assert(vm->xe, op->subop == XE_VMA_SUBOP_MAP_RANGE);
> -
> -             if (!xe_svm_range_pages_valid(range)) {
> -                     xe_svm_range_debug(range, "PRE-COMMIT - RETRY");
> -                     xe_svm_notifier_unlock(vm);
> -                     return -EAGAIN;
> -             }
> -     }
> -
> -     return 0;
> -}
> -#endif
> -
>  struct invalidation_fence {
>       struct xe_gt_tlb_invalidation_fence base;
>       struct xe_gt *gt;
> @@ -1858,7 +1840,7 @@ static int bind_op_prepare(struct xe_vm *vm, struct 
> xe_tile *tile,
>                                                xe_vma_start(vma),
>                                                xe_vma_end(vma));
>               ++pt_update_ops->current_op;
> -             pt_update_ops->needs_userptr_lock |= xe_vma_is_userptr(vma);
> +             pt_update_ops->needs_svm_lock |= xe_vma_is_userptr(vma);
>  
>               /*
>                * If rebind, we have to invalidate TLB on !LR vms to invalidate
> @@ -1966,7 +1948,7 @@ static int unbind_op_prepare(struct xe_tile *tile,
>       xe_pt_update_ops_rfence_interval(pt_update_ops, xe_vma_start(vma),
>                                        xe_vma_end(vma));
>       ++pt_update_ops->current_op;
> -     pt_update_ops->needs_userptr_lock |= xe_vma_is_userptr(vma);
> +     pt_update_ops->needs_svm_lock |= xe_vma_is_userptr(vma);
>       pt_update_ops->needs_invalidation = true;
>  
>       xe_pt_commit_prepare_unbind(vma, pt_op->entries, pt_op->num_entries);
> @@ -2289,22 +2271,12 @@ static const struct xe_migrate_pt_update_ops 
> migrate_ops = {
>       .pre_commit = xe_pt_pre_commit,
>  };
>  
> -static const struct xe_migrate_pt_update_ops userptr_migrate_ops = {
> +static const struct xe_migrate_pt_update_ops svm_userptr_migrate_ops = {
>       .populate = xe_vm_populate_pgtable,
>       .clear = xe_migrate_clear_pgtable_callback,
> -     .pre_commit = xe_pt_userptr_pre_commit,
> +     .pre_commit = xe_pt_svm_userptr_pre_commit,
>  };
>  
> -#if IS_ENABLED(CONFIG_DRM_XE_GPUSVM)
> -static const struct xe_migrate_pt_update_ops svm_migrate_ops = {
> -     .populate = xe_vm_populate_pgtable,
> -     .clear = xe_migrate_clear_pgtable_callback,
> -     .pre_commit = xe_pt_svm_pre_commit,
> -};
> -#else
> -static const struct xe_migrate_pt_update_ops svm_migrate_ops;
> -#endif
> -
>  /**
>   * xe_pt_update_ops_run() - Run PT update operations
>   * @tile: Tile of PT update operations
> @@ -2331,9 +2303,7 @@ xe_pt_update_ops_run(struct xe_tile *tile, struct 
> xe_vma_ops *vops)
>       int err = 0, i;
>       struct xe_migrate_pt_update update = {
>               .ops = pt_update_ops->needs_svm_lock ?
> -                     &svm_migrate_ops :
> -                     pt_update_ops->needs_userptr_lock ?
> -                     &userptr_migrate_ops :
> +                     &svm_userptr_migrate_ops :
>                       &migrate_ops,
>               .vops = vops,
>               .tile_id = tile->id,
> @@ -2455,8 +2425,6 @@ xe_pt_update_ops_run(struct xe_tile *tile, struct 
> xe_vma_ops *vops)
>  
>       if (pt_update_ops->needs_svm_lock)
>               xe_svm_notifier_unlock(vm);
> -     if (pt_update_ops->needs_userptr_lock)
> -             up_read(&vm->svm.gpusvm.notifier_lock);
>  
>       return fence;
>  
> diff --git a/drivers/gpu/drm/xe/xe_pt_types.h 
> b/drivers/gpu/drm/xe/xe_pt_types.h
> index 69eab6f37cfe..dc0b2d8c3af8 100644
> --- a/drivers/gpu/drm/xe/xe_pt_types.h
> +++ b/drivers/gpu/drm/xe/xe_pt_types.h
> @@ -106,8 +106,6 @@ struct xe_vm_pgtable_update_ops {
>       u32 current_op;
>       /** @needs_svm_lock: Needs SVM lock */
>       bool needs_svm_lock;
> -     /** @needs_userptr_lock: Needs userptr lock */
> -     bool needs_userptr_lock;
>       /** @needs_invalidation: Needs invalidation */
>       bool needs_invalidation;
>       /**
> diff --git a/drivers/gpu/drm/xe/xe_svm.h b/drivers/gpu/drm/xe/xe_svm.h
> index 923cb074d0cb..696496b52465 100644
> --- a/drivers/gpu/drm/xe/xe_svm.h
> +++ b/drivers/gpu/drm/xe/xe_svm.h
> @@ -87,15 +87,6 @@ static inline bool xe_svm_range_has_dma_mapping(struct 
> xe_svm_range *range)
>       return range->base.pages.flags.has_dma_mapping;
>  }
>  
> -#define xe_svm_assert_in_notifier(vm__) \
> -     lockdep_assert_held_write(&(vm__)->svm.gpusvm.notifier_lock)
> -
> -#define xe_svm_notifier_lock(vm__)   \
> -     drm_gpusvm_notifier_lock(&(vm__)->svm.gpusvm)
> -
> -#define xe_svm_notifier_unlock(vm__) \
> -     drm_gpusvm_notifier_unlock(&(vm__)->svm.gpusvm)
> -
>  #else
>  #include <linux/interval_tree.h>
>  
> @@ -187,4 +178,14 @@ static inline void xe_svm_notifier_unlock(struct xe_vm 
> *vm)
>  {
>  }
>  #endif
> +
> +#define xe_svm_assert_in_notifier(vm__) \
> +     lockdep_assert_held_write(&(vm__)->svm.gpusvm.notifier_lock)
> +
> +#define xe_svm_notifier_lock(vm__)   \
> +     drm_gpusvm_notifier_lock(&(vm__)->svm.gpusvm)
> +
> +#define xe_svm_notifier_unlock(vm__) \
> +     drm_gpusvm_notifier_unlock(&(vm__)->svm.gpusvm)
> +
>  #endif
> -- 
> 2.49.0
> 

Reply via email to