On Fri, Jul 19, 2024 at 11:55:06AM +0200, Nirmoy Das wrote:

Not a complete review, just a few comments.

> On LNL because of flat CCS, driver creates a migrate job to clear
> CCS meta data. Extend that to also clear system pages using GPU.
> Inform TTM to allocate pages without __GFP_ZERO to avoid double page
> clearing by clearing out TTM_TT_FLAG_ZERO_ALLOC flag and set
> TTM_TT_FLAG_CLEARED_ON_FREE while freeing to skip ttm pool's
> clearn-on-free as XE now takes care of clearing pages. If a bo
> is in system placement and there is a cpu map then for such BO gpu
> clear will be avoided as there is no dma mapping for such BO at
> that moment.
> 
> To test the patch, created a small test that tries to submit a job
> after binding various sizes of buffer which shows good gains for larger
> buffer. For lower buffer sizes, the result is not very reliable as the
> results vary a lot for both CPU and GPU page clearing.
> 

I could see at smaller BO sizes the CPU being faster and results varying
a bit, certainly once BOs are larger I'd expect to see reliable perf
improvements. See my comment below on with / without as that info might
be helpful.

> With the patch
> sudo  ~/igt-gpu-tools/build/tests/xe_exec_store --run
> basic-store-benchmark
> IGT-Version: 1.28-g2ed908c0b (x86_64) (Linux: 6.10.0-rc2-xe+ x86_64)
> Using IGT_SRANDOM=1719237905 for randomisation
> Opened device: /dev/dri/card0
> Starting subtest: basic-store-benchmark
> Starting dynamic subtest: WC
> Dynamic subtest WC: SUCCESS (0.000s)
> Time taken for size SZ_4K: 9493 us
> Time taken for size SZ_2M: 5503 us
> Time taken for size SZ_64M: 13016 us
> Time taken for size SZ_128M: 29464 us
> Time taken for size SZ_256M: 38408 us
> Time taken for size SZ_1G: 148758 us
> Starting dynamic subtest: WB
> Dynamic subtest WB: SUCCESS (0.000s)
> Time taken for size SZ_4K: 3889 us
> Time taken for size SZ_2M: 6091 us
> Time taken for size SZ_64M: 20920 us
> Time taken for size SZ_128M: 32394 us
> Time taken for size SZ_256M: 61710 us
> Time taken for size SZ_1G: 215437 us
> Subtest basic-store-benchmark: SUCCESS (0.589s)
> 
> With the patch:

Both results 'with the patch'? What does it look like without?

> sudo  ~/igt-gpu-tools/build/tests/xe_exec_store --run
> basic-store-benchmark
> IGT-Version: 1.28-g2ed908c0b (x86_64) (Linux: 6.10.0-rc2-xe+ x86_64)
> Using IGT_SRANDOM=1719238062 for randomisation
> Opened device: /dev/dri/card0
> Starting subtest: basic-store-benchmark
> Starting dynamic subtest: WC
> Dynamic subtest WC: SUCCESS (0.000s)
> Time taken for size SZ_4K: 11803 us
> Time taken for size SZ_2M: 4237 us
> Time taken for size SZ_64M: 8649 us
> Time taken for size SZ_128M: 14682 us
> Time taken for size SZ_256M: 22156 us
> Time taken for size SZ_1G: 74457 us
> Starting dynamic subtest: WB
> Dynamic subtest WB: SUCCESS (0.000s)
> Time taken for size SZ_4K: 5129 us
> Time taken for size SZ_2M: 12563 us
> Time taken for size SZ_64M: 14860 us
> Time taken for size SZ_128M: 26064 us
> Time taken for size SZ_256M: 47167 us
> Time taken for size SZ_1G: 170304 us
> Subtest basic-store-benchmark: SUCCESS (0.417s)
> 
> With the patch and init_on_alloc=0

Can you explain what init_on_alloc means in this context? A GPU clear is
skipped?

> sudo  ~/igt-gpu-tools/build/tests/xe_exec_store --run
> basic-store-benchmark
> IGT-Version: 1.28-g2ed908c0b (x86_64) (Linux: 6.10.0-rc2-xe+ x86_64)
> Using IGT_SRANDOM=1719238219 for randomisation
> Opened device: /dev/dri/card0
> Starting subtest: basic-store-benchmark
> Starting dynamic subtest: WC
> Dynamic subtest WC: SUCCESS (0.000s)
> Time taken for size SZ_4K: 4803 us
> Time taken for size SZ_2M: 9212 us
> Time taken for size SZ_64M: 9643 us
> Time taken for size SZ_128M: 13479 us
> Time taken for size SZ_256M: 22429 us
> Time taken for size SZ_1G: 83110 us
> Starting dynamic subtest: WB
> Dynamic subtest WB: SUCCESS (0.000s)
> Time taken for size SZ_4K: 4003 us
> Time taken for size SZ_2M: 4443 us
> Time taken for size SZ_64M: 12960 us
> Time taken for size SZ_128M: 13741 us
> Time taken for size SZ_256M: 26841 us
> Time taken for size SZ_1G: 84746 us
> Subtest basic-store-benchmark: SUCCESS (0.290s)
> 
> v2: Handle regression on dgfx(Himal)
>     Update commit message as no ttm API changes needed.
> v3: Fix Kunit test.
> v4: handle data leak on cpu mmap(Thomas)
> v5: s/gpu_page_clear/gpu_page_clear_sys and move setting
>     it to xe_ttm_sys_mgr_init() and other nits (Matt Auld)
> 
> Cc: Himal Prasad Ghimiray <himal.prasad.ghimi...@intel.com>
> Cc: Matthew Auld <matthew.a...@intel.com>
> Cc: Matthew Brost <matthew.br...@intel.com>
> Cc: "Thomas Hellström" <thomas.hellst...@linux.intel.com>
> Signed-off-by: Nirmoy Das <nirmoy....@intel.com>
> ---
>  drivers/gpu/drm/xe/xe_bo.c           | 26 ++++++++++++++++++++++++--
>  drivers/gpu/drm/xe/xe_device_types.h |  2 ++
>  drivers/gpu/drm/xe/xe_ttm_sys_mgr.c  |  8 ++++++++
>  3 files changed, 34 insertions(+), 2 deletions(-)
> 
> diff --git a/drivers/gpu/drm/xe/xe_bo.c b/drivers/gpu/drm/xe/xe_bo.c
> index 0b7202dd4c9e..668c4c0f5ff0 100644
> --- a/drivers/gpu/drm/xe/xe_bo.c
> +++ b/drivers/gpu/drm/xe/xe_bo.c
> @@ -396,6 +396,14 @@ static struct ttm_tt *xe_ttm_tt_create(struct 
> ttm_buffer_object *ttm_bo,
>               caching = ttm_uncached;
>       }
>  
> +     /*
> +      * If the device can support gpu clear system pages then set proper ttm
> +      * flag. Zeroed pages are only required for ttm_bo_type_device so
> +      * unwanted data is not leaked to userspace.
> +      */
> +     if (ttm_bo->type == ttm_bo_type_device && xe->mem.gpu_page_clear_sys)
> +             page_flags |= TTM_TT_FLAG_CLEARED_ON_FREE;
> +
>       err = ttm_tt_init(&tt->ttm, &bo->ttm, page_flags, caching, extra_pages);
>       if (err) {
>               kfree(tt);
> @@ -417,6 +425,10 @@ static int xe_ttm_tt_populate(struct ttm_device 
> *ttm_dev, struct ttm_tt *tt,
>       if (tt->page_flags & TTM_TT_FLAG_EXTERNAL)
>               return 0;
>  
> +     /* Clear TTM_TT_FLAG_ZERO_ALLOC when GPU is set to clear system pages */
> +     if (tt->page_flags & TTM_TT_FLAG_CLEARED_ON_FREE)
> +             tt->page_flags &= ~TTM_TT_FLAG_ZERO_ALLOC;
> +
>       err = ttm_pool_alloc(&ttm_dev->pool, tt, ctx);
>       if (err)
>               return err;
> @@ -659,8 +671,16 @@ static int xe_bo_move(struct ttm_buffer_object *ttm_bo, 
> bool evict,
>       bool needs_clear;
>       bool handle_system_ccs = (!IS_DGFX(xe) && xe_bo_needs_ccs_pages(bo) &&
>                                 ttm && ttm_tt_is_populated(ttm)) ? true : 
> false;
> +     bool clear_system_pages;
>       int ret = 0;
>  
> +     /*
> +      * Clear TTM_TT_FLAG_CLEARED_ON_FREE on bo creation path when
> +      * moving to system as the bo doesn't have dma_mapping.
> +      */
> +     if (!old_mem && ttm && !ttm_tt_is_populated(ttm))
> +             ttm->page_flags &= ~TTM_TT_FLAG_CLEARED_ON_FREE;
> +
>       /* Bo creation path, moving to system or TT. */
>       if ((!old_mem && ttm) && !handle_system_ccs) {
>               if (new_mem->mem_type == XE_PL_TT)
> @@ -683,8 +703,10 @@ static int xe_bo_move(struct ttm_buffer_object *ttm_bo, 
> bool evict,
>       move_lacks_source = handle_system_ccs ? (!bo->ccs_cleared)  :
>                                               
> (!mem_type_is_vram(old_mem_type) && !tt_has_data);
>  
> +     clear_system_pages = ttm && (ttm->page_flags & 
> TTM_TT_FLAG_CLEARED_ON_FREE);
>       needs_clear = (ttm && ttm->page_flags & TTM_TT_FLAG_ZERO_ALLOC) ||
> -             (!ttm && ttm_bo->type == ttm_bo_type_device);
> +             (!ttm && ttm_bo->type == ttm_bo_type_device) ||
> +             clear_system_pages;
>  
>       if (new_mem->mem_type == XE_PL_TT) {
>               ret = xe_tt_map_sg(ttm);
> @@ -796,7 +818,7 @@ static int xe_bo_move(struct ttm_buffer_object *ttm_bo, 
> bool evict,
>               if (move_lacks_source) {
>                       u32 flags = 0;
>  
> -                     if (mem_type_is_vram(new_mem->mem_type))
> +                     if (mem_type_is_vram(new_mem->mem_type) || 
> clear_system_pages)
>                               flags |= XE_MIGRATE_CLEAR_FLAG_FULL;
>                       else if (handle_system_ccs)
>                               flags |= XE_MIGRATE_CLEAR_FLAG_CCS_DATA;
> diff --git a/drivers/gpu/drm/xe/xe_device_types.h 
> b/drivers/gpu/drm/xe/xe_device_types.h
> index 36252d5b1663..203ad50c894e 100644
> --- a/drivers/gpu/drm/xe/xe_device_types.h
> +++ b/drivers/gpu/drm/xe/xe_device_types.h
> @@ -332,6 +332,8 @@ struct xe_device {
>               struct xe_mem_region vram;
>               /** @mem.sys_mgr: system TTM manager */
>               struct ttm_resource_manager sys_mgr;
> +             /** @mem.gpu_page_clear_sys: clear system pages offloaded to 
> GPU */
> +             bool gpu_page_clear_sys;
>       } mem;
>  
>       /** @sriov: device level virtualization data */
> diff --git a/drivers/gpu/drm/xe/xe_ttm_sys_mgr.c 
> b/drivers/gpu/drm/xe/xe_ttm_sys_mgr.c
> index 9844a8edbfe1..fde0512a838a 100644
> --- a/drivers/gpu/drm/xe/xe_ttm_sys_mgr.c
> +++ b/drivers/gpu/drm/xe/xe_ttm_sys_mgr.c
> @@ -117,5 +117,13 @@ int xe_ttm_sys_mgr_init(struct xe_device *xe)
>       ttm_resource_manager_init(man, &xe->ttm, gtt_size >> PAGE_SHIFT);
>       ttm_set_driver_manager(&xe->ttm, XE_PL_TT, man);
>       ttm_resource_manager_set_used(man, true);
> +
> +     /*
> +      * On iGFX device with flat CCS, we clear CCS metadata, let's extend 
> that
> +      * and use GPU to clear pages as well.
> +      */
> +     if (xe_device_has_flat_ccs(xe) && !IS_DGFX(xe))
> +             xe->mem.gpu_page_clear_sys = true;

Seems like a xe->info might be better place with this being a platform
thing setup in xe_pci.c.

Matt

> +
>       return drmm_add_action_or_reset(&xe->drm, ttm_sys_mgr_fini, xe);
>  }
> -- 
> 2.42.0
> 

Reply via email to