On 3/31/2025 6:31 PM, Christian König wrote:
This reverts commit c2cc3648ba517a6c270500b5447d5a1efdad5936.

Turned out that this has some negative consequences for some workloads.
Instead check if the cleaner shader should run directly.

Signed-off-by: Christian König <christian.koe...@amd.com>
---
  drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c | 4 ++--
  drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 2 +-
  2 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
index 802743efa3b3..5eab1c1a380c 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ib.c
@@ -191,8 +191,8 @@ int amdgpu_ib_schedule(struct amdgpu_ring *ring, unsigned 
int num_ibs,
        need_ctx_switch = ring->current_ctx != fence_ctx;
        if (ring->funcs->emit_pipeline_sync && job &&
            ((tmp = amdgpu_sync_get_fence(&job->explicit_sync)) ||
-            need_ctx_switch || amdgpu_vm_need_pipeline_sync(ring, job))) {
-
+            (amdgpu_sriov_vf(adev) && need_ctx_switch) ||
Should we need to, do this context switch, only on SRIOV cases "amdgpu_sriov_vf(adev)" or even in normal BM use cases also?
+            amdgpu_vm_need_pipeline_sync(ring, job))) {
                need_pipe_sync = true;
if (tmp)
If yes, could we split this patch into two 1. Actuall revert 2. below part is new changes?
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 
b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
index b5ddfcbbc9fc..5f0f9e4beea9 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
@@ -689,7 +689,7 @@ int amdgpu_vm_flush(struct amdgpu_ring *ring, struct 
amdgpu_job *job,
                patch = amdgpu_ring_init_cond_exec(ring,
                                                   ring->cond_exe_gpu_addr);
- if (need_pipe_sync)
+       if (need_pipe_sync || cleaner_shader_needed)

Here now, this pipe line synchronization was usually meant for GPU jobs? and not for client level switching? may I kno please, why it was OR'ed  for even "cleaner_shader_needed"? Is that do we have any usecases like where we don't need pipeline sync in between jobs but we need to emit pipeline sync only when "cleaner_shader_needed"  (ie., wrt new enforce_isolation feature)? - but even though in this "new_enforce_isolation feature" case - we would be skipping the GPU jobs level pipe line synchronization within a client? and do we forsee any synchronization/disruption issues in between jobs within a same client wrt new enforce_ioslation feature?

Best regards,

Srini

                amdgpu_ring_emit_pipeline_sync(ring);
if (cleaner_shader_needed)

Reply via email to