On 28/06/2021 08:42, Boris Brezillon wrote:
> Now that we can pass our own workqueue to drm_sched_init(), we can use
> an ordered workqueue on for both the scheduler timeout tdr and our own
> reset work (which we use when the reset is not caused by a fault/timeout
> on a specific job, like when we have AS_ACTIVE bit stuck). This
> guarantees that the timeout handlers and reset handler can't run
> concurrently which drastically simplifies the locking.
> 
> v4:
> * Actually pass the reset workqueue to drm_sched_init()
> * Don't call cancel_work_sync() in panfrost_reset(). It will deadlock
>   since it might be called from the reset work, which is executing and
>   cancel_work_sync() will wait for the handler to return. Checking the
>   reset pending status should avoid spurious resets
> 
> v3:
> * New patch
> 
> Suggested-by: Daniel Vetter <daniel.vet...@ffwll.ch>
> Signed-off-by: Boris Brezillon <boris.brezil...@collabora.com>

While I'm very much not keen on the hack to drop out of the
dma_fence_signalling section - I don't believe this is introducing the
bug just making it more visible. What we really need is someone to find
time to actually fix up the lock ordering. But until then:

Reviewed-by: Steven Price <steven.pr...@arm.com>

> ---
>  drivers/gpu/drm/panfrost/panfrost_device.h |   6 +-
>  drivers/gpu/drm/panfrost/panfrost_job.c    | 187 ++++++++-------------
>  2 files changed, 72 insertions(+), 121 deletions(-)
> 
> diff --git a/drivers/gpu/drm/panfrost/panfrost_device.h 
> b/drivers/gpu/drm/panfrost/panfrost_device.h
> index f2190f90be75..59a487e8aba3 100644
> --- a/drivers/gpu/drm/panfrost/panfrost_device.h
> +++ b/drivers/gpu/drm/panfrost/panfrost_device.h
> @@ -108,6 +108,7 @@ struct panfrost_device {
>       struct mutex sched_lock;
>  
>       struct {
> +             struct workqueue_struct *wq;
>               struct work_struct work;
>               atomic_t pending;
>       } reset;
> @@ -246,9 +247,8 @@ const char *panfrost_exception_name(u32 exception_code);
>  static inline void
>  panfrost_device_schedule_reset(struct panfrost_device *pfdev)
>  {
> -     /* Schedule a reset if there's no reset in progress. */
> -     if (!atomic_xchg(&pfdev->reset.pending, 1))
> -             schedule_work(&pfdev->reset.work);
> +     atomic_set(&pfdev->reset.pending, 1);
> +     queue_work(pfdev->reset.wq, &pfdev->reset.work);
>  }
>  
>  #endif
> diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c 
> b/drivers/gpu/drm/panfrost/panfrost_job.c
> index e0c479e67304..98193a557a2d 100644
> --- a/drivers/gpu/drm/panfrost/panfrost_job.c
> +++ b/drivers/gpu/drm/panfrost/panfrost_job.c
> @@ -25,17 +25,8 @@
>  #define job_write(dev, reg, data) writel(data, dev->iomem + (reg))
>  #define job_read(dev, reg) readl(dev->iomem + (reg))
>  
> -enum panfrost_queue_status {
> -     PANFROST_QUEUE_STATUS_ACTIVE,
> -     PANFROST_QUEUE_STATUS_STOPPED,
> -     PANFROST_QUEUE_STATUS_STARTING,
> -     PANFROST_QUEUE_STATUS_FAULT_PENDING,
> -};
> -
>  struct panfrost_queue_state {
>       struct drm_gpu_scheduler sched;
> -     atomic_t status;
> -     struct mutex lock;
>       u64 fence_context;
>       u64 emit_seqno;
>  };
> @@ -379,57 +370,72 @@ void panfrost_job_enable_interrupts(struct 
> panfrost_device *pfdev)
>       job_write(pfdev, JOB_INT_MASK, irq_mask);
>  }
>  
> -static bool panfrost_scheduler_stop(struct panfrost_queue_state *queue,
> -                                 struct drm_sched_job *bad)
> +static void panfrost_reset(struct panfrost_device *pfdev,
> +                        struct drm_sched_job *bad)
>  {
> -     enum panfrost_queue_status old_status;
> -     bool stopped = false;
> +     unsigned int i;
> +     bool cookie;
>  
> -     mutex_lock(&queue->lock);
> -     old_status = atomic_xchg(&queue->status,
> -                              PANFROST_QUEUE_STATUS_STOPPED);
> -     if (old_status == PANFROST_QUEUE_STATUS_STOPPED)
> -             goto out;
> +     if (!atomic_read(&pfdev->reset.pending))
> +             return;
> +
> +     /* Stop the schedulers.
> +      *
> +      * FIXME: We temporarily get out of the dma_fence_signalling section
> +      * because the cleanup path generate lockdep splats when taking locks
> +      * to release job resources. We should rework the code to follow this
> +      * pattern:
> +      *
> +      *      try_lock
> +      *      if (locked)
> +      *              release
> +      *      else
> +      *              schedule_work_to_release_later
> +      */
> +     for (i = 0; i < NUM_JOB_SLOTS; i++)
> +             drm_sched_stop(&pfdev->js->queue[i].sched, bad);
> +
> +     cookie = dma_fence_begin_signalling();
>  
> -     WARN_ON(old_status != PANFROST_QUEUE_STATUS_ACTIVE);
> -     drm_sched_stop(&queue->sched, bad);
>       if (bad)
>               drm_sched_increase_karma(bad);
>  
> -     stopped = true;
> +     spin_lock(&pfdev->js->job_lock);
> +     for (i = 0; i < NUM_JOB_SLOTS; i++) {
> +             if (pfdev->jobs[i]) {
> +                     pm_runtime_put_noidle(pfdev->dev);
> +                     panfrost_devfreq_record_idle(&pfdev->pfdevfreq);
> +                     pfdev->jobs[i] = NULL;
> +             }
> +     }
> +     spin_unlock(&pfdev->js->job_lock);
>  
> -     /*
> -      * Set the timeout to max so the timer doesn't get started
> -      * when we return from the timeout handler (restored in
> -      * panfrost_scheduler_start()).
> +     panfrost_device_reset(pfdev);
> +
> +     /* GPU has been reset, we can cancel timeout/fault work that may have
> +      * been queued in the meantime and clear the reset pending bit.
>        */
> -     queue->sched.timeout = MAX_SCHEDULE_TIMEOUT;
> +     atomic_set(&pfdev->reset.pending, 0);
> +     for (i = 0; i < NUM_JOB_SLOTS; i++)
> +             cancel_delayed_work(&pfdev->js->queue[i].sched.work_tdr);
>  
> -out:
> -     mutex_unlock(&queue->lock);
>  
> -     return stopped;
> -}
> +     /* Now resubmit jobs that were previously queued but didn't have a
> +      * chance to finish.
> +      * FIXME: We temporarily get out of the DMA fence signalling section
> +      * while resubmitting jobs because the job submission logic will
> +      * allocate memory with the GFP_KERNEL flag which can trigger memory
> +      * reclaim and exposes a lock ordering issue.
> +      */
> +     dma_fence_end_signalling(cookie);
> +     for (i = 0; i < NUM_JOB_SLOTS; i++)
> +             drm_sched_resubmit_jobs(&pfdev->js->queue[i].sched);
> +     cookie = dma_fence_begin_signalling();
>  
> -static void panfrost_scheduler_start(struct panfrost_queue_state *queue)
> -{
> -     enum panfrost_queue_status old_status;
> +     for (i = 0; i < NUM_JOB_SLOTS; i++)
> +             drm_sched_start(&pfdev->js->queue[i].sched, true);
>  
> -     mutex_lock(&queue->lock);
> -     old_status = atomic_xchg(&queue->status,
> -                              PANFROST_QUEUE_STATUS_STARTING);
> -     WARN_ON(old_status != PANFROST_QUEUE_STATUS_STOPPED);
> -
> -     /* Restore the original timeout before starting the scheduler. */
> -     queue->sched.timeout = msecs_to_jiffies(JOB_TIMEOUT_MS);
> -     drm_sched_resubmit_jobs(&queue->sched);
> -     drm_sched_start(&queue->sched, true);
> -     old_status = atomic_xchg(&queue->status,
> -                              PANFROST_QUEUE_STATUS_ACTIVE);
> -     if (old_status == PANFROST_QUEUE_STATUS_FAULT_PENDING)
> -             drm_sched_fault(&queue->sched);
> -
> -     mutex_unlock(&queue->lock);
> +     dma_fence_end_signalling(cookie);
>  }
>  
>  static enum drm_gpu_sched_stat panfrost_job_timedout(struct drm_sched_job
> @@ -454,11 +460,8 @@ static enum drm_gpu_sched_stat 
> panfrost_job_timedout(struct drm_sched_job
>               job_read(pfdev, JS_TAIL_LO(js)),
>               sched_job);
>  
> -     /* Scheduler is already stopped, nothing to do. */
> -     if (!panfrost_scheduler_stop(&pfdev->js->queue[js], sched_job))
> -             return DRM_GPU_SCHED_STAT_NOMINAL;
> -
> -     panfrost_device_schedule_reset(pfdev);
> +     atomic_set(&pfdev->reset.pending, 1);
> +     panfrost_reset(pfdev, sched_job);
>  
>       return DRM_GPU_SCHED_STAT_NOMINAL;
>  }
> @@ -485,8 +488,6 @@ static void panfrost_job_handle_irq(struct 
> panfrost_device *pfdev, u32 status)
>               job_write(pfdev, JOB_INT_CLEAR, mask);
>  
>               if (status & JOB_INT_MASK_ERR(j)) {
> -                     enum panfrost_queue_status old_status;
> -
>                       job_write(pfdev, JS_COMMAND_NEXT(j), JS_COMMAND_NOP);
>  
>                       dev_err(pfdev->dev, "js fault, js=%d, status=%s, 
> head=0x%x, tail=0x%x",
> @@ -494,19 +495,7 @@ static void panfrost_job_handle_irq(struct 
> panfrost_device *pfdev, u32 status)
>                               panfrost_exception_name(job_read(pfdev, 
> JS_STATUS(j))),
>                               job_read(pfdev, JS_HEAD_LO(j)),
>                               job_read(pfdev, JS_TAIL_LO(j)));
> -
> -                     /*
> -                      * When the queue is being restarted we don't report
> -                      * faults directly to avoid races between the timeout
> -                      * and reset handlers. panfrost_scheduler_start() will
> -                      * call drm_sched_fault() after the queue has been
> -                      * started if status == FAULT_PENDING.
> -                      */
> -                     old_status = atomic_cmpxchg(&pfdev->js->queue[j].status,
> -                                                 
> PANFROST_QUEUE_STATUS_STARTING,
> -                                                 
> PANFROST_QUEUE_STATUS_FAULT_PENDING);
> -                     if (old_status == PANFROST_QUEUE_STATUS_ACTIVE)
> -                             drm_sched_fault(&pfdev->js->queue[j].sched);
> +                     drm_sched_fault(&pfdev->js->queue[j].sched);
>               }
>  
>               if (status & JOB_INT_MASK_DONE(j)) {
> @@ -561,56 +550,13 @@ static irqreturn_t panfrost_job_irq_handler(int irq, 
> void *data)
>       return IRQ_WAKE_THREAD;
>  }
>  
> -static void panfrost_reset(struct work_struct *work)
> +static void panfrost_reset_work(struct work_struct *work)
>  {
>       struct panfrost_device *pfdev = container_of(work,
>                                                    struct panfrost_device,
>                                                    reset.work);
> -     unsigned int i;
> -     bool cookie;
>  
> -     cookie = dma_fence_begin_signalling();
> -     for (i = 0; i < NUM_JOB_SLOTS; i++) {
> -             /*
> -              * We want pending timeouts to be handled before we attempt
> -              * to stop the scheduler. If we don't do that and the timeout
> -              * handler is in flight, it might have removed the bad job
> -              * from the list, and we'll lose this job if the reset handler
> -              * enters the critical section in panfrost_scheduler_stop()
> -              * before the timeout handler.
> -              *
> -              * Timeout is set to MAX_SCHEDULE_TIMEOUT - 1 because we need
> -              * something big enough to make sure the timer will not expire
> -              * before we manage to stop the scheduler, but we can't use
> -              * MAX_SCHEDULE_TIMEOUT because drm_sched_get_cleanup_job()
> -              * considers that as 'timer is not running' and will dequeue
> -              * the job without making sure the timeout handler is not
> -              * running.
> -              */
> -             pfdev->js->queue[i].sched.timeout = MAX_SCHEDULE_TIMEOUT - 1;
> -             cancel_delayed_work_sync(&pfdev->js->queue[i].sched.work_tdr);
> -             panfrost_scheduler_stop(&pfdev->js->queue[i], NULL);
> -     }
> -
> -     /* All timers have been stopped, we can safely reset the pending state. 
> */
> -     atomic_set(&pfdev->reset.pending, 0);
> -
> -     spin_lock(&pfdev->js->job_lock);
> -     for (i = 0; i < NUM_JOB_SLOTS; i++) {
> -             if (pfdev->jobs[i]) {
> -                     pm_runtime_put_noidle(pfdev->dev);
> -                     panfrost_devfreq_record_idle(&pfdev->pfdevfreq);
> -                     pfdev->jobs[i] = NULL;
> -             }
> -     }
> -     spin_unlock(&pfdev->js->job_lock);
> -
> -     panfrost_device_reset(pfdev);
> -
> -     for (i = 0; i < NUM_JOB_SLOTS; i++)
> -             panfrost_scheduler_start(&pfdev->js->queue[i]);
> -
> -     dma_fence_end_signalling(cookie);
> +     panfrost_reset(pfdev, NULL);
>  }
>  
>  int panfrost_job_init(struct panfrost_device *pfdev)
> @@ -618,7 +564,7 @@ int panfrost_job_init(struct panfrost_device *pfdev)
>       struct panfrost_job_slot *js;
>       int ret, j, irq;
>  
> -     INIT_WORK(&pfdev->reset.work, panfrost_reset);
> +     INIT_WORK(&pfdev->reset.work, panfrost_reset_work);
>  
>       pfdev->js = js = devm_kzalloc(pfdev->dev, sizeof(*js), GFP_KERNEL);
>       if (!js)
> @@ -640,15 +586,18 @@ int panfrost_job_init(struct panfrost_device *pfdev)
>               return ret;
>       }
>  
> -     for (j = 0; j < NUM_JOB_SLOTS; j++) {
> -             mutex_init(&js->queue[j].lock);
> +     pfdev->reset.wq = alloc_ordered_workqueue("panfrost-reset", 0);
> +     if (!pfdev->reset.wq)
> +             return -ENOMEM;
>  
> +     for (j = 0; j < NUM_JOB_SLOTS; j++) {
>               js->queue[j].fence_context = dma_fence_context_alloc(1);
>  
>               ret = drm_sched_init(&js->queue[j].sched,
>                                    &panfrost_sched_ops,
>                                    1, 0,
> -                                  msecs_to_jiffies(JOB_TIMEOUT_MS), NULL,
> +                                  msecs_to_jiffies(JOB_TIMEOUT_MS),
> +                                  pfdev->reset.wq,
>                                    NULL, "pan_js");
>               if (ret) {
>                       dev_err(pfdev->dev, "Failed to create scheduler: %d.", 
> ret);
> @@ -664,6 +613,7 @@ int panfrost_job_init(struct panfrost_device *pfdev)
>       for (j--; j >= 0; j--)
>               drm_sched_fini(&js->queue[j].sched);
>  
> +     destroy_workqueue(pfdev->reset.wq);
>       return ret;
>  }
>  
> @@ -676,9 +626,10 @@ void panfrost_job_fini(struct panfrost_device *pfdev)
>  
>       for (j = 0; j < NUM_JOB_SLOTS; j++) {
>               drm_sched_fini(&js->queue[j].sched);
> -             mutex_destroy(&js->queue[j].lock);
>       }
>  
> +     cancel_work_sync(&pfdev->reset.work);
> +     destroy_workqueue(pfdev->reset.wq);
>  }
>  
>  int panfrost_job_open(struct panfrost_file_priv *panfrost_priv)
> 

Reply via email to