Chris Wilson <ch...@chris-wilson.co.uk> writes:

> Since we process schedule-in of a context after submitting the request,
> if we decide to reset the context at that time, we also have to cancel
> the requets we have marked for submission.
>
> Signed-off-by: Chris Wilson <ch...@chris-wilson.co.uk>

Reviewed-by: Mika Kuoppala <mika.kuopp...@linux.intel.com>

> ---
>  .../drm/i915/gt/intel_execlists_submission.c  | 22 ++++++++++++++-----
>  drivers/gpu/drm/i915/i915_request.c           |  2 ++
>  2 files changed, 18 insertions(+), 6 deletions(-)
>
> diff --git a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c 
> b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> index b79365b5159a..18b23a332835 100644
> --- a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> +++ b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
> @@ -215,22 +215,32 @@ static void mark_eio(struct i915_request *rq)
>  }
>  
>  static struct i915_request *
> -active_request(const struct intel_timeline * const tl, struct i915_request 
> *rq)
> +__active_request(const struct intel_timeline * const tl,
> +              struct i915_request *rq,
> +              int error)
>  {
>       struct i915_request *active = rq;
>  
> -     rcu_read_lock();
> -     list_for_each_entry_continue_reverse(rq, &tl->requests, link) {
> +     list_for_each_entry_from_reverse(rq, &tl->requests, link) {
>               if (__i915_request_is_complete(rq))
>                       break;
>  
> +             if (error) {
> +                     i915_request_set_error_once(rq, error);
> +                     __i915_request_skip(rq);
> +             }
>               active = rq;
>       }
> -     rcu_read_unlock();
>  
>       return active;
>  }
>  
> +static struct i915_request *
> +active_request(const struct intel_timeline * const tl, struct i915_request 
> *rq)
> +{
> +     return __active_request(tl, rq, 0);
> +}
> +
>  static inline void
>  ring_set_paused(const struct intel_engine_cs *engine, int state)
>  {
> @@ -487,14 +497,14 @@ static void reset_active(struct i915_request *rq,
>        * remain correctly ordered. And we defer to __i915_request_submit()
>        * so that all asynchronous waits are correctly handled.
>        */
> -     ENGINE_TRACE(engine, "{ rq=%llx:%lld }\n",
> +     ENGINE_TRACE(engine, "{ reset rq=%llx:%lld }\n",
>                    rq->fence.context, rq->fence.seqno);
>  
>       /* On resubmission of the active request, payload will be scrubbed */
>       if (__i915_request_is_complete(rq))
>               head = rq->tail;
>       else
> -             head = active_request(ce->timeline, rq)->head;
> +             head = __active_request(ce->timeline, rq, -EIO)->head;
>       head = intel_ring_wrap(ce->ring, head);
>  
>       /* Scrub the context image to prevent replaying the previous batch */
> diff --git a/drivers/gpu/drm/i915/i915_request.c 
> b/drivers/gpu/drm/i915/i915_request.c
> index de434697dccd..03ac6eead4db 100644
> --- a/drivers/gpu/drm/i915/i915_request.c
> +++ b/drivers/gpu/drm/i915/i915_request.c
> @@ -490,6 +490,8 @@ void __i915_request_skip(struct i915_request *rq)
>       if (rq->infix == rq->postfix)
>               return;
>  
> +     RQ_TRACE(rq, "error: %d\n", rq->fence.error);
> +
>       /*
>        * As this request likely depends on state from the lost
>        * context, clear out all the user operations leaving the
> -- 
> 2.20.1
>
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx@lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/intel-gfx
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

Reply via email to