On 06/02/2025 09:51, Philipp Stanner wrote:
On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote:
Implement a mock scheduler backend and add some basic test to
exercise the
core scheduler code paths.

Mock backend (kind of like a very simple mock GPU) can either process
jobs
by tests manually advancing the "timeline" job at a time, or
alternatively
jobs can be configured with a time duration in which case they get
completed asynchronously from the unit test code.

Core scheduler classes are subclassed to support this mock
implementation.

The tests added are just a few simple submission patterns.

Signed-off-by: Tvrtko Ursulin <tvrtko.ursu...@igalia.com>

Could add a Suggested-by: Philipp :)

Will do.

Cc: Christian König <christian.koe...@amd.com>
Cc: Danilo Krummrich <d...@kernel.org>
Cc: Matthew Brost <matthew.br...@intel.com>
Cc: Philipp Stanner <pha...@kernel.org>
---
  drivers/gpu/drm/Kconfig.debug                 |  12 +
  drivers/gpu/drm/scheduler/.kunitconfig        |  12 +
  drivers/gpu/drm/scheduler/Makefile            |   1 +
  drivers/gpu/drm/scheduler/tests/Makefile      |   4 +
  .../gpu/drm/scheduler/tests/drm_mock_entity.c |  29 ++
  .../gpu/drm/scheduler/tests/drm_mock_job.c    |   3 +
  .../drm/scheduler/tests/drm_mock_scheduler.c  | 254
++++++++++++++++++
  .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++
  .../scheduler/tests/drm_sched_tests_basic.c   | 247
+++++++++++++++++
  9 files changed, 686 insertions(+)
  create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig
  create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile
  create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
  create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c
  create mode 100644
drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
  create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
  create mode 100644
drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c

diff --git a/drivers/gpu/drm/Kconfig.debug
b/drivers/gpu/drm/Kconfig.debug
index a35d74171b7b..89f777f21e95 100644
--- a/drivers/gpu/drm/Kconfig.debug
+++ b/drivers/gpu/drm/Kconfig.debug
@@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST
    If in doubt, say "N". +config DRM_SCHED_KUNIT_TEST
+       tristate "KUnit tests for the DRM scheduler" if
!KUNIT_ALL_TESTS
+       select DRM_SCHED
+       depends on DRM && KUNIT
+       default KUNIT_ALL_TESTS
+       help
+         Choose this option to build unit tests for the DRM
scheduler.

nit: Might say "DRM GPU scheduler" so readers not familiar with all
that get a better idea of what it's about

+
+         Recommended for driver developers only.

nit: s/driver developers/DRM developers
?

+
+         If in doubt, say "N".
+
  config DRM_EXPORT_FOR_TESTS
        bool
diff --git a/drivers/gpu/drm/scheduler/.kunitconfig
b/drivers/gpu/drm/scheduler/.kunitconfig
new file mode 100644
index 000000000000..cece53609fcf
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/.kunitconfig
@@ -0,0 +1,12 @@
+CONFIG_KUNIT=y
+CONFIG_DRM=y
+CONFIG_DRM_SCHED_KUNIT_TEST=y
+CONFIG_EXPERT=y
+CONFIG_DEBUG_SPINLOCK=y
+CONFIG_DEBUG_MUTEXES=y
+CONFIG_DEBUG_ATOMIC_SLEEP=y
+CONFIG_LOCK_DEBUGGING_SUPPORT=y
+CONFIG_PROVE_LOCKING=y
+CONFIG_LOCKDEP=y
+CONFIG_DEBUG_LOCKDEP=y
+CONFIG_DEBUG_LIST=y
diff --git a/drivers/gpu/drm/scheduler/Makefile
b/drivers/gpu/drm/scheduler/Makefile
index 53863621829f..46dfdca0758a 100644
--- a/drivers/gpu/drm/scheduler/Makefile
+++ b/drivers/gpu/drm/scheduler/Makefile
@@ -23,3 +23,4 @@
  gpu-sched-y := sched_main.o sched_fence.o sched_entity.o
 obj-$(CONFIG_DRM_SCHED) += gpu-sched.o
+obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/
diff --git a/drivers/gpu/drm/scheduler/tests/Makefile
b/drivers/gpu/drm/scheduler/tests/Makefile
new file mode 100644
index 000000000000..d69eab6a2e9b
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/Makefile
@@ -0,0 +1,4 @@
+
+obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \
+        drm_mock_scheduler.o \
+        drm_sched_tests_basic.o
diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
new file mode 100644
index 000000000000..c9205f9ff524
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
@@ -0,0 +1,29 @@
+
+#include "drm_sched_tests.h"
+
+struct drm_mock_sched_entity *
+drm_mock_sched_entity_new(struct kunit *test,
+                         enum drm_sched_priority priority,
+                         struct drm_mock_scheduler *sched)


I personally do like this function head style and


+{
+       struct drm_sched_mock_entity *entity;
+       int ret;
+
+       entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL);
+       KUNIT_ASSERT_NOT_NULL(test, entity);
+
+       ret = drm_sched_entity_init(&entity->base,
+                                   priority,
+                                   &sched->base, 1,
+                                   NULL);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       entity->test = test;
+
+       return entity;
+}
+
+void drm_mock_sched_entity_free(struct drm_mock_sched_entity
*entity)

do believe it should then be used consistently everywhere, regardless
of length.


+{
+       drm_sched_entity_fini(&entity->base);
+}
diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
new file mode 100644
index 000000000000..d94568cf3da9
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
@@ -0,0 +1,3 @@
+
+#include "drm_sched_tests.h"
+
diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
new file mode 100644
index 000000000000..f1985900a6ba
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
@@ -0,0 +1,254 @@
+
+#include "drm_sched_tests.h"
+
+struct drm_mock_sched_entity *
+drm_mock_new_sched_entity(struct kunit *test,
+                         enum drm_sched_priority priority,
+                         struct drm_mock_scheduler *sched)
+{
+       struct drm_mock_sched_entity *entity;
+       struct drm_gpu_scheduler *drm_sched;
+       int ret;
+
+       entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL);
+       KUNIT_ASSERT_NOT_NULL(test, entity);
+
+       drm_sched = &sched->base;
+       ret = drm_sched_entity_init(&entity->base,
+                                   priority,
+                                   &drm_sched, 1,
+                                   NULL);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       entity->test = test;
+
+       return entity;
+}
+
+void drm_mock_sched_entity_free(struct drm_mock_sched_entity
*entity)
+{
+       drm_sched_entity_destroy(&entity->base);
+}
+
+static enum hrtimer_restart
+drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer)

I think we should get up with a consistent naming convention. Some
things are called drm_mock_sched_, some others drm_sched_mock_.
As far as I saw, drm_mock_* does not yet exist.
So do you want to introduce drm_mock as something generic for drm? or
just for drm/sched?

This one is literally the only when where I tranposed the two. But it is also local and I am not too bothered about naming conventions of a local functions. If it were to me I would favour brevity and emit the long drm_.. prefixes which IMO do not help readability. If anything, seeing something with a long drm_.. prefix can only be confusing since one can assume it is some sort of exported interface.

I will change this instance.

+{
+       struct drm_mock_sched_job *upto =
+               container_of(hrtimer, typeof(*upto), timer);
+       struct drm_mock_scheduler *sched =
+               drm_sched_to_mock_sched(upto->base.sched);
+       struct drm_mock_sched_job *job, *next;
+       ktime_t now = ktime_get();
+       unsigned long flags;
+       LIST_HEAD(signal);
+
+       spin_lock_irqsave(&sched->lock, flags);
+       list_for_each_entry_safe(job, next, &sched->job_list, link)
{
+               if (!job->duration_us)
+                       break;
+
+               if (ktime_before(now, job->finish_at))
+                       break;
+
+               list_move_tail(&job->link, &signal);
+               sched->seqno = job->hw_fence.seqno;
+       }
+       spin_unlock_irqrestore(&sched->lock, flags);
+
+       list_for_each_entry(job, &signal, link) {
+               dma_fence_signal(&job->hw_fence);
+               dma_fence_put(&job->hw_fence);
+       }
+
+       return HRTIMER_NORESTART;
+}
+
+struct drm_mock_sched_job *
+drm_mock_new_sched_job(struct kunit *test,
+                      struct drm_mock_sched_entity *entity)
+{
+       struct drm_mock_sched_job *job;
+       int ret;
+
+       job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL);
+       KUNIT_ASSERT_NOT_NULL(test, job);
+
+       ret = drm_sched_job_init(&job->base,
+                                &entity->base,
+                                1,
+                                NULL);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       job->test = test;
+
+       spin_lock_init(&job->lock);
+       INIT_LIST_HEAD(&job->link);
+       hrtimer_init(&job->timer, CLOCK_MONOTONIC,
HRTIMER_MODE_ABS);
+       job->timer.function = drm_sched_mock_job_signal_timer;
+
+       return job;
+}
+
+static const char *drm_mock_sched_hw_fence_driver_name(struct
dma_fence *fence)
+{
+       return "drm_mock_sched";
+}
+
+static const char *
+drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence)
+{
+       struct drm_mock_sched_job *job =
+               container_of(fence, typeof(*job), hw_fence);
+
+       return (const char *)job->base.sched->name;
+}
+static void drm_mock_sched_hw_fence_release(struct dma_fence *fence)
+{
+       struct drm_mock_sched_job *job =
+               container_of(fence, typeof(*job), hw_fence);
+
+       hrtimer_cancel(&job->timer);
+
+       /* Freed by the kunit framework */
+}
+
+static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = {
+       .get_driver_name = drm_mock_sched_hw_fence_driver_name,
+       .get_timeline_name = drm_mock_sched_hw_fence_timeline_name,
+       .release = drm_mock_sched_hw_fence_release,
+};
+
+static struct dma_fence *mock_sched_run_job(struct drm_sched_job
*sched_job)
+{
+       struct drm_mock_scheduler *sched =
+               drm_sched_to_mock_sched(sched_job->sched);
+       struct drm_mock_sched_job *job =
drm_sched_job_to_mock_job(sched_job);
+
+       dma_fence_init(&job->hw_fence,
+                      &drm_mock_sched_hw_fence_ops,
+                      &job->lock,
+                      sched->hw_fence.context,
+                      atomic_inc_return(&sched->hw_fence.seqno));
+
+       dma_fence_get(&job->hw_fence); /* Reference for the job_list
*/
+
+       spin_lock_irq(&sched->lock);
+       if (job->duration_us) {
+               ktime_t prev_finish_at = 0;
+
+               if (!list_empty(&sched->job_list)) {
+                       struct drm_mock_sched_job *prev =
+                               list_last_entry(&sched->job_list,
typeof(*prev),
+                                               link);
+
+                       prev_finish_at = prev->finish_at;
+               }
+
+               if (!prev_finish_at)
+                       prev_finish_at = ktime_get();
+
+               job->finish_at = ktime_add_us(prev_finish_at, job-
duration_us);
+       }
+       list_add_tail(&job->link, &sched->job_list);
+       if (job->finish_at)
+               hrtimer_start(&job->timer, job->finish_at,
HRTIMER_MODE_ABS);
+       spin_unlock_irq(&sched->lock);
+
+       return &job->hw_fence;
+}
+
+static enum drm_gpu_sched_stat
+mock_sched_timedout_job(struct drm_sched_job *sched_job)
+{
+       return DRM_GPU_SCHED_STAT_ENODEV;
+}
+
+static void mock_sched_free_job(struct drm_sched_job *sched_job)
+{
+       drm_sched_job_cleanup(sched_job);
+}
+
+static const struct drm_sched_backend_ops drm_mock_scheduler_ops = {
+       .run_job = mock_sched_run_job,
+       .timedout_job = mock_sched_timedout_job,
+       .free_job = mock_sched_free_job
+};
+
+struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
*test)
+{
+       struct drm_mock_scheduler *sched;
+       int ret;
+
+       sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL);
+       KUNIT_ASSERT_NOT_NULL(test, sched);
+
+       ret = drm_sched_init(&sched->base,
+                            &drm_mock_scheduler_ops,
+                            NULL, /* wq */
+                            DRM_SCHED_PRIORITY_COUNT,
+                            U32_MAX, /* max credits */
+                            UINT_MAX, /* hang limit */
+                            MAX_SCHEDULE_TIMEOUT, /* timeout */
+                            NULL, /* timeout wq */
+                            NULL, /* score */
+                            "drm-mock-scheduler",
+                            NULL /* dev */);
+       KUNIT_ASSERT_EQ(test, ret, 0);
+
+       sched->test = test;
+       sched->hw_fence.context = dma_fence_context_alloc(1);
+       atomic_set(&sched->hw_fence.seqno, 0);
+       INIT_LIST_HEAD(&sched->job_list);
+       spin_lock_init(&sched->lock);
+
+       return sched;
+}
+
+void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched)
+{
+       struct drm_mock_sched_job *job, *next;
+       unsigned long flags;
+       LIST_HEAD(signal);
+
+       spin_lock_irqsave(&sched->lock, flags);
+       list_for_each_entry_safe(job, next, &sched->job_list, link)
+               list_move_tail(&job->link, &signal);
+       spin_unlock_irqrestore(&sched->lock, flags);

So maybe you can help me get up to speed here a bit. What is the
purpose behind job->link? Is the general idea documented somewhere?

List versus link suffix convention I use to distinguish struct list_head which is a list versus struct list_head which is used to put something on the list.

In this case job->link is for the mock GPU driver to keep track of jobs which have been submitted to the "hardware" for "execution".

I will of course document these things once the high level design is settled.

+
+       list_for_each_entry(job, &signal, link) {
+               hrtimer_cancel(&job->timer);
+               dma_fence_put(&job->hw_fence);
+       }
+
+       drm_sched_fini(&sched->base);
+}
+
+unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
*sched,
+                                   unsigned int num)
+{
+       struct drm_mock_sched_job *job, *next;
+       unsigned int found = 0;
+       unsigned long flags;
+       LIST_HEAD(signal);
+
+       spin_lock_irqsave(&sched->lock, flags);
+       if (WARN_ON_ONCE(sched->seqno + num < sched->seqno))
+               goto unlock;
+       sched->seqno += num;
+       list_for_each_entry_safe(job, next, &sched->job_list, link)
{
+               if (sched->seqno < job->hw_fence.seqno)
+                       break;
+
+               list_move_tail(&job->link, &signal);
+               found++;
+       }
+unlock:
+       spin_unlock_irqrestore(&sched->lock, flags);
+
+       list_for_each_entry(job, &signal, link) {
+               dma_fence_signal(&job->hw_fence);
+               dma_fence_put(&job->hw_fence);
+       }
+
+       return found;
+}
diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
new file mode 100644
index 000000000000..421ee2712985
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
@@ -0,0 +1,124 @@
+
+#include <kunit/test.h>
+#include <linux/atomic.h>
+#include <linux/dma-fence.h>
+#include <linux/hrtimer.h>
+#include <linux/ktime.h>
+#include <linux/list.h>
+#include <linux/atomic.h>
+#include <linux/mutex.h>
+#include <linux/types.h>
+
+#include <drm/gpu_scheduler.h>
+
+struct drm_mock_scheduler {
+       struct drm_gpu_scheduler base;
+
+       struct kunit            *test;
+
+       struct {
+               u64             context;
+               atomic_t        seqno;
+       } hw_fence;

Hm, well, so this is confusing. drm_mock_sched_job below contains an
actual struct dma_fence that is also called hw_fence, whereas this here
seems to be some pseudo-hw_fence?

What is its purpose?

I believe we agreed that "Hardware fence" should always mean a fence
per job that is signaled by the hardware (driver interrupt) once the
job is done.

If this hw_fence here is the same, why is it per scheduler? That's
confusing.

Mock_job->hw_fence is what the mock GPU driver returns from the sched->run_job().

Mock_sched->hw_fence is what the mock GPU driver uses to track execution of jobs submitted to it for execution. If Irename this to hw_timeline will it make it clearer?

+
+       spinlock_t              lock;

Maybe document the lock's purpose


+       unsigned int            seqno;
+       struct list_head        job_list;
+};
+
+struct drm_mock_sched_entity {
+       struct drm_sched_entity base;
+
+       struct kunit            *test;
+};
+
+struct drm_mock_sched_job {
+       struct drm_sched_job    base;
+
+       struct list_head        link;
+       struct hrtimer          timer;
+
+       unsigned int            duration_us;
+       ktime_t                 finish_at;
+
+       spinlock_t              lock;

Same

+       struct dma_fence        hw_fence;
+
+       struct kunit            *test;
+};
+
+static inline struct drm_mock_scheduler *
+drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched)
+{
+       return container_of(sched, struct drm_mock_scheduler, base);
+};
+
+static inline struct drm_mock_sched_entity *
+drm_sched_entity_to_mock_entity(struct drm_sched_entity
*sched_entity)
+{
+       return container_of(sched_entity, struct
drm_mock_sched_entity, base);
+};
+
+static inline struct drm_mock_sched_job *
+drm_sched_job_to_mock_job(struct drm_sched_job *sched_job)
+{
+       return container_of(sched_job, struct drm_mock_sched_job,
base);
+};
+
+struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
*test);
+void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched);
+unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
*sched,
+                                   unsigned int num);
+
+struct drm_mock_sched_entity *
+drm_mock_new_sched_entity(struct kunit *test,
+                         enum drm_sched_priority priority,
+                         struct drm_mock_scheduler *sched);
+void drm_mock_sched_entity_free(struct drm_mock_sched_entity
*entity);
+
+struct drm_mock_sched_job *
+drm_mock_new_sched_job(struct kunit *test,
+                      struct drm_mock_sched_entity *entity);
+
+static inline void drm_mock_sched_job_submit(struct
drm_mock_sched_job *job)
+{
+       drm_sched_job_arm(&job->base);
+       drm_sched_entity_push_job(&job->base);
+}
+
+static inline void
+drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job,
+                                  unsigned int duration_us)
+{
+       job->duration_us = duration_us;
+}
+
+static inline bool
+drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job)
+{
+       return dma_fence_is_signaled(&job->base.s_fence->finished);
+}
+
+static inline bool
+drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job,
long timeout)
+{
+       long ret;
+
+       ret = dma_fence_wait_timeout(&job->base.s_fence->finished,
+                                     false,
+                                     timeout);
+
+       return ret != 0;
+}
+
+static inline long
+drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job,
long timeout)
+{
+       long ret;
+
+       ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled,
+                                     false,
+                                     timeout);
+
+       return ret != 0;
+}
diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
new file mode 100644
index 000000000000..6fd39bea95b1
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
@@ -0,0 +1,247 @@
+
+#include "drm_sched_tests.h"
+
+static int drm_sched_basic_init(struct kunit *test)
+{
+       test->priv = drm_mock_new_scheduler(test);
+
+       return 0;
+}
+
+static void drm_sched_basic_exit(struct kunit *test)
+{
+       struct drm_mock_scheduler *sched = test->priv;
+
+       drm_mock_scheduler_fini(sched);
+}
+
+static void drm_sched_basic_submit(struct kunit *test)
+{
+       struct drm_mock_scheduler *sched = test->priv;
+       struct drm_mock_sched_entity *entity;
+       struct drm_mock_sched_job *job;
+       unsigned int i;
+       bool done;
+
+       /*
+        * Submit one job to the scheduler and verify that it gets
scheduled
+        * and completed only when the mock hw backend processes it.
+        */
+
+       entity = drm_mock_new_sched_entity(test,
+                                       
DRM_SCHED_PRIORITY_NORMAL,
+                                          sched);
+       job = drm_mock_new_sched_job(test, entity);
+
+       drm_mock_sched_job_submit(job);
+
+       done = drm_mock_sched_job_wait_scheduled(job, HZ);
+       KUNIT_ASSERT_EQ(test, done, true);
+
+       done = drm_mock_sched_job_wait_finished(job, HZ / 2);
+       KUNIT_ASSERT_EQ(test, done, false);
+
+       i = drm_mock_sched_advance(sched, 1);
+       KUNIT_ASSERT_EQ(test, i, 1);
+
+       done = drm_mock_sched_job_wait_finished(job, HZ);
+       KUNIT_ASSERT_EQ(test, done, true);
+
+       drm_mock_sched_entity_free(entity);
+}
+
+static void drm_sched_basic_queue(struct kunit *test)
+{
+       struct drm_mock_scheduler *sched = test->priv;
+       struct drm_mock_sched_entity *entity;
+       struct drm_mock_sched_job *job;
+       const unsigned int qd = 100;

Not the best variable name – is this "queue depth"? Why is it 100? ->
global define & document

I wouldn't promote this to global and TBH if you look how small this test function is it feels pretty self documenting.



+       unsigned int i;
+       bool done;
+
+       /*
+        * Submit a queue of jobs on the same entity, let them be
completed by
+        * the mock hw backend and check the status of the last job.
+        */
+
+       entity = drm_mock_new_sched_entity(test,
+                                       
DRM_SCHED_PRIORITY_NORMAL,
+                                          sched);
+
+       for (i = 0; i < qd; i++) {
+               job = drm_mock_new_sched_job(test, entity);
+               drm_mock_sched_job_set_duration_us(job, 1000);
+               drm_mock_sched_job_submit(job);
+       }
+
+       done = drm_mock_sched_job_wait_finished(job, HZ);
+       KUNIT_ASSERT_EQ(test, done, true);
+
+       drm_mock_sched_entity_free(entity);
+}
+
+static void drm_sched_basic_chain(struct kunit *test)
+{
+       struct drm_mock_scheduler *sched = test->priv;
+       struct drm_mock_sched_job *job, *prev = NULL;
+       struct drm_mock_sched_entity *entity;
+       const unsigned int qd = 100;
+       unsigned int i;
+       bool done;
+
+       /*
+        * Submit a queue of jobs on the same entity but with an
explicit
+        * chain of dependencies between them. Let the jobs be
completed by
+        * the mock hw backend and check the status of the last job.
+        */
+
+       entity = drm_mock_new_sched_entity(test,
+                                       
DRM_SCHED_PRIORITY_NORMAL,
+                                          sched);
+
+       for (i = 0; i < qd; i++) {
+               job = drm_mock_new_sched_job(test, entity);
+               drm_mock_sched_job_set_duration_us(job, 1000);
+               if (prev)
+                       drm_sched_job_add_dependency(&job->base,
+                                               
dma_fence_get(&prev->base.s_fence->finished));
+               drm_mock_sched_job_submit(job);
+               prev = job;
+       }
+
+       done = drm_mock_sched_job_wait_finished(job, HZ);
+       KUNIT_ASSERT_EQ(test, done, true);
+
+       drm_mock_sched_entity_free(entity);
+}
+
+static void drm_sched_basic_entities(struct kunit *test)
+{
+       struct drm_mock_scheduler *sched = test->priv;
+       struct drm_mock_sched_entity *entity[4];
+       struct drm_mock_sched_job *job;
+       const unsigned int qd = 100;
+       unsigned int i, cur_ent = 0;
+       bool done;
+
+       /*
+        * Submit a queue of jobs across different entities, let
them be
+        * completed by the mock hw backend and check the status of
the last
+        * job.
+        */
+
+       for (i = 0; i < ARRAY_SIZE(entity); i++)
+               entity[i] = drm_mock_new_sched_entity(test,
+                                               
DRM_SCHED_PRIORITY_NORMAL,
+                                                     sched);
+
+       for (i = 0; i < qd; i++) {
+               job = drm_mock_new_sched_job(test,
entity[cur_ent++]);
+               cur_ent %= ARRAY_SIZE(entity);
+               drm_mock_sched_job_set_duration_us(job, 1000);
+               drm_mock_sched_job_submit(job);
+       }
+
+       done = drm_mock_sched_job_wait_finished(job, HZ);
+       KUNIT_ASSERT_EQ(test, done, true);
+
+       for (i = 0; i < ARRAY_SIZE(entity); i++)
+               drm_mock_sched_entity_free(entity[i]);
+}
+
+static void drm_sched_basic_entities_chain(struct kunit *test)
+{
+       struct drm_mock_scheduler *sched = test->priv;
+       struct drm_mock_sched_job *job, *prev = NULL;
+       struct drm_mock_sched_entity *entity[4];
+       const unsigned int qd = 100;
+       unsigned int i, cur_ent = 0;
+       bool done;
+
+       /*
+        * Submit a queue of jobs across different entities with an
explicit
+        * chain of dependencies between them. Let the jobs be
completed by
+        * the mock hw backend and check the status of the last job.
+        */
+
+       for (i = 0; i < ARRAY_SIZE(entity); i++)
+               entity[i] = drm_mock_new_sched_entity(test,
+                                               
DRM_SCHED_PRIORITY_NORMAL,
+                                                     sched);
+
+       for (i = 0; i < qd; i++) {
+               job = drm_mock_new_sched_job(test,
entity[cur_ent++]);
+               cur_ent %= ARRAY_SIZE(entity);
+               drm_mock_sched_job_set_duration_us(job, 1000);
+               if (prev)
+                       drm_sched_job_add_dependency(&job->base,
+                                               
dma_fence_get(&prev->base.s_fence->finished));
+               drm_mock_sched_job_submit(job);
+               prev = job;
+       }
+
+       done = drm_mock_sched_job_wait_finished(job, HZ);
+       KUNIT_ASSERT_EQ(test, done, true);
+
+       for (i = 0; i < ARRAY_SIZE(entity); i++)
+               drm_mock_sched_entity_free(entity[i]);
+}
+
+static void drm_sched_basic_entity_cleanup(struct kunit *test)
+{
+       struct drm_mock_sched_job *job, *mid, *prev = NULL;
+       struct drm_mock_scheduler *sched = test->priv;
+       struct drm_mock_sched_entity *entity[4];

4 is used in several places, so could be defined globally. In case
there's a special reason for why it's 4, that could be mentioned, too

It's completely arbitrary. In this test we just need a bunch of entities to be active on the scheduler as we trigger a client exit in the middle of it. IMO the comment few lines below should be good enough to explain the idea. I fear that promoting this to a global define would just give it more weight that what it has. Then some test would want a different number etc.

Regards,

Tvrtko

+       const unsigned int qd = 100;
+       unsigned int i, cur_ent = 0;
+       bool done;
+
+       /*
+        * Submit a queue of jobs across different entities with an
explicit
+        * chain of dependencies between them and trigger entity
cleanup while
+        * the queue is still being processed.
+        */
+
+       for (i = 0; i < ARRAY_SIZE(entity); i++)
+               entity[i] = drm_mock_new_sched_entity(test,
+                                               
DRM_SCHED_PRIORITY_NORMAL,
+                                                     sched);
+
+       for (i = 0; i < qd; i++) {
+               job = drm_mock_new_sched_job(test,
entity[cur_ent++]);
+               cur_ent %= ARRAY_SIZE(entity);
+               drm_mock_sched_job_set_duration_us(job, 1000);
+               if (prev)
+                       drm_sched_job_add_dependency(&job->base,
+                                               
dma_fence_get(&prev->base.s_fence->finished));
+               drm_mock_sched_job_submit(job);
+               if (i == qd / 2)
+                       mid = job;
+               prev = job;
+       }
+
+       done = drm_mock_sched_job_wait_finished(mid, HZ);
+       KUNIT_ASSERT_EQ(test, done, true);
+
+       /* Exit with half of the queue still pending to be executed.
*/
+       for (i = 0; i < ARRAY_SIZE(entity); i++)
+               drm_mock_sched_entity_free(entity[i]);}
+
+static struct kunit_case drm_sched_basic_tests[] = {
+       KUNIT_CASE(drm_sched_basic_submit),
+       KUNIT_CASE(drm_sched_basic_queue),
+       KUNIT_CASE(drm_sched_basic_chain),
+       KUNIT_CASE(drm_sched_basic_entities),
+       KUNIT_CASE(drm_sched_basic_entities_chain),
+       KUNIT_CASE(drm_sched_basic_entity_cleanup),
+       {}
+};
+
+static struct kunit_suite drm_sched_basic = {
+       .name = "drm_sched_basic_tests",
+       .init = drm_sched_basic_init,
+       .exit = drm_sched_basic_exit,
+       .test_cases = drm_sched_basic_tests,
+};
+
+kunit_test_suite(drm_sched_basic);

Reply via email to