Message ID | 20250203153007.63400-3-tvrtko.ursulin@igalia.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | DRM scheduler kunit tests | expand |
On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote: > Implement a mock scheduler backend and add some basic test to > exercise the > core scheduler code paths. > > Mock backend (kind of like a very simple mock GPU) can either process > jobs > by tests manually advancing the "timeline" job at a time, or > alternatively > jobs can be configured with a time duration in which case they get > completed asynchronously from the unit test code. > > Core scheduler classes are subclassed to support this mock > implementation. > > The tests added are just a few simple submission patterns. > > Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com> Could add a Suggested-by: Philipp :) > Cc: Christian König <christian.koenig@amd.com> > Cc: Danilo Krummrich <dakr@kernel.org> > Cc: Matthew Brost <matthew.brost@intel.com> > Cc: Philipp Stanner <phasta@kernel.org> > --- > drivers/gpu/drm/Kconfig.debug | 12 + > drivers/gpu/drm/scheduler/.kunitconfig | 12 + > drivers/gpu/drm/scheduler/Makefile | 1 + > drivers/gpu/drm/scheduler/tests/Makefile | 4 + > .../gpu/drm/scheduler/tests/drm_mock_entity.c | 29 ++ > .../gpu/drm/scheduler/tests/drm_mock_job.c | 3 + > .../drm/scheduler/tests/drm_mock_scheduler.c | 254 > ++++++++++++++++++ > .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++ > .../scheduler/tests/drm_sched_tests_basic.c | 247 > +++++++++++++++++ > 9 files changed, 686 insertions(+) > create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig > create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile > create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c > create mode 100644 > drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > create mode 100644 > drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > > diff --git a/drivers/gpu/drm/Kconfig.debug > b/drivers/gpu/drm/Kconfig.debug > index a35d74171b7b..89f777f21e95 100644 > --- a/drivers/gpu/drm/Kconfig.debug > +++ b/drivers/gpu/drm/Kconfig.debug > @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST > > If in doubt, say "N". > > +config DRM_SCHED_KUNIT_TEST > + tristate "KUnit tests for the DRM scheduler" if > !KUNIT_ALL_TESTS > + select DRM_SCHED > + depends on DRM && KUNIT > + default KUNIT_ALL_TESTS > + help > + Choose this option to build unit tests for the DRM > scheduler. nit: Might say "DRM GPU scheduler" so readers not familiar with all that get a better idea of what it's about > + > + Recommended for driver developers only. nit: s/driver developers/DRM developers ? > + > + If in doubt, say "N". > + > config DRM_EXPORT_FOR_TESTS > bool > diff --git a/drivers/gpu/drm/scheduler/.kunitconfig > b/drivers/gpu/drm/scheduler/.kunitconfig > new file mode 100644 > index 000000000000..cece53609fcf > --- /dev/null > +++ b/drivers/gpu/drm/scheduler/.kunitconfig > @@ -0,0 +1,12 @@ > +CONFIG_KUNIT=y > +CONFIG_DRM=y > +CONFIG_DRM_SCHED_KUNIT_TEST=y > +CONFIG_EXPERT=y > +CONFIG_DEBUG_SPINLOCK=y > +CONFIG_DEBUG_MUTEXES=y > +CONFIG_DEBUG_ATOMIC_SLEEP=y > +CONFIG_LOCK_DEBUGGING_SUPPORT=y > +CONFIG_PROVE_LOCKING=y > +CONFIG_LOCKDEP=y > +CONFIG_DEBUG_LOCKDEP=y > +CONFIG_DEBUG_LIST=y > diff --git a/drivers/gpu/drm/scheduler/Makefile > b/drivers/gpu/drm/scheduler/Makefile > index 53863621829f..46dfdca0758a 100644 > --- a/drivers/gpu/drm/scheduler/Makefile > +++ b/drivers/gpu/drm/scheduler/Makefile > @@ -23,3 +23,4 @@ > gpu-sched-y := sched_main.o sched_fence.o sched_entity.o > > obj-$(CONFIG_DRM_SCHED) += gpu-sched.o > +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/ > diff --git a/drivers/gpu/drm/scheduler/tests/Makefile > b/drivers/gpu/drm/scheduler/tests/Makefile > new file mode 100644 > index 000000000000..d69eab6a2e9b > --- /dev/null > +++ b/drivers/gpu/drm/scheduler/tests/Makefile > @@ -0,0 +1,4 @@ > + > +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \ > + drm_mock_scheduler.o \ > + drm_sched_tests_basic.o > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > new file mode 100644 > index 000000000000..c9205f9ff524 > --- /dev/null > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > @@ -0,0 +1,29 @@ > + > +#include "drm_sched_tests.h" > + > +struct drm_mock_sched_entity * > +drm_mock_sched_entity_new(struct kunit *test, > + enum drm_sched_priority priority, > + struct drm_mock_scheduler *sched) I personally do like this function head style and > +{ > + struct drm_sched_mock_entity *entity; > + int ret; > + > + entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL); > + KUNIT_ASSERT_NOT_NULL(test, entity); > + > + ret = drm_sched_entity_init(&entity->base, > + priority, > + &sched->base, 1, > + NULL); > + KUNIT_ASSERT_EQ(test, ret, 0); > + > + entity->test = test; > + > + return entity; > +} > + > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity > *entity) do believe it should then be used consistently everywhere, regardless of length. > +{ > + drm_sched_entity_fini(&entity->base); > +} > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c > b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c > new file mode 100644 > index 000000000000..d94568cf3da9 > --- /dev/null > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c > @@ -0,0 +1,3 @@ > + > +#include "drm_sched_tests.h" > + > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > new file mode 100644 > index 000000000000..f1985900a6ba > --- /dev/null > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > @@ -0,0 +1,254 @@ > + > +#include "drm_sched_tests.h" > + > +struct drm_mock_sched_entity * > +drm_mock_new_sched_entity(struct kunit *test, > + enum drm_sched_priority priority, > + struct drm_mock_scheduler *sched) > +{ > + struct drm_mock_sched_entity *entity; > + struct drm_gpu_scheduler *drm_sched; > + int ret; > + > + entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL); > + KUNIT_ASSERT_NOT_NULL(test, entity); > + > + drm_sched = &sched->base; > + ret = drm_sched_entity_init(&entity->base, > + priority, > + &drm_sched, 1, > + NULL); > + KUNIT_ASSERT_EQ(test, ret, 0); > + > + entity->test = test; > + > + return entity; > +} > + > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity > *entity) > +{ > + drm_sched_entity_destroy(&entity->base); > +} > + > +static enum hrtimer_restart > +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer) I think we should get up with a consistent naming convention. Some things are called drm_mock_sched_, some others drm_sched_mock_. As far as I saw, drm_mock_* does not yet exist. So do you want to introduce drm_mock as something generic for drm? or just for drm/sched? > +{ > + struct drm_mock_sched_job *upto = > + container_of(hrtimer, typeof(*upto), timer); > + struct drm_mock_scheduler *sched = > + drm_sched_to_mock_sched(upto->base.sched); > + struct drm_mock_sched_job *job, *next; > + ktime_t now = ktime_get(); > + unsigned long flags; > + LIST_HEAD(signal); > + > + spin_lock_irqsave(&sched->lock, flags); > + list_for_each_entry_safe(job, next, &sched->job_list, link) > { > + if (!job->duration_us) > + break; > + > + if (ktime_before(now, job->finish_at)) > + break; > + > + list_move_tail(&job->link, &signal); > + sched->seqno = job->hw_fence.seqno; > + } > + spin_unlock_irqrestore(&sched->lock, flags); > + > + list_for_each_entry(job, &signal, link) { > + dma_fence_signal(&job->hw_fence); > + dma_fence_put(&job->hw_fence); > + } > + > + return HRTIMER_NORESTART; > +} > + > +struct drm_mock_sched_job * > +drm_mock_new_sched_job(struct kunit *test, > + struct drm_mock_sched_entity *entity) > +{ > + struct drm_mock_sched_job *job; > + int ret; > + > + job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL); > + KUNIT_ASSERT_NOT_NULL(test, job); > + > + ret = drm_sched_job_init(&job->base, > + &entity->base, > + 1, > + NULL); > + KUNIT_ASSERT_EQ(test, ret, 0); > + > + job->test = test; > + > + spin_lock_init(&job->lock); > + INIT_LIST_HEAD(&job->link); > + hrtimer_init(&job->timer, CLOCK_MONOTONIC, > HRTIMER_MODE_ABS); > + job->timer.function = drm_sched_mock_job_signal_timer; > + > + return job; > +} > + > +static const char *drm_mock_sched_hw_fence_driver_name(struct > dma_fence *fence) > +{ > + return "drm_mock_sched"; > +} > + > +static const char * > +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence) > +{ > + struct drm_mock_sched_job *job = > + container_of(fence, typeof(*job), hw_fence); > + > + return (const char *)job->base.sched->name; > +} > +static void drm_mock_sched_hw_fence_release(struct dma_fence *fence) > +{ > + struct drm_mock_sched_job *job = > + container_of(fence, typeof(*job), hw_fence); > + > + hrtimer_cancel(&job->timer); > + > + /* Freed by the kunit framework */ > +} > + > +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = { > + .get_driver_name = drm_mock_sched_hw_fence_driver_name, > + .get_timeline_name = drm_mock_sched_hw_fence_timeline_name, > + .release = drm_mock_sched_hw_fence_release, > +}; > + > +static struct dma_fence *mock_sched_run_job(struct drm_sched_job > *sched_job) > +{ > + struct drm_mock_scheduler *sched = > + drm_sched_to_mock_sched(sched_job->sched); > + struct drm_mock_sched_job *job = > drm_sched_job_to_mock_job(sched_job); > + > + dma_fence_init(&job->hw_fence, > + &drm_mock_sched_hw_fence_ops, > + &job->lock, > + sched->hw_fence.context, > + atomic_inc_return(&sched->hw_fence.seqno)); > + > + dma_fence_get(&job->hw_fence); /* Reference for the job_list > */ > + > + spin_lock_irq(&sched->lock); > + if (job->duration_us) { > + ktime_t prev_finish_at = 0; > + > + if (!list_empty(&sched->job_list)) { > + struct drm_mock_sched_job *prev = > + list_last_entry(&sched->job_list, > typeof(*prev), > + link); > + > + prev_finish_at = prev->finish_at; > + } > + > + if (!prev_finish_at) > + prev_finish_at = ktime_get(); > + > + job->finish_at = ktime_add_us(prev_finish_at, job- > >duration_us); > + } > + list_add_tail(&job->link, &sched->job_list); > + if (job->finish_at) > + hrtimer_start(&job->timer, job->finish_at, > HRTIMER_MODE_ABS); > + spin_unlock_irq(&sched->lock); > + > + return &job->hw_fence; > +} > + > +static enum drm_gpu_sched_stat > +mock_sched_timedout_job(struct drm_sched_job *sched_job) > +{ > + return DRM_GPU_SCHED_STAT_ENODEV; > +} > + > +static void mock_sched_free_job(struct drm_sched_job *sched_job) > +{ > + drm_sched_job_cleanup(sched_job); > +} > + > +static const struct drm_sched_backend_ops drm_mock_scheduler_ops = { > + .run_job = mock_sched_run_job, > + .timedout_job = mock_sched_timedout_job, > + .free_job = mock_sched_free_job > +}; > + > +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit > *test) > +{ > + struct drm_mock_scheduler *sched; > + int ret; > + > + sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL); > + KUNIT_ASSERT_NOT_NULL(test, sched); > + > + ret = drm_sched_init(&sched->base, > + &drm_mock_scheduler_ops, > + NULL, /* wq */ > + DRM_SCHED_PRIORITY_COUNT, > + U32_MAX, /* max credits */ > + UINT_MAX, /* hang limit */ > + MAX_SCHEDULE_TIMEOUT, /* timeout */ > + NULL, /* timeout wq */ > + NULL, /* score */ > + "drm-mock-scheduler", > + NULL /* dev */); > + KUNIT_ASSERT_EQ(test, ret, 0); > + > + sched->test = test; > + sched->hw_fence.context = dma_fence_context_alloc(1); > + atomic_set(&sched->hw_fence.seqno, 0); > + INIT_LIST_HEAD(&sched->job_list); > + spin_lock_init(&sched->lock); > + > + return sched; > +} > + > +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched) > +{ > + struct drm_mock_sched_job *job, *next; > + unsigned long flags; > + LIST_HEAD(signal); > + > + spin_lock_irqsave(&sched->lock, flags); > + list_for_each_entry_safe(job, next, &sched->job_list, link) > + list_move_tail(&job->link, &signal); > + spin_unlock_irqrestore(&sched->lock, flags); So maybe you can help me get up to speed here a bit. What is the purpose behind job->link? Is the general idea documented somewhere? > + > + list_for_each_entry(job, &signal, link) { > + hrtimer_cancel(&job->timer); > + dma_fence_put(&job->hw_fence); > + } > + > + drm_sched_fini(&sched->base); > +} > + > +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler > *sched, > + unsigned int num) > +{ > + struct drm_mock_sched_job *job, *next; > + unsigned int found = 0; > + unsigned long flags; > + LIST_HEAD(signal); > + > + spin_lock_irqsave(&sched->lock, flags); > + if (WARN_ON_ONCE(sched->seqno + num < sched->seqno)) > + goto unlock; > + sched->seqno += num; > + list_for_each_entry_safe(job, next, &sched->job_list, link) > { > + if (sched->seqno < job->hw_fence.seqno) > + break; > + > + list_move_tail(&job->link, &signal); > + found++; > + } > +unlock: > + spin_unlock_irqrestore(&sched->lock, flags); > + > + list_for_each_entry(job, &signal, link) { > + dma_fence_signal(&job->hw_fence); > + dma_fence_put(&job->hw_fence); > + } > + > + return found; > +} > diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > new file mode 100644 > index 000000000000..421ee2712985 > --- /dev/null > +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > @@ -0,0 +1,124 @@ > + > +#include <kunit/test.h> > +#include <linux/atomic.h> > +#include <linux/dma-fence.h> > +#include <linux/hrtimer.h> > +#include <linux/ktime.h> > +#include <linux/list.h> > +#include <linux/atomic.h> > +#include <linux/mutex.h> > +#include <linux/types.h> > + > +#include <drm/gpu_scheduler.h> > + > +struct drm_mock_scheduler { > + struct drm_gpu_scheduler base; > + > + struct kunit *test; > + > + struct { > + u64 context; > + atomic_t seqno; > + } hw_fence; Hm, well, so this is confusing. drm_mock_sched_job below contains an actual struct dma_fence that is also called hw_fence, whereas this here seems to be some pseudo-hw_fence? What is its purpose? I believe we agreed that "Hardware fence" should always mean a fence per job that is signaled by the hardware (driver interrupt) once the job is done. If this hw_fence here is the same, why is it per scheduler? That's confusing. > + > + spinlock_t lock; Maybe document the lock's purpose > + unsigned int seqno; > + struct list_head job_list; > +}; > + > +struct drm_mock_sched_entity { > + struct drm_sched_entity base; > + > + struct kunit *test; > +}; > + > +struct drm_mock_sched_job { > + struct drm_sched_job base; > + > + struct list_head link; > + struct hrtimer timer; > + > + unsigned int duration_us; > + ktime_t finish_at; > + > + spinlock_t lock; Same > + struct dma_fence hw_fence; > + > + struct kunit *test; > +}; > + > +static inline struct drm_mock_scheduler * > +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched) > +{ > + return container_of(sched, struct drm_mock_scheduler, base); > +}; > + > +static inline struct drm_mock_sched_entity * > +drm_sched_entity_to_mock_entity(struct drm_sched_entity > *sched_entity) > +{ > + return container_of(sched_entity, struct > drm_mock_sched_entity, base); > +}; > + > +static inline struct drm_mock_sched_job * > +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job) > +{ > + return container_of(sched_job, struct drm_mock_sched_job, > base); > +}; > + > +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit > *test); > +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched); > +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler > *sched, > + unsigned int num); > + > +struct drm_mock_sched_entity * > +drm_mock_new_sched_entity(struct kunit *test, > + enum drm_sched_priority priority, > + struct drm_mock_scheduler *sched); > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity > *entity); > + > +struct drm_mock_sched_job * > +drm_mock_new_sched_job(struct kunit *test, > + struct drm_mock_sched_entity *entity); > + > +static inline void drm_mock_sched_job_submit(struct > drm_mock_sched_job *job) > +{ > + drm_sched_job_arm(&job->base); > + drm_sched_entity_push_job(&job->base); > +} > + > +static inline void > +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job, > + unsigned int duration_us) > +{ > + job->duration_us = duration_us; > +} > + > +static inline bool > +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job) > +{ > + return dma_fence_is_signaled(&job->base.s_fence->finished); > +} > + > +static inline bool > +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job, > long timeout) > +{ > + long ret; > + > + ret = dma_fence_wait_timeout(&job->base.s_fence->finished, > + false, > + timeout); > + > + return ret != 0; > +} > + > +static inline long > +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job, > long timeout) > +{ > + long ret; > + > + ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled, > + false, > + timeout); > + > + return ret != 0; > +} > diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > new file mode 100644 > index 000000000000..6fd39bea95b1 > --- /dev/null > +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > @@ -0,0 +1,247 @@ > + > +#include "drm_sched_tests.h" > + > +static int drm_sched_basic_init(struct kunit *test) > +{ > + test->priv = drm_mock_new_scheduler(test); > + > + return 0; > +} > + > +static void drm_sched_basic_exit(struct kunit *test) > +{ > + struct drm_mock_scheduler *sched = test->priv; > + > + drm_mock_scheduler_fini(sched); > +} > + > +static void drm_sched_basic_submit(struct kunit *test) > +{ > + struct drm_mock_scheduler *sched = test->priv; > + struct drm_mock_sched_entity *entity; > + struct drm_mock_sched_job *job; > + unsigned int i; > + bool done; > + > + /* > + * Submit one job to the scheduler and verify that it gets > scheduled > + * and completed only when the mock hw backend processes it. > + */ > + > + entity = drm_mock_new_sched_entity(test, > + > DRM_SCHED_PRIORITY_NORMAL, > + sched); > + job = drm_mock_new_sched_job(test, entity); > + > + drm_mock_sched_job_submit(job); > + > + done = drm_mock_sched_job_wait_scheduled(job, HZ); > + KUNIT_ASSERT_EQ(test, done, true); > + > + done = drm_mock_sched_job_wait_finished(job, HZ / 2); > + KUNIT_ASSERT_EQ(test, done, false); > + > + i = drm_mock_sched_advance(sched, 1); > + KUNIT_ASSERT_EQ(test, i, 1); > + > + done = drm_mock_sched_job_wait_finished(job, HZ); > + KUNIT_ASSERT_EQ(test, done, true); > + > + drm_mock_sched_entity_free(entity); > +} > + > +static void drm_sched_basic_queue(struct kunit *test) > +{ > + struct drm_mock_scheduler *sched = test->priv; > + struct drm_mock_sched_entity *entity; > + struct drm_mock_sched_job *job; > + const unsigned int qd = 100; Not the best variable name – is this "queue depth"? Why is it 100? -> global define & document > + unsigned int i; > + bool done; > + > + /* > + * Submit a queue of jobs on the same entity, let them be > completed by > + * the mock hw backend and check the status of the last job. > + */ > + > + entity = drm_mock_new_sched_entity(test, > + > DRM_SCHED_PRIORITY_NORMAL, > + sched); > + > + for (i = 0; i < qd; i++) { > + job = drm_mock_new_sched_job(test, entity); > + drm_mock_sched_job_set_duration_us(job, 1000); > + drm_mock_sched_job_submit(job); > + } > + > + done = drm_mock_sched_job_wait_finished(job, HZ); > + KUNIT_ASSERT_EQ(test, done, true); > + > + drm_mock_sched_entity_free(entity); > +} > + > +static void drm_sched_basic_chain(struct kunit *test) > +{ > + struct drm_mock_scheduler *sched = test->priv; > + struct drm_mock_sched_job *job, *prev = NULL; > + struct drm_mock_sched_entity *entity; > + const unsigned int qd = 100; > + unsigned int i; > + bool done; > + > + /* > + * Submit a queue of jobs on the same entity but with an > explicit > + * chain of dependencies between them. Let the jobs be > completed by > + * the mock hw backend and check the status of the last job. > + */ > + > + entity = drm_mock_new_sched_entity(test, > + > DRM_SCHED_PRIORITY_NORMAL, > + sched); > + > + for (i = 0; i < qd; i++) { > + job = drm_mock_new_sched_job(test, entity); > + drm_mock_sched_job_set_duration_us(job, 1000); > + if (prev) > + drm_sched_job_add_dependency(&job->base, > + > dma_fence_get(&prev->base.s_fence->finished)); > + drm_mock_sched_job_submit(job); > + prev = job; > + } > + > + done = drm_mock_sched_job_wait_finished(job, HZ); > + KUNIT_ASSERT_EQ(test, done, true); > + > + drm_mock_sched_entity_free(entity); > +} > + > +static void drm_sched_basic_entities(struct kunit *test) > +{ > + struct drm_mock_scheduler *sched = test->priv; > + struct drm_mock_sched_entity *entity[4]; > + struct drm_mock_sched_job *job; > + const unsigned int qd = 100; > + unsigned int i, cur_ent = 0; > + bool done; > + > + /* > + * Submit a queue of jobs across different entities, let > them be > + * completed by the mock hw backend and check the status of > the last > + * job. > + */ > + > + for (i = 0; i < ARRAY_SIZE(entity); i++) > + entity[i] = drm_mock_new_sched_entity(test, > + > DRM_SCHED_PRIORITY_NORMAL, > + sched); > + > + for (i = 0; i < qd; i++) { > + job = drm_mock_new_sched_job(test, > entity[cur_ent++]); > + cur_ent %= ARRAY_SIZE(entity); > + drm_mock_sched_job_set_duration_us(job, 1000); > + drm_mock_sched_job_submit(job); > + } > + > + done = drm_mock_sched_job_wait_finished(job, HZ); > + KUNIT_ASSERT_EQ(test, done, true); > + > + for (i = 0; i < ARRAY_SIZE(entity); i++) > + drm_mock_sched_entity_free(entity[i]); > +} > + > +static void drm_sched_basic_entities_chain(struct kunit *test) > +{ > + struct drm_mock_scheduler *sched = test->priv; > + struct drm_mock_sched_job *job, *prev = NULL; > + struct drm_mock_sched_entity *entity[4]; > + const unsigned int qd = 100; > + unsigned int i, cur_ent = 0; > + bool done; > + > + /* > + * Submit a queue of jobs across different entities with an > explicit > + * chain of dependencies between them. Let the jobs be > completed by > + * the mock hw backend and check the status of the last job. > + */ > + > + for (i = 0; i < ARRAY_SIZE(entity); i++) > + entity[i] = drm_mock_new_sched_entity(test, > + > DRM_SCHED_PRIORITY_NORMAL, > + sched); > + > + for (i = 0; i < qd; i++) { > + job = drm_mock_new_sched_job(test, > entity[cur_ent++]); > + cur_ent %= ARRAY_SIZE(entity); > + drm_mock_sched_job_set_duration_us(job, 1000); > + if (prev) > + drm_sched_job_add_dependency(&job->base, > + > dma_fence_get(&prev->base.s_fence->finished)); > + drm_mock_sched_job_submit(job); > + prev = job; > + } > + > + done = drm_mock_sched_job_wait_finished(job, HZ); > + KUNIT_ASSERT_EQ(test, done, true); > + > + for (i = 0; i < ARRAY_SIZE(entity); i++) > + drm_mock_sched_entity_free(entity[i]); > +} > + > +static void drm_sched_basic_entity_cleanup(struct kunit *test) > +{ > + struct drm_mock_sched_job *job, *mid, *prev = NULL; > + struct drm_mock_scheduler *sched = test->priv; > + struct drm_mock_sched_entity *entity[4]; 4 is used in several places, so could be defined globally. In case there's a special reason for why it's 4, that could be mentioned, too P. > + const unsigned int qd = 100; > + unsigned int i, cur_ent = 0; > + bool done; > + > + /* > + * Submit a queue of jobs across different entities with an > explicit > + * chain of dependencies between them and trigger entity > cleanup while > + * the queue is still being processed. > + */ > + > + for (i = 0; i < ARRAY_SIZE(entity); i++) > + entity[i] = drm_mock_new_sched_entity(test, > + > DRM_SCHED_PRIORITY_NORMAL, > + sched); > + > + for (i = 0; i < qd; i++) { > + job = drm_mock_new_sched_job(test, > entity[cur_ent++]); > + cur_ent %= ARRAY_SIZE(entity); > + drm_mock_sched_job_set_duration_us(job, 1000); > + if (prev) > + drm_sched_job_add_dependency(&job->base, > + > dma_fence_get(&prev->base.s_fence->finished)); > + drm_mock_sched_job_submit(job); > + if (i == qd / 2) > + mid = job; > + prev = job; > + } > + > + done = drm_mock_sched_job_wait_finished(mid, HZ); > + KUNIT_ASSERT_EQ(test, done, true); > + > + /* Exit with half of the queue still pending to be executed. > */ > + for (i = 0; i < ARRAY_SIZE(entity); i++) > + drm_mock_sched_entity_free(entity[i]);} > + > +static struct kunit_case drm_sched_basic_tests[] = { > + KUNIT_CASE(drm_sched_basic_submit), > + KUNIT_CASE(drm_sched_basic_queue), > + KUNIT_CASE(drm_sched_basic_chain), > + KUNIT_CASE(drm_sched_basic_entities), > + KUNIT_CASE(drm_sched_basic_entities_chain), > + KUNIT_CASE(drm_sched_basic_entity_cleanup), > + {} > +}; > + > +static struct kunit_suite drm_sched_basic = { > + .name = "drm_sched_basic_tests", > + .init = drm_sched_basic_init, > + .exit = drm_sched_basic_exit, > + .test_cases = drm_sched_basic_tests, > +}; > + > +kunit_test_suite(drm_sched_basic);
On 06/02/2025 09:51, Philipp Stanner wrote: > On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote: >> Implement a mock scheduler backend and add some basic test to >> exercise the >> core scheduler code paths. >> >> Mock backend (kind of like a very simple mock GPU) can either process >> jobs >> by tests manually advancing the "timeline" job at a time, or >> alternatively >> jobs can be configured with a time duration in which case they get >> completed asynchronously from the unit test code. >> >> Core scheduler classes are subclassed to support this mock >> implementation. >> >> The tests added are just a few simple submission patterns. >> >> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com> > > Could add a Suggested-by: Philipp :) Will do. >> Cc: Christian König <christian.koenig@amd.com> >> Cc: Danilo Krummrich <dakr@kernel.org> >> Cc: Matthew Brost <matthew.brost@intel.com> >> Cc: Philipp Stanner <phasta@kernel.org> >> --- >> drivers/gpu/drm/Kconfig.debug | 12 + >> drivers/gpu/drm/scheduler/.kunitconfig | 12 + >> drivers/gpu/drm/scheduler/Makefile | 1 + >> drivers/gpu/drm/scheduler/tests/Makefile | 4 + >> .../gpu/drm/scheduler/tests/drm_mock_entity.c | 29 ++ >> .../gpu/drm/scheduler/tests/drm_mock_job.c | 3 + >> .../drm/scheduler/tests/drm_mock_scheduler.c | 254 >> ++++++++++++++++++ >> .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++ >> .../scheduler/tests/drm_sched_tests_basic.c | 247 >> +++++++++++++++++ >> 9 files changed, 686 insertions(+) >> create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig >> create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile >> create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >> create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c >> create mode 100644 >> drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >> create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >> create mode 100644 >> drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >> >> diff --git a/drivers/gpu/drm/Kconfig.debug >> b/drivers/gpu/drm/Kconfig.debug >> index a35d74171b7b..89f777f21e95 100644 >> --- a/drivers/gpu/drm/Kconfig.debug >> +++ b/drivers/gpu/drm/Kconfig.debug >> @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST >> >> If in doubt, say "N". >> >> +config DRM_SCHED_KUNIT_TEST >> + tristate "KUnit tests for the DRM scheduler" if >> !KUNIT_ALL_TESTS >> + select DRM_SCHED >> + depends on DRM && KUNIT >> + default KUNIT_ALL_TESTS >> + help >> + Choose this option to build unit tests for the DRM >> scheduler. > > nit: Might say "DRM GPU scheduler" so readers not familiar with all > that get a better idea of what it's about > >> + >> + Recommended for driver developers only. > > nit: s/driver developers/DRM developers > ? > >> + >> + If in doubt, say "N". >> + >> config DRM_EXPORT_FOR_TESTS >> bool >> diff --git a/drivers/gpu/drm/scheduler/.kunitconfig >> b/drivers/gpu/drm/scheduler/.kunitconfig >> new file mode 100644 >> index 000000000000..cece53609fcf >> --- /dev/null >> +++ b/drivers/gpu/drm/scheduler/.kunitconfig >> @@ -0,0 +1,12 @@ >> +CONFIG_KUNIT=y >> +CONFIG_DRM=y >> +CONFIG_DRM_SCHED_KUNIT_TEST=y >> +CONFIG_EXPERT=y >> +CONFIG_DEBUG_SPINLOCK=y >> +CONFIG_DEBUG_MUTEXES=y >> +CONFIG_DEBUG_ATOMIC_SLEEP=y >> +CONFIG_LOCK_DEBUGGING_SUPPORT=y >> +CONFIG_PROVE_LOCKING=y >> +CONFIG_LOCKDEP=y >> +CONFIG_DEBUG_LOCKDEP=y >> +CONFIG_DEBUG_LIST=y >> diff --git a/drivers/gpu/drm/scheduler/Makefile >> b/drivers/gpu/drm/scheduler/Makefile >> index 53863621829f..46dfdca0758a 100644 >> --- a/drivers/gpu/drm/scheduler/Makefile >> +++ b/drivers/gpu/drm/scheduler/Makefile >> @@ -23,3 +23,4 @@ >> gpu-sched-y := sched_main.o sched_fence.o sched_entity.o >> >> obj-$(CONFIG_DRM_SCHED) += gpu-sched.o >> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/ >> diff --git a/drivers/gpu/drm/scheduler/tests/Makefile >> b/drivers/gpu/drm/scheduler/tests/Makefile >> new file mode 100644 >> index 000000000000..d69eab6a2e9b >> --- /dev/null >> +++ b/drivers/gpu/drm/scheduler/tests/Makefile >> @@ -0,0 +1,4 @@ >> + >> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \ >> + drm_mock_scheduler.o \ >> + drm_sched_tests_basic.o >> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >> b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >> new file mode 100644 >> index 000000000000..c9205f9ff524 >> --- /dev/null >> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >> @@ -0,0 +1,29 @@ >> + >> +#include "drm_sched_tests.h" >> + >> +struct drm_mock_sched_entity * >> +drm_mock_sched_entity_new(struct kunit *test, >> + enum drm_sched_priority priority, >> + struct drm_mock_scheduler *sched) > > > I personally do like this function head style and > > >> +{ >> + struct drm_sched_mock_entity *entity; >> + int ret; >> + >> + entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL); >> + KUNIT_ASSERT_NOT_NULL(test, entity); >> + >> + ret = drm_sched_entity_init(&entity->base, >> + priority, >> + &sched->base, 1, >> + NULL); >> + KUNIT_ASSERT_EQ(test, ret, 0); >> + >> + entity->test = test; >> + >> + return entity; >> +} >> + >> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity >> *entity) > > do believe it should then be used consistently everywhere, regardless > of length. > > >> +{ >> + drm_sched_entity_fini(&entity->base); >> +} >> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c >> b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c >> new file mode 100644 >> index 000000000000..d94568cf3da9 >> --- /dev/null >> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c >> @@ -0,0 +1,3 @@ >> + >> +#include "drm_sched_tests.h" >> + >> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >> b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >> new file mode 100644 >> index 000000000000..f1985900a6ba >> --- /dev/null >> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >> @@ -0,0 +1,254 @@ >> + >> +#include "drm_sched_tests.h" >> + >> +struct drm_mock_sched_entity * >> +drm_mock_new_sched_entity(struct kunit *test, >> + enum drm_sched_priority priority, >> + struct drm_mock_scheduler *sched) >> +{ >> + struct drm_mock_sched_entity *entity; >> + struct drm_gpu_scheduler *drm_sched; >> + int ret; >> + >> + entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL); >> + KUNIT_ASSERT_NOT_NULL(test, entity); >> + >> + drm_sched = &sched->base; >> + ret = drm_sched_entity_init(&entity->base, >> + priority, >> + &drm_sched, 1, >> + NULL); >> + KUNIT_ASSERT_EQ(test, ret, 0); >> + >> + entity->test = test; >> + >> + return entity; >> +} >> + >> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity >> *entity) >> +{ >> + drm_sched_entity_destroy(&entity->base); >> +} >> + >> +static enum hrtimer_restart >> +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer) > > I think we should get up with a consistent naming convention. Some > things are called drm_mock_sched_, some others drm_sched_mock_. > As far as I saw, drm_mock_* does not yet exist. > So do you want to introduce drm_mock as something generic for drm? or > just for drm/sched? This one is literally the only when where I tranposed the two. But it is also local and I am not too bothered about naming conventions of a local functions. If it were to me I would favour brevity and emit the long drm_.. prefixes which IMO do not help readability. If anything, seeing something with a long drm_.. prefix can only be confusing since one can assume it is some sort of exported interface. I will change this instance. >> +{ >> + struct drm_mock_sched_job *upto = >> + container_of(hrtimer, typeof(*upto), timer); >> + struct drm_mock_scheduler *sched = >> + drm_sched_to_mock_sched(upto->base.sched); >> + struct drm_mock_sched_job *job, *next; >> + ktime_t now = ktime_get(); >> + unsigned long flags; >> + LIST_HEAD(signal); >> + >> + spin_lock_irqsave(&sched->lock, flags); >> + list_for_each_entry_safe(job, next, &sched->job_list, link) >> { >> + if (!job->duration_us) >> + break; >> + >> + if (ktime_before(now, job->finish_at)) >> + break; >> + >> + list_move_tail(&job->link, &signal); >> + sched->seqno = job->hw_fence.seqno; >> + } >> + spin_unlock_irqrestore(&sched->lock, flags); >> + >> + list_for_each_entry(job, &signal, link) { >> + dma_fence_signal(&job->hw_fence); >> + dma_fence_put(&job->hw_fence); >> + } >> + >> + return HRTIMER_NORESTART; >> +} >> + >> +struct drm_mock_sched_job * >> +drm_mock_new_sched_job(struct kunit *test, >> + struct drm_mock_sched_entity *entity) >> +{ >> + struct drm_mock_sched_job *job; >> + int ret; >> + >> + job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL); >> + KUNIT_ASSERT_NOT_NULL(test, job); >> + >> + ret = drm_sched_job_init(&job->base, >> + &entity->base, >> + 1, >> + NULL); >> + KUNIT_ASSERT_EQ(test, ret, 0); >> + >> + job->test = test; >> + >> + spin_lock_init(&job->lock); >> + INIT_LIST_HEAD(&job->link); >> + hrtimer_init(&job->timer, CLOCK_MONOTONIC, >> HRTIMER_MODE_ABS); >> + job->timer.function = drm_sched_mock_job_signal_timer; >> + >> + return job; >> +} >> + >> +static const char *drm_mock_sched_hw_fence_driver_name(struct >> dma_fence *fence) >> +{ >> + return "drm_mock_sched"; >> +} >> + >> +static const char * >> +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence) >> +{ >> + struct drm_mock_sched_job *job = >> + container_of(fence, typeof(*job), hw_fence); >> + >> + return (const char *)job->base.sched->name; >> +} >> +static void drm_mock_sched_hw_fence_release(struct dma_fence *fence) >> +{ >> + struct drm_mock_sched_job *job = >> + container_of(fence, typeof(*job), hw_fence); >> + >> + hrtimer_cancel(&job->timer); >> + >> + /* Freed by the kunit framework */ >> +} >> + >> +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = { >> + .get_driver_name = drm_mock_sched_hw_fence_driver_name, >> + .get_timeline_name = drm_mock_sched_hw_fence_timeline_name, >> + .release = drm_mock_sched_hw_fence_release, >> +}; >> + >> +static struct dma_fence *mock_sched_run_job(struct drm_sched_job >> *sched_job) >> +{ >> + struct drm_mock_scheduler *sched = >> + drm_sched_to_mock_sched(sched_job->sched); >> + struct drm_mock_sched_job *job = >> drm_sched_job_to_mock_job(sched_job); >> + >> + dma_fence_init(&job->hw_fence, >> + &drm_mock_sched_hw_fence_ops, >> + &job->lock, >> + sched->hw_fence.context, >> + atomic_inc_return(&sched->hw_fence.seqno)); >> + >> + dma_fence_get(&job->hw_fence); /* Reference for the job_list >> */ >> + >> + spin_lock_irq(&sched->lock); >> + if (job->duration_us) { >> + ktime_t prev_finish_at = 0; >> + >> + if (!list_empty(&sched->job_list)) { >> + struct drm_mock_sched_job *prev = >> + list_last_entry(&sched->job_list, >> typeof(*prev), >> + link); >> + >> + prev_finish_at = prev->finish_at; >> + } >> + >> + if (!prev_finish_at) >> + prev_finish_at = ktime_get(); >> + >> + job->finish_at = ktime_add_us(prev_finish_at, job- >>> duration_us); >> + } >> + list_add_tail(&job->link, &sched->job_list); >> + if (job->finish_at) >> + hrtimer_start(&job->timer, job->finish_at, >> HRTIMER_MODE_ABS); >> + spin_unlock_irq(&sched->lock); >> + >> + return &job->hw_fence; >> +} >> + >> +static enum drm_gpu_sched_stat >> +mock_sched_timedout_job(struct drm_sched_job *sched_job) >> +{ >> + return DRM_GPU_SCHED_STAT_ENODEV; >> +} >> + >> +static void mock_sched_free_job(struct drm_sched_job *sched_job) >> +{ >> + drm_sched_job_cleanup(sched_job); >> +} >> + >> +static const struct drm_sched_backend_ops drm_mock_scheduler_ops = { >> + .run_job = mock_sched_run_job, >> + .timedout_job = mock_sched_timedout_job, >> + .free_job = mock_sched_free_job >> +}; >> + >> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit >> *test) >> +{ >> + struct drm_mock_scheduler *sched; >> + int ret; >> + >> + sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL); >> + KUNIT_ASSERT_NOT_NULL(test, sched); >> + >> + ret = drm_sched_init(&sched->base, >> + &drm_mock_scheduler_ops, >> + NULL, /* wq */ >> + DRM_SCHED_PRIORITY_COUNT, >> + U32_MAX, /* max credits */ >> + UINT_MAX, /* hang limit */ >> + MAX_SCHEDULE_TIMEOUT, /* timeout */ >> + NULL, /* timeout wq */ >> + NULL, /* score */ >> + "drm-mock-scheduler", >> + NULL /* dev */); >> + KUNIT_ASSERT_EQ(test, ret, 0); >> + >> + sched->test = test; >> + sched->hw_fence.context = dma_fence_context_alloc(1); >> + atomic_set(&sched->hw_fence.seqno, 0); >> + INIT_LIST_HEAD(&sched->job_list); >> + spin_lock_init(&sched->lock); >> + >> + return sched; >> +} >> + >> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched) >> +{ >> + struct drm_mock_sched_job *job, *next; >> + unsigned long flags; >> + LIST_HEAD(signal); >> + >> + spin_lock_irqsave(&sched->lock, flags); >> + list_for_each_entry_safe(job, next, &sched->job_list, link) >> + list_move_tail(&job->link, &signal); >> + spin_unlock_irqrestore(&sched->lock, flags); > > So maybe you can help me get up to speed here a bit. What is the > purpose behind job->link? Is the general idea documented somewhere? List versus link suffix convention I use to distinguish struct list_head which is a list versus struct list_head which is used to put something on the list. In this case job->link is for the mock GPU driver to keep track of jobs which have been submitted to the "hardware" for "execution". I will of course document these things once the high level design is settled. >> + >> + list_for_each_entry(job, &signal, link) { >> + hrtimer_cancel(&job->timer); >> + dma_fence_put(&job->hw_fence); >> + } >> + >> + drm_sched_fini(&sched->base); >> +} >> + >> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler >> *sched, >> + unsigned int num) >> +{ >> + struct drm_mock_sched_job *job, *next; >> + unsigned int found = 0; >> + unsigned long flags; >> + LIST_HEAD(signal); >> + >> + spin_lock_irqsave(&sched->lock, flags); >> + if (WARN_ON_ONCE(sched->seqno + num < sched->seqno)) >> + goto unlock; >> + sched->seqno += num; >> + list_for_each_entry_safe(job, next, &sched->job_list, link) >> { >> + if (sched->seqno < job->hw_fence.seqno) >> + break; >> + >> + list_move_tail(&job->link, &signal); >> + found++; >> + } >> +unlock: >> + spin_unlock_irqrestore(&sched->lock, flags); >> + >> + list_for_each_entry(job, &signal, link) { >> + dma_fence_signal(&job->hw_fence); >> + dma_fence_put(&job->hw_fence); >> + } >> + >> + return found; >> +} >> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >> new file mode 100644 >> index 000000000000..421ee2712985 >> --- /dev/null >> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >> @@ -0,0 +1,124 @@ >> + >> +#include <kunit/test.h> >> +#include <linux/atomic.h> >> +#include <linux/dma-fence.h> >> +#include <linux/hrtimer.h> >> +#include <linux/ktime.h> >> +#include <linux/list.h> >> +#include <linux/atomic.h> >> +#include <linux/mutex.h> >> +#include <linux/types.h> >> + >> +#include <drm/gpu_scheduler.h> >> + >> +struct drm_mock_scheduler { >> + struct drm_gpu_scheduler base; >> + >> + struct kunit *test; >> + >> + struct { >> + u64 context; >> + atomic_t seqno; >> + } hw_fence; > > Hm, well, so this is confusing. drm_mock_sched_job below contains an > actual struct dma_fence that is also called hw_fence, whereas this here > seems to be some pseudo-hw_fence? > > What is its purpose? > > I believe we agreed that "Hardware fence" should always mean a fence > per job that is signaled by the hardware (driver interrupt) once the > job is done. > > If this hw_fence here is the same, why is it per scheduler? That's > confusing. Mock_job->hw_fence is what the mock GPU driver returns from the sched->run_job(). Mock_sched->hw_fence is what the mock GPU driver uses to track execution of jobs submitted to it for execution. If Irename this to hw_timeline will it make it clearer? >> + >> + spinlock_t lock; > > Maybe document the lock's purpose > > >> + unsigned int seqno; >> + struct list_head job_list; >> +}; >> + >> +struct drm_mock_sched_entity { >> + struct drm_sched_entity base; >> + >> + struct kunit *test; >> +}; >> + >> +struct drm_mock_sched_job { >> + struct drm_sched_job base; >> + >> + struct list_head link; >> + struct hrtimer timer; >> + >> + unsigned int duration_us; >> + ktime_t finish_at; >> + >> + spinlock_t lock; > > Same > >> + struct dma_fence hw_fence; >> + >> + struct kunit *test; >> +}; >> + >> +static inline struct drm_mock_scheduler * >> +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched) >> +{ >> + return container_of(sched, struct drm_mock_scheduler, base); >> +}; >> + >> +static inline struct drm_mock_sched_entity * >> +drm_sched_entity_to_mock_entity(struct drm_sched_entity >> *sched_entity) >> +{ >> + return container_of(sched_entity, struct >> drm_mock_sched_entity, base); >> +}; >> + >> +static inline struct drm_mock_sched_job * >> +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job) >> +{ >> + return container_of(sched_job, struct drm_mock_sched_job, >> base); >> +}; >> + >> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit >> *test); >> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched); >> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler >> *sched, >> + unsigned int num); >> + >> +struct drm_mock_sched_entity * >> +drm_mock_new_sched_entity(struct kunit *test, >> + enum drm_sched_priority priority, >> + struct drm_mock_scheduler *sched); >> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity >> *entity); >> + >> +struct drm_mock_sched_job * >> +drm_mock_new_sched_job(struct kunit *test, >> + struct drm_mock_sched_entity *entity); >> + >> +static inline void drm_mock_sched_job_submit(struct >> drm_mock_sched_job *job) >> +{ >> + drm_sched_job_arm(&job->base); >> + drm_sched_entity_push_job(&job->base); >> +} >> + >> +static inline void >> +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job, >> + unsigned int duration_us) >> +{ >> + job->duration_us = duration_us; >> +} >> + >> +static inline bool >> +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job) >> +{ >> + return dma_fence_is_signaled(&job->base.s_fence->finished); >> +} >> + >> +static inline bool >> +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job, >> long timeout) >> +{ >> + long ret; >> + >> + ret = dma_fence_wait_timeout(&job->base.s_fence->finished, >> + false, >> + timeout); >> + >> + return ret != 0; >> +} >> + >> +static inline long >> +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job, >> long timeout) >> +{ >> + long ret; >> + >> + ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled, >> + false, >> + timeout); >> + >> + return ret != 0; >> +} >> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >> new file mode 100644 >> index 000000000000..6fd39bea95b1 >> --- /dev/null >> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >> @@ -0,0 +1,247 @@ >> + >> +#include "drm_sched_tests.h" >> + >> +static int drm_sched_basic_init(struct kunit *test) >> +{ >> + test->priv = drm_mock_new_scheduler(test); >> + >> + return 0; >> +} >> + >> +static void drm_sched_basic_exit(struct kunit *test) >> +{ >> + struct drm_mock_scheduler *sched = test->priv; >> + >> + drm_mock_scheduler_fini(sched); >> +} >> + >> +static void drm_sched_basic_submit(struct kunit *test) >> +{ >> + struct drm_mock_scheduler *sched = test->priv; >> + struct drm_mock_sched_entity *entity; >> + struct drm_mock_sched_job *job; >> + unsigned int i; >> + bool done; >> + >> + /* >> + * Submit one job to the scheduler and verify that it gets >> scheduled >> + * and completed only when the mock hw backend processes it. >> + */ >> + >> + entity = drm_mock_new_sched_entity(test, >> + >> DRM_SCHED_PRIORITY_NORMAL, >> + sched); >> + job = drm_mock_new_sched_job(test, entity); >> + >> + drm_mock_sched_job_submit(job); >> + >> + done = drm_mock_sched_job_wait_scheduled(job, HZ); >> + KUNIT_ASSERT_EQ(test, done, true); >> + >> + done = drm_mock_sched_job_wait_finished(job, HZ / 2); >> + KUNIT_ASSERT_EQ(test, done, false); >> + >> + i = drm_mock_sched_advance(sched, 1); >> + KUNIT_ASSERT_EQ(test, i, 1); >> + >> + done = drm_mock_sched_job_wait_finished(job, HZ); >> + KUNIT_ASSERT_EQ(test, done, true); >> + >> + drm_mock_sched_entity_free(entity); >> +} >> + >> +static void drm_sched_basic_queue(struct kunit *test) >> +{ >> + struct drm_mock_scheduler *sched = test->priv; >> + struct drm_mock_sched_entity *entity; >> + struct drm_mock_sched_job *job; >> + const unsigned int qd = 100; > > Not the best variable name – is this "queue depth"? Why is it 100? -> > global define & document I wouldn't promote this to global and TBH if you look how small this test function is it feels pretty self documenting. > > >> + unsigned int i; >> + bool done; >> + >> + /* >> + * Submit a queue of jobs on the same entity, let them be >> completed by >> + * the mock hw backend and check the status of the last job. >> + */ >> + >> + entity = drm_mock_new_sched_entity(test, >> + >> DRM_SCHED_PRIORITY_NORMAL, >> + sched); >> + >> + for (i = 0; i < qd; i++) { >> + job = drm_mock_new_sched_job(test, entity); >> + drm_mock_sched_job_set_duration_us(job, 1000); >> + drm_mock_sched_job_submit(job); >> + } >> + >> + done = drm_mock_sched_job_wait_finished(job, HZ); >> + KUNIT_ASSERT_EQ(test, done, true); >> + >> + drm_mock_sched_entity_free(entity); >> +} >> + >> +static void drm_sched_basic_chain(struct kunit *test) >> +{ >> + struct drm_mock_scheduler *sched = test->priv; >> + struct drm_mock_sched_job *job, *prev = NULL; >> + struct drm_mock_sched_entity *entity; >> + const unsigned int qd = 100; >> + unsigned int i; >> + bool done; >> + >> + /* >> + * Submit a queue of jobs on the same entity but with an >> explicit >> + * chain of dependencies between them. Let the jobs be >> completed by >> + * the mock hw backend and check the status of the last job. >> + */ >> + >> + entity = drm_mock_new_sched_entity(test, >> + >> DRM_SCHED_PRIORITY_NORMAL, >> + sched); >> + >> + for (i = 0; i < qd; i++) { >> + job = drm_mock_new_sched_job(test, entity); >> + drm_mock_sched_job_set_duration_us(job, 1000); >> + if (prev) >> + drm_sched_job_add_dependency(&job->base, >> + >> dma_fence_get(&prev->base.s_fence->finished)); >> + drm_mock_sched_job_submit(job); >> + prev = job; >> + } >> + >> + done = drm_mock_sched_job_wait_finished(job, HZ); >> + KUNIT_ASSERT_EQ(test, done, true); >> + >> + drm_mock_sched_entity_free(entity); >> +} >> + >> +static void drm_sched_basic_entities(struct kunit *test) >> +{ >> + struct drm_mock_scheduler *sched = test->priv; >> + struct drm_mock_sched_entity *entity[4]; >> + struct drm_mock_sched_job *job; >> + const unsigned int qd = 100; >> + unsigned int i, cur_ent = 0; >> + bool done; >> + >> + /* >> + * Submit a queue of jobs across different entities, let >> them be >> + * completed by the mock hw backend and check the status of >> the last >> + * job. >> + */ >> + >> + for (i = 0; i < ARRAY_SIZE(entity); i++) >> + entity[i] = drm_mock_new_sched_entity(test, >> + >> DRM_SCHED_PRIORITY_NORMAL, >> + sched); >> + >> + for (i = 0; i < qd; i++) { >> + job = drm_mock_new_sched_job(test, >> entity[cur_ent++]); >> + cur_ent %= ARRAY_SIZE(entity); >> + drm_mock_sched_job_set_duration_us(job, 1000); >> + drm_mock_sched_job_submit(job); >> + } >> + >> + done = drm_mock_sched_job_wait_finished(job, HZ); >> + KUNIT_ASSERT_EQ(test, done, true); >> + >> + for (i = 0; i < ARRAY_SIZE(entity); i++) >> + drm_mock_sched_entity_free(entity[i]); >> +} >> + >> +static void drm_sched_basic_entities_chain(struct kunit *test) >> +{ >> + struct drm_mock_scheduler *sched = test->priv; >> + struct drm_mock_sched_job *job, *prev = NULL; >> + struct drm_mock_sched_entity *entity[4]; >> + const unsigned int qd = 100; >> + unsigned int i, cur_ent = 0; >> + bool done; >> + >> + /* >> + * Submit a queue of jobs across different entities with an >> explicit >> + * chain of dependencies between them. Let the jobs be >> completed by >> + * the mock hw backend and check the status of the last job. >> + */ >> + >> + for (i = 0; i < ARRAY_SIZE(entity); i++) >> + entity[i] = drm_mock_new_sched_entity(test, >> + >> DRM_SCHED_PRIORITY_NORMAL, >> + sched); >> + >> + for (i = 0; i < qd; i++) { >> + job = drm_mock_new_sched_job(test, >> entity[cur_ent++]); >> + cur_ent %= ARRAY_SIZE(entity); >> + drm_mock_sched_job_set_duration_us(job, 1000); >> + if (prev) >> + drm_sched_job_add_dependency(&job->base, >> + >> dma_fence_get(&prev->base.s_fence->finished)); >> + drm_mock_sched_job_submit(job); >> + prev = job; >> + } >> + >> + done = drm_mock_sched_job_wait_finished(job, HZ); >> + KUNIT_ASSERT_EQ(test, done, true); >> + >> + for (i = 0; i < ARRAY_SIZE(entity); i++) >> + drm_mock_sched_entity_free(entity[i]); >> +} >> + >> +static void drm_sched_basic_entity_cleanup(struct kunit *test) >> +{ >> + struct drm_mock_sched_job *job, *mid, *prev = NULL; >> + struct drm_mock_scheduler *sched = test->priv; >> + struct drm_mock_sched_entity *entity[4]; > > 4 is used in several places, so could be defined globally. In case > there's a special reason for why it's 4, that could be mentioned, too It's completely arbitrary. In this test we just need a bunch of entities to be active on the scheduler as we trigger a client exit in the middle of it. IMO the comment few lines below should be good enough to explain the idea. I fear that promoting this to a global define would just give it more weight that what it has. Then some test would want a different number etc. Regards, Tvrtko >> + const unsigned int qd = 100; >> + unsigned int i, cur_ent = 0; >> + bool done; >> + >> + /* >> + * Submit a queue of jobs across different entities with an >> explicit >> + * chain of dependencies between them and trigger entity >> cleanup while >> + * the queue is still being processed. >> + */ >> + >> + for (i = 0; i < ARRAY_SIZE(entity); i++) >> + entity[i] = drm_mock_new_sched_entity(test, >> + >> DRM_SCHED_PRIORITY_NORMAL, >> + sched); >> + >> + for (i = 0; i < qd; i++) { >> + job = drm_mock_new_sched_job(test, >> entity[cur_ent++]); >> + cur_ent %= ARRAY_SIZE(entity); >> + drm_mock_sched_job_set_duration_us(job, 1000); >> + if (prev) >> + drm_sched_job_add_dependency(&job->base, >> + >> dma_fence_get(&prev->base.s_fence->finished)); >> + drm_mock_sched_job_submit(job); >> + if (i == qd / 2) >> + mid = job; >> + prev = job; >> + } >> + >> + done = drm_mock_sched_job_wait_finished(mid, HZ); >> + KUNIT_ASSERT_EQ(test, done, true); >> + >> + /* Exit with half of the queue still pending to be executed. >> */ >> + for (i = 0; i < ARRAY_SIZE(entity); i++) >> + drm_mock_sched_entity_free(entity[i]);} >> + >> +static struct kunit_case drm_sched_basic_tests[] = { >> + KUNIT_CASE(drm_sched_basic_submit), >> + KUNIT_CASE(drm_sched_basic_queue), >> + KUNIT_CASE(drm_sched_basic_chain), >> + KUNIT_CASE(drm_sched_basic_entities), >> + KUNIT_CASE(drm_sched_basic_entities_chain), >> + KUNIT_CASE(drm_sched_basic_entity_cleanup), >> + {} >> +}; >> + >> +static struct kunit_suite drm_sched_basic = { >> + .name = "drm_sched_basic_tests", >> + .init = drm_sched_basic_init, >> + .exit = drm_sched_basic_exit, >> + .test_cases = drm_sched_basic_tests, >> +}; >> + >> +kunit_test_suite(drm_sched_basic); >
On Thu, 2025-02-06 at 10:42 +0000, Tvrtko Ursulin wrote: > > On 06/02/2025 09:51, Philipp Stanner wrote: > > On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote: > > > Implement a mock scheduler backend and add some basic test to > > > exercise the > > > core scheduler code paths. > > > > > > Mock backend (kind of like a very simple mock GPU) can either > > > process > > > jobs > > > by tests manually advancing the "timeline" job at a time, or > > > alternatively > > > jobs can be configured with a time duration in which case they > > > get > > > completed asynchronously from the unit test code. > > > > > > Core scheduler classes are subclassed to support this mock > > > implementation. > > > > > > The tests added are just a few simple submission patterns. > > > > > > Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com> > > > > Could add a Suggested-by: Philipp :) > > Will do. > > > > Cc: Christian König <christian.koenig@amd.com> > > > Cc: Danilo Krummrich <dakr@kernel.org> > > > Cc: Matthew Brost <matthew.brost@intel.com> > > > Cc: Philipp Stanner <phasta@kernel.org> > > > --- > > > drivers/gpu/drm/Kconfig.debug | 12 + > > > drivers/gpu/drm/scheduler/.kunitconfig | 12 + > > > drivers/gpu/drm/scheduler/Makefile | 1 + > > > drivers/gpu/drm/scheduler/tests/Makefile | 4 + > > > .../gpu/drm/scheduler/tests/drm_mock_entity.c | 29 ++ > > > .../gpu/drm/scheduler/tests/drm_mock_job.c | 3 + > > > .../drm/scheduler/tests/drm_mock_scheduler.c | 254 > > > ++++++++++++++++++ > > > .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++ > > > .../scheduler/tests/drm_sched_tests_basic.c | 247 > > > +++++++++++++++++ > > > 9 files changed, 686 insertions(+) > > > create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig > > > create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile > > > create mode 100644 > > > drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > > > create mode 100644 > > > drivers/gpu/drm/scheduler/tests/drm_mock_job.c > > > create mode 100644 > > > drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > > > create mode 100644 > > > drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > > > create mode 100644 > > > drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > > > > > > diff --git a/drivers/gpu/drm/Kconfig.debug > > > b/drivers/gpu/drm/Kconfig.debug > > > index a35d74171b7b..89f777f21e95 100644 > > > --- a/drivers/gpu/drm/Kconfig.debug > > > +++ b/drivers/gpu/drm/Kconfig.debug > > > @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST > > > > > > If in doubt, say "N". > > > > > > +config DRM_SCHED_KUNIT_TEST > > > + tristate "KUnit tests for the DRM scheduler" if > > > !KUNIT_ALL_TESTS > > > + select DRM_SCHED > > > + depends on DRM && KUNIT > > > + default KUNIT_ALL_TESTS > > > + help > > > + Choose this option to build unit tests for the DRM > > > scheduler. > > > > nit: Might say "DRM GPU scheduler" so readers not familiar with all > > that get a better idea of what it's about > > > > > + > > > + Recommended for driver developers only. > > > > nit: s/driver developers/DRM developers > > ? > > > > > + > > > + If in doubt, say "N". > > > + > > > config DRM_EXPORT_FOR_TESTS > > > bool > > > diff --git a/drivers/gpu/drm/scheduler/.kunitconfig > > > b/drivers/gpu/drm/scheduler/.kunitconfig > > > new file mode 100644 > > > index 000000000000..cece53609fcf > > > --- /dev/null > > > +++ b/drivers/gpu/drm/scheduler/.kunitconfig > > > @@ -0,0 +1,12 @@ > > > +CONFIG_KUNIT=y > > > +CONFIG_DRM=y > > > +CONFIG_DRM_SCHED_KUNIT_TEST=y > > > +CONFIG_EXPERT=y > > > +CONFIG_DEBUG_SPINLOCK=y > > > +CONFIG_DEBUG_MUTEXES=y > > > +CONFIG_DEBUG_ATOMIC_SLEEP=y > > > +CONFIG_LOCK_DEBUGGING_SUPPORT=y > > > +CONFIG_PROVE_LOCKING=y > > > +CONFIG_LOCKDEP=y > > > +CONFIG_DEBUG_LOCKDEP=y > > > +CONFIG_DEBUG_LIST=y > > > diff --git a/drivers/gpu/drm/scheduler/Makefile > > > b/drivers/gpu/drm/scheduler/Makefile > > > index 53863621829f..46dfdca0758a 100644 > > > --- a/drivers/gpu/drm/scheduler/Makefile > > > +++ b/drivers/gpu/drm/scheduler/Makefile > > > @@ -23,3 +23,4 @@ > > > gpu-sched-y := sched_main.o sched_fence.o sched_entity.o > > > > > > obj-$(CONFIG_DRM_SCHED) += gpu-sched.o > > > +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/ > > > diff --git a/drivers/gpu/drm/scheduler/tests/Makefile > > > b/drivers/gpu/drm/scheduler/tests/Makefile > > > new file mode 100644 > > > index 000000000000..d69eab6a2e9b > > > --- /dev/null > > > +++ b/drivers/gpu/drm/scheduler/tests/Makefile > > > @@ -0,0 +1,4 @@ > > > + > > > +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \ > > > + drm_mock_scheduler.o \ > > > + drm_sched_tests_basic.o > > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > > > b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > > > new file mode 100644 > > > index 000000000000..c9205f9ff524 > > > --- /dev/null > > > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c > > > @@ -0,0 +1,29 @@ > > > + > > > +#include "drm_sched_tests.h" > > > + > > > +struct drm_mock_sched_entity * > > > +drm_mock_sched_entity_new(struct kunit *test, > > > + enum drm_sched_priority priority, > > > + struct drm_mock_scheduler *sched) > > > > > > I personally do like this function head style and > > > > > > > +{ > > > + struct drm_sched_mock_entity *entity; > > > + int ret; > > > + > > > + entity = kunit_kmalloc(test, sizeof(*entity), > > > GFP_KERNEL); > > > + KUNIT_ASSERT_NOT_NULL(test, entity); > > > + > > > + ret = drm_sched_entity_init(&entity->base, > > > + priority, > > > + &sched->base, 1, > > > + NULL); > > > + KUNIT_ASSERT_EQ(test, ret, 0); > > > + > > > + entity->test = test; > > > + > > > + return entity; > > > +} > > > + > > > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity > > > *entity) > > > > do believe it should then be used consistently everywhere, > > regardless > > of length. > > > > > > > +{ > > > + drm_sched_entity_fini(&entity->base); > > > +} > > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c > > > b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c > > > new file mode 100644 > > > index 000000000000..d94568cf3da9 > > > --- /dev/null > > > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c > > > @@ -0,0 +1,3 @@ > > > + > > > +#include "drm_sched_tests.h" > > > + > > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > > > b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > > > new file mode 100644 > > > index 000000000000..f1985900a6ba > > > --- /dev/null > > > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c > > > @@ -0,0 +1,254 @@ > > > + > > > +#include "drm_sched_tests.h" > > > + > > > +struct drm_mock_sched_entity * > > > +drm_mock_new_sched_entity(struct kunit *test, > > > + enum drm_sched_priority priority, > > > + struct drm_mock_scheduler *sched) > > > +{ > > > + struct drm_mock_sched_entity *entity; > > > + struct drm_gpu_scheduler *drm_sched; > > > + int ret; > > > + > > > + entity = kunit_kzalloc(test, sizeof(*entity), > > > GFP_KERNEL); > > > + KUNIT_ASSERT_NOT_NULL(test, entity); > > > + > > > + drm_sched = &sched->base; > > > + ret = drm_sched_entity_init(&entity->base, > > > + priority, > > > + &drm_sched, 1, > > > + NULL); > > > + KUNIT_ASSERT_EQ(test, ret, 0); > > > + > > > + entity->test = test; > > > + > > > + return entity; > > > +} > > > + > > > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity > > > *entity) > > > +{ > > > + drm_sched_entity_destroy(&entity->base); > > > +} > > > + > > > +static enum hrtimer_restart > > > +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer) > > > > I think we should get up with a consistent naming convention. Some > > things are called drm_mock_sched_, some others drm_sched_mock_. > > As far as I saw, drm_mock_* does not yet exist. > > So do you want to introduce drm_mock as something generic for drm? > > or > > just for drm/sched? > > This one is literally the only when where I tranposed the two. But it > is > also local and I am not too bothered about naming conventions of a > local functions. If it were to me I would favour brevity and emit the > long drm_.. prefixes which IMO do not help readability. I tentatively agree with that. I just think whatever the style is, it should be consistent. Same applies for new files' names > If anything, > seeing something with a long drm_.. prefix can only be confusing > since > one can assume it is some sort of exported interface. > > I will change this instance. > > > > +{ > > > + struct drm_mock_sched_job *upto = > > > + container_of(hrtimer, typeof(*upto), timer); > > > + struct drm_mock_scheduler *sched = > > > + drm_sched_to_mock_sched(upto->base.sched); > > > + struct drm_mock_sched_job *job, *next; > > > + ktime_t now = ktime_get(); > > > + unsigned long flags; > > > + LIST_HEAD(signal); > > > + > > > + spin_lock_irqsave(&sched->lock, flags); > > > + list_for_each_entry_safe(job, next, &sched->job_list, > > > link) > > > { > > > + if (!job->duration_us) > > > + break; > > > + > > > + if (ktime_before(now, job->finish_at)) > > > + break; > > > + > > > + list_move_tail(&job->link, &signal); > > > + sched->seqno = job->hw_fence.seqno; > > > + } > > > + spin_unlock_irqrestore(&sched->lock, flags); > > > + > > > + list_for_each_entry(job, &signal, link) { > > > + dma_fence_signal(&job->hw_fence); > > > + dma_fence_put(&job->hw_fence); > > > + } > > > + > > > + return HRTIMER_NORESTART; > > > +} > > > + > > > +struct drm_mock_sched_job * > > > +drm_mock_new_sched_job(struct kunit *test, > > > + struct drm_mock_sched_entity *entity) > > > +{ > > > + struct drm_mock_sched_job *job; > > > + int ret; > > > + > > > + job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL); > > > + KUNIT_ASSERT_NOT_NULL(test, job); > > > + > > > + ret = drm_sched_job_init(&job->base, > > > + &entity->base, > > > + 1, > > > + NULL); > > > + KUNIT_ASSERT_EQ(test, ret, 0); > > > + > > > + job->test = test; > > > + > > > + spin_lock_init(&job->lock); > > > + INIT_LIST_HEAD(&job->link); > > > + hrtimer_init(&job->timer, CLOCK_MONOTONIC, > > > HRTIMER_MODE_ABS); > > > + job->timer.function = drm_sched_mock_job_signal_timer; > > > + > > > + return job; > > > +} > > > + > > > +static const char *drm_mock_sched_hw_fence_driver_name(struct > > > dma_fence *fence) > > > +{ > > > + return "drm_mock_sched"; > > > +} > > > + > > > +static const char * > > > +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence) > > > +{ > > > + struct drm_mock_sched_job *job = > > > + container_of(fence, typeof(*job), hw_fence); > > > + > > > + return (const char *)job->base.sched->name; > > > +} > > > +static void drm_mock_sched_hw_fence_release(struct dma_fence > > > *fence) > > > +{ > > > + struct drm_mock_sched_job *job = > > > + container_of(fence, typeof(*job), hw_fence); > > > + > > > + hrtimer_cancel(&job->timer); > > > + > > > + /* Freed by the kunit framework */ > > > +} > > > + > > > +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = > > > { > > > + .get_driver_name = drm_mock_sched_hw_fence_driver_name, > > > + .get_timeline_name = > > > drm_mock_sched_hw_fence_timeline_name, > > > + .release = drm_mock_sched_hw_fence_release, > > > +}; > > > + > > > +static struct dma_fence *mock_sched_run_job(struct drm_sched_job > > > *sched_job) > > > +{ > > > + struct drm_mock_scheduler *sched = > > > + drm_sched_to_mock_sched(sched_job->sched); > > > + struct drm_mock_sched_job *job = > > > drm_sched_job_to_mock_job(sched_job); > > > + > > > + dma_fence_init(&job->hw_fence, > > > + &drm_mock_sched_hw_fence_ops, > > > + &job->lock, > > > + sched->hw_fence.context, > > > + atomic_inc_return(&sched- > > > >hw_fence.seqno)); > > > + > > > + dma_fence_get(&job->hw_fence); /* Reference for the > > > job_list > > > */ > > > + > > > + spin_lock_irq(&sched->lock); > > > + if (job->duration_us) { > > > + ktime_t prev_finish_at = 0; > > > + > > > + if (!list_empty(&sched->job_list)) { > > > + struct drm_mock_sched_job *prev = > > > + list_last_entry(&sched- > > > >job_list, > > > typeof(*prev), > > > + link); > > > + > > > + prev_finish_at = prev->finish_at; > > > + } > > > + > > > + if (!prev_finish_at) > > > + prev_finish_at = ktime_get(); > > > + > > > + job->finish_at = ktime_add_us(prev_finish_at, > > > job- > > > > duration_us); > > > + } > > > + list_add_tail(&job->link, &sched->job_list); > > > + if (job->finish_at) > > > + hrtimer_start(&job->timer, job->finish_at, > > > HRTIMER_MODE_ABS); > > > + spin_unlock_irq(&sched->lock); > > > + > > > + return &job->hw_fence; > > > +} > > > + > > > +static enum drm_gpu_sched_stat > > > +mock_sched_timedout_job(struct drm_sched_job *sched_job) > > > +{ > > > + return DRM_GPU_SCHED_STAT_ENODEV; > > > +} > > > + > > > +static void mock_sched_free_job(struct drm_sched_job *sched_job) > > > +{ > > > + drm_sched_job_cleanup(sched_job); > > > +} > > > + > > > +static const struct drm_sched_backend_ops drm_mock_scheduler_ops > > > = { > > > + .run_job = mock_sched_run_job, > > > + .timedout_job = mock_sched_timedout_job, > > > + .free_job = mock_sched_free_job > > > +}; > > > + > > > +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit > > > *test) > > > +{ > > > + struct drm_mock_scheduler *sched; > > > + int ret; > > > + > > > + sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL); > > > + KUNIT_ASSERT_NOT_NULL(test, sched); > > > + > > > + ret = drm_sched_init(&sched->base, > > > + &drm_mock_scheduler_ops, > > > + NULL, /* wq */ > > > + DRM_SCHED_PRIORITY_COUNT, > > > + U32_MAX, /* max credits */ > > > + UINT_MAX, /* hang limit */ > > > + MAX_SCHEDULE_TIMEOUT, /* timeout */ > > > + NULL, /* timeout wq */ > > > + NULL, /* score */ > > > + "drm-mock-scheduler", > > > + NULL /* dev */); > > > + KUNIT_ASSERT_EQ(test, ret, 0); > > > + > > > + sched->test = test; > > > + sched->hw_fence.context = dma_fence_context_alloc(1); > > > + atomic_set(&sched->hw_fence.seqno, 0); > > > + INIT_LIST_HEAD(&sched->job_list); > > > + spin_lock_init(&sched->lock); > > > + > > > + return sched; > > > +} > > > + > > > +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched) > > > +{ > > > + struct drm_mock_sched_job *job, *next; > > > + unsigned long flags; > > > + LIST_HEAD(signal); > > > + > > > + spin_lock_irqsave(&sched->lock, flags); > > > + list_for_each_entry_safe(job, next, &sched->job_list, > > > link) > > > + list_move_tail(&job->link, &signal); > > > + spin_unlock_irqrestore(&sched->lock, flags); > > > > So maybe you can help me get up to speed here a bit. What is the > > purpose behind job->link? Is the general idea documented somewhere? > > List versus link suffix convention I use to distinguish struct > list_head > which is a list versus struct list_head which is used to put > something > on the list. > > In this case job->link is for the mock GPU driver to keep track of > jobs > which have been submitted to the "hardware" for "execution". > > I will of course document these things once the high level design is > settled. > > > > + > > > + list_for_each_entry(job, &signal, link) { > > > + hrtimer_cancel(&job->timer); > > > + dma_fence_put(&job->hw_fence); > > > + } > > > + > > > + drm_sched_fini(&sched->base); > > > +} > > > + > > > +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler > > > *sched, > > > + unsigned int num) > > > +{ > > > + struct drm_mock_sched_job *job, *next; > > > + unsigned int found = 0; > > > + unsigned long flags; > > > + LIST_HEAD(signal); > > > + > > > + spin_lock_irqsave(&sched->lock, flags); > > > + if (WARN_ON_ONCE(sched->seqno + num < sched->seqno)) > > > + goto unlock; > > > + sched->seqno += num; > > > + list_for_each_entry_safe(job, next, &sched->job_list, > > > link) > > > { > > > + if (sched->seqno < job->hw_fence.seqno) > > > + break; > > > + > > > + list_move_tail(&job->link, &signal); > > > + found++; > > > + } > > > +unlock: > > > + spin_unlock_irqrestore(&sched->lock, flags); > > > + > > > + list_for_each_entry(job, &signal, link) { > > > + dma_fence_signal(&job->hw_fence); > > > + dma_fence_put(&job->hw_fence); > > > + } > > > + > > > + return found; > > > +} > > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > > > b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > > > new file mode 100644 > > > index 000000000000..421ee2712985 > > > --- /dev/null > > > +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h > > > @@ -0,0 +1,124 @@ > > > + > > > +#include <kunit/test.h> > > > +#include <linux/atomic.h> > > > +#include <linux/dma-fence.h> > > > +#include <linux/hrtimer.h> > > > +#include <linux/ktime.h> > > > +#include <linux/list.h> > > > +#include <linux/atomic.h> > > > +#include <linux/mutex.h> > > > +#include <linux/types.h> > > > + > > > +#include <drm/gpu_scheduler.h> > > > + > > > +struct drm_mock_scheduler { > > > + struct drm_gpu_scheduler base; > > > + > > > + struct kunit *test; > > > + > > > + struct { > > > + u64 context; > > > + atomic_t seqno; > > > + } hw_fence; > > > > Hm, well, so this is confusing. drm_mock_sched_job below contains > > an > > actual struct dma_fence that is also called hw_fence, whereas this > > here > > seems to be some pseudo-hw_fence? > > > > What is its purpose? > > > > I believe we agreed that "Hardware fence" should always mean a > > fence > > per job that is signaled by the hardware (driver interrupt) once > > the > > job is done. > > > > If this hw_fence here is the same, why is it per scheduler? That's > > confusing. > > Mock_job->hw_fence is what the mock GPU driver returns from the > sched->run_job(). > > Mock_sched->hw_fence is what the mock GPU driver uses to track > execution > of jobs submitted to it for execution. If Irename this to hw_timeline > will it make it clearer? How about "current_hw_fence"? > > > > + > > > + spinlock_t lock; > > > > Maybe document the lock's purpose > > > > > > > + unsigned int seqno; > > > + struct list_head job_list; > > > +}; > > > + > > > +struct drm_mock_sched_entity { > > > + struct drm_sched_entity base; > > > + > > > + struct kunit *test; > > > +}; > > > + > > > +struct drm_mock_sched_job { > > > + struct drm_sched_job base; > > > + > > > + struct list_head link; > > > + struct hrtimer timer; > > > + > > > + unsigned int duration_us; > > > + ktime_t finish_at; > > > + > > > + spinlock_t lock; > > > > Same > > > > > + struct dma_fence hw_fence; > > > + > > > + struct kunit *test; > > > +}; > > > + > > > +static inline struct drm_mock_scheduler * > > > +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched) > > > +{ > > > + return container_of(sched, struct drm_mock_scheduler, > > > base); > > > +}; > > > + > > > +static inline struct drm_mock_sched_entity * > > > +drm_sched_entity_to_mock_entity(struct drm_sched_entity > > > *sched_entity) > > > +{ > > > + return container_of(sched_entity, struct > > > drm_mock_sched_entity, base); > > > +}; > > > + > > > +static inline struct drm_mock_sched_job * > > > +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job) > > > +{ > > > + return container_of(sched_job, struct > > > drm_mock_sched_job, > > > base); > > > +}; > > > + > > > +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit > > > *test); > > > +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched); > > > +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler > > > *sched, > > > + unsigned int num); > > > + > > > +struct drm_mock_sched_entity * > > > +drm_mock_new_sched_entity(struct kunit *test, > > > + enum drm_sched_priority priority, > > > + struct drm_mock_scheduler *sched); > > > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity > > > *entity); > > > + > > > +struct drm_mock_sched_job * > > > +drm_mock_new_sched_job(struct kunit *test, > > > + struct drm_mock_sched_entity *entity); > > > + > > > +static inline void drm_mock_sched_job_submit(struct > > > drm_mock_sched_job *job) > > > +{ > > > + drm_sched_job_arm(&job->base); > > > + drm_sched_entity_push_job(&job->base); > > > +} > > > + > > > +static inline void > > > +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job > > > *job, > > > + unsigned int duration_us) > > > +{ > > > + job->duration_us = duration_us; > > > +} > > > + > > > +static inline bool > > > +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job) > > > +{ > > > + return dma_fence_is_signaled(&job->base.s_fence- > > > >finished); > > > +} > > > + > > > +static inline bool > > > +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job, > > > long timeout) > > > +{ > > > + long ret; > > > + > > > + ret = dma_fence_wait_timeout(&job->base.s_fence- > > > >finished, > > > + false, > > > + timeout); > > > + > > > + return ret != 0; > > > +} > > > + > > > +static inline long > > > +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job > > > *job, > > > long timeout) > > > +{ > > > + long ret; > > > + > > > + ret = dma_fence_wait_timeout(&job->base.s_fence- > > > >scheduled, > > > + false, > > > + timeout); > > > + > > > + return ret != 0; > > > +} > > > diff --git > > > a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > > > b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > > > new file mode 100644 > > > index 000000000000..6fd39bea95b1 > > > --- /dev/null > > > +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c > > > @@ -0,0 +1,247 @@ > > > + > > > +#include "drm_sched_tests.h" > > > + > > > +static int drm_sched_basic_init(struct kunit *test) > > > +{ > > > + test->priv = drm_mock_new_scheduler(test); > > > + > > > + return 0; > > > +} > > > + > > > +static void drm_sched_basic_exit(struct kunit *test) > > > +{ > > > + struct drm_mock_scheduler *sched = test->priv; > > > + > > > + drm_mock_scheduler_fini(sched); > > > +} > > > + > > > +static void drm_sched_basic_submit(struct kunit *test) > > > +{ > > > + struct drm_mock_scheduler *sched = test->priv; > > > + struct drm_mock_sched_entity *entity; > > > + struct drm_mock_sched_job *job; > > > + unsigned int i; > > > + bool done; > > > + > > > + /* > > > + * Submit one job to the scheduler and verify that it > > > gets > > > scheduled > > > + * and completed only when the mock hw backend processes > > > it. > > > + */ > > > + > > > + entity = drm_mock_new_sched_entity(test, > > > + > > > DRM_SCHED_PRIORITY_NORMAL, > > > + sched); > > > + job = drm_mock_new_sched_job(test, entity); > > > + > > > + drm_mock_sched_job_submit(job); > > > + > > > + done = drm_mock_sched_job_wait_scheduled(job, HZ); > > > + KUNIT_ASSERT_EQ(test, done, true); > > > + > > > + done = drm_mock_sched_job_wait_finished(job, HZ / 2); > > > + KUNIT_ASSERT_EQ(test, done, false); > > > + > > > + i = drm_mock_sched_advance(sched, 1); > > > + KUNIT_ASSERT_EQ(test, i, 1); > > > + > > > + done = drm_mock_sched_job_wait_finished(job, HZ); > > > + KUNIT_ASSERT_EQ(test, done, true); > > > + > > > + drm_mock_sched_entity_free(entity); > > > +} > > > + > > > +static void drm_sched_basic_queue(struct kunit *test) > > > +{ > > > + struct drm_mock_scheduler *sched = test->priv; > > > + struct drm_mock_sched_entity *entity; > > > + struct drm_mock_sched_job *job; > > > + const unsigned int qd = 100; > > > > Not the best variable name – is this "queue depth"? Why is it 100? > > -> > > global define & document > > I wouldn't promote this to global and TBH if you look how small this > test function is it feels pretty self documenting. I meant global in the sense of at the top of the file as a constant. Wouldn't you agree that it would be good to have test parameters at a place where you can easily modify them, in case you want to compile the tests a bit differently? P. > > > > > > > > + unsigned int i; > > > + bool done; > > > + > > > + /* > > > + * Submit a queue of jobs on the same entity, let them > > > be > > > completed by > > > + * the mock hw backend and check the status of the last > > > job. > > > + */ > > > + > > > + entity = drm_mock_new_sched_entity(test, > > > + > > > DRM_SCHED_PRIORITY_NORMAL, > > > + sched); > > > + > > > + for (i = 0; i < qd; i++) { > > > + job = drm_mock_new_sched_job(test, entity); > > > + drm_mock_sched_job_set_duration_us(job, 1000); > > > + drm_mock_sched_job_submit(job); > > > + } > > > + > > > + done = drm_mock_sched_job_wait_finished(job, HZ); > > > + KUNIT_ASSERT_EQ(test, done, true); > > > + > > > + drm_mock_sched_entity_free(entity); > > > +} > > > + > > > +static void drm_sched_basic_chain(struct kunit *test) > > > +{ > > > + struct drm_mock_scheduler *sched = test->priv; > > > + struct drm_mock_sched_job *job, *prev = NULL; > > > + struct drm_mock_sched_entity *entity; > > > + const unsigned int qd = 100; > > > + unsigned int i; > > > + bool done; > > > + > > > + /* > > > + * Submit a queue of jobs on the same entity but with an > > > explicit > > > + * chain of dependencies between them. Let the jobs be > > > completed by > > > + * the mock hw backend and check the status of the last > > > job. > > > + */ > > > + > > > + entity = drm_mock_new_sched_entity(test, > > > + > > > DRM_SCHED_PRIORITY_NORMAL, > > > + sched); > > > + > > > + for (i = 0; i < qd; i++) { > > > + job = drm_mock_new_sched_job(test, entity); > > > + drm_mock_sched_job_set_duration_us(job, 1000); > > > + if (prev) > > > + drm_sched_job_add_dependency(&job->base, > > > + > > > dma_fence_get(&prev->base.s_fence->finished)); > > > + drm_mock_sched_job_submit(job); > > > + prev = job; > > > + } > > > + > > > + done = drm_mock_sched_job_wait_finished(job, HZ); > > > + KUNIT_ASSERT_EQ(test, done, true); > > > + > > > + drm_mock_sched_entity_free(entity); > > > +} > > > + > > > +static void drm_sched_basic_entities(struct kunit *test) > > > +{ > > > + struct drm_mock_scheduler *sched = test->priv; > > > + struct drm_mock_sched_entity *entity[4]; > > > + struct drm_mock_sched_job *job; > > > + const unsigned int qd = 100; > > > + unsigned int i, cur_ent = 0; > > > + bool done; > > > + > > > + /* > > > + * Submit a queue of jobs across different entities, let > > > them be > > > + * completed by the mock hw backend and check the status > > > of > > > the last > > > + * job. > > > + */ > > > + > > > + for (i = 0; i < ARRAY_SIZE(entity); i++) > > > + entity[i] = drm_mock_new_sched_entity(test, > > > + > > > DRM_SCHED_PRIORITY_NORMAL, > > > + sched); > > > + > > > + for (i = 0; i < qd; i++) { > > > + job = drm_mock_new_sched_job(test, > > > entity[cur_ent++]); > > > + cur_ent %= ARRAY_SIZE(entity); > > > + drm_mock_sched_job_set_duration_us(job, 1000); > > > + drm_mock_sched_job_submit(job); > > > + } > > > + > > > + done = drm_mock_sched_job_wait_finished(job, HZ); > > > + KUNIT_ASSERT_EQ(test, done, true); > > > + > > > + for (i = 0; i < ARRAY_SIZE(entity); i++) > > > + drm_mock_sched_entity_free(entity[i]); > > > +} > > > + > > > +static void drm_sched_basic_entities_chain(struct kunit *test) > > > +{ > > > + struct drm_mock_scheduler *sched = test->priv; > > > + struct drm_mock_sched_job *job, *prev = NULL; > > > + struct drm_mock_sched_entity *entity[4]; > > > + const unsigned int qd = 100; > > > + unsigned int i, cur_ent = 0; > > > + bool done; > > > + > > > + /* > > > + * Submit a queue of jobs across different entities with > > > an > > > explicit > > > + * chain of dependencies between them. Let the jobs be > > > completed by > > > + * the mock hw backend and check the status of the last > > > job. > > > + */ > > > + > > > + for (i = 0; i < ARRAY_SIZE(entity); i++) > > > + entity[i] = drm_mock_new_sched_entity(test, > > > + > > > DRM_SCHED_PRIORITY_NORMAL, > > > + sched); > > > + > > > + for (i = 0; i < qd; i++) { > > > + job = drm_mock_new_sched_job(test, > > > entity[cur_ent++]); > > > + cur_ent %= ARRAY_SIZE(entity); > > > + drm_mock_sched_job_set_duration_us(job, 1000); > > > + if (prev) > > > + drm_sched_job_add_dependency(&job->base, > > > + > > > dma_fence_get(&prev->base.s_fence->finished)); > > > + drm_mock_sched_job_submit(job); > > > + prev = job; > > > + } > > > + > > > + done = drm_mock_sched_job_wait_finished(job, HZ); > > > + KUNIT_ASSERT_EQ(test, done, true); > > > + > > > + for (i = 0; i < ARRAY_SIZE(entity); i++) > > > + drm_mock_sched_entity_free(entity[i]); > > > +} > > > + > > > +static void drm_sched_basic_entity_cleanup(struct kunit *test) > > > +{ > > > + struct drm_mock_sched_job *job, *mid, *prev = NULL; > > > + struct drm_mock_scheduler *sched = test->priv; > > > + struct drm_mock_sched_entity *entity[4]; > > > > 4 is used in several places, so could be defined globally. In case > > there's a special reason for why it's 4, that could be mentioned, > > too > > It's completely arbitrary. In this test we just need a bunch of > entities > to be active on the scheduler as we trigger a client exit in the > middle > of it. IMO the comment few lines below should be good enough to > explain > the idea. I fear that promoting this to a global define would just > give > it more weight that what it has. Then some test would want a > different > number etc. > > Regards, > > Tvrtko > > > > + const unsigned int qd = 100; > > > + unsigned int i, cur_ent = 0; > > > + bool done; > > > + > > > + /* > > > + * Submit a queue of jobs across different entities with > > > an > > > explicit > > > + * chain of dependencies between them and trigger entity > > > cleanup while > > > + * the queue is still being processed. > > > + */ > > > + > > > + for (i = 0; i < ARRAY_SIZE(entity); i++) > > > + entity[i] = drm_mock_new_sched_entity(test, > > > + > > > DRM_SCHED_PRIORITY_NORMAL, > > > + sched); > > > + > > > + for (i = 0; i < qd; i++) { > > > + job = drm_mock_new_sched_job(test, > > > entity[cur_ent++]); > > > + cur_ent %= ARRAY_SIZE(entity); > > > + drm_mock_sched_job_set_duration_us(job, 1000); > > > + if (prev) > > > + drm_sched_job_add_dependency(&job->base, > > > + > > > dma_fence_get(&prev->base.s_fence->finished)); > > > + drm_mock_sched_job_submit(job); > > > + if (i == qd / 2) > > > + mid = job; > > > + prev = job; > > > + } > > > + > > > + done = drm_mock_sched_job_wait_finished(mid, HZ); > > > + KUNIT_ASSERT_EQ(test, done, true); > > > + > > > + /* Exit with half of the queue still pending to be > > > executed. > > > */ > > > + for (i = 0; i < ARRAY_SIZE(entity); i++) > > > + drm_mock_sched_entity_free(entity[i]);} > > > + > > > +static struct kunit_case drm_sched_basic_tests[] = { > > > + KUNIT_CASE(drm_sched_basic_submit), > > > + KUNIT_CASE(drm_sched_basic_queue), > > > + KUNIT_CASE(drm_sched_basic_chain), > > > + KUNIT_CASE(drm_sched_basic_entities), > > > + KUNIT_CASE(drm_sched_basic_entities_chain), > > > + KUNIT_CASE(drm_sched_basic_entity_cleanup), > > > + {} > > > +}; > > > + > > > +static struct kunit_suite drm_sched_basic = { > > > + .name = "drm_sched_basic_tests", > > > + .init = drm_sched_basic_init, > > > + .exit = drm_sched_basic_exit, > > > + .test_cases = drm_sched_basic_tests, > > > +}; > > > + > > > +kunit_test_suite(drm_sched_basic); > >
On 06/02/2025 12:33, Philipp Stanner wrote: > On Thu, 2025-02-06 at 10:42 +0000, Tvrtko Ursulin wrote: >> >> On 06/02/2025 09:51, Philipp Stanner wrote: >>> On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote: >>>> Implement a mock scheduler backend and add some basic test to >>>> exercise the >>>> core scheduler code paths. >>>> >>>> Mock backend (kind of like a very simple mock GPU) can either >>>> process >>>> jobs >>>> by tests manually advancing the "timeline" job at a time, or >>>> alternatively >>>> jobs can be configured with a time duration in which case they >>>> get >>>> completed asynchronously from the unit test code. >>>> >>>> Core scheduler classes are subclassed to support this mock >>>> implementation. >>>> >>>> The tests added are just a few simple submission patterns. >>>> >>>> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com> >>> >>> Could add a Suggested-by: Philipp :) >> >> Will do. >> >>>> Cc: Christian König <christian.koenig@amd.com> >>>> Cc: Danilo Krummrich <dakr@kernel.org> >>>> Cc: Matthew Brost <matthew.brost@intel.com> >>>> Cc: Philipp Stanner <phasta@kernel.org> >>>> --- >>>> drivers/gpu/drm/Kconfig.debug | 12 + >>>> drivers/gpu/drm/scheduler/.kunitconfig | 12 + >>>> drivers/gpu/drm/scheduler/Makefile | 1 + >>>> drivers/gpu/drm/scheduler/tests/Makefile | 4 + >>>> .../gpu/drm/scheduler/tests/drm_mock_entity.c | 29 ++ >>>> .../gpu/drm/scheduler/tests/drm_mock_job.c | 3 + >>>> .../drm/scheduler/tests/drm_mock_scheduler.c | 254 >>>> ++++++++++++++++++ >>>> .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++ >>>> .../scheduler/tests/drm_sched_tests_basic.c | 247 >>>> +++++++++++++++++ >>>> 9 files changed, 686 insertions(+) >>>> create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig >>>> create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile >>>> create mode 100644 >>>> drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >>>> create mode 100644 >>>> drivers/gpu/drm/scheduler/tests/drm_mock_job.c >>>> create mode 100644 >>>> drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >>>> create mode 100644 >>>> drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >>>> create mode 100644 >>>> drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >>>> >>>> diff --git a/drivers/gpu/drm/Kconfig.debug >>>> b/drivers/gpu/drm/Kconfig.debug >>>> index a35d74171b7b..89f777f21e95 100644 >>>> --- a/drivers/gpu/drm/Kconfig.debug >>>> +++ b/drivers/gpu/drm/Kconfig.debug >>>> @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST >>>> >>>> If in doubt, say "N". >>>> >>>> +config DRM_SCHED_KUNIT_TEST >>>> + tristate "KUnit tests for the DRM scheduler" if >>>> !KUNIT_ALL_TESTS >>>> + select DRM_SCHED >>>> + depends on DRM && KUNIT >>>> + default KUNIT_ALL_TESTS >>>> + help >>>> + Choose this option to build unit tests for the DRM >>>> scheduler. >>> >>> nit: Might say "DRM GPU scheduler" so readers not familiar with all >>> that get a better idea of what it's about >>> >>>> + >>>> + Recommended for driver developers only. >>> >>> nit: s/driver developers/DRM developers >>> ? >>> >>>> + >>>> + If in doubt, say "N". >>>> + >>>> config DRM_EXPORT_FOR_TESTS >>>> bool >>>> diff --git a/drivers/gpu/drm/scheduler/.kunitconfig >>>> b/drivers/gpu/drm/scheduler/.kunitconfig >>>> new file mode 100644 >>>> index 000000000000..cece53609fcf >>>> --- /dev/null >>>> +++ b/drivers/gpu/drm/scheduler/.kunitconfig >>>> @@ -0,0 +1,12 @@ >>>> +CONFIG_KUNIT=y >>>> +CONFIG_DRM=y >>>> +CONFIG_DRM_SCHED_KUNIT_TEST=y >>>> +CONFIG_EXPERT=y >>>> +CONFIG_DEBUG_SPINLOCK=y >>>> +CONFIG_DEBUG_MUTEXES=y >>>> +CONFIG_DEBUG_ATOMIC_SLEEP=y >>>> +CONFIG_LOCK_DEBUGGING_SUPPORT=y >>>> +CONFIG_PROVE_LOCKING=y >>>> +CONFIG_LOCKDEP=y >>>> +CONFIG_DEBUG_LOCKDEP=y >>>> +CONFIG_DEBUG_LIST=y >>>> diff --git a/drivers/gpu/drm/scheduler/Makefile >>>> b/drivers/gpu/drm/scheduler/Makefile >>>> index 53863621829f..46dfdca0758a 100644 >>>> --- a/drivers/gpu/drm/scheduler/Makefile >>>> +++ b/drivers/gpu/drm/scheduler/Makefile >>>> @@ -23,3 +23,4 @@ >>>> gpu-sched-y := sched_main.o sched_fence.o sched_entity.o >>>> >>>> obj-$(CONFIG_DRM_SCHED) += gpu-sched.o >>>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/ >>>> diff --git a/drivers/gpu/drm/scheduler/tests/Makefile >>>> b/drivers/gpu/drm/scheduler/tests/Makefile >>>> new file mode 100644 >>>> index 000000000000..d69eab6a2e9b >>>> --- /dev/null >>>> +++ b/drivers/gpu/drm/scheduler/tests/Makefile >>>> @@ -0,0 +1,4 @@ >>>> + >>>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \ >>>> + drm_mock_scheduler.o \ >>>> + drm_sched_tests_basic.o >>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >>>> b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >>>> new file mode 100644 >>>> index 000000000000..c9205f9ff524 >>>> --- /dev/null >>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c >>>> @@ -0,0 +1,29 @@ >>>> + >>>> +#include "drm_sched_tests.h" >>>> + >>>> +struct drm_mock_sched_entity * >>>> +drm_mock_sched_entity_new(struct kunit *test, >>>> + enum drm_sched_priority priority, >>>> + struct drm_mock_scheduler *sched) >>> >>> >>> I personally do like this function head style and >>> >>> >>>> +{ >>>> + struct drm_sched_mock_entity *entity; >>>> + int ret; >>>> + >>>> + entity = kunit_kmalloc(test, sizeof(*entity), >>>> GFP_KERNEL); >>>> + KUNIT_ASSERT_NOT_NULL(test, entity); >>>> + >>>> + ret = drm_sched_entity_init(&entity->base, >>>> + priority, >>>> + &sched->base, 1, >>>> + NULL); >>>> + KUNIT_ASSERT_EQ(test, ret, 0); >>>> + >>>> + entity->test = test; >>>> + >>>> + return entity; >>>> +} >>>> + >>>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity >>>> *entity) >>> >>> do believe it should then be used consistently everywhere, >>> regardless >>> of length. >>> >>> >>>> +{ >>>> + drm_sched_entity_fini(&entity->base); >>>> +} >>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c >>>> b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c >>>> new file mode 100644 >>>> index 000000000000..d94568cf3da9 >>>> --- /dev/null >>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c >>>> @@ -0,0 +1,3 @@ >>>> + >>>> +#include "drm_sched_tests.h" >>>> + >>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >>>> b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >>>> new file mode 100644 >>>> index 000000000000..f1985900a6ba >>>> --- /dev/null >>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c >>>> @@ -0,0 +1,254 @@ >>>> + >>>> +#include "drm_sched_tests.h" >>>> + >>>> +struct drm_mock_sched_entity * >>>> +drm_mock_new_sched_entity(struct kunit *test, >>>> + enum drm_sched_priority priority, >>>> + struct drm_mock_scheduler *sched) >>>> +{ >>>> + struct drm_mock_sched_entity *entity; >>>> + struct drm_gpu_scheduler *drm_sched; >>>> + int ret; >>>> + >>>> + entity = kunit_kzalloc(test, sizeof(*entity), >>>> GFP_KERNEL); >>>> + KUNIT_ASSERT_NOT_NULL(test, entity); >>>> + >>>> + drm_sched = &sched->base; >>>> + ret = drm_sched_entity_init(&entity->base, >>>> + priority, >>>> + &drm_sched, 1, >>>> + NULL); >>>> + KUNIT_ASSERT_EQ(test, ret, 0); >>>> + >>>> + entity->test = test; >>>> + >>>> + return entity; >>>> +} >>>> + >>>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity >>>> *entity) >>>> +{ >>>> + drm_sched_entity_destroy(&entity->base); >>>> +} >>>> + >>>> +static enum hrtimer_restart >>>> +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer) >>> >>> I think we should get up with a consistent naming convention. Some >>> things are called drm_mock_sched_, some others drm_sched_mock_. >>> As far as I saw, drm_mock_* does not yet exist. >>> So do you want to introduce drm_mock as something generic for drm? >>> or >>> just for drm/sched? >> >> This one is literally the only when where I tranposed the two. But it >> is >> also local and I am not too bothered about naming conventions of a >> local functions. If it were to me I would favour brevity and emit the >> long drm_.. prefixes which IMO do not help readability. > > I tentatively agree with that. I just think whatever the style is, it > should be consistent. > > Same applies for new files' names > > >> If anything, >> seeing something with a long drm_.. prefix can only be confusing >> since >> one can assume it is some sort of exported interface. >> >> I will change this instance. >> >>>> +{ >>>> + struct drm_mock_sched_job *upto = >>>> + container_of(hrtimer, typeof(*upto), timer); >>>> + struct drm_mock_scheduler *sched = >>>> + drm_sched_to_mock_sched(upto->base.sched); >>>> + struct drm_mock_sched_job *job, *next; >>>> + ktime_t now = ktime_get(); >>>> + unsigned long flags; >>>> + LIST_HEAD(signal); >>>> + >>>> + spin_lock_irqsave(&sched->lock, flags); >>>> + list_for_each_entry_safe(job, next, &sched->job_list, >>>> link) >>>> { >>>> + if (!job->duration_us) >>>> + break; >>>> + >>>> + if (ktime_before(now, job->finish_at)) >>>> + break; >>>> + >>>> + list_move_tail(&job->link, &signal); >>>> + sched->seqno = job->hw_fence.seqno; >>>> + } >>>> + spin_unlock_irqrestore(&sched->lock, flags); >>>> + >>>> + list_for_each_entry(job, &signal, link) { >>>> + dma_fence_signal(&job->hw_fence); >>>> + dma_fence_put(&job->hw_fence); >>>> + } >>>> + >>>> + return HRTIMER_NORESTART; >>>> +} >>>> + >>>> +struct drm_mock_sched_job * >>>> +drm_mock_new_sched_job(struct kunit *test, >>>> + struct drm_mock_sched_entity *entity) >>>> +{ >>>> + struct drm_mock_sched_job *job; >>>> + int ret; >>>> + >>>> + job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL); >>>> + KUNIT_ASSERT_NOT_NULL(test, job); >>>> + >>>> + ret = drm_sched_job_init(&job->base, >>>> + &entity->base, >>>> + 1, >>>> + NULL); >>>> + KUNIT_ASSERT_EQ(test, ret, 0); >>>> + >>>> + job->test = test; >>>> + >>>> + spin_lock_init(&job->lock); >>>> + INIT_LIST_HEAD(&job->link); >>>> + hrtimer_init(&job->timer, CLOCK_MONOTONIC, >>>> HRTIMER_MODE_ABS); >>>> + job->timer.function = drm_sched_mock_job_signal_timer; >>>> + >>>> + return job; >>>> +} >>>> + >>>> +static const char *drm_mock_sched_hw_fence_driver_name(struct >>>> dma_fence *fence) >>>> +{ >>>> + return "drm_mock_sched"; >>>> +} >>>> + >>>> +static const char * >>>> +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence) >>>> +{ >>>> + struct drm_mock_sched_job *job = >>>> + container_of(fence, typeof(*job), hw_fence); >>>> + >>>> + return (const char *)job->base.sched->name; >>>> +} >>>> +static void drm_mock_sched_hw_fence_release(struct dma_fence >>>> *fence) >>>> +{ >>>> + struct drm_mock_sched_job *job = >>>> + container_of(fence, typeof(*job), hw_fence); >>>> + >>>> + hrtimer_cancel(&job->timer); >>>> + >>>> + /* Freed by the kunit framework */ >>>> +} >>>> + >>>> +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = >>>> { >>>> + .get_driver_name = drm_mock_sched_hw_fence_driver_name, >>>> + .get_timeline_name = >>>> drm_mock_sched_hw_fence_timeline_name, >>>> + .release = drm_mock_sched_hw_fence_release, >>>> +}; >>>> + >>>> +static struct dma_fence *mock_sched_run_job(struct drm_sched_job >>>> *sched_job) >>>> +{ >>>> + struct drm_mock_scheduler *sched = >>>> + drm_sched_to_mock_sched(sched_job->sched); >>>> + struct drm_mock_sched_job *job = >>>> drm_sched_job_to_mock_job(sched_job); >>>> + >>>> + dma_fence_init(&job->hw_fence, >>>> + &drm_mock_sched_hw_fence_ops, >>>> + &job->lock, >>>> + sched->hw_fence.context, >>>> + atomic_inc_return(&sched- >>>>> hw_fence.seqno)); >>>> + >>>> + dma_fence_get(&job->hw_fence); /* Reference for the >>>> job_list >>>> */ >>>> + >>>> + spin_lock_irq(&sched->lock); >>>> + if (job->duration_us) { >>>> + ktime_t prev_finish_at = 0; >>>> + >>>> + if (!list_empty(&sched->job_list)) { >>>> + struct drm_mock_sched_job *prev = >>>> + list_last_entry(&sched- >>>>> job_list, >>>> typeof(*prev), >>>> + link); >>>> + >>>> + prev_finish_at = prev->finish_at; >>>> + } >>>> + >>>> + if (!prev_finish_at) >>>> + prev_finish_at = ktime_get(); >>>> + >>>> + job->finish_at = ktime_add_us(prev_finish_at, >>>> job- >>>>> duration_us); >>>> + } >>>> + list_add_tail(&job->link, &sched->job_list); >>>> + if (job->finish_at) >>>> + hrtimer_start(&job->timer, job->finish_at, >>>> HRTIMER_MODE_ABS); >>>> + spin_unlock_irq(&sched->lock); >>>> + >>>> + return &job->hw_fence; >>>> +} >>>> + >>>> +static enum drm_gpu_sched_stat >>>> +mock_sched_timedout_job(struct drm_sched_job *sched_job) >>>> +{ >>>> + return DRM_GPU_SCHED_STAT_ENODEV; >>>> +} >>>> + >>>> +static void mock_sched_free_job(struct drm_sched_job *sched_job) >>>> +{ >>>> + drm_sched_job_cleanup(sched_job); >>>> +} >>>> + >>>> +static const struct drm_sched_backend_ops drm_mock_scheduler_ops >>>> = { >>>> + .run_job = mock_sched_run_job, >>>> + .timedout_job = mock_sched_timedout_job, >>>> + .free_job = mock_sched_free_job >>>> +}; >>>> + >>>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit >>>> *test) >>>> +{ >>>> + struct drm_mock_scheduler *sched; >>>> + int ret; >>>> + >>>> + sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL); >>>> + KUNIT_ASSERT_NOT_NULL(test, sched); >>>> + >>>> + ret = drm_sched_init(&sched->base, >>>> + &drm_mock_scheduler_ops, >>>> + NULL, /* wq */ >>>> + DRM_SCHED_PRIORITY_COUNT, >>>> + U32_MAX, /* max credits */ >>>> + UINT_MAX, /* hang limit */ >>>> + MAX_SCHEDULE_TIMEOUT, /* timeout */ >>>> + NULL, /* timeout wq */ >>>> + NULL, /* score */ >>>> + "drm-mock-scheduler", >>>> + NULL /* dev */); >>>> + KUNIT_ASSERT_EQ(test, ret, 0); >>>> + >>>> + sched->test = test; >>>> + sched->hw_fence.context = dma_fence_context_alloc(1); >>>> + atomic_set(&sched->hw_fence.seqno, 0); >>>> + INIT_LIST_HEAD(&sched->job_list); >>>> + spin_lock_init(&sched->lock); >>>> + >>>> + return sched; >>>> +} >>>> + >>>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched) >>>> +{ >>>> + struct drm_mock_sched_job *job, *next; >>>> + unsigned long flags; >>>> + LIST_HEAD(signal); >>>> + >>>> + spin_lock_irqsave(&sched->lock, flags); >>>> + list_for_each_entry_safe(job, next, &sched->job_list, >>>> link) >>>> + list_move_tail(&job->link, &signal); >>>> + spin_unlock_irqrestore(&sched->lock, flags); >>> >>> So maybe you can help me get up to speed here a bit. What is the >>> purpose behind job->link? Is the general idea documented somewhere? >> >> List versus link suffix convention I use to distinguish struct >> list_head >> which is a list versus struct list_head which is used to put >> something >> on the list. >> >> In this case job->link is for the mock GPU driver to keep track of >> jobs >> which have been submitted to the "hardware" for "execution". >> >> I will of course document these things once the high level design is >> settled. >> >>>> + >>>> + list_for_each_entry(job, &signal, link) { >>>> + hrtimer_cancel(&job->timer); >>>> + dma_fence_put(&job->hw_fence); >>>> + } >>>> + >>>> + drm_sched_fini(&sched->base); >>>> +} >>>> + >>>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler >>>> *sched, >>>> + unsigned int num) >>>> +{ >>>> + struct drm_mock_sched_job *job, *next; >>>> + unsigned int found = 0; >>>> + unsigned long flags; >>>> + LIST_HEAD(signal); >>>> + >>>> + spin_lock_irqsave(&sched->lock, flags); >>>> + if (WARN_ON_ONCE(sched->seqno + num < sched->seqno)) >>>> + goto unlock; >>>> + sched->seqno += num; >>>> + list_for_each_entry_safe(job, next, &sched->job_list, >>>> link) >>>> { >>>> + if (sched->seqno < job->hw_fence.seqno) >>>> + break; >>>> + >>>> + list_move_tail(&job->link, &signal); >>>> + found++; >>>> + } >>>> +unlock: >>>> + spin_unlock_irqrestore(&sched->lock, flags); >>>> + >>>> + list_for_each_entry(job, &signal, link) { >>>> + dma_fence_signal(&job->hw_fence); >>>> + dma_fence_put(&job->hw_fence); >>>> + } >>>> + >>>> + return found; >>>> +} >>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >>>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >>>> new file mode 100644 >>>> index 000000000000..421ee2712985 >>>> --- /dev/null >>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h >>>> @@ -0,0 +1,124 @@ >>>> + >>>> +#include <kunit/test.h> >>>> +#include <linux/atomic.h> >>>> +#include <linux/dma-fence.h> >>>> +#include <linux/hrtimer.h> >>>> +#include <linux/ktime.h> >>>> +#include <linux/list.h> >>>> +#include <linux/atomic.h> >>>> +#include <linux/mutex.h> >>>> +#include <linux/types.h> >>>> + >>>> +#include <drm/gpu_scheduler.h> >>>> + >>>> +struct drm_mock_scheduler { >>>> + struct drm_gpu_scheduler base; >>>> + >>>> + struct kunit *test; >>>> + >>>> + struct { >>>> + u64 context; >>>> + atomic_t seqno; >>>> + } hw_fence; >>> >>> Hm, well, so this is confusing. drm_mock_sched_job below contains >>> an >>> actual struct dma_fence that is also called hw_fence, whereas this >>> here >>> seems to be some pseudo-hw_fence? >>> >>> What is its purpose? >>> >>> I believe we agreed that "Hardware fence" should always mean a >>> fence >>> per job that is signaled by the hardware (driver interrupt) once >>> the >>> job is done. >>> >>> If this hw_fence here is the same, why is it per scheduler? That's >>> confusing. >> >> Mock_job->hw_fence is what the mock GPU driver returns from the >> sched->run_job(). >> >> Mock_sched->hw_fence is what the mock GPU driver uses to track >> execution >> of jobs submitted to it for execution. If Irename this to hw_timeline >> will it make it clearer? > > How about "current_hw_fence"? I prefer timeline. The current working version is: struct drm_mock_scheduler { struct drm_gpu_scheduler base; struct kunit *test; spinlock_t lock; struct list_head job_list; /* Protected by the lock */ struct { u64 context; atomic_t next_seqno; unsigned int cur_seqno; /* Protected by the lock */ } hw_timeline; }; >>>> + >>>> + spinlock_t lock; >>> >>> Maybe document the lock's purpose >>> >>> >>>> + unsigned int seqno; >>>> + struct list_head job_list; >>>> +}; >>>> + >>>> +struct drm_mock_sched_entity { >>>> + struct drm_sched_entity base; >>>> + >>>> + struct kunit *test; >>>> +}; >>>> + >>>> +struct drm_mock_sched_job { >>>> + struct drm_sched_job base; >>>> + >>>> + struct list_head link; >>>> + struct hrtimer timer; >>>> + >>>> + unsigned int duration_us; >>>> + ktime_t finish_at; >>>> + >>>> + spinlock_t lock; >>> >>> Same >>> >>>> + struct dma_fence hw_fence; >>>> + >>>> + struct kunit *test; >>>> +}; >>>> + >>>> +static inline struct drm_mock_scheduler * >>>> +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched) >>>> +{ >>>> + return container_of(sched, struct drm_mock_scheduler, >>>> base); >>>> +}; >>>> + >>>> +static inline struct drm_mock_sched_entity * >>>> +drm_sched_entity_to_mock_entity(struct drm_sched_entity >>>> *sched_entity) >>>> +{ >>>> + return container_of(sched_entity, struct >>>> drm_mock_sched_entity, base); >>>> +}; >>>> + >>>> +static inline struct drm_mock_sched_job * >>>> +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job) >>>> +{ >>>> + return container_of(sched_job, struct >>>> drm_mock_sched_job, >>>> base); >>>> +}; >>>> + >>>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit >>>> *test); >>>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched); >>>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler >>>> *sched, >>>> + unsigned int num); >>>> + >>>> +struct drm_mock_sched_entity * >>>> +drm_mock_new_sched_entity(struct kunit *test, >>>> + enum drm_sched_priority priority, >>>> + struct drm_mock_scheduler *sched); >>>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity >>>> *entity); >>>> + >>>> +struct drm_mock_sched_job * >>>> +drm_mock_new_sched_job(struct kunit *test, >>>> + struct drm_mock_sched_entity *entity); >>>> + >>>> +static inline void drm_mock_sched_job_submit(struct >>>> drm_mock_sched_job *job) >>>> +{ >>>> + drm_sched_job_arm(&job->base); >>>> + drm_sched_entity_push_job(&job->base); >>>> +} >>>> + >>>> +static inline void >>>> +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job >>>> *job, >>>> + unsigned int duration_us) >>>> +{ >>>> + job->duration_us = duration_us; >>>> +} >>>> + >>>> +static inline bool >>>> +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job) >>>> +{ >>>> + return dma_fence_is_signaled(&job->base.s_fence- >>>>> finished); >>>> +} >>>> + >>>> +static inline bool >>>> +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job, >>>> long timeout) >>>> +{ >>>> + long ret; >>>> + >>>> + ret = dma_fence_wait_timeout(&job->base.s_fence- >>>>> finished, >>>> + false, >>>> + timeout); >>>> + >>>> + return ret != 0; >>>> +} >>>> + >>>> +static inline long >>>> +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job >>>> *job, >>>> long timeout) >>>> +{ >>>> + long ret; >>>> + >>>> + ret = dma_fence_wait_timeout(&job->base.s_fence- >>>>> scheduled, >>>> + false, >>>> + timeout); >>>> + >>>> + return ret != 0; >>>> +} >>>> diff --git >>>> a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >>>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >>>> new file mode 100644 >>>> index 000000000000..6fd39bea95b1 >>>> --- /dev/null >>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c >>>> @@ -0,0 +1,247 @@ >>>> + >>>> +#include "drm_sched_tests.h" >>>> + >>>> +static int drm_sched_basic_init(struct kunit *test) >>>> +{ >>>> + test->priv = drm_mock_new_scheduler(test); >>>> + >>>> + return 0; >>>> +} >>>> + >>>> +static void drm_sched_basic_exit(struct kunit *test) >>>> +{ >>>> + struct drm_mock_scheduler *sched = test->priv; >>>> + >>>> + drm_mock_scheduler_fini(sched); >>>> +} >>>> + >>>> +static void drm_sched_basic_submit(struct kunit *test) >>>> +{ >>>> + struct drm_mock_scheduler *sched = test->priv; >>>> + struct drm_mock_sched_entity *entity; >>>> + struct drm_mock_sched_job *job; >>>> + unsigned int i; >>>> + bool done; >>>> + >>>> + /* >>>> + * Submit one job to the scheduler and verify that it >>>> gets >>>> scheduled >>>> + * and completed only when the mock hw backend processes >>>> it. >>>> + */ >>>> + >>>> + entity = drm_mock_new_sched_entity(test, >>>> + >>>> DRM_SCHED_PRIORITY_NORMAL, >>>> + sched); >>>> + job = drm_mock_new_sched_job(test, entity); >>>> + >>>> + drm_mock_sched_job_submit(job); >>>> + >>>> + done = drm_mock_sched_job_wait_scheduled(job, HZ); >>>> + KUNIT_ASSERT_EQ(test, done, true); >>>> + >>>> + done = drm_mock_sched_job_wait_finished(job, HZ / 2); >>>> + KUNIT_ASSERT_EQ(test, done, false); >>>> + >>>> + i = drm_mock_sched_advance(sched, 1); >>>> + KUNIT_ASSERT_EQ(test, i, 1); >>>> + >>>> + done = drm_mock_sched_job_wait_finished(job, HZ); >>>> + KUNIT_ASSERT_EQ(test, done, true); >>>> + >>>> + drm_mock_sched_entity_free(entity); >>>> +} >>>> + >>>> +static void drm_sched_basic_queue(struct kunit *test) >>>> +{ >>>> + struct drm_mock_scheduler *sched = test->priv; >>>> + struct drm_mock_sched_entity *entity; >>>> + struct drm_mock_sched_job *job; >>>> + const unsigned int qd = 100; >>> >>> Not the best variable name – is this "queue depth"? Why is it 100? >>> -> >>> global define & document >> >> I wouldn't promote this to global and TBH if you look how small this >> test function is it feels pretty self documenting. > > I meant global in the sense of at the top of the file as a constant. > > Wouldn't you agree that it would be good to have test parameters at a > place where you can easily modify them, in case you want to compile the > tests a bit differently? I agree in principle, just don't want to go with one define for all tests. I changed it locally to a parameterized test: static const struct drm_sched_basic_params drm_sched_basic_cases[] = { { .description = "A queue of jobs in a single entity", .queue_depth = 100, .job_us = 1000, .num_entities = 1, }, { .description = "A chain of dependent jobs across multiple entities", .queue_depth = 100, .job_us = 1000, .num_entities = 1, .dep_chain = true, }, { .description = "Multiple independent job queues", .queue_depth = 100, .job_us = 1000, .num_entities = 4, }, { .description = "Multiple inter-dependent job queues", .queue_depth = 100, .job_us = 1000, .num_entities = 4, .dep_chain = true, }, }; ... KUNIT_CASE_PARAM(drm_sched_basic_test, drm_sched_basic_gen_params), ... So now it is a single test body. Hopefully this satisfies both your desire to be able to modify easily, and my annoyance that there was too much of code duplication in v1. Regards, Tvrtko > > > P. > > >> >>> >>> >>>> + unsigned int i; >>>> + bool done; >>>> + >>>> + /* >>>> + * Submit a queue of jobs on the same entity, let them >>>> be >>>> completed by >>>> + * the mock hw backend and check the status of the last >>>> job. >>>> + */ >>>> + >>>> + entity = drm_mock_new_sched_entity(test, >>>> + >>>> DRM_SCHED_PRIORITY_NORMAL, >>>> + sched); >>>> + >>>> + for (i = 0; i < qd; i++) { >>>> + job = drm_mock_new_sched_job(test, entity); >>>> + drm_mock_sched_job_set_duration_us(job, 1000); >>>> + drm_mock_sched_job_submit(job); >>>> + } >>>> + >>>> + done = drm_mock_sched_job_wait_finished(job, HZ); >>>> + KUNIT_ASSERT_EQ(test, done, true); >>>> + >>>> + drm_mock_sched_entity_free(entity); >>>> +} >>>> + >>>> +static void drm_sched_basic_chain(struct kunit *test) >>>> +{ >>>> + struct drm_mock_scheduler *sched = test->priv; >>>> + struct drm_mock_sched_job *job, *prev = NULL; >>>> + struct drm_mock_sched_entity *entity; >>>> + const unsigned int qd = 100; >>>> + unsigned int i; >>>> + bool done; >>>> + >>>> + /* >>>> + * Submit a queue of jobs on the same entity but with an >>>> explicit >>>> + * chain of dependencies between them. Let the jobs be >>>> completed by >>>> + * the mock hw backend and check the status of the last >>>> job. >>>> + */ >>>> + >>>> + entity = drm_mock_new_sched_entity(test, >>>> + >>>> DRM_SCHED_PRIORITY_NORMAL, >>>> + sched); >>>> + >>>> + for (i = 0; i < qd; i++) { >>>> + job = drm_mock_new_sched_job(test, entity); >>>> + drm_mock_sched_job_set_duration_us(job, 1000); >>>> + if (prev) >>>> + drm_sched_job_add_dependency(&job->base, >>>> + >>>> dma_fence_get(&prev->base.s_fence->finished)); >>>> + drm_mock_sched_job_submit(job); >>>> + prev = job; >>>> + } >>>> + >>>> + done = drm_mock_sched_job_wait_finished(job, HZ); >>>> + KUNIT_ASSERT_EQ(test, done, true); >>>> + >>>> + drm_mock_sched_entity_free(entity); >>>> +} >>>> + >>>> +static void drm_sched_basic_entities(struct kunit *test) >>>> +{ >>>> + struct drm_mock_scheduler *sched = test->priv; >>>> + struct drm_mock_sched_entity *entity[4]; >>>> + struct drm_mock_sched_job *job; >>>> + const unsigned int qd = 100; >>>> + unsigned int i, cur_ent = 0; >>>> + bool done; >>>> + >>>> + /* >>>> + * Submit a queue of jobs across different entities, let >>>> them be >>>> + * completed by the mock hw backend and check the status >>>> of >>>> the last >>>> + * job. >>>> + */ >>>> + >>>> + for (i = 0; i < ARRAY_SIZE(entity); i++) >>>> + entity[i] = drm_mock_new_sched_entity(test, >>>> + >>>> DRM_SCHED_PRIORITY_NORMAL, >>>> + sched); >>>> + >>>> + for (i = 0; i < qd; i++) { >>>> + job = drm_mock_new_sched_job(test, >>>> entity[cur_ent++]); >>>> + cur_ent %= ARRAY_SIZE(entity); >>>> + drm_mock_sched_job_set_duration_us(job, 1000); >>>> + drm_mock_sched_job_submit(job); >>>> + } >>>> + >>>> + done = drm_mock_sched_job_wait_finished(job, HZ); >>>> + KUNIT_ASSERT_EQ(test, done, true); >>>> + >>>> + for (i = 0; i < ARRAY_SIZE(entity); i++) >>>> + drm_mock_sched_entity_free(entity[i]); >>>> +} >>>> + >>>> +static void drm_sched_basic_entities_chain(struct kunit *test) >>>> +{ >>>> + struct drm_mock_scheduler *sched = test->priv; >>>> + struct drm_mock_sched_job *job, *prev = NULL; >>>> + struct drm_mock_sched_entity *entity[4]; >>>> + const unsigned int qd = 100; >>>> + unsigned int i, cur_ent = 0; >>>> + bool done; >>>> + >>>> + /* >>>> + * Submit a queue of jobs across different entities with >>>> an >>>> explicit >>>> + * chain of dependencies between them. Let the jobs be >>>> completed by >>>> + * the mock hw backend and check the status of the last >>>> job. >>>> + */ >>>> + >>>> + for (i = 0; i < ARRAY_SIZE(entity); i++) >>>> + entity[i] = drm_mock_new_sched_entity(test, >>>> + >>>> DRM_SCHED_PRIORITY_NORMAL, >>>> + sched); >>>> + >>>> + for (i = 0; i < qd; i++) { >>>> + job = drm_mock_new_sched_job(test, >>>> entity[cur_ent++]); >>>> + cur_ent %= ARRAY_SIZE(entity); >>>> + drm_mock_sched_job_set_duration_us(job, 1000); >>>> + if (prev) >>>> + drm_sched_job_add_dependency(&job->base, >>>> + >>>> dma_fence_get(&prev->base.s_fence->finished)); >>>> + drm_mock_sched_job_submit(job); >>>> + prev = job; >>>> + } >>>> + >>>> + done = drm_mock_sched_job_wait_finished(job, HZ); >>>> + KUNIT_ASSERT_EQ(test, done, true); >>>> + >>>> + for (i = 0; i < ARRAY_SIZE(entity); i++) >>>> + drm_mock_sched_entity_free(entity[i]); >>>> +} >>>> + >>>> +static void drm_sched_basic_entity_cleanup(struct kunit *test) >>>> +{ >>>> + struct drm_mock_sched_job *job, *mid, *prev = NULL; >>>> + struct drm_mock_scheduler *sched = test->priv; >>>> + struct drm_mock_sched_entity *entity[4]; >>> >>> 4 is used in several places, so could be defined globally. In case >>> there's a special reason for why it's 4, that could be mentioned, >>> too >> >> It's completely arbitrary. In this test we just need a bunch of >> entities >> to be active on the scheduler as we trigger a client exit in the >> middle >> of it. IMO the comment few lines below should be good enough to >> explain >> the idea. I fear that promoting this to a global define would just >> give >> it more weight that what it has. Then some test would want a >> different >> number etc. >> >> Regards, >> >> Tvrtko >> >>>> + const unsigned int qd = 100; >>>> + unsigned int i, cur_ent = 0; >>>> + bool done; >>>> + >>>> + /* >>>> + * Submit a queue of jobs across different entities with >>>> an >>>> explicit >>>> + * chain of dependencies between them and trigger entity >>>> cleanup while >>>> + * the queue is still being processed. >>>> + */ >>>> + >>>> + for (i = 0; i < ARRAY_SIZE(entity); i++) >>>> + entity[i] = drm_mock_new_sched_entity(test, >>>> + >>>> DRM_SCHED_PRIORITY_NORMAL, >>>> + sched); >>>> + >>>> + for (i = 0; i < qd; i++) { >>>> + job = drm_mock_new_sched_job(test, >>>> entity[cur_ent++]); >>>> + cur_ent %= ARRAY_SIZE(entity); >>>> + drm_mock_sched_job_set_duration_us(job, 1000); >>>> + if (prev) >>>> + drm_sched_job_add_dependency(&job->base, >>>> + >>>> dma_fence_get(&prev->base.s_fence->finished)); >>>> + drm_mock_sched_job_submit(job); >>>> + if (i == qd / 2) >>>> + mid = job; >>>> + prev = job; >>>> + } >>>> + >>>> + done = drm_mock_sched_job_wait_finished(mid, HZ); >>>> + KUNIT_ASSERT_EQ(test, done, true); >>>> + >>>> + /* Exit with half of the queue still pending to be >>>> executed. >>>> */ >>>> + for (i = 0; i < ARRAY_SIZE(entity); i++) >>>> + drm_mock_sched_entity_free(entity[i]);} >>>> + >>>> +static struct kunit_case drm_sched_basic_tests[] = { >>>> + KUNIT_CASE(drm_sched_basic_submit), >>>> + KUNIT_CASE(drm_sched_basic_queue), >>>> + KUNIT_CASE(drm_sched_basic_chain), >>>> + KUNIT_CASE(drm_sched_basic_entities), >>>> + KUNIT_CASE(drm_sched_basic_entities_chain), >>>> + KUNIT_CASE(drm_sched_basic_entity_cleanup), >>>> + {} >>>> +}; >>>> + >>>> +static struct kunit_suite drm_sched_basic = { >>>> + .name = "drm_sched_basic_tests", >>>> + .init = drm_sched_basic_init, >>>> + .exit = drm_sched_basic_exit, >>>> + .test_cases = drm_sched_basic_tests, >>>> +}; >>>> + >>>> +kunit_test_suite(drm_sched_basic); >>> >
diff --git a/drivers/gpu/drm/Kconfig.debug b/drivers/gpu/drm/Kconfig.debug index a35d74171b7b..89f777f21e95 100644 --- a/drivers/gpu/drm/Kconfig.debug +++ b/drivers/gpu/drm/Kconfig.debug @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST If in doubt, say "N". +config DRM_SCHED_KUNIT_TEST + tristate "KUnit tests for the DRM scheduler" if !KUNIT_ALL_TESTS + select DRM_SCHED + depends on DRM && KUNIT + default KUNIT_ALL_TESTS + help + Choose this option to build unit tests for the DRM scheduler. + + Recommended for driver developers only. + + If in doubt, say "N". + config DRM_EXPORT_FOR_TESTS bool diff --git a/drivers/gpu/drm/scheduler/.kunitconfig b/drivers/gpu/drm/scheduler/.kunitconfig new file mode 100644 index 000000000000..cece53609fcf --- /dev/null +++ b/drivers/gpu/drm/scheduler/.kunitconfig @@ -0,0 +1,12 @@ +CONFIG_KUNIT=y +CONFIG_DRM=y +CONFIG_DRM_SCHED_KUNIT_TEST=y +CONFIG_EXPERT=y +CONFIG_DEBUG_SPINLOCK=y +CONFIG_DEBUG_MUTEXES=y +CONFIG_DEBUG_ATOMIC_SLEEP=y +CONFIG_LOCK_DEBUGGING_SUPPORT=y +CONFIG_PROVE_LOCKING=y +CONFIG_LOCKDEP=y +CONFIG_DEBUG_LOCKDEP=y +CONFIG_DEBUG_LIST=y diff --git a/drivers/gpu/drm/scheduler/Makefile b/drivers/gpu/drm/scheduler/Makefile index 53863621829f..46dfdca0758a 100644 --- a/drivers/gpu/drm/scheduler/Makefile +++ b/drivers/gpu/drm/scheduler/Makefile @@ -23,3 +23,4 @@ gpu-sched-y := sched_main.o sched_fence.o sched_entity.o obj-$(CONFIG_DRM_SCHED) += gpu-sched.o +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/ diff --git a/drivers/gpu/drm/scheduler/tests/Makefile b/drivers/gpu/drm/scheduler/tests/Makefile new file mode 100644 index 000000000000..d69eab6a2e9b --- /dev/null +++ b/drivers/gpu/drm/scheduler/tests/Makefile @@ -0,0 +1,4 @@ + +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \ + drm_mock_scheduler.o \ + drm_sched_tests_basic.o diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c new file mode 100644 index 000000000000..c9205f9ff524 --- /dev/null +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c @@ -0,0 +1,29 @@ + +#include "drm_sched_tests.h" + +struct drm_mock_sched_entity * +drm_mock_sched_entity_new(struct kunit *test, + enum drm_sched_priority priority, + struct drm_mock_scheduler *sched) +{ + struct drm_sched_mock_entity *entity; + int ret; + + entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL); + KUNIT_ASSERT_NOT_NULL(test, entity); + + ret = drm_sched_entity_init(&entity->base, + priority, + &sched->base, 1, + NULL); + KUNIT_ASSERT_EQ(test, ret, 0); + + entity->test = test; + + return entity; +} + +void drm_mock_sched_entity_free(struct drm_mock_sched_entity *entity) +{ + drm_sched_entity_fini(&entity->base); +} diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c new file mode 100644 index 000000000000..d94568cf3da9 --- /dev/null +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c @@ -0,0 +1,3 @@ + +#include "drm_sched_tests.h" + diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c new file mode 100644 index 000000000000..f1985900a6ba --- /dev/null +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c @@ -0,0 +1,254 @@ + +#include "drm_sched_tests.h" + +struct drm_mock_sched_entity * +drm_mock_new_sched_entity(struct kunit *test, + enum drm_sched_priority priority, + struct drm_mock_scheduler *sched) +{ + struct drm_mock_sched_entity *entity; + struct drm_gpu_scheduler *drm_sched; + int ret; + + entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL); + KUNIT_ASSERT_NOT_NULL(test, entity); + + drm_sched = &sched->base; + ret = drm_sched_entity_init(&entity->base, + priority, + &drm_sched, 1, + NULL); + KUNIT_ASSERT_EQ(test, ret, 0); + + entity->test = test; + + return entity; +} + +void drm_mock_sched_entity_free(struct drm_mock_sched_entity *entity) +{ + drm_sched_entity_destroy(&entity->base); +} + +static enum hrtimer_restart +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer) +{ + struct drm_mock_sched_job *upto = + container_of(hrtimer, typeof(*upto), timer); + struct drm_mock_scheduler *sched = + drm_sched_to_mock_sched(upto->base.sched); + struct drm_mock_sched_job *job, *next; + ktime_t now = ktime_get(); + unsigned long flags; + LIST_HEAD(signal); + + spin_lock_irqsave(&sched->lock, flags); + list_for_each_entry_safe(job, next, &sched->job_list, link) { + if (!job->duration_us) + break; + + if (ktime_before(now, job->finish_at)) + break; + + list_move_tail(&job->link, &signal); + sched->seqno = job->hw_fence.seqno; + } + spin_unlock_irqrestore(&sched->lock, flags); + + list_for_each_entry(job, &signal, link) { + dma_fence_signal(&job->hw_fence); + dma_fence_put(&job->hw_fence); + } + + return HRTIMER_NORESTART; +} + +struct drm_mock_sched_job * +drm_mock_new_sched_job(struct kunit *test, + struct drm_mock_sched_entity *entity) +{ + struct drm_mock_sched_job *job; + int ret; + + job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL); + KUNIT_ASSERT_NOT_NULL(test, job); + + ret = drm_sched_job_init(&job->base, + &entity->base, + 1, + NULL); + KUNIT_ASSERT_EQ(test, ret, 0); + + job->test = test; + + spin_lock_init(&job->lock); + INIT_LIST_HEAD(&job->link); + hrtimer_init(&job->timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); + job->timer.function = drm_sched_mock_job_signal_timer; + + return job; +} + +static const char *drm_mock_sched_hw_fence_driver_name(struct dma_fence *fence) +{ + return "drm_mock_sched"; +} + +static const char * +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence) +{ + struct drm_mock_sched_job *job = + container_of(fence, typeof(*job), hw_fence); + + return (const char *)job->base.sched->name; +} +static void drm_mock_sched_hw_fence_release(struct dma_fence *fence) +{ + struct drm_mock_sched_job *job = + container_of(fence, typeof(*job), hw_fence); + + hrtimer_cancel(&job->timer); + + /* Freed by the kunit framework */ +} + +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = { + .get_driver_name = drm_mock_sched_hw_fence_driver_name, + .get_timeline_name = drm_mock_sched_hw_fence_timeline_name, + .release = drm_mock_sched_hw_fence_release, +}; + +static struct dma_fence *mock_sched_run_job(struct drm_sched_job *sched_job) +{ + struct drm_mock_scheduler *sched = + drm_sched_to_mock_sched(sched_job->sched); + struct drm_mock_sched_job *job = drm_sched_job_to_mock_job(sched_job); + + dma_fence_init(&job->hw_fence, + &drm_mock_sched_hw_fence_ops, + &job->lock, + sched->hw_fence.context, + atomic_inc_return(&sched->hw_fence.seqno)); + + dma_fence_get(&job->hw_fence); /* Reference for the job_list */ + + spin_lock_irq(&sched->lock); + if (job->duration_us) { + ktime_t prev_finish_at = 0; + + if (!list_empty(&sched->job_list)) { + struct drm_mock_sched_job *prev = + list_last_entry(&sched->job_list, typeof(*prev), + link); + + prev_finish_at = prev->finish_at; + } + + if (!prev_finish_at) + prev_finish_at = ktime_get(); + + job->finish_at = ktime_add_us(prev_finish_at, job->duration_us); + } + list_add_tail(&job->link, &sched->job_list); + if (job->finish_at) + hrtimer_start(&job->timer, job->finish_at, HRTIMER_MODE_ABS); + spin_unlock_irq(&sched->lock); + + return &job->hw_fence; +} + +static enum drm_gpu_sched_stat +mock_sched_timedout_job(struct drm_sched_job *sched_job) +{ + return DRM_GPU_SCHED_STAT_ENODEV; +} + +static void mock_sched_free_job(struct drm_sched_job *sched_job) +{ + drm_sched_job_cleanup(sched_job); +} + +static const struct drm_sched_backend_ops drm_mock_scheduler_ops = { + .run_job = mock_sched_run_job, + .timedout_job = mock_sched_timedout_job, + .free_job = mock_sched_free_job +}; + +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit *test) +{ + struct drm_mock_scheduler *sched; + int ret; + + sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL); + KUNIT_ASSERT_NOT_NULL(test, sched); + + ret = drm_sched_init(&sched->base, + &drm_mock_scheduler_ops, + NULL, /* wq */ + DRM_SCHED_PRIORITY_COUNT, + U32_MAX, /* max credits */ + UINT_MAX, /* hang limit */ + MAX_SCHEDULE_TIMEOUT, /* timeout */ + NULL, /* timeout wq */ + NULL, /* score */ + "drm-mock-scheduler", + NULL /* dev */); + KUNIT_ASSERT_EQ(test, ret, 0); + + sched->test = test; + sched->hw_fence.context = dma_fence_context_alloc(1); + atomic_set(&sched->hw_fence.seqno, 0); + INIT_LIST_HEAD(&sched->job_list); + spin_lock_init(&sched->lock); + + return sched; +} + +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched) +{ + struct drm_mock_sched_job *job, *next; + unsigned long flags; + LIST_HEAD(signal); + + spin_lock_irqsave(&sched->lock, flags); + list_for_each_entry_safe(job, next, &sched->job_list, link) + list_move_tail(&job->link, &signal); + spin_unlock_irqrestore(&sched->lock, flags); + + list_for_each_entry(job, &signal, link) { + hrtimer_cancel(&job->timer); + dma_fence_put(&job->hw_fence); + } + + drm_sched_fini(&sched->base); +} + +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler *sched, + unsigned int num) +{ + struct drm_mock_sched_job *job, *next; + unsigned int found = 0; + unsigned long flags; + LIST_HEAD(signal); + + spin_lock_irqsave(&sched->lock, flags); + if (WARN_ON_ONCE(sched->seqno + num < sched->seqno)) + goto unlock; + sched->seqno += num; + list_for_each_entry_safe(job, next, &sched->job_list, link) { + if (sched->seqno < job->hw_fence.seqno) + break; + + list_move_tail(&job->link, &signal); + found++; + } +unlock: + spin_unlock_irqrestore(&sched->lock, flags); + + list_for_each_entry(job, &signal, link) { + dma_fence_signal(&job->hw_fence); + dma_fence_put(&job->hw_fence); + } + + return found; +} diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h new file mode 100644 index 000000000000..421ee2712985 --- /dev/null +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h @@ -0,0 +1,124 @@ + +#include <kunit/test.h> +#include <linux/atomic.h> +#include <linux/dma-fence.h> +#include <linux/hrtimer.h> +#include <linux/ktime.h> +#include <linux/list.h> +#include <linux/atomic.h> +#include <linux/mutex.h> +#include <linux/types.h> + +#include <drm/gpu_scheduler.h> + +struct drm_mock_scheduler { + struct drm_gpu_scheduler base; + + struct kunit *test; + + struct { + u64 context; + atomic_t seqno; + } hw_fence; + + spinlock_t lock; + unsigned int seqno; + struct list_head job_list; +}; + +struct drm_mock_sched_entity { + struct drm_sched_entity base; + + struct kunit *test; +}; + +struct drm_mock_sched_job { + struct drm_sched_job base; + + struct list_head link; + struct hrtimer timer; + + unsigned int duration_us; + ktime_t finish_at; + + spinlock_t lock; + struct dma_fence hw_fence; + + struct kunit *test; +}; + +static inline struct drm_mock_scheduler * +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched) +{ + return container_of(sched, struct drm_mock_scheduler, base); +}; + +static inline struct drm_mock_sched_entity * +drm_sched_entity_to_mock_entity(struct drm_sched_entity *sched_entity) +{ + return container_of(sched_entity, struct drm_mock_sched_entity, base); +}; + +static inline struct drm_mock_sched_job * +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job) +{ + return container_of(sched_job, struct drm_mock_sched_job, base); +}; + +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit *test); +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched); +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler *sched, + unsigned int num); + +struct drm_mock_sched_entity * +drm_mock_new_sched_entity(struct kunit *test, + enum drm_sched_priority priority, + struct drm_mock_scheduler *sched); +void drm_mock_sched_entity_free(struct drm_mock_sched_entity *entity); + +struct drm_mock_sched_job * +drm_mock_new_sched_job(struct kunit *test, + struct drm_mock_sched_entity *entity); + +static inline void drm_mock_sched_job_submit(struct drm_mock_sched_job *job) +{ + drm_sched_job_arm(&job->base); + drm_sched_entity_push_job(&job->base); +} + +static inline void +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job, + unsigned int duration_us) +{ + job->duration_us = duration_us; +} + +static inline bool +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job) +{ + return dma_fence_is_signaled(&job->base.s_fence->finished); +} + +static inline bool +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job, long timeout) +{ + long ret; + + ret = dma_fence_wait_timeout(&job->base.s_fence->finished, + false, + timeout); + + return ret != 0; +} + +static inline long +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job, long timeout) +{ + long ret; + + ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled, + false, + timeout); + + return ret != 0; +} diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c new file mode 100644 index 000000000000..6fd39bea95b1 --- /dev/null +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c @@ -0,0 +1,247 @@ + +#include "drm_sched_tests.h" + +static int drm_sched_basic_init(struct kunit *test) +{ + test->priv = drm_mock_new_scheduler(test); + + return 0; +} + +static void drm_sched_basic_exit(struct kunit *test) +{ + struct drm_mock_scheduler *sched = test->priv; + + drm_mock_scheduler_fini(sched); +} + +static void drm_sched_basic_submit(struct kunit *test) +{ + struct drm_mock_scheduler *sched = test->priv; + struct drm_mock_sched_entity *entity; + struct drm_mock_sched_job *job; + unsigned int i; + bool done; + + /* + * Submit one job to the scheduler and verify that it gets scheduled + * and completed only when the mock hw backend processes it. + */ + + entity = drm_mock_new_sched_entity(test, + DRM_SCHED_PRIORITY_NORMAL, + sched); + job = drm_mock_new_sched_job(test, entity); + + drm_mock_sched_job_submit(job); + + done = drm_mock_sched_job_wait_scheduled(job, HZ); + KUNIT_ASSERT_EQ(test, done, true); + + done = drm_mock_sched_job_wait_finished(job, HZ / 2); + KUNIT_ASSERT_EQ(test, done, false); + + i = drm_mock_sched_advance(sched, 1); + KUNIT_ASSERT_EQ(test, i, 1); + + done = drm_mock_sched_job_wait_finished(job, HZ); + KUNIT_ASSERT_EQ(test, done, true); + + drm_mock_sched_entity_free(entity); +} + +static void drm_sched_basic_queue(struct kunit *test) +{ + struct drm_mock_scheduler *sched = test->priv; + struct drm_mock_sched_entity *entity; + struct drm_mock_sched_job *job; + const unsigned int qd = 100; + unsigned int i; + bool done; + + /* + * Submit a queue of jobs on the same entity, let them be completed by + * the mock hw backend and check the status of the last job. + */ + + entity = drm_mock_new_sched_entity(test, + DRM_SCHED_PRIORITY_NORMAL, + sched); + + for (i = 0; i < qd; i++) { + job = drm_mock_new_sched_job(test, entity); + drm_mock_sched_job_set_duration_us(job, 1000); + drm_mock_sched_job_submit(job); + } + + done = drm_mock_sched_job_wait_finished(job, HZ); + KUNIT_ASSERT_EQ(test, done, true); + + drm_mock_sched_entity_free(entity); +} + +static void drm_sched_basic_chain(struct kunit *test) +{ + struct drm_mock_scheduler *sched = test->priv; + struct drm_mock_sched_job *job, *prev = NULL; + struct drm_mock_sched_entity *entity; + const unsigned int qd = 100; + unsigned int i; + bool done; + + /* + * Submit a queue of jobs on the same entity but with an explicit + * chain of dependencies between them. Let the jobs be completed by + * the mock hw backend and check the status of the last job. + */ + + entity = drm_mock_new_sched_entity(test, + DRM_SCHED_PRIORITY_NORMAL, + sched); + + for (i = 0; i < qd; i++) { + job = drm_mock_new_sched_job(test, entity); + drm_mock_sched_job_set_duration_us(job, 1000); + if (prev) + drm_sched_job_add_dependency(&job->base, + dma_fence_get(&prev->base.s_fence->finished)); + drm_mock_sched_job_submit(job); + prev = job; + } + + done = drm_mock_sched_job_wait_finished(job, HZ); + KUNIT_ASSERT_EQ(test, done, true); + + drm_mock_sched_entity_free(entity); +} + +static void drm_sched_basic_entities(struct kunit *test) +{ + struct drm_mock_scheduler *sched = test->priv; + struct drm_mock_sched_entity *entity[4]; + struct drm_mock_sched_job *job; + const unsigned int qd = 100; + unsigned int i, cur_ent = 0; + bool done; + + /* + * Submit a queue of jobs across different entities, let them be + * completed by the mock hw backend and check the status of the last + * job. + */ + + for (i = 0; i < ARRAY_SIZE(entity); i++) + entity[i] = drm_mock_new_sched_entity(test, + DRM_SCHED_PRIORITY_NORMAL, + sched); + + for (i = 0; i < qd; i++) { + job = drm_mock_new_sched_job(test, entity[cur_ent++]); + cur_ent %= ARRAY_SIZE(entity); + drm_mock_sched_job_set_duration_us(job, 1000); + drm_mock_sched_job_submit(job); + } + + done = drm_mock_sched_job_wait_finished(job, HZ); + KUNIT_ASSERT_EQ(test, done, true); + + for (i = 0; i < ARRAY_SIZE(entity); i++) + drm_mock_sched_entity_free(entity[i]); +} + +static void drm_sched_basic_entities_chain(struct kunit *test) +{ + struct drm_mock_scheduler *sched = test->priv; + struct drm_mock_sched_job *job, *prev = NULL; + struct drm_mock_sched_entity *entity[4]; + const unsigned int qd = 100; + unsigned int i, cur_ent = 0; + bool done; + + /* + * Submit a queue of jobs across different entities with an explicit + * chain of dependencies between them. Let the jobs be completed by + * the mock hw backend and check the status of the last job. + */ + + for (i = 0; i < ARRAY_SIZE(entity); i++) + entity[i] = drm_mock_new_sched_entity(test, + DRM_SCHED_PRIORITY_NORMAL, + sched); + + for (i = 0; i < qd; i++) { + job = drm_mock_new_sched_job(test, entity[cur_ent++]); + cur_ent %= ARRAY_SIZE(entity); + drm_mock_sched_job_set_duration_us(job, 1000); + if (prev) + drm_sched_job_add_dependency(&job->base, + dma_fence_get(&prev->base.s_fence->finished)); + drm_mock_sched_job_submit(job); + prev = job; + } + + done = drm_mock_sched_job_wait_finished(job, HZ); + KUNIT_ASSERT_EQ(test, done, true); + + for (i = 0; i < ARRAY_SIZE(entity); i++) + drm_mock_sched_entity_free(entity[i]); +} + +static void drm_sched_basic_entity_cleanup(struct kunit *test) +{ + struct drm_mock_sched_job *job, *mid, *prev = NULL; + struct drm_mock_scheduler *sched = test->priv; + struct drm_mock_sched_entity *entity[4]; + const unsigned int qd = 100; + unsigned int i, cur_ent = 0; + bool done; + + /* + * Submit a queue of jobs across different entities with an explicit + * chain of dependencies between them and trigger entity cleanup while + * the queue is still being processed. + */ + + for (i = 0; i < ARRAY_SIZE(entity); i++) + entity[i] = drm_mock_new_sched_entity(test, + DRM_SCHED_PRIORITY_NORMAL, + sched); + + for (i = 0; i < qd; i++) { + job = drm_mock_new_sched_job(test, entity[cur_ent++]); + cur_ent %= ARRAY_SIZE(entity); + drm_mock_sched_job_set_duration_us(job, 1000); + if (prev) + drm_sched_job_add_dependency(&job->base, + dma_fence_get(&prev->base.s_fence->finished)); + drm_mock_sched_job_submit(job); + if (i == qd / 2) + mid = job; + prev = job; + } + + done = drm_mock_sched_job_wait_finished(mid, HZ); + KUNIT_ASSERT_EQ(test, done, true); + + /* Exit with half of the queue still pending to be executed. */ + for (i = 0; i < ARRAY_SIZE(entity); i++) + drm_mock_sched_entity_free(entity[i]);} + +static struct kunit_case drm_sched_basic_tests[] = { + KUNIT_CASE(drm_sched_basic_submit), + KUNIT_CASE(drm_sched_basic_queue), + KUNIT_CASE(drm_sched_basic_chain), + KUNIT_CASE(drm_sched_basic_entities), + KUNIT_CASE(drm_sched_basic_entities_chain), + KUNIT_CASE(drm_sched_basic_entity_cleanup), + {} +}; + +static struct kunit_suite drm_sched_basic = { + .name = "drm_sched_basic_tests", + .init = drm_sched_basic_init, + .exit = drm_sched_basic_exit, + .test_cases = drm_sched_basic_tests, +}; + +kunit_test_suite(drm_sched_basic);
Implement a mock scheduler backend and add some basic test to exercise the core scheduler code paths. Mock backend (kind of like a very simple mock GPU) can either process jobs by tests manually advancing the "timeline" job at a time, or alternatively jobs can be configured with a time duration in which case they get completed asynchronously from the unit test code. Core scheduler classes are subclassed to support this mock implementation. The tests added are just a few simple submission patterns. Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com> Cc: Christian König <christian.koenig@amd.com> Cc: Danilo Krummrich <dakr@kernel.org> Cc: Matthew Brost <matthew.brost@intel.com> Cc: Philipp Stanner <phasta@kernel.org> --- drivers/gpu/drm/Kconfig.debug | 12 + drivers/gpu/drm/scheduler/.kunitconfig | 12 + drivers/gpu/drm/scheduler/Makefile | 1 + drivers/gpu/drm/scheduler/tests/Makefile | 4 + .../gpu/drm/scheduler/tests/drm_mock_entity.c | 29 ++ .../gpu/drm/scheduler/tests/drm_mock_job.c | 3 + .../drm/scheduler/tests/drm_mock_scheduler.c | 254 ++++++++++++++++++ .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++ .../scheduler/tests/drm_sched_tests_basic.c | 247 +++++++++++++++++ 9 files changed, 686 insertions(+) create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c