diff mbox series

[RFC,2/5] drm/scheduler: Add scheduler unit testing infrastructure and some basic tests

Message ID 20250203153007.63400-3-tvrtko.ursulin@igalia.com (mailing list archive)
State New, archived
Headers show
Series DRM scheduler kunit tests | expand

Commit Message

Tvrtko Ursulin Feb. 3, 2025, 3:30 p.m. UTC
Implement a mock scheduler backend and add some basic test to exercise the
core scheduler code paths.

Mock backend (kind of like a very simple mock GPU) can either process jobs
by tests manually advancing the "timeline" job at a time, or alternatively
jobs can be configured with a time duration in which case they get
completed asynchronously from the unit test code.

Core scheduler classes are subclassed to support this mock implementation.

The tests added are just a few simple submission patterns.

Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com>
Cc: Christian König <christian.koenig@amd.com>
Cc: Danilo Krummrich <dakr@kernel.org>
Cc: Matthew Brost <matthew.brost@intel.com>
Cc: Philipp Stanner <phasta@kernel.org>
---
 drivers/gpu/drm/Kconfig.debug                 |  12 +
 drivers/gpu/drm/scheduler/.kunitconfig        |  12 +
 drivers/gpu/drm/scheduler/Makefile            |   1 +
 drivers/gpu/drm/scheduler/tests/Makefile      |   4 +
 .../gpu/drm/scheduler/tests/drm_mock_entity.c |  29 ++
 .../gpu/drm/scheduler/tests/drm_mock_job.c    |   3 +
 .../drm/scheduler/tests/drm_mock_scheduler.c  | 254 ++++++++++++++++++
 .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++
 .../scheduler/tests/drm_sched_tests_basic.c   | 247 +++++++++++++++++
 9 files changed, 686 insertions(+)
 create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig
 create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile
 create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
 create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c
 create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
 create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
 create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c

Comments

Philipp Stanner Feb. 6, 2025, 9:51 a.m. UTC | #1
On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote:
> Implement a mock scheduler backend and add some basic test to
> exercise the
> core scheduler code paths.
> 
> Mock backend (kind of like a very simple mock GPU) can either process
> jobs
> by tests manually advancing the "timeline" job at a time, or
> alternatively
> jobs can be configured with a time duration in which case they get
> completed asynchronously from the unit test code.
> 
> Core scheduler classes are subclassed to support this mock
> implementation.
> 
> The tests added are just a few simple submission patterns.
> 
> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com>

Could add a Suggested-by: Philipp :)


> Cc: Christian König <christian.koenig@amd.com>
> Cc: Danilo Krummrich <dakr@kernel.org>
> Cc: Matthew Brost <matthew.brost@intel.com>
> Cc: Philipp Stanner <phasta@kernel.org>
> ---
>  drivers/gpu/drm/Kconfig.debug                 |  12 +
>  drivers/gpu/drm/scheduler/.kunitconfig        |  12 +
>  drivers/gpu/drm/scheduler/Makefile            |   1 +
>  drivers/gpu/drm/scheduler/tests/Makefile      |   4 +
>  .../gpu/drm/scheduler/tests/drm_mock_entity.c |  29 ++
>  .../gpu/drm/scheduler/tests/drm_mock_job.c    |   3 +
>  .../drm/scheduler/tests/drm_mock_scheduler.c  | 254
> ++++++++++++++++++
>  .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++
>  .../scheduler/tests/drm_sched_tests_basic.c   | 247
> +++++++++++++++++
>  9 files changed, 686 insertions(+)
>  create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig
>  create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile
>  create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>  create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>  create mode 100644
> drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>  create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>  create mode 100644
> drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> 
> diff --git a/drivers/gpu/drm/Kconfig.debug
> b/drivers/gpu/drm/Kconfig.debug
> index a35d74171b7b..89f777f21e95 100644
> --- a/drivers/gpu/drm/Kconfig.debug
> +++ b/drivers/gpu/drm/Kconfig.debug
> @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST
>  
>  	  If in doubt, say "N".
>  
> +config DRM_SCHED_KUNIT_TEST
> +	tristate "KUnit tests for the DRM scheduler" if
> !KUNIT_ALL_TESTS
> +	select DRM_SCHED
> +	depends on DRM && KUNIT
> +	default KUNIT_ALL_TESTS
> +	help
> +	  Choose this option to build unit tests for the DRM
> scheduler.

nit: Might say "DRM GPU scheduler" so readers not familiar with all
that get a better idea of what it's about

> +
> +	  Recommended for driver developers only.

nit: s/driver developers/DRM developers
?

> +
> +	  If in doubt, say "N".
> +
>  config DRM_EXPORT_FOR_TESTS
>  	bool
> diff --git a/drivers/gpu/drm/scheduler/.kunitconfig
> b/drivers/gpu/drm/scheduler/.kunitconfig
> new file mode 100644
> index 000000000000..cece53609fcf
> --- /dev/null
> +++ b/drivers/gpu/drm/scheduler/.kunitconfig
> @@ -0,0 +1,12 @@
> +CONFIG_KUNIT=y
> +CONFIG_DRM=y
> +CONFIG_DRM_SCHED_KUNIT_TEST=y
> +CONFIG_EXPERT=y
> +CONFIG_DEBUG_SPINLOCK=y
> +CONFIG_DEBUG_MUTEXES=y
> +CONFIG_DEBUG_ATOMIC_SLEEP=y
> +CONFIG_LOCK_DEBUGGING_SUPPORT=y
> +CONFIG_PROVE_LOCKING=y
> +CONFIG_LOCKDEP=y
> +CONFIG_DEBUG_LOCKDEP=y
> +CONFIG_DEBUG_LIST=y
> diff --git a/drivers/gpu/drm/scheduler/Makefile
> b/drivers/gpu/drm/scheduler/Makefile
> index 53863621829f..46dfdca0758a 100644
> --- a/drivers/gpu/drm/scheduler/Makefile
> +++ b/drivers/gpu/drm/scheduler/Makefile
> @@ -23,3 +23,4 @@
>  gpu-sched-y := sched_main.o sched_fence.o sched_entity.o
>  
>  obj-$(CONFIG_DRM_SCHED) += gpu-sched.o
> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/
> diff --git a/drivers/gpu/drm/scheduler/tests/Makefile
> b/drivers/gpu/drm/scheduler/tests/Makefile
> new file mode 100644
> index 000000000000..d69eab6a2e9b
> --- /dev/null
> +++ b/drivers/gpu/drm/scheduler/tests/Makefile
> @@ -0,0 +1,4 @@
> +
> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \
> +        drm_mock_scheduler.o \
> +        drm_sched_tests_basic.o
> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
> b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
> new file mode 100644
> index 000000000000..c9205f9ff524
> --- /dev/null
> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
> @@ -0,0 +1,29 @@
> +
> +#include "drm_sched_tests.h"
> +
> +struct drm_mock_sched_entity *
> +drm_mock_sched_entity_new(struct kunit *test,
> +			  enum drm_sched_priority priority,
> +			  struct drm_mock_scheduler *sched)


I personally do like this function head style and


> +{
> +	struct drm_sched_mock_entity *entity;
> +	int ret;
> +
> +	entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL);
> +	KUNIT_ASSERT_NOT_NULL(test, entity);
> +
> +	ret = drm_sched_entity_init(&entity->base,
> +				    priority,
> +				    &sched->base, 1,
> +				    NULL);
> +	KUNIT_ASSERT_EQ(test, ret, 0);
> +
> +	entity->test = test;
> +
> +	return entity;
> +}
> +
> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
> *entity)

do believe it should then be used consistently everywhere, regardless
of length.


> +{
> +	drm_sched_entity_fini(&entity->base);
> +}
> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
> b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
> new file mode 100644
> index 000000000000..d94568cf3da9
> --- /dev/null
> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
> @@ -0,0 +1,3 @@
> +
> +#include "drm_sched_tests.h"
> +
> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
> b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
> new file mode 100644
> index 000000000000..f1985900a6ba
> --- /dev/null
> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
> @@ -0,0 +1,254 @@
> +
> +#include "drm_sched_tests.h"
> +
> +struct drm_mock_sched_entity *
> +drm_mock_new_sched_entity(struct kunit *test,
> +			  enum drm_sched_priority priority,
> +			  struct drm_mock_scheduler *sched)
> +{
> +	struct drm_mock_sched_entity *entity;
> +	struct drm_gpu_scheduler *drm_sched;
> +	int ret;
> +
> +	entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL);
> +	KUNIT_ASSERT_NOT_NULL(test, entity);
> +
> +	drm_sched = &sched->base;
> +	ret = drm_sched_entity_init(&entity->base,
> +				    priority,
> +				    &drm_sched, 1,
> +				    NULL);
> +	KUNIT_ASSERT_EQ(test, ret, 0);
> +
> +	entity->test = test;
> +
> +	return entity;
> +}
> +
> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
> *entity)
> +{
> +	drm_sched_entity_destroy(&entity->base);
> +}
> +
> +static enum hrtimer_restart
> +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer)

I think we should get up with a consistent naming convention. Some
things are called drm_mock_sched_, some others drm_sched_mock_.
As far as I saw, drm_mock_* does not yet exist.
So do you want to introduce drm_mock as something generic for drm? or
just for drm/sched?


> +{
> +	struct drm_mock_sched_job *upto =
> +		container_of(hrtimer, typeof(*upto), timer);
> +	struct drm_mock_scheduler *sched =
> +		drm_sched_to_mock_sched(upto->base.sched);
> +	struct drm_mock_sched_job *job, *next;
> +	ktime_t now = ktime_get();
> +	unsigned long flags;
> +	LIST_HEAD(signal);
> +
> +	spin_lock_irqsave(&sched->lock, flags);
> +	list_for_each_entry_safe(job, next, &sched->job_list, link)
> {
> +		if (!job->duration_us)
> +			break;
> +
> +		if (ktime_before(now, job->finish_at))
> +			break;
> +
> +		list_move_tail(&job->link, &signal);
> +		sched->seqno = job->hw_fence.seqno;
> +	}
> +	spin_unlock_irqrestore(&sched->lock, flags);
> +
> +	list_for_each_entry(job, &signal, link) {
> +		dma_fence_signal(&job->hw_fence);
> +		dma_fence_put(&job->hw_fence);
> +	}
> +
> +	return HRTIMER_NORESTART;
> +}
> +
> +struct drm_mock_sched_job *
> +drm_mock_new_sched_job(struct kunit *test,
> +		       struct drm_mock_sched_entity *entity)
> +{
> +	struct drm_mock_sched_job *job;
> +	int ret;
> +
> +	job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL);
> +	KUNIT_ASSERT_NOT_NULL(test, job);
> +
> +	ret = drm_sched_job_init(&job->base,
> +				 &entity->base,
> +				 1,
> +				 NULL);
> +	KUNIT_ASSERT_EQ(test, ret, 0);
> +
> +	job->test = test;
> +
> +	spin_lock_init(&job->lock);
> +	INIT_LIST_HEAD(&job->link);
> +	hrtimer_init(&job->timer, CLOCK_MONOTONIC,
> HRTIMER_MODE_ABS);
> +	job->timer.function = drm_sched_mock_job_signal_timer;
> +
> +	return job;
> +}
> +
> +static const char *drm_mock_sched_hw_fence_driver_name(struct
> dma_fence *fence)
> +{
> +	return "drm_mock_sched";
> +}
> +
> +static const char *
> +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence)
> +{
> +	struct drm_mock_sched_job *job =
> +		container_of(fence, typeof(*job), hw_fence);
> +
> +	return (const char *)job->base.sched->name;
> +}
> +static void drm_mock_sched_hw_fence_release(struct dma_fence *fence)
> +{
> +	struct drm_mock_sched_job *job =
> +		container_of(fence, typeof(*job), hw_fence);
> +
> +	hrtimer_cancel(&job->timer);
> +
> +	/* Freed by the kunit framework */
> +}
> +
> +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = {
> +	.get_driver_name = drm_mock_sched_hw_fence_driver_name,
> +	.get_timeline_name = drm_mock_sched_hw_fence_timeline_name,
> +	.release = drm_mock_sched_hw_fence_release,
> +};
> +
> +static struct dma_fence *mock_sched_run_job(struct drm_sched_job
> *sched_job)
> +{
> +	struct drm_mock_scheduler *sched =
> +		drm_sched_to_mock_sched(sched_job->sched);
> +	struct drm_mock_sched_job *job =
> drm_sched_job_to_mock_job(sched_job);
> +
> +	dma_fence_init(&job->hw_fence,
> +		       &drm_mock_sched_hw_fence_ops,
> +		       &job->lock,
> +		       sched->hw_fence.context,
> +		       atomic_inc_return(&sched->hw_fence.seqno));
> +
> +	dma_fence_get(&job->hw_fence); /* Reference for the job_list
> */
> +
> +	spin_lock_irq(&sched->lock);
> +	if (job->duration_us) {
> +		ktime_t prev_finish_at = 0;
> +
> +		if (!list_empty(&sched->job_list)) {
> +			struct drm_mock_sched_job *prev =
> +				list_last_entry(&sched->job_list,
> typeof(*prev),
> +						link);
> +
> +			prev_finish_at = prev->finish_at;
> +		}
> +
> +		if (!prev_finish_at)
> +			prev_finish_at = ktime_get();
> +
> +		job->finish_at = ktime_add_us(prev_finish_at, job-
> >duration_us);
> +	}
> +	list_add_tail(&job->link, &sched->job_list);
> +	if (job->finish_at)
> +		hrtimer_start(&job->timer, job->finish_at,
> HRTIMER_MODE_ABS);
> +	spin_unlock_irq(&sched->lock);
> +
> +	return &job->hw_fence;
> +}
> +
> +static enum drm_gpu_sched_stat
> +mock_sched_timedout_job(struct drm_sched_job *sched_job)
> +{
> +	return DRM_GPU_SCHED_STAT_ENODEV;
> +}
> +
> +static void mock_sched_free_job(struct drm_sched_job *sched_job)
> +{
> +	drm_sched_job_cleanup(sched_job);
> +}
> +
> +static const struct drm_sched_backend_ops drm_mock_scheduler_ops = {
> +	.run_job = mock_sched_run_job,
> +	.timedout_job = mock_sched_timedout_job,
> +	.free_job = mock_sched_free_job
> +};
> +
> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
> *test)
> +{
> +	struct drm_mock_scheduler *sched;
> +	int ret;
> +
> +	sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL);
> +	KUNIT_ASSERT_NOT_NULL(test, sched);
> +
> +	ret = drm_sched_init(&sched->base,
> +			     &drm_mock_scheduler_ops,
> +			     NULL, /* wq */
> +			     DRM_SCHED_PRIORITY_COUNT,
> +			     U32_MAX, /* max credits */
> +			     UINT_MAX, /* hang limit */
> +			     MAX_SCHEDULE_TIMEOUT, /* timeout */
> +			     NULL, /* timeout wq */
> +			     NULL, /* score */
> +			     "drm-mock-scheduler",
> +			     NULL /* dev */);
> +	KUNIT_ASSERT_EQ(test, ret, 0);
> +
> +	sched->test = test;
> +	sched->hw_fence.context = dma_fence_context_alloc(1);
> +	atomic_set(&sched->hw_fence.seqno, 0);
> +	INIT_LIST_HEAD(&sched->job_list);
> +	spin_lock_init(&sched->lock);
> +
> +	return sched;
> +}
> +
> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched)
> +{
> +	struct drm_mock_sched_job *job, *next;
> +	unsigned long flags;
> +	LIST_HEAD(signal);
> +
> +	spin_lock_irqsave(&sched->lock, flags);
> +	list_for_each_entry_safe(job, next, &sched->job_list, link)
> +		list_move_tail(&job->link, &signal);
> +	spin_unlock_irqrestore(&sched->lock, flags);

So maybe you can help me get up to speed here a bit. What is the
purpose behind job->link? Is the general idea documented somewhere?


> +
> +	list_for_each_entry(job, &signal, link) {
> +		hrtimer_cancel(&job->timer);
> +		dma_fence_put(&job->hw_fence);
> +	}
> +
> +	drm_sched_fini(&sched->base);
> +}
> +
> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
> *sched,
> +				    unsigned int num)
> +{
> +	struct drm_mock_sched_job *job, *next;
> +	unsigned int found = 0;
> +	unsigned long flags;
> +	LIST_HEAD(signal);
> +
> +	spin_lock_irqsave(&sched->lock, flags);
> +	if (WARN_ON_ONCE(sched->seqno + num < sched->seqno))
> +		goto unlock;
> +	sched->seqno += num;
> +	list_for_each_entry_safe(job, next, &sched->job_list, link)
> {
> +		if (sched->seqno < job->hw_fence.seqno)
> +			break;
> +
> +		list_move_tail(&job->link, &signal);
> +		found++;
> +	}
> +unlock:
> +	spin_unlock_irqrestore(&sched->lock, flags);
> +
> +	list_for_each_entry(job, &signal, link) {
> +		dma_fence_signal(&job->hw_fence);
> +		dma_fence_put(&job->hw_fence);
> +	}
> +
> +	return found;
> +}
> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
> new file mode 100644
> index 000000000000..421ee2712985
> --- /dev/null
> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
> @@ -0,0 +1,124 @@
> +
> +#include <kunit/test.h>
> +#include <linux/atomic.h>
> +#include <linux/dma-fence.h>
> +#include <linux/hrtimer.h>
> +#include <linux/ktime.h>
> +#include <linux/list.h>
> +#include <linux/atomic.h>
> +#include <linux/mutex.h>
> +#include <linux/types.h>
> +
> +#include <drm/gpu_scheduler.h>
> +
> +struct drm_mock_scheduler {
> +	struct drm_gpu_scheduler base;
> +
> +	struct kunit		*test;
> +
> +	struct {
> +		u64		context;
> +		atomic_t	seqno;
> +	} hw_fence;

Hm, well, so this is confusing. drm_mock_sched_job below contains an
actual struct dma_fence that is also called hw_fence, whereas this here
seems to be some pseudo-hw_fence?

What is its purpose?

I believe we agreed that "Hardware fence" should always mean a fence
per job that is signaled by the hardware (driver interrupt) once the
job is done.

If this hw_fence here is the same, why is it per scheduler? That's
confusing.

> +
> +	spinlock_t		lock;

Maybe document the lock's purpose


> +	unsigned int		seqno;
> +	struct list_head	job_list;
> +};
> +
> +struct drm_mock_sched_entity {
> +	struct drm_sched_entity base;
> +
> +	struct kunit		*test;
> +};
> +
> +struct drm_mock_sched_job {
> +	struct drm_sched_job	base;
> +
> +	struct list_head	link;
> +	struct hrtimer		timer;
> +
> +	unsigned int		duration_us;
> +	ktime_t			finish_at;
> +
> +	spinlock_t		lock;

Same

> +	struct dma_fence	hw_fence;
> +
> +	struct kunit		*test;
> +};
> +
> +static inline struct drm_mock_scheduler *
> +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched)
> +{
> +	return container_of(sched, struct drm_mock_scheduler, base);
> +};
> +
> +static inline struct drm_mock_sched_entity *
> +drm_sched_entity_to_mock_entity(struct drm_sched_entity
> *sched_entity)
> +{
> +	return container_of(sched_entity, struct
> drm_mock_sched_entity, base);
> +};
> +
> +static inline struct drm_mock_sched_job *
> +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job)
> +{
> +	return container_of(sched_job, struct drm_mock_sched_job,
> base);
> +};
> +
> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
> *test);
> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched);
> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
> *sched,
> +				    unsigned int num);
> +
> +struct drm_mock_sched_entity *
> +drm_mock_new_sched_entity(struct kunit *test,
> +			  enum drm_sched_priority priority,
> +			  struct drm_mock_scheduler *sched);
> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
> *entity);
> +
> +struct drm_mock_sched_job *
> +drm_mock_new_sched_job(struct kunit *test,
> +		       struct drm_mock_sched_entity *entity);
> +
> +static inline void drm_mock_sched_job_submit(struct
> drm_mock_sched_job *job)
> +{
> +	drm_sched_job_arm(&job->base);
> +	drm_sched_entity_push_job(&job->base);
> +}
> +
> +static inline void
> +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job,
> +				   unsigned int duration_us)
> +{
> +	job->duration_us = duration_us;
> +}
> +
> +static inline bool
> +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job)
> +{
> +	return dma_fence_is_signaled(&job->base.s_fence->finished);
> +}
> +
> +static inline bool
> +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job,
> long timeout)
> +{
> +	long ret;
> +
> +	ret = dma_fence_wait_timeout(&job->base.s_fence->finished,
> +				      false,
> +				      timeout);
> +
> +	return ret != 0;
> +}
> +
> +static inline long
> +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job,
> long timeout)
> +{
> +	long ret;
> +
> +	ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled,
> +				      false,
> +				      timeout);
> +
> +	return ret != 0;
> +}
> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> new file mode 100644
> index 000000000000..6fd39bea95b1
> --- /dev/null
> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> @@ -0,0 +1,247 @@
> +
> +#include "drm_sched_tests.h"
> +
> +static int drm_sched_basic_init(struct kunit *test)
> +{
> +	test->priv = drm_mock_new_scheduler(test);
> +
> +	return 0;
> +}
> +
> +static void drm_sched_basic_exit(struct kunit *test)
> +{
> +	struct drm_mock_scheduler *sched = test->priv;
> +
> +	drm_mock_scheduler_fini(sched);
> +}
> +
> +static void drm_sched_basic_submit(struct kunit *test)
> +{
> +	struct drm_mock_scheduler *sched = test->priv;
> +	struct drm_mock_sched_entity *entity;
> +	struct drm_mock_sched_job *job;
> +	unsigned int i;
> +	bool done;
> +
> +	/*
> +	 * Submit one job to the scheduler and verify that it gets
> scheduled
> +	 * and completed only when the mock hw backend processes it.
> +	 */
> +
> +	entity = drm_mock_new_sched_entity(test,
> +					  
> DRM_SCHED_PRIORITY_NORMAL,
> +					   sched);
> +	job = drm_mock_new_sched_job(test, entity);
> +
> +	drm_mock_sched_job_submit(job);
> +
> +	done = drm_mock_sched_job_wait_scheduled(job, HZ);
> +	KUNIT_ASSERT_EQ(test, done, true);
> +
> +	done = drm_mock_sched_job_wait_finished(job, HZ / 2);
> +	KUNIT_ASSERT_EQ(test, done, false);
> +
> +	i = drm_mock_sched_advance(sched, 1);
> +	KUNIT_ASSERT_EQ(test, i, 1);
> +
> +	done = drm_mock_sched_job_wait_finished(job, HZ);
> +	KUNIT_ASSERT_EQ(test, done, true);
> +
> +	drm_mock_sched_entity_free(entity);
> +}
> +
> +static void drm_sched_basic_queue(struct kunit *test)
> +{
> +	struct drm_mock_scheduler *sched = test->priv;
> +	struct drm_mock_sched_entity *entity;
> +	struct drm_mock_sched_job *job;
> +	const unsigned int qd = 100;

Not the best variable name – is this "queue depth"? Why is it 100? ->
global define & document


> +	unsigned int i;
> +	bool done;
> +
> +	/*
> +	 * Submit a queue of jobs on the same entity, let them be
> completed by
> +	 * the mock hw backend and check the status of the last job.
> +	 */
> +
> +	entity = drm_mock_new_sched_entity(test,
> +					  
> DRM_SCHED_PRIORITY_NORMAL,
> +					   sched);
> +
> +	for (i = 0; i < qd; i++) {
> +		job = drm_mock_new_sched_job(test, entity);
> +		drm_mock_sched_job_set_duration_us(job, 1000);
> +		drm_mock_sched_job_submit(job);
> +	}
> +
> +	done = drm_mock_sched_job_wait_finished(job, HZ);
> +	KUNIT_ASSERT_EQ(test, done, true);
> +
> +	drm_mock_sched_entity_free(entity);
> +}
> +
> +static void drm_sched_basic_chain(struct kunit *test)
> +{
> +	struct drm_mock_scheduler *sched = test->priv;
> +	struct drm_mock_sched_job *job, *prev = NULL;
> +	struct drm_mock_sched_entity *entity;
> +	const unsigned int qd = 100;
> +	unsigned int i;
> +	bool done;
> +
> +	/*
> +	 * Submit a queue of jobs on the same entity but with an
> explicit
> +	 * chain of dependencies between them. Let the jobs be
> completed by
> +	 * the mock hw backend and check the status of the last job.
> +	 */
> +
> +	entity = drm_mock_new_sched_entity(test,
> +					  
> DRM_SCHED_PRIORITY_NORMAL,
> +					   sched);
> +
> +	for (i = 0; i < qd; i++) {
> +		job = drm_mock_new_sched_job(test, entity);
> +		drm_mock_sched_job_set_duration_us(job, 1000);
> +		if (prev)
> +			drm_sched_job_add_dependency(&job->base,
> +						    
> dma_fence_get(&prev->base.s_fence->finished));
> +		drm_mock_sched_job_submit(job);
> +		prev = job;
> +	}
> +
> +	done = drm_mock_sched_job_wait_finished(job, HZ);
> +	KUNIT_ASSERT_EQ(test, done, true);
> +
> +	drm_mock_sched_entity_free(entity);
> +}
> +
> +static void drm_sched_basic_entities(struct kunit *test)
> +{
> +	struct drm_mock_scheduler *sched = test->priv;
> +	struct drm_mock_sched_entity *entity[4];
> +	struct drm_mock_sched_job *job;
> +	const unsigned int qd = 100;
> +	unsigned int i, cur_ent = 0;
> +	bool done;
> +
> +	/*
> +	 * Submit a queue of jobs across different entities, let
> them be
> +	 * completed by the mock hw backend and check the status of
> the last
> +	 * job.
> +	 */
> +
> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> +		entity[i] = drm_mock_new_sched_entity(test,
> +						     
> DRM_SCHED_PRIORITY_NORMAL,
> +						      sched);
> +
> +	for (i = 0; i < qd; i++) {
> +		job = drm_mock_new_sched_job(test,
> entity[cur_ent++]);
> +		cur_ent %= ARRAY_SIZE(entity);
> +		drm_mock_sched_job_set_duration_us(job, 1000);
> +		drm_mock_sched_job_submit(job);
> +	}
> +
> +	done = drm_mock_sched_job_wait_finished(job, HZ);
> +	KUNIT_ASSERT_EQ(test, done, true);
> +
> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> +		drm_mock_sched_entity_free(entity[i]);
> +}
> +
> +static void drm_sched_basic_entities_chain(struct kunit *test)
> +{
> +	struct drm_mock_scheduler *sched = test->priv;
> +	struct drm_mock_sched_job *job, *prev = NULL;
> +	struct drm_mock_sched_entity *entity[4];
> +	const unsigned int qd = 100;
> +	unsigned int i, cur_ent = 0;
> +	bool done;
> +
> +	/*
> +	 * Submit a queue of jobs across different entities with an
> explicit
> +	 * chain of dependencies between them. Let the jobs be
> completed by
> +	 * the mock hw backend and check the status of the last job.
> +	 */
> +
> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> +		entity[i] = drm_mock_new_sched_entity(test,
> +						     
> DRM_SCHED_PRIORITY_NORMAL,
> +						      sched);
> +
> +	for (i = 0; i < qd; i++) {
> +		job = drm_mock_new_sched_job(test,
> entity[cur_ent++]);
> +		cur_ent %= ARRAY_SIZE(entity);
> +		drm_mock_sched_job_set_duration_us(job, 1000);
> +		if (prev)
> +			drm_sched_job_add_dependency(&job->base,
> +						    
> dma_fence_get(&prev->base.s_fence->finished));
> +		drm_mock_sched_job_submit(job);
> +		prev = job;
> +	}
> +
> +	done = drm_mock_sched_job_wait_finished(job, HZ);
> +	KUNIT_ASSERT_EQ(test, done, true);
> +
> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> +		drm_mock_sched_entity_free(entity[i]);
> +}
> +
> +static void drm_sched_basic_entity_cleanup(struct kunit *test)
> +{
> +	struct drm_mock_sched_job *job, *mid, *prev = NULL;
> +	struct drm_mock_scheduler *sched = test->priv;
> +	struct drm_mock_sched_entity *entity[4];

4 is used in several places, so could be defined globally. In case
there's a special reason for why it's 4, that could be mentioned, too


P.

> +	const unsigned int qd = 100;
> +	unsigned int i, cur_ent = 0;
> +	bool done;
> +
> +	/*
> +	 * Submit a queue of jobs across different entities with an
> explicit
> +	 * chain of dependencies between them and trigger entity
> cleanup while
> +	 * the queue is still being processed.
> +	 */
> +
> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> +		entity[i] = drm_mock_new_sched_entity(test,
> +						     
> DRM_SCHED_PRIORITY_NORMAL,
> +						      sched);
> +
> +	for (i = 0; i < qd; i++) {
> +		job = drm_mock_new_sched_job(test,
> entity[cur_ent++]);
> +		cur_ent %= ARRAY_SIZE(entity);
> +		drm_mock_sched_job_set_duration_us(job, 1000);
> +		if (prev)
> +			drm_sched_job_add_dependency(&job->base,
> +						    
> dma_fence_get(&prev->base.s_fence->finished));
> +		drm_mock_sched_job_submit(job);
> +		if (i == qd / 2)
> +			mid = job;
> +		prev = job;
> +	}
> +
> +	done = drm_mock_sched_job_wait_finished(mid, HZ);
> +	KUNIT_ASSERT_EQ(test, done, true);
> +
> +	/* Exit with half of the queue still pending to be executed.
> */
> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> +		drm_mock_sched_entity_free(entity[i]);}
> +
> +static struct kunit_case drm_sched_basic_tests[] = {
> +	KUNIT_CASE(drm_sched_basic_submit),
> +	KUNIT_CASE(drm_sched_basic_queue),
> +	KUNIT_CASE(drm_sched_basic_chain),
> +	KUNIT_CASE(drm_sched_basic_entities),
> +	KUNIT_CASE(drm_sched_basic_entities_chain),
> +	KUNIT_CASE(drm_sched_basic_entity_cleanup),
> +	{}
> +};
> +
> +static struct kunit_suite drm_sched_basic = {
> +	.name = "drm_sched_basic_tests",
> +	.init = drm_sched_basic_init,
> +	.exit = drm_sched_basic_exit,
> +	.test_cases = drm_sched_basic_tests,
> +};
> +
> +kunit_test_suite(drm_sched_basic);
Tvrtko Ursulin Feb. 6, 2025, 10:42 a.m. UTC | #2
On 06/02/2025 09:51, Philipp Stanner wrote:
> On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote:
>> Implement a mock scheduler backend and add some basic test to
>> exercise the
>> core scheduler code paths.
>>
>> Mock backend (kind of like a very simple mock GPU) can either process
>> jobs
>> by tests manually advancing the "timeline" job at a time, or
>> alternatively
>> jobs can be configured with a time duration in which case they get
>> completed asynchronously from the unit test code.
>>
>> Core scheduler classes are subclassed to support this mock
>> implementation.
>>
>> The tests added are just a few simple submission patterns.
>>
>> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com>
> 
> Could add a Suggested-by: Philipp :)

Will do.

>> Cc: Christian König <christian.koenig@amd.com>
>> Cc: Danilo Krummrich <dakr@kernel.org>
>> Cc: Matthew Brost <matthew.brost@intel.com>
>> Cc: Philipp Stanner <phasta@kernel.org>
>> ---
>>   drivers/gpu/drm/Kconfig.debug                 |  12 +
>>   drivers/gpu/drm/scheduler/.kunitconfig        |  12 +
>>   drivers/gpu/drm/scheduler/Makefile            |   1 +
>>   drivers/gpu/drm/scheduler/tests/Makefile      |   4 +
>>   .../gpu/drm/scheduler/tests/drm_mock_entity.c |  29 ++
>>   .../gpu/drm/scheduler/tests/drm_mock_job.c    |   3 +
>>   .../drm/scheduler/tests/drm_mock_scheduler.c  | 254
>> ++++++++++++++++++
>>   .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++
>>   .../scheduler/tests/drm_sched_tests_basic.c   | 247
>> +++++++++++++++++
>>   9 files changed, 686 insertions(+)
>>   create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig
>>   create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile
>>   create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>>   create mode 100644 drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>>   create mode 100644
>> drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>>   create mode 100644 drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>>   create mode 100644
>> drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>>
>> diff --git a/drivers/gpu/drm/Kconfig.debug
>> b/drivers/gpu/drm/Kconfig.debug
>> index a35d74171b7b..89f777f21e95 100644
>> --- a/drivers/gpu/drm/Kconfig.debug
>> +++ b/drivers/gpu/drm/Kconfig.debug
>> @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST
>>   
>>   	  If in doubt, say "N".
>>   
>> +config DRM_SCHED_KUNIT_TEST
>> +	tristate "KUnit tests for the DRM scheduler" if
>> !KUNIT_ALL_TESTS
>> +	select DRM_SCHED
>> +	depends on DRM && KUNIT
>> +	default KUNIT_ALL_TESTS
>> +	help
>> +	  Choose this option to build unit tests for the DRM
>> scheduler.
> 
> nit: Might say "DRM GPU scheduler" so readers not familiar with all
> that get a better idea of what it's about
> 
>> +
>> +	  Recommended for driver developers only.
> 
> nit: s/driver developers/DRM developers
> ?
> 
>> +
>> +	  If in doubt, say "N".
>> +
>>   config DRM_EXPORT_FOR_TESTS
>>   	bool
>> diff --git a/drivers/gpu/drm/scheduler/.kunitconfig
>> b/drivers/gpu/drm/scheduler/.kunitconfig
>> new file mode 100644
>> index 000000000000..cece53609fcf
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/.kunitconfig
>> @@ -0,0 +1,12 @@
>> +CONFIG_KUNIT=y
>> +CONFIG_DRM=y
>> +CONFIG_DRM_SCHED_KUNIT_TEST=y
>> +CONFIG_EXPERT=y
>> +CONFIG_DEBUG_SPINLOCK=y
>> +CONFIG_DEBUG_MUTEXES=y
>> +CONFIG_DEBUG_ATOMIC_SLEEP=y
>> +CONFIG_LOCK_DEBUGGING_SUPPORT=y
>> +CONFIG_PROVE_LOCKING=y
>> +CONFIG_LOCKDEP=y
>> +CONFIG_DEBUG_LOCKDEP=y
>> +CONFIG_DEBUG_LIST=y
>> diff --git a/drivers/gpu/drm/scheduler/Makefile
>> b/drivers/gpu/drm/scheduler/Makefile
>> index 53863621829f..46dfdca0758a 100644
>> --- a/drivers/gpu/drm/scheduler/Makefile
>> +++ b/drivers/gpu/drm/scheduler/Makefile
>> @@ -23,3 +23,4 @@
>>   gpu-sched-y := sched_main.o sched_fence.o sched_entity.o
>>   
>>   obj-$(CONFIG_DRM_SCHED) += gpu-sched.o
>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/
>> diff --git a/drivers/gpu/drm/scheduler/tests/Makefile
>> b/drivers/gpu/drm/scheduler/tests/Makefile
>> new file mode 100644
>> index 000000000000..d69eab6a2e9b
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/Makefile
>> @@ -0,0 +1,4 @@
>> +
>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \
>> +        drm_mock_scheduler.o \
>> +        drm_sched_tests_basic.o
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>> b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>> new file mode 100644
>> index 000000000000..c9205f9ff524
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>> @@ -0,0 +1,29 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> +struct drm_mock_sched_entity *
>> +drm_mock_sched_entity_new(struct kunit *test,
>> +			  enum drm_sched_priority priority,
>> +			  struct drm_mock_scheduler *sched)
> 
> 
> I personally do like this function head style and
> 
> 
>> +{
>> +	struct drm_sched_mock_entity *entity;
>> +	int ret;
>> +
>> +	entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL);
>> +	KUNIT_ASSERT_NOT_NULL(test, entity);
>> +
>> +	ret = drm_sched_entity_init(&entity->base,
>> +				    priority,
>> +				    &sched->base, 1,
>> +				    NULL);
>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> +	entity->test = test;
>> +
>> +	return entity;
>> +}
>> +
>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>> *entity)
> 
> do believe it should then be used consistently everywhere, regardless
> of length.
> 
> 
>> +{
>> +	drm_sched_entity_fini(&entity->base);
>> +}
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>> b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>> new file mode 100644
>> index 000000000000..d94568cf3da9
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>> @@ -0,0 +1,3 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>> b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>> new file mode 100644
>> index 000000000000..f1985900a6ba
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>> @@ -0,0 +1,254 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> +struct drm_mock_sched_entity *
>> +drm_mock_new_sched_entity(struct kunit *test,
>> +			  enum drm_sched_priority priority,
>> +			  struct drm_mock_scheduler *sched)
>> +{
>> +	struct drm_mock_sched_entity *entity;
>> +	struct drm_gpu_scheduler *drm_sched;
>> +	int ret;
>> +
>> +	entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL);
>> +	KUNIT_ASSERT_NOT_NULL(test, entity);
>> +
>> +	drm_sched = &sched->base;
>> +	ret = drm_sched_entity_init(&entity->base,
>> +				    priority,
>> +				    &drm_sched, 1,
>> +				    NULL);
>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> +	entity->test = test;
>> +
>> +	return entity;
>> +}
>> +
>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>> *entity)
>> +{
>> +	drm_sched_entity_destroy(&entity->base);
>> +}
>> +
>> +static enum hrtimer_restart
>> +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer)
> 
> I think we should get up with a consistent naming convention. Some
> things are called drm_mock_sched_, some others drm_sched_mock_.
> As far as I saw, drm_mock_* does not yet exist.
> So do you want to introduce drm_mock as something generic for drm? or
> just for drm/sched?

This one is literally the only when where I tranposed the two. But it is 
also local and I am not too bothered about naming conventions of a 
local functions. If it were to me I would favour brevity and emit the 
long drm_.. prefixes which IMO do not help readability. If anything, 
seeing something with a long drm_.. prefix can only be confusing since 
one can assume it is some sort of exported interface.

I will change this instance.

>> +{
>> +	struct drm_mock_sched_job *upto =
>> +		container_of(hrtimer, typeof(*upto), timer);
>> +	struct drm_mock_scheduler *sched =
>> +		drm_sched_to_mock_sched(upto->base.sched);
>> +	struct drm_mock_sched_job *job, *next;
>> +	ktime_t now = ktime_get();
>> +	unsigned long flags;
>> +	LIST_HEAD(signal);
>> +
>> +	spin_lock_irqsave(&sched->lock, flags);
>> +	list_for_each_entry_safe(job, next, &sched->job_list, link)
>> {
>> +		if (!job->duration_us)
>> +			break;
>> +
>> +		if (ktime_before(now, job->finish_at))
>> +			break;
>> +
>> +		list_move_tail(&job->link, &signal);
>> +		sched->seqno = job->hw_fence.seqno;
>> +	}
>> +	spin_unlock_irqrestore(&sched->lock, flags);
>> +
>> +	list_for_each_entry(job, &signal, link) {
>> +		dma_fence_signal(&job->hw_fence);
>> +		dma_fence_put(&job->hw_fence);
>> +	}
>> +
>> +	return HRTIMER_NORESTART;
>> +}
>> +
>> +struct drm_mock_sched_job *
>> +drm_mock_new_sched_job(struct kunit *test,
>> +		       struct drm_mock_sched_entity *entity)
>> +{
>> +	struct drm_mock_sched_job *job;
>> +	int ret;
>> +
>> +	job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL);
>> +	KUNIT_ASSERT_NOT_NULL(test, job);
>> +
>> +	ret = drm_sched_job_init(&job->base,
>> +				 &entity->base,
>> +				 1,
>> +				 NULL);
>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> +	job->test = test;
>> +
>> +	spin_lock_init(&job->lock);
>> +	INIT_LIST_HEAD(&job->link);
>> +	hrtimer_init(&job->timer, CLOCK_MONOTONIC,
>> HRTIMER_MODE_ABS);
>> +	job->timer.function = drm_sched_mock_job_signal_timer;
>> +
>> +	return job;
>> +}
>> +
>> +static const char *drm_mock_sched_hw_fence_driver_name(struct
>> dma_fence *fence)
>> +{
>> +	return "drm_mock_sched";
>> +}
>> +
>> +static const char *
>> +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence)
>> +{
>> +	struct drm_mock_sched_job *job =
>> +		container_of(fence, typeof(*job), hw_fence);
>> +
>> +	return (const char *)job->base.sched->name;
>> +}
>> +static void drm_mock_sched_hw_fence_release(struct dma_fence *fence)
>> +{
>> +	struct drm_mock_sched_job *job =
>> +		container_of(fence, typeof(*job), hw_fence);
>> +
>> +	hrtimer_cancel(&job->timer);
>> +
>> +	/* Freed by the kunit framework */
>> +}
>> +
>> +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = {
>> +	.get_driver_name = drm_mock_sched_hw_fence_driver_name,
>> +	.get_timeline_name = drm_mock_sched_hw_fence_timeline_name,
>> +	.release = drm_mock_sched_hw_fence_release,
>> +};
>> +
>> +static struct dma_fence *mock_sched_run_job(struct drm_sched_job
>> *sched_job)
>> +{
>> +	struct drm_mock_scheduler *sched =
>> +		drm_sched_to_mock_sched(sched_job->sched);
>> +	struct drm_mock_sched_job *job =
>> drm_sched_job_to_mock_job(sched_job);
>> +
>> +	dma_fence_init(&job->hw_fence,
>> +		       &drm_mock_sched_hw_fence_ops,
>> +		       &job->lock,
>> +		       sched->hw_fence.context,
>> +		       atomic_inc_return(&sched->hw_fence.seqno));
>> +
>> +	dma_fence_get(&job->hw_fence); /* Reference for the job_list
>> */
>> +
>> +	spin_lock_irq(&sched->lock);
>> +	if (job->duration_us) {
>> +		ktime_t prev_finish_at = 0;
>> +
>> +		if (!list_empty(&sched->job_list)) {
>> +			struct drm_mock_sched_job *prev =
>> +				list_last_entry(&sched->job_list,
>> typeof(*prev),
>> +						link);
>> +
>> +			prev_finish_at = prev->finish_at;
>> +		}
>> +
>> +		if (!prev_finish_at)
>> +			prev_finish_at = ktime_get();
>> +
>> +		job->finish_at = ktime_add_us(prev_finish_at, job-
>>> duration_us);
>> +	}
>> +	list_add_tail(&job->link, &sched->job_list);
>> +	if (job->finish_at)
>> +		hrtimer_start(&job->timer, job->finish_at,
>> HRTIMER_MODE_ABS);
>> +	spin_unlock_irq(&sched->lock);
>> +
>> +	return &job->hw_fence;
>> +}
>> +
>> +static enum drm_gpu_sched_stat
>> +mock_sched_timedout_job(struct drm_sched_job *sched_job)
>> +{
>> +	return DRM_GPU_SCHED_STAT_ENODEV;
>> +}
>> +
>> +static void mock_sched_free_job(struct drm_sched_job *sched_job)
>> +{
>> +	drm_sched_job_cleanup(sched_job);
>> +}
>> +
>> +static const struct drm_sched_backend_ops drm_mock_scheduler_ops = {
>> +	.run_job = mock_sched_run_job,
>> +	.timedout_job = mock_sched_timedout_job,
>> +	.free_job = mock_sched_free_job
>> +};
>> +
>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
>> *test)
>> +{
>> +	struct drm_mock_scheduler *sched;
>> +	int ret;
>> +
>> +	sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL);
>> +	KUNIT_ASSERT_NOT_NULL(test, sched);
>> +
>> +	ret = drm_sched_init(&sched->base,
>> +			     &drm_mock_scheduler_ops,
>> +			     NULL, /* wq */
>> +			     DRM_SCHED_PRIORITY_COUNT,
>> +			     U32_MAX, /* max credits */
>> +			     UINT_MAX, /* hang limit */
>> +			     MAX_SCHEDULE_TIMEOUT, /* timeout */
>> +			     NULL, /* timeout wq */
>> +			     NULL, /* score */
>> +			     "drm-mock-scheduler",
>> +			     NULL /* dev */);
>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>> +
>> +	sched->test = test;
>> +	sched->hw_fence.context = dma_fence_context_alloc(1);
>> +	atomic_set(&sched->hw_fence.seqno, 0);
>> +	INIT_LIST_HEAD(&sched->job_list);
>> +	spin_lock_init(&sched->lock);
>> +
>> +	return sched;
>> +}
>> +
>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched)
>> +{
>> +	struct drm_mock_sched_job *job, *next;
>> +	unsigned long flags;
>> +	LIST_HEAD(signal);
>> +
>> +	spin_lock_irqsave(&sched->lock, flags);
>> +	list_for_each_entry_safe(job, next, &sched->job_list, link)
>> +		list_move_tail(&job->link, &signal);
>> +	spin_unlock_irqrestore(&sched->lock, flags);
> 
> So maybe you can help me get up to speed here a bit. What is the
> purpose behind job->link? Is the general idea documented somewhere?

List versus link suffix convention I use to distinguish struct list_head 
which is a list versus struct list_head which is used to put something 
on the list.

In this case job->link is for the mock GPU driver to keep track of jobs 
which have been submitted to the "hardware" for "execution".

I will of course document these things once the high level design is 
settled.

>> +
>> +	list_for_each_entry(job, &signal, link) {
>> +		hrtimer_cancel(&job->timer);
>> +		dma_fence_put(&job->hw_fence);
>> +	}
>> +
>> +	drm_sched_fini(&sched->base);
>> +}
>> +
>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
>> *sched,
>> +				    unsigned int num)
>> +{
>> +	struct drm_mock_sched_job *job, *next;
>> +	unsigned int found = 0;
>> +	unsigned long flags;
>> +	LIST_HEAD(signal);
>> +
>> +	spin_lock_irqsave(&sched->lock, flags);
>> +	if (WARN_ON_ONCE(sched->seqno + num < sched->seqno))
>> +		goto unlock;
>> +	sched->seqno += num;
>> +	list_for_each_entry_safe(job, next, &sched->job_list, link)
>> {
>> +		if (sched->seqno < job->hw_fence.seqno)
>> +			break;
>> +
>> +		list_move_tail(&job->link, &signal);
>> +		found++;
>> +	}
>> +unlock:
>> +	spin_unlock_irqrestore(&sched->lock, flags);
>> +
>> +	list_for_each_entry(job, &signal, link) {
>> +		dma_fence_signal(&job->hw_fence);
>> +		dma_fence_put(&job->hw_fence);
>> +	}
>> +
>> +	return found;
>> +}
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>> new file mode 100644
>> index 000000000000..421ee2712985
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>> @@ -0,0 +1,124 @@
>> +
>> +#include <kunit/test.h>
>> +#include <linux/atomic.h>
>> +#include <linux/dma-fence.h>
>> +#include <linux/hrtimer.h>
>> +#include <linux/ktime.h>
>> +#include <linux/list.h>
>> +#include <linux/atomic.h>
>> +#include <linux/mutex.h>
>> +#include <linux/types.h>
>> +
>> +#include <drm/gpu_scheduler.h>
>> +
>> +struct drm_mock_scheduler {
>> +	struct drm_gpu_scheduler base;
>> +
>> +	struct kunit		*test;
>> +
>> +	struct {
>> +		u64		context;
>> +		atomic_t	seqno;
>> +	} hw_fence;
> 
> Hm, well, so this is confusing. drm_mock_sched_job below contains an
> actual struct dma_fence that is also called hw_fence, whereas this here
> seems to be some pseudo-hw_fence?
> 
> What is its purpose?
> 
> I believe we agreed that "Hardware fence" should always mean a fence
> per job that is signaled by the hardware (driver interrupt) once the
> job is done.
> 
> If this hw_fence here is the same, why is it per scheduler? That's
> confusing.

Mock_job->hw_fence is what the mock GPU driver returns from the 
sched->run_job().

Mock_sched->hw_fence is what the mock GPU driver uses to track execution 
of jobs submitted to it for execution. If Irename this to hw_timeline 
will it make it clearer?

>> +
>> +	spinlock_t		lock;
> 
> Maybe document the lock's purpose
> 
> 
>> +	unsigned int		seqno;
>> +	struct list_head	job_list;
>> +};
>> +
>> +struct drm_mock_sched_entity {
>> +	struct drm_sched_entity base;
>> +
>> +	struct kunit		*test;
>> +};
>> +
>> +struct drm_mock_sched_job {
>> +	struct drm_sched_job	base;
>> +
>> +	struct list_head	link;
>> +	struct hrtimer		timer;
>> +
>> +	unsigned int		duration_us;
>> +	ktime_t			finish_at;
>> +
>> +	spinlock_t		lock;
> 
> Same
> 
>> +	struct dma_fence	hw_fence;
>> +
>> +	struct kunit		*test;
>> +};
>> +
>> +static inline struct drm_mock_scheduler *
>> +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched)
>> +{
>> +	return container_of(sched, struct drm_mock_scheduler, base);
>> +};
>> +
>> +static inline struct drm_mock_sched_entity *
>> +drm_sched_entity_to_mock_entity(struct drm_sched_entity
>> *sched_entity)
>> +{
>> +	return container_of(sched_entity, struct
>> drm_mock_sched_entity, base);
>> +};
>> +
>> +static inline struct drm_mock_sched_job *
>> +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job)
>> +{
>> +	return container_of(sched_job, struct drm_mock_sched_job,
>> base);
>> +};
>> +
>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
>> *test);
>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched);
>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
>> *sched,
>> +				    unsigned int num);
>> +
>> +struct drm_mock_sched_entity *
>> +drm_mock_new_sched_entity(struct kunit *test,
>> +			  enum drm_sched_priority priority,
>> +			  struct drm_mock_scheduler *sched);
>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>> *entity);
>> +
>> +struct drm_mock_sched_job *
>> +drm_mock_new_sched_job(struct kunit *test,
>> +		       struct drm_mock_sched_entity *entity);
>> +
>> +static inline void drm_mock_sched_job_submit(struct
>> drm_mock_sched_job *job)
>> +{
>> +	drm_sched_job_arm(&job->base);
>> +	drm_sched_entity_push_job(&job->base);
>> +}
>> +
>> +static inline void
>> +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job,
>> +				   unsigned int duration_us)
>> +{
>> +	job->duration_us = duration_us;
>> +}
>> +
>> +static inline bool
>> +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job)
>> +{
>> +	return dma_fence_is_signaled(&job->base.s_fence->finished);
>> +}
>> +
>> +static inline bool
>> +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job,
>> long timeout)
>> +{
>> +	long ret;
>> +
>> +	ret = dma_fence_wait_timeout(&job->base.s_fence->finished,
>> +				      false,
>> +				      timeout);
>> +
>> +	return ret != 0;
>> +}
>> +
>> +static inline long
>> +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job,
>> long timeout)
>> +{
>> +	long ret;
>> +
>> +	ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled,
>> +				      false,
>> +				      timeout);
>> +
>> +	return ret != 0;
>> +}
>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>> new file mode 100644
>> index 000000000000..6fd39bea95b1
>> --- /dev/null
>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>> @@ -0,0 +1,247 @@
>> +
>> +#include "drm_sched_tests.h"
>> +
>> +static int drm_sched_basic_init(struct kunit *test)
>> +{
>> +	test->priv = drm_mock_new_scheduler(test);
>> +
>> +	return 0;
>> +}
>> +
>> +static void drm_sched_basic_exit(struct kunit *test)
>> +{
>> +	struct drm_mock_scheduler *sched = test->priv;
>> +
>> +	drm_mock_scheduler_fini(sched);
>> +}
>> +
>> +static void drm_sched_basic_submit(struct kunit *test)
>> +{
>> +	struct drm_mock_scheduler *sched = test->priv;
>> +	struct drm_mock_sched_entity *entity;
>> +	struct drm_mock_sched_job *job;
>> +	unsigned int i;
>> +	bool done;
>> +
>> +	/*
>> +	 * Submit one job to the scheduler and verify that it gets
>> scheduled
>> +	 * and completed only when the mock hw backend processes it.
>> +	 */
>> +
>> +	entity = drm_mock_new_sched_entity(test,
>> +					
>> DRM_SCHED_PRIORITY_NORMAL,
>> +					   sched);
>> +	job = drm_mock_new_sched_job(test, entity);
>> +
>> +	drm_mock_sched_job_submit(job);
>> +
>> +	done = drm_mock_sched_job_wait_scheduled(job, HZ);
>> +	KUNIT_ASSERT_EQ(test, done, true);
>> +
>> +	done = drm_mock_sched_job_wait_finished(job, HZ / 2);
>> +	KUNIT_ASSERT_EQ(test, done, false);
>> +
>> +	i = drm_mock_sched_advance(sched, 1);
>> +	KUNIT_ASSERT_EQ(test, i, 1);
>> +
>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>> +	KUNIT_ASSERT_EQ(test, done, true);
>> +
>> +	drm_mock_sched_entity_free(entity);
>> +}
>> +
>> +static void drm_sched_basic_queue(struct kunit *test)
>> +{
>> +	struct drm_mock_scheduler *sched = test->priv;
>> +	struct drm_mock_sched_entity *entity;
>> +	struct drm_mock_sched_job *job;
>> +	const unsigned int qd = 100;
> 
> Not the best variable name – is this "queue depth"? Why is it 100? ->
> global define & document

I wouldn't promote this to global and TBH if you look how small this 
test function is it feels pretty self documenting.

> 
> 
>> +	unsigned int i;
>> +	bool done;
>> +
>> +	/*
>> +	 * Submit a queue of jobs on the same entity, let them be
>> completed by
>> +	 * the mock hw backend and check the status of the last job.
>> +	 */
>> +
>> +	entity = drm_mock_new_sched_entity(test,
>> +					
>> DRM_SCHED_PRIORITY_NORMAL,
>> +					   sched);
>> +
>> +	for (i = 0; i < qd; i++) {
>> +		job = drm_mock_new_sched_job(test, entity);
>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>> +		drm_mock_sched_job_submit(job);
>> +	}
>> +
>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>> +	KUNIT_ASSERT_EQ(test, done, true);
>> +
>> +	drm_mock_sched_entity_free(entity);
>> +}
>> +
>> +static void drm_sched_basic_chain(struct kunit *test)
>> +{
>> +	struct drm_mock_scheduler *sched = test->priv;
>> +	struct drm_mock_sched_job *job, *prev = NULL;
>> +	struct drm_mock_sched_entity *entity;
>> +	const unsigned int qd = 100;
>> +	unsigned int i;
>> +	bool done;
>> +
>> +	/*
>> +	 * Submit a queue of jobs on the same entity but with an
>> explicit
>> +	 * chain of dependencies between them. Let the jobs be
>> completed by
>> +	 * the mock hw backend and check the status of the last job.
>> +	 */
>> +
>> +	entity = drm_mock_new_sched_entity(test,
>> +					
>> DRM_SCHED_PRIORITY_NORMAL,
>> +					   sched);
>> +
>> +	for (i = 0; i < qd; i++) {
>> +		job = drm_mock_new_sched_job(test, entity);
>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>> +		if (prev)
>> +			drm_sched_job_add_dependency(&job->base,
>> +						
>> dma_fence_get(&prev->base.s_fence->finished));
>> +		drm_mock_sched_job_submit(job);
>> +		prev = job;
>> +	}
>> +
>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>> +	KUNIT_ASSERT_EQ(test, done, true);
>> +
>> +	drm_mock_sched_entity_free(entity);
>> +}
>> +
>> +static void drm_sched_basic_entities(struct kunit *test)
>> +{
>> +	struct drm_mock_scheduler *sched = test->priv;
>> +	struct drm_mock_sched_entity *entity[4];
>> +	struct drm_mock_sched_job *job;
>> +	const unsigned int qd = 100;
>> +	unsigned int i, cur_ent = 0;
>> +	bool done;
>> +
>> +	/*
>> +	 * Submit a queue of jobs across different entities, let
>> them be
>> +	 * completed by the mock hw backend and check the status of
>> the last
>> +	 * job.
>> +	 */
>> +
>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>> +		entity[i] = drm_mock_new_sched_entity(test,
>> +						
>> DRM_SCHED_PRIORITY_NORMAL,
>> +						      sched);
>> +
>> +	for (i = 0; i < qd; i++) {
>> +		job = drm_mock_new_sched_job(test,
>> entity[cur_ent++]);
>> +		cur_ent %= ARRAY_SIZE(entity);
>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>> +		drm_mock_sched_job_submit(job);
>> +	}
>> +
>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>> +	KUNIT_ASSERT_EQ(test, done, true);
>> +
>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>> +		drm_mock_sched_entity_free(entity[i]);
>> +}
>> +
>> +static void drm_sched_basic_entities_chain(struct kunit *test)
>> +{
>> +	struct drm_mock_scheduler *sched = test->priv;
>> +	struct drm_mock_sched_job *job, *prev = NULL;
>> +	struct drm_mock_sched_entity *entity[4];
>> +	const unsigned int qd = 100;
>> +	unsigned int i, cur_ent = 0;
>> +	bool done;
>> +
>> +	/*
>> +	 * Submit a queue of jobs across different entities with an
>> explicit
>> +	 * chain of dependencies between them. Let the jobs be
>> completed by
>> +	 * the mock hw backend and check the status of the last job.
>> +	 */
>> +
>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>> +		entity[i] = drm_mock_new_sched_entity(test,
>> +						
>> DRM_SCHED_PRIORITY_NORMAL,
>> +						      sched);
>> +
>> +	for (i = 0; i < qd; i++) {
>> +		job = drm_mock_new_sched_job(test,
>> entity[cur_ent++]);
>> +		cur_ent %= ARRAY_SIZE(entity);
>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>> +		if (prev)
>> +			drm_sched_job_add_dependency(&job->base,
>> +						
>> dma_fence_get(&prev->base.s_fence->finished));
>> +		drm_mock_sched_job_submit(job);
>> +		prev = job;
>> +	}
>> +
>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>> +	KUNIT_ASSERT_EQ(test, done, true);
>> +
>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>> +		drm_mock_sched_entity_free(entity[i]);
>> +}
>> +
>> +static void drm_sched_basic_entity_cleanup(struct kunit *test)
>> +{
>> +	struct drm_mock_sched_job *job, *mid, *prev = NULL;
>> +	struct drm_mock_scheduler *sched = test->priv;
>> +	struct drm_mock_sched_entity *entity[4];
> 
> 4 is used in several places, so could be defined globally. In case
> there's a special reason for why it's 4, that could be mentioned, too

It's completely arbitrary. In this test we just need a bunch of entities 
to be active on the scheduler as we trigger a client exit in the middle 
of it. IMO the comment few lines below should be good enough to explain 
the idea. I fear that promoting this to a global define would just give 
it more weight that what it has. Then some test would want a different 
number etc.

Regards,

Tvrtko

>> +	const unsigned int qd = 100;
>> +	unsigned int i, cur_ent = 0;
>> +	bool done;
>> +
>> +	/*
>> +	 * Submit a queue of jobs across different entities with an
>> explicit
>> +	 * chain of dependencies between them and trigger entity
>> cleanup while
>> +	 * the queue is still being processed.
>> +	 */
>> +
>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>> +		entity[i] = drm_mock_new_sched_entity(test,
>> +						
>> DRM_SCHED_PRIORITY_NORMAL,
>> +						      sched);
>> +
>> +	for (i = 0; i < qd; i++) {
>> +		job = drm_mock_new_sched_job(test,
>> entity[cur_ent++]);
>> +		cur_ent %= ARRAY_SIZE(entity);
>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>> +		if (prev)
>> +			drm_sched_job_add_dependency(&job->base,
>> +						
>> dma_fence_get(&prev->base.s_fence->finished));
>> +		drm_mock_sched_job_submit(job);
>> +		if (i == qd / 2)
>> +			mid = job;
>> +		prev = job;
>> +	}
>> +
>> +	done = drm_mock_sched_job_wait_finished(mid, HZ);
>> +	KUNIT_ASSERT_EQ(test, done, true);
>> +
>> +	/* Exit with half of the queue still pending to be executed.
>> */
>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>> +		drm_mock_sched_entity_free(entity[i]);}
>> +
>> +static struct kunit_case drm_sched_basic_tests[] = {
>> +	KUNIT_CASE(drm_sched_basic_submit),
>> +	KUNIT_CASE(drm_sched_basic_queue),
>> +	KUNIT_CASE(drm_sched_basic_chain),
>> +	KUNIT_CASE(drm_sched_basic_entities),
>> +	KUNIT_CASE(drm_sched_basic_entities_chain),
>> +	KUNIT_CASE(drm_sched_basic_entity_cleanup),
>> +	{}
>> +};
>> +
>> +static struct kunit_suite drm_sched_basic = {
>> +	.name = "drm_sched_basic_tests",
>> +	.init = drm_sched_basic_init,
>> +	.exit = drm_sched_basic_exit,
>> +	.test_cases = drm_sched_basic_tests,
>> +};
>> +
>> +kunit_test_suite(drm_sched_basic);
>
Philipp Stanner Feb. 6, 2025, 12:33 p.m. UTC | #3
On Thu, 2025-02-06 at 10:42 +0000, Tvrtko Ursulin wrote:
> 
> On 06/02/2025 09:51, Philipp Stanner wrote:
> > On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote:
> > > Implement a mock scheduler backend and add some basic test to
> > > exercise the
> > > core scheduler code paths.
> > > 
> > > Mock backend (kind of like a very simple mock GPU) can either
> > > process
> > > jobs
> > > by tests manually advancing the "timeline" job at a time, or
> > > alternatively
> > > jobs can be configured with a time duration in which case they
> > > get
> > > completed asynchronously from the unit test code.
> > > 
> > > Core scheduler classes are subclassed to support this mock
> > > implementation.
> > > 
> > > The tests added are just a few simple submission patterns.
> > > 
> > > Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com>
> > 
> > Could add a Suggested-by: Philipp :)
> 
> Will do.
> 
> > > Cc: Christian König <christian.koenig@amd.com>
> > > Cc: Danilo Krummrich <dakr@kernel.org>
> > > Cc: Matthew Brost <matthew.brost@intel.com>
> > > Cc: Philipp Stanner <phasta@kernel.org>
> > > ---
> > >   drivers/gpu/drm/Kconfig.debug                 |  12 +
> > >   drivers/gpu/drm/scheduler/.kunitconfig        |  12 +
> > >   drivers/gpu/drm/scheduler/Makefile            |   1 +
> > >   drivers/gpu/drm/scheduler/tests/Makefile      |   4 +
> > >   .../gpu/drm/scheduler/tests/drm_mock_entity.c |  29 ++
> > >   .../gpu/drm/scheduler/tests/drm_mock_job.c    |   3 +
> > >   .../drm/scheduler/tests/drm_mock_scheduler.c  | 254
> > > ++++++++++++++++++
> > >   .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++
> > >   .../scheduler/tests/drm_sched_tests_basic.c   | 247
> > > +++++++++++++++++
> > >   9 files changed, 686 insertions(+)
> > >   create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig
> > >   create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile
> > >   create mode 100644
> > > drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
> > >   create mode 100644
> > > drivers/gpu/drm/scheduler/tests/drm_mock_job.c
> > >   create mode 100644
> > > drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
> > >   create mode 100644
> > > drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
> > >   create mode 100644
> > > drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> > > 
> > > diff --git a/drivers/gpu/drm/Kconfig.debug
> > > b/drivers/gpu/drm/Kconfig.debug
> > > index a35d74171b7b..89f777f21e95 100644
> > > --- a/drivers/gpu/drm/Kconfig.debug
> > > +++ b/drivers/gpu/drm/Kconfig.debug
> > > @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST
> > >   
> > >   	  If in doubt, say "N".
> > >   
> > > +config DRM_SCHED_KUNIT_TEST
> > > +	tristate "KUnit tests for the DRM scheduler" if
> > > !KUNIT_ALL_TESTS
> > > +	select DRM_SCHED
> > > +	depends on DRM && KUNIT
> > > +	default KUNIT_ALL_TESTS
> > > +	help
> > > +	  Choose this option to build unit tests for the DRM
> > > scheduler.
> > 
> > nit: Might say "DRM GPU scheduler" so readers not familiar with all
> > that get a better idea of what it's about
> > 
> > > +
> > > +	  Recommended for driver developers only.
> > 
> > nit: s/driver developers/DRM developers
> > ?
> > 
> > > +
> > > +	  If in doubt, say "N".
> > > +
> > >   config DRM_EXPORT_FOR_TESTS
> > >   	bool
> > > diff --git a/drivers/gpu/drm/scheduler/.kunitconfig
> > > b/drivers/gpu/drm/scheduler/.kunitconfig
> > > new file mode 100644
> > > index 000000000000..cece53609fcf
> > > --- /dev/null
> > > +++ b/drivers/gpu/drm/scheduler/.kunitconfig
> > > @@ -0,0 +1,12 @@
> > > +CONFIG_KUNIT=y
> > > +CONFIG_DRM=y
> > > +CONFIG_DRM_SCHED_KUNIT_TEST=y
> > > +CONFIG_EXPERT=y
> > > +CONFIG_DEBUG_SPINLOCK=y
> > > +CONFIG_DEBUG_MUTEXES=y
> > > +CONFIG_DEBUG_ATOMIC_SLEEP=y
> > > +CONFIG_LOCK_DEBUGGING_SUPPORT=y
> > > +CONFIG_PROVE_LOCKING=y
> > > +CONFIG_LOCKDEP=y
> > > +CONFIG_DEBUG_LOCKDEP=y
> > > +CONFIG_DEBUG_LIST=y
> > > diff --git a/drivers/gpu/drm/scheduler/Makefile
> > > b/drivers/gpu/drm/scheduler/Makefile
> > > index 53863621829f..46dfdca0758a 100644
> > > --- a/drivers/gpu/drm/scheduler/Makefile
> > > +++ b/drivers/gpu/drm/scheduler/Makefile
> > > @@ -23,3 +23,4 @@
> > >   gpu-sched-y := sched_main.o sched_fence.o sched_entity.o
> > >   
> > >   obj-$(CONFIG_DRM_SCHED) += gpu-sched.o
> > > +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/
> > > diff --git a/drivers/gpu/drm/scheduler/tests/Makefile
> > > b/drivers/gpu/drm/scheduler/tests/Makefile
> > > new file mode 100644
> > > index 000000000000..d69eab6a2e9b
> > > --- /dev/null
> > > +++ b/drivers/gpu/drm/scheduler/tests/Makefile
> > > @@ -0,0 +1,4 @@
> > > +
> > > +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \
> > > +        drm_mock_scheduler.o \
> > > +        drm_sched_tests_basic.o
> > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
> > > b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
> > > new file mode 100644
> > > index 000000000000..c9205f9ff524
> > > --- /dev/null
> > > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
> > > @@ -0,0 +1,29 @@
> > > +
> > > +#include "drm_sched_tests.h"
> > > +
> > > +struct drm_mock_sched_entity *
> > > +drm_mock_sched_entity_new(struct kunit *test,
> > > +			  enum drm_sched_priority priority,
> > > +			  struct drm_mock_scheduler *sched)
> > 
> > 
> > I personally do like this function head style and
> > 
> > 
> > > +{
> > > +	struct drm_sched_mock_entity *entity;
> > > +	int ret;
> > > +
> > > +	entity = kunit_kmalloc(test, sizeof(*entity),
> > > GFP_KERNEL);
> > > +	KUNIT_ASSERT_NOT_NULL(test, entity);
> > > +
> > > +	ret = drm_sched_entity_init(&entity->base,
> > > +				    priority,
> > > +				    &sched->base, 1,
> > > +				    NULL);
> > > +	KUNIT_ASSERT_EQ(test, ret, 0);
> > > +
> > > +	entity->test = test;
> > > +
> > > +	return entity;
> > > +}
> > > +
> > > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
> > > *entity)
> > 
> > do believe it should then be used consistently everywhere,
> > regardless
> > of length.
> > 
> > 
> > > +{
> > > +	drm_sched_entity_fini(&entity->base);
> > > +}
> > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
> > > b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
> > > new file mode 100644
> > > index 000000000000..d94568cf3da9
> > > --- /dev/null
> > > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
> > > @@ -0,0 +1,3 @@
> > > +
> > > +#include "drm_sched_tests.h"
> > > +
> > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
> > > b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
> > > new file mode 100644
> > > index 000000000000..f1985900a6ba
> > > --- /dev/null
> > > +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
> > > @@ -0,0 +1,254 @@
> > > +
> > > +#include "drm_sched_tests.h"
> > > +
> > > +struct drm_mock_sched_entity *
> > > +drm_mock_new_sched_entity(struct kunit *test,
> > > +			  enum drm_sched_priority priority,
> > > +			  struct drm_mock_scheduler *sched)
> > > +{
> > > +	struct drm_mock_sched_entity *entity;
> > > +	struct drm_gpu_scheduler *drm_sched;
> > > +	int ret;
> > > +
> > > +	entity = kunit_kzalloc(test, sizeof(*entity),
> > > GFP_KERNEL);
> > > +	KUNIT_ASSERT_NOT_NULL(test, entity);
> > > +
> > > +	drm_sched = &sched->base;
> > > +	ret = drm_sched_entity_init(&entity->base,
> > > +				    priority,
> > > +				    &drm_sched, 1,
> > > +				    NULL);
> > > +	KUNIT_ASSERT_EQ(test, ret, 0);
> > > +
> > > +	entity->test = test;
> > > +
> > > +	return entity;
> > > +}
> > > +
> > > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
> > > *entity)
> > > +{
> > > +	drm_sched_entity_destroy(&entity->base);
> > > +}
> > > +
> > > +static enum hrtimer_restart
> > > +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer)
> > 
> > I think we should get up with a consistent naming convention. Some
> > things are called drm_mock_sched_, some others drm_sched_mock_.
> > As far as I saw, drm_mock_* does not yet exist.
> > So do you want to introduce drm_mock as something generic for drm?
> > or
> > just for drm/sched?
> 
> This one is literally the only when where I tranposed the two. But it
> is 
> also local and I am not too bothered about naming conventions of a 
> local functions. If it were to me I would favour brevity and emit the
> long drm_.. prefixes which IMO do not help readability.

I tentatively agree with that. I just think whatever the style is, it
should be consistent.

Same applies for new files' names


>  If anything, 
> seeing something with a long drm_.. prefix can only be confusing
> since 
> one can assume it is some sort of exported interface.
> 
> I will change this instance.
> 
> > > +{
> > > +	struct drm_mock_sched_job *upto =
> > > +		container_of(hrtimer, typeof(*upto), timer);
> > > +	struct drm_mock_scheduler *sched =
> > > +		drm_sched_to_mock_sched(upto->base.sched);
> > > +	struct drm_mock_sched_job *job, *next;
> > > +	ktime_t now = ktime_get();
> > > +	unsigned long flags;
> > > +	LIST_HEAD(signal);
> > > +
> > > +	spin_lock_irqsave(&sched->lock, flags);
> > > +	list_for_each_entry_safe(job, next, &sched->job_list,
> > > link)
> > > {
> > > +		if (!job->duration_us)
> > > +			break;
> > > +
> > > +		if (ktime_before(now, job->finish_at))
> > > +			break;
> > > +
> > > +		list_move_tail(&job->link, &signal);
> > > +		sched->seqno = job->hw_fence.seqno;
> > > +	}
> > > +	spin_unlock_irqrestore(&sched->lock, flags);
> > > +
> > > +	list_for_each_entry(job, &signal, link) {
> > > +		dma_fence_signal(&job->hw_fence);
> > > +		dma_fence_put(&job->hw_fence);
> > > +	}
> > > +
> > > +	return HRTIMER_NORESTART;
> > > +}
> > > +
> > > +struct drm_mock_sched_job *
> > > +drm_mock_new_sched_job(struct kunit *test,
> > > +		       struct drm_mock_sched_entity *entity)
> > > +{
> > > +	struct drm_mock_sched_job *job;
> > > +	int ret;
> > > +
> > > +	job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL);
> > > +	KUNIT_ASSERT_NOT_NULL(test, job);
> > > +
> > > +	ret = drm_sched_job_init(&job->base,
> > > +				 &entity->base,
> > > +				 1,
> > > +				 NULL);
> > > +	KUNIT_ASSERT_EQ(test, ret, 0);
> > > +
> > > +	job->test = test;
> > > +
> > > +	spin_lock_init(&job->lock);
> > > +	INIT_LIST_HEAD(&job->link);
> > > +	hrtimer_init(&job->timer, CLOCK_MONOTONIC,
> > > HRTIMER_MODE_ABS);
> > > +	job->timer.function = drm_sched_mock_job_signal_timer;
> > > +
> > > +	return job;
> > > +}
> > > +
> > > +static const char *drm_mock_sched_hw_fence_driver_name(struct
> > > dma_fence *fence)
> > > +{
> > > +	return "drm_mock_sched";
> > > +}
> > > +
> > > +static const char *
> > > +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence)
> > > +{
> > > +	struct drm_mock_sched_job *job =
> > > +		container_of(fence, typeof(*job), hw_fence);
> > > +
> > > +	return (const char *)job->base.sched->name;
> > > +}
> > > +static void drm_mock_sched_hw_fence_release(struct dma_fence
> > > *fence)
> > > +{
> > > +	struct drm_mock_sched_job *job =
> > > +		container_of(fence, typeof(*job), hw_fence);
> > > +
> > > +	hrtimer_cancel(&job->timer);
> > > +
> > > +	/* Freed by the kunit framework */
> > > +}
> > > +
> > > +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops =
> > > {
> > > +	.get_driver_name = drm_mock_sched_hw_fence_driver_name,
> > > +	.get_timeline_name =
> > > drm_mock_sched_hw_fence_timeline_name,
> > > +	.release = drm_mock_sched_hw_fence_release,
> > > +};
> > > +
> > > +static struct dma_fence *mock_sched_run_job(struct drm_sched_job
> > > *sched_job)
> > > +{
> > > +	struct drm_mock_scheduler *sched =
> > > +		drm_sched_to_mock_sched(sched_job->sched);
> > > +	struct drm_mock_sched_job *job =
> > > drm_sched_job_to_mock_job(sched_job);
> > > +
> > > +	dma_fence_init(&job->hw_fence,
> > > +		       &drm_mock_sched_hw_fence_ops,
> > > +		       &job->lock,
> > > +		       sched->hw_fence.context,
> > > +		       atomic_inc_return(&sched-
> > > >hw_fence.seqno));
> > > +
> > > +	dma_fence_get(&job->hw_fence); /* Reference for the
> > > job_list
> > > */
> > > +
> > > +	spin_lock_irq(&sched->lock);
> > > +	if (job->duration_us) {
> > > +		ktime_t prev_finish_at = 0;
> > > +
> > > +		if (!list_empty(&sched->job_list)) {
> > > +			struct drm_mock_sched_job *prev =
> > > +				list_last_entry(&sched-
> > > >job_list,
> > > typeof(*prev),
> > > +						link);
> > > +
> > > +			prev_finish_at = prev->finish_at;
> > > +		}
> > > +
> > > +		if (!prev_finish_at)
> > > +			prev_finish_at = ktime_get();
> > > +
> > > +		job->finish_at = ktime_add_us(prev_finish_at,
> > > job-
> > > > duration_us);
> > > +	}
> > > +	list_add_tail(&job->link, &sched->job_list);
> > > +	if (job->finish_at)
> > > +		hrtimer_start(&job->timer, job->finish_at,
> > > HRTIMER_MODE_ABS);
> > > +	spin_unlock_irq(&sched->lock);
> > > +
> > > +	return &job->hw_fence;
> > > +}
> > > +
> > > +static enum drm_gpu_sched_stat
> > > +mock_sched_timedout_job(struct drm_sched_job *sched_job)
> > > +{
> > > +	return DRM_GPU_SCHED_STAT_ENODEV;
> > > +}
> > > +
> > > +static void mock_sched_free_job(struct drm_sched_job *sched_job)
> > > +{
> > > +	drm_sched_job_cleanup(sched_job);
> > > +}
> > > +
> > > +static const struct drm_sched_backend_ops drm_mock_scheduler_ops
> > > = {
> > > +	.run_job = mock_sched_run_job,
> > > +	.timedout_job = mock_sched_timedout_job,
> > > +	.free_job = mock_sched_free_job
> > > +};
> > > +
> > > +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
> > > *test)
> > > +{
> > > +	struct drm_mock_scheduler *sched;
> > > +	int ret;
> > > +
> > > +	sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL);
> > > +	KUNIT_ASSERT_NOT_NULL(test, sched);
> > > +
> > > +	ret = drm_sched_init(&sched->base,
> > > +			     &drm_mock_scheduler_ops,
> > > +			     NULL, /* wq */
> > > +			     DRM_SCHED_PRIORITY_COUNT,
> > > +			     U32_MAX, /* max credits */
> > > +			     UINT_MAX, /* hang limit */
> > > +			     MAX_SCHEDULE_TIMEOUT, /* timeout */
> > > +			     NULL, /* timeout wq */
> > > +			     NULL, /* score */
> > > +			     "drm-mock-scheduler",
> > > +			     NULL /* dev */);
> > > +	KUNIT_ASSERT_EQ(test, ret, 0);
> > > +
> > > +	sched->test = test;
> > > +	sched->hw_fence.context = dma_fence_context_alloc(1);
> > > +	atomic_set(&sched->hw_fence.seqno, 0);
> > > +	INIT_LIST_HEAD(&sched->job_list);
> > > +	spin_lock_init(&sched->lock);
> > > +
> > > +	return sched;
> > > +}
> > > +
> > > +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched)
> > > +{
> > > +	struct drm_mock_sched_job *job, *next;
> > > +	unsigned long flags;
> > > +	LIST_HEAD(signal);
> > > +
> > > +	spin_lock_irqsave(&sched->lock, flags);
> > > +	list_for_each_entry_safe(job, next, &sched->job_list,
> > > link)
> > > +		list_move_tail(&job->link, &signal);
> > > +	spin_unlock_irqrestore(&sched->lock, flags);
> > 
> > So maybe you can help me get up to speed here a bit. What is the
> > purpose behind job->link? Is the general idea documented somewhere?
> 
> List versus link suffix convention I use to distinguish struct
> list_head 
> which is a list versus struct list_head which is used to put
> something 
> on the list.
> 
> In this case job->link is for the mock GPU driver to keep track of
> jobs 
> which have been submitted to the "hardware" for "execution".
> 
> I will of course document these things once the high level design is 
> settled.
> 
> > > +
> > > +	list_for_each_entry(job, &signal, link) {
> > > +		hrtimer_cancel(&job->timer);
> > > +		dma_fence_put(&job->hw_fence);
> > > +	}
> > > +
> > > +	drm_sched_fini(&sched->base);
> > > +}
> > > +
> > > +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
> > > *sched,
> > > +				    unsigned int num)
> > > +{
> > > +	struct drm_mock_sched_job *job, *next;
> > > +	unsigned int found = 0;
> > > +	unsigned long flags;
> > > +	LIST_HEAD(signal);
> > > +
> > > +	spin_lock_irqsave(&sched->lock, flags);
> > > +	if (WARN_ON_ONCE(sched->seqno + num < sched->seqno))
> > > +		goto unlock;
> > > +	sched->seqno += num;
> > > +	list_for_each_entry_safe(job, next, &sched->job_list,
> > > link)
> > > {
> > > +		if (sched->seqno < job->hw_fence.seqno)
> > > +			break;
> > > +
> > > +		list_move_tail(&job->link, &signal);
> > > +		found++;
> > > +	}
> > > +unlock:
> > > +	spin_unlock_irqrestore(&sched->lock, flags);
> > > +
> > > +	list_for_each_entry(job, &signal, link) {
> > > +		dma_fence_signal(&job->hw_fence);
> > > +		dma_fence_put(&job->hw_fence);
> > > +	}
> > > +
> > > +	return found;
> > > +}
> > > diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
> > > b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
> > > new file mode 100644
> > > index 000000000000..421ee2712985
> > > --- /dev/null
> > > +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
> > > @@ -0,0 +1,124 @@
> > > +
> > > +#include <kunit/test.h>
> > > +#include <linux/atomic.h>
> > > +#include <linux/dma-fence.h>
> > > +#include <linux/hrtimer.h>
> > > +#include <linux/ktime.h>
> > > +#include <linux/list.h>
> > > +#include <linux/atomic.h>
> > > +#include <linux/mutex.h>
> > > +#include <linux/types.h>
> > > +
> > > +#include <drm/gpu_scheduler.h>
> > > +
> > > +struct drm_mock_scheduler {
> > > +	struct drm_gpu_scheduler base;
> > > +
> > > +	struct kunit		*test;
> > > +
> > > +	struct {
> > > +		u64		context;
> > > +		atomic_t	seqno;
> > > +	} hw_fence;
> > 
> > Hm, well, so this is confusing. drm_mock_sched_job below contains
> > an
> > actual struct dma_fence that is also called hw_fence, whereas this
> > here
> > seems to be some pseudo-hw_fence?
> > 
> > What is its purpose?
> > 
> > I believe we agreed that "Hardware fence" should always mean a
> > fence
> > per job that is signaled by the hardware (driver interrupt) once
> > the
> > job is done.
> > 
> > If this hw_fence here is the same, why is it per scheduler? That's
> > confusing.
> 
> Mock_job->hw_fence is what the mock GPU driver returns from the 
> sched->run_job().
> 
> Mock_sched->hw_fence is what the mock GPU driver uses to track
> execution 
> of jobs submitted to it for execution. If Irename this to hw_timeline
> will it make it clearer?

How about "current_hw_fence"?


> 
> > > +
> > > +	spinlock_t		lock;
> > 
> > Maybe document the lock's purpose
> > 
> > 
> > > +	unsigned int		seqno;
> > > +	struct list_head	job_list;
> > > +};
> > > +
> > > +struct drm_mock_sched_entity {
> > > +	struct drm_sched_entity base;
> > > +
> > > +	struct kunit		*test;
> > > +};
> > > +
> > > +struct drm_mock_sched_job {
> > > +	struct drm_sched_job	base;
> > > +
> > > +	struct list_head	link;
> > > +	struct hrtimer		timer;
> > > +
> > > +	unsigned int		duration_us;
> > > +	ktime_t			finish_at;
> > > +
> > > +	spinlock_t		lock;
> > 
> > Same
> > 
> > > +	struct dma_fence	hw_fence;
> > > +
> > > +	struct kunit		*test;
> > > +};
> > > +
> > > +static inline struct drm_mock_scheduler *
> > > +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched)
> > > +{
> > > +	return container_of(sched, struct drm_mock_scheduler,
> > > base);
> > > +};
> > > +
> > > +static inline struct drm_mock_sched_entity *
> > > +drm_sched_entity_to_mock_entity(struct drm_sched_entity
> > > *sched_entity)
> > > +{
> > > +	return container_of(sched_entity, struct
> > > drm_mock_sched_entity, base);
> > > +};
> > > +
> > > +static inline struct drm_mock_sched_job *
> > > +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job)
> > > +{
> > > +	return container_of(sched_job, struct
> > > drm_mock_sched_job,
> > > base);
> > > +};
> > > +
> > > +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
> > > *test);
> > > +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched);
> > > +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
> > > *sched,
> > > +				    unsigned int num);
> > > +
> > > +struct drm_mock_sched_entity *
> > > +drm_mock_new_sched_entity(struct kunit *test,
> > > +			  enum drm_sched_priority priority,
> > > +			  struct drm_mock_scheduler *sched);
> > > +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
> > > *entity);
> > > +
> > > +struct drm_mock_sched_job *
> > > +drm_mock_new_sched_job(struct kunit *test,
> > > +		       struct drm_mock_sched_entity *entity);
> > > +
> > > +static inline void drm_mock_sched_job_submit(struct
> > > drm_mock_sched_job *job)
> > > +{
> > > +	drm_sched_job_arm(&job->base);
> > > +	drm_sched_entity_push_job(&job->base);
> > > +}
> > > +
> > > +static inline void
> > > +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job
> > > *job,
> > > +				   unsigned int duration_us)
> > > +{
> > > +	job->duration_us = duration_us;
> > > +}
> > > +
> > > +static inline bool
> > > +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job)
> > > +{
> > > +	return dma_fence_is_signaled(&job->base.s_fence-
> > > >finished);
> > > +}
> > > +
> > > +static inline bool
> > > +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job,
> > > long timeout)
> > > +{
> > > +	long ret;
> > > +
> > > +	ret = dma_fence_wait_timeout(&job->base.s_fence-
> > > >finished,
> > > +				      false,
> > > +				      timeout);
> > > +
> > > +	return ret != 0;
> > > +}
> > > +
> > > +static inline long
> > > +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job
> > > *job,
> > > long timeout)
> > > +{
> > > +	long ret;
> > > +
> > > +	ret = dma_fence_wait_timeout(&job->base.s_fence-
> > > >scheduled,
> > > +				      false,
> > > +				      timeout);
> > > +
> > > +	return ret != 0;
> > > +}
> > > diff --git
> > > a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> > > b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> > > new file mode 100644
> > > index 000000000000..6fd39bea95b1
> > > --- /dev/null
> > > +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
> > > @@ -0,0 +1,247 @@
> > > +
> > > +#include "drm_sched_tests.h"
> > > +
> > > +static int drm_sched_basic_init(struct kunit *test)
> > > +{
> > > +	test->priv = drm_mock_new_scheduler(test);
> > > +
> > > +	return 0;
> > > +}
> > > +
> > > +static void drm_sched_basic_exit(struct kunit *test)
> > > +{
> > > +	struct drm_mock_scheduler *sched = test->priv;
> > > +
> > > +	drm_mock_scheduler_fini(sched);
> > > +}
> > > +
> > > +static void drm_sched_basic_submit(struct kunit *test)
> > > +{
> > > +	struct drm_mock_scheduler *sched = test->priv;
> > > +	struct drm_mock_sched_entity *entity;
> > > +	struct drm_mock_sched_job *job;
> > > +	unsigned int i;
> > > +	bool done;
> > > +
> > > +	/*
> > > +	 * Submit one job to the scheduler and verify that it
> > > gets
> > > scheduled
> > > +	 * and completed only when the mock hw backend processes
> > > it.
> > > +	 */
> > > +
> > > +	entity = drm_mock_new_sched_entity(test,
> > > +					
> > > DRM_SCHED_PRIORITY_NORMAL,
> > > +					   sched);
> > > +	job = drm_mock_new_sched_job(test, entity);
> > > +
> > > +	drm_mock_sched_job_submit(job);
> > > +
> > > +	done = drm_mock_sched_job_wait_scheduled(job, HZ);
> > > +	KUNIT_ASSERT_EQ(test, done, true);
> > > +
> > > +	done = drm_mock_sched_job_wait_finished(job, HZ / 2);
> > > +	KUNIT_ASSERT_EQ(test, done, false);
> > > +
> > > +	i = drm_mock_sched_advance(sched, 1);
> > > +	KUNIT_ASSERT_EQ(test, i, 1);
> > > +
> > > +	done = drm_mock_sched_job_wait_finished(job, HZ);
> > > +	KUNIT_ASSERT_EQ(test, done, true);
> > > +
> > > +	drm_mock_sched_entity_free(entity);
> > > +}
> > > +
> > > +static void drm_sched_basic_queue(struct kunit *test)
> > > +{
> > > +	struct drm_mock_scheduler *sched = test->priv;
> > > +	struct drm_mock_sched_entity *entity;
> > > +	struct drm_mock_sched_job *job;
> > > +	const unsigned int qd = 100;
> > 
> > Not the best variable name – is this "queue depth"? Why is it 100?
> > ->
> > global define & document
> 
> I wouldn't promote this to global and TBH if you look how small this 
> test function is it feels pretty self documenting.

I meant global in the sense of at the top of the file as a constant.

Wouldn't you agree that it would be good to have test parameters at a
place where you can easily modify them, in case you want to compile the
tests a bit differently?


P.


> 
> > 
> > 
> > > +	unsigned int i;
> > > +	bool done;
> > > +
> > > +	/*
> > > +	 * Submit a queue of jobs on the same entity, let them
> > > be
> > > completed by
> > > +	 * the mock hw backend and check the status of the last
> > > job.
> > > +	 */
> > > +
> > > +	entity = drm_mock_new_sched_entity(test,
> > > +					
> > > DRM_SCHED_PRIORITY_NORMAL,
> > > +					   sched);
> > > +
> > > +	for (i = 0; i < qd; i++) {
> > > +		job = drm_mock_new_sched_job(test, entity);
> > > +		drm_mock_sched_job_set_duration_us(job, 1000);
> > > +		drm_mock_sched_job_submit(job);
> > > +	}
> > > +
> > > +	done = drm_mock_sched_job_wait_finished(job, HZ);
> > > +	KUNIT_ASSERT_EQ(test, done, true);
> > > +
> > > +	drm_mock_sched_entity_free(entity);
> > > +}
> > > +
> > > +static void drm_sched_basic_chain(struct kunit *test)
> > > +{
> > > +	struct drm_mock_scheduler *sched = test->priv;
> > > +	struct drm_mock_sched_job *job, *prev = NULL;
> > > +	struct drm_mock_sched_entity *entity;
> > > +	const unsigned int qd = 100;
> > > +	unsigned int i;
> > > +	bool done;
> > > +
> > > +	/*
> > > +	 * Submit a queue of jobs on the same entity but with an
> > > explicit
> > > +	 * chain of dependencies between them. Let the jobs be
> > > completed by
> > > +	 * the mock hw backend and check the status of the last
> > > job.
> > > +	 */
> > > +
> > > +	entity = drm_mock_new_sched_entity(test,
> > > +					
> > > DRM_SCHED_PRIORITY_NORMAL,
> > > +					   sched);
> > > +
> > > +	for (i = 0; i < qd; i++) {
> > > +		job = drm_mock_new_sched_job(test, entity);
> > > +		drm_mock_sched_job_set_duration_us(job, 1000);
> > > +		if (prev)
> > > +			drm_sched_job_add_dependency(&job->base,
> > > +						
> > > dma_fence_get(&prev->base.s_fence->finished));
> > > +		drm_mock_sched_job_submit(job);
> > > +		prev = job;
> > > +	}
> > > +
> > > +	done = drm_mock_sched_job_wait_finished(job, HZ);
> > > +	KUNIT_ASSERT_EQ(test, done, true);
> > > +
> > > +	drm_mock_sched_entity_free(entity);
> > > +}
> > > +
> > > +static void drm_sched_basic_entities(struct kunit *test)
> > > +{
> > > +	struct drm_mock_scheduler *sched = test->priv;
> > > +	struct drm_mock_sched_entity *entity[4];
> > > +	struct drm_mock_sched_job *job;
> > > +	const unsigned int qd = 100;
> > > +	unsigned int i, cur_ent = 0;
> > > +	bool done;
> > > +
> > > +	/*
> > > +	 * Submit a queue of jobs across different entities, let
> > > them be
> > > +	 * completed by the mock hw backend and check the status
> > > of
> > > the last
> > > +	 * job.
> > > +	 */
> > > +
> > > +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> > > +		entity[i] = drm_mock_new_sched_entity(test,
> > > +						
> > > DRM_SCHED_PRIORITY_NORMAL,
> > > +						      sched);
> > > +
> > > +	for (i = 0; i < qd; i++) {
> > > +		job = drm_mock_new_sched_job(test,
> > > entity[cur_ent++]);
> > > +		cur_ent %= ARRAY_SIZE(entity);
> > > +		drm_mock_sched_job_set_duration_us(job, 1000);
> > > +		drm_mock_sched_job_submit(job);
> > > +	}
> > > +
> > > +	done = drm_mock_sched_job_wait_finished(job, HZ);
> > > +	KUNIT_ASSERT_EQ(test, done, true);
> > > +
> > > +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> > > +		drm_mock_sched_entity_free(entity[i]);
> > > +}
> > > +
> > > +static void drm_sched_basic_entities_chain(struct kunit *test)
> > > +{
> > > +	struct drm_mock_scheduler *sched = test->priv;
> > > +	struct drm_mock_sched_job *job, *prev = NULL;
> > > +	struct drm_mock_sched_entity *entity[4];
> > > +	const unsigned int qd = 100;
> > > +	unsigned int i, cur_ent = 0;
> > > +	bool done;
> > > +
> > > +	/*
> > > +	 * Submit a queue of jobs across different entities with
> > > an
> > > explicit
> > > +	 * chain of dependencies between them. Let the jobs be
> > > completed by
> > > +	 * the mock hw backend and check the status of the last
> > > job.
> > > +	 */
> > > +
> > > +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> > > +		entity[i] = drm_mock_new_sched_entity(test,
> > > +						
> > > DRM_SCHED_PRIORITY_NORMAL,
> > > +						      sched);
> > > +
> > > +	for (i = 0; i < qd; i++) {
> > > +		job = drm_mock_new_sched_job(test,
> > > entity[cur_ent++]);
> > > +		cur_ent %= ARRAY_SIZE(entity);
> > > +		drm_mock_sched_job_set_duration_us(job, 1000);
> > > +		if (prev)
> > > +			drm_sched_job_add_dependency(&job->base,
> > > +						
> > > dma_fence_get(&prev->base.s_fence->finished));
> > > +		drm_mock_sched_job_submit(job);
> > > +		prev = job;
> > > +	}
> > > +
> > > +	done = drm_mock_sched_job_wait_finished(job, HZ);
> > > +	KUNIT_ASSERT_EQ(test, done, true);
> > > +
> > > +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> > > +		drm_mock_sched_entity_free(entity[i]);
> > > +}
> > > +
> > > +static void drm_sched_basic_entity_cleanup(struct kunit *test)
> > > +{
> > > +	struct drm_mock_sched_job *job, *mid, *prev = NULL;
> > > +	struct drm_mock_scheduler *sched = test->priv;
> > > +	struct drm_mock_sched_entity *entity[4];
> > 
> > 4 is used in several places, so could be defined globally. In case
> > there's a special reason for why it's 4, that could be mentioned,
> > too
> 
> It's completely arbitrary. In this test we just need a bunch of
> entities 
> to be active on the scheduler as we trigger a client exit in the
> middle 
> of it. IMO the comment few lines below should be good enough to
> explain 
> the idea. I fear that promoting this to a global define would just
> give 
> it more weight that what it has. Then some test would want a
> different 
> number etc.
> 
> Regards,
> 
> Tvrtko
> 
> > > +	const unsigned int qd = 100;
> > > +	unsigned int i, cur_ent = 0;
> > > +	bool done;
> > > +
> > > +	/*
> > > +	 * Submit a queue of jobs across different entities with
> > > an
> > > explicit
> > > +	 * chain of dependencies between them and trigger entity
> > > cleanup while
> > > +	 * the queue is still being processed.
> > > +	 */
> > > +
> > > +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> > > +		entity[i] = drm_mock_new_sched_entity(test,
> > > +						
> > > DRM_SCHED_PRIORITY_NORMAL,
> > > +						      sched);
> > > +
> > > +	for (i = 0; i < qd; i++) {
> > > +		job = drm_mock_new_sched_job(test,
> > > entity[cur_ent++]);
> > > +		cur_ent %= ARRAY_SIZE(entity);
> > > +		drm_mock_sched_job_set_duration_us(job, 1000);
> > > +		if (prev)
> > > +			drm_sched_job_add_dependency(&job->base,
> > > +						
> > > dma_fence_get(&prev->base.s_fence->finished));
> > > +		drm_mock_sched_job_submit(job);
> > > +		if (i == qd / 2)
> > > +			mid = job;
> > > +		prev = job;
> > > +	}
> > > +
> > > +	done = drm_mock_sched_job_wait_finished(mid, HZ);
> > > +	KUNIT_ASSERT_EQ(test, done, true);
> > > +
> > > +	/* Exit with half of the queue still pending to be
> > > executed.
> > > */
> > > +	for (i = 0; i < ARRAY_SIZE(entity); i++)
> > > +		drm_mock_sched_entity_free(entity[i]);}
> > > +
> > > +static struct kunit_case drm_sched_basic_tests[] = {
> > > +	KUNIT_CASE(drm_sched_basic_submit),
> > > +	KUNIT_CASE(drm_sched_basic_queue),
> > > +	KUNIT_CASE(drm_sched_basic_chain),
> > > +	KUNIT_CASE(drm_sched_basic_entities),
> > > +	KUNIT_CASE(drm_sched_basic_entities_chain),
> > > +	KUNIT_CASE(drm_sched_basic_entity_cleanup),
> > > +	{}
> > > +};
> > > +
> > > +static struct kunit_suite drm_sched_basic = {
> > > +	.name = "drm_sched_basic_tests",
> > > +	.init = drm_sched_basic_init,
> > > +	.exit = drm_sched_basic_exit,
> > > +	.test_cases = drm_sched_basic_tests,
> > > +};
> > > +
> > > +kunit_test_suite(drm_sched_basic);
> >
Tvrtko Ursulin Feb. 6, 2025, 1:38 p.m. UTC | #4
On 06/02/2025 12:33, Philipp Stanner wrote:
> On Thu, 2025-02-06 at 10:42 +0000, Tvrtko Ursulin wrote:
>>
>> On 06/02/2025 09:51, Philipp Stanner wrote:
>>> On Mon, 2025-02-03 at 15:30 +0000, Tvrtko Ursulin wrote:
>>>> Implement a mock scheduler backend and add some basic test to
>>>> exercise the
>>>> core scheduler code paths.
>>>>
>>>> Mock backend (kind of like a very simple mock GPU) can either
>>>> process
>>>> jobs
>>>> by tests manually advancing the "timeline" job at a time, or
>>>> alternatively
>>>> jobs can be configured with a time duration in which case they
>>>> get
>>>> completed asynchronously from the unit test code.
>>>>
>>>> Core scheduler classes are subclassed to support this mock
>>>> implementation.
>>>>
>>>> The tests added are just a few simple submission patterns.
>>>>
>>>> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@igalia.com>
>>>
>>> Could add a Suggested-by: Philipp :)
>>
>> Will do.
>>
>>>> Cc: Christian König <christian.koenig@amd.com>
>>>> Cc: Danilo Krummrich <dakr@kernel.org>
>>>> Cc: Matthew Brost <matthew.brost@intel.com>
>>>> Cc: Philipp Stanner <phasta@kernel.org>
>>>> ---
>>>>    drivers/gpu/drm/Kconfig.debug                 |  12 +
>>>>    drivers/gpu/drm/scheduler/.kunitconfig        |  12 +
>>>>    drivers/gpu/drm/scheduler/Makefile            |   1 +
>>>>    drivers/gpu/drm/scheduler/tests/Makefile      |   4 +
>>>>    .../gpu/drm/scheduler/tests/drm_mock_entity.c |  29 ++
>>>>    .../gpu/drm/scheduler/tests/drm_mock_job.c    |   3 +
>>>>    .../drm/scheduler/tests/drm_mock_scheduler.c  | 254
>>>> ++++++++++++++++++
>>>>    .../gpu/drm/scheduler/tests/drm_sched_tests.h | 124 +++++++++
>>>>    .../scheduler/tests/drm_sched_tests_basic.c   | 247
>>>> +++++++++++++++++
>>>>    9 files changed, 686 insertions(+)
>>>>    create mode 100644 drivers/gpu/drm/scheduler/.kunitconfig
>>>>    create mode 100644 drivers/gpu/drm/scheduler/tests/Makefile
>>>>    create mode 100644
>>>> drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>>>>    create mode 100644
>>>> drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>>>>    create mode 100644
>>>> drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>>>>    create mode 100644
>>>> drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>>>>    create mode 100644
>>>> drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>>>>
>>>> diff --git a/drivers/gpu/drm/Kconfig.debug
>>>> b/drivers/gpu/drm/Kconfig.debug
>>>> index a35d74171b7b..89f777f21e95 100644
>>>> --- a/drivers/gpu/drm/Kconfig.debug
>>>> +++ b/drivers/gpu/drm/Kconfig.debug
>>>> @@ -88,5 +88,17 @@ config DRM_TTM_KUNIT_TEST
>>>>    
>>>>    	  If in doubt, say "N".
>>>>    
>>>> +config DRM_SCHED_KUNIT_TEST
>>>> +	tristate "KUnit tests for the DRM scheduler" if
>>>> !KUNIT_ALL_TESTS
>>>> +	select DRM_SCHED
>>>> +	depends on DRM && KUNIT
>>>> +	default KUNIT_ALL_TESTS
>>>> +	help
>>>> +	  Choose this option to build unit tests for the DRM
>>>> scheduler.
>>>
>>> nit: Might say "DRM GPU scheduler" so readers not familiar with all
>>> that get a better idea of what it's about
>>>
>>>> +
>>>> +	  Recommended for driver developers only.
>>>
>>> nit: s/driver developers/DRM developers
>>> ?
>>>
>>>> +
>>>> +	  If in doubt, say "N".
>>>> +
>>>>    config DRM_EXPORT_FOR_TESTS
>>>>    	bool
>>>> diff --git a/drivers/gpu/drm/scheduler/.kunitconfig
>>>> b/drivers/gpu/drm/scheduler/.kunitconfig
>>>> new file mode 100644
>>>> index 000000000000..cece53609fcf
>>>> --- /dev/null
>>>> +++ b/drivers/gpu/drm/scheduler/.kunitconfig
>>>> @@ -0,0 +1,12 @@
>>>> +CONFIG_KUNIT=y
>>>> +CONFIG_DRM=y
>>>> +CONFIG_DRM_SCHED_KUNIT_TEST=y
>>>> +CONFIG_EXPERT=y
>>>> +CONFIG_DEBUG_SPINLOCK=y
>>>> +CONFIG_DEBUG_MUTEXES=y
>>>> +CONFIG_DEBUG_ATOMIC_SLEEP=y
>>>> +CONFIG_LOCK_DEBUGGING_SUPPORT=y
>>>> +CONFIG_PROVE_LOCKING=y
>>>> +CONFIG_LOCKDEP=y
>>>> +CONFIG_DEBUG_LOCKDEP=y
>>>> +CONFIG_DEBUG_LIST=y
>>>> diff --git a/drivers/gpu/drm/scheduler/Makefile
>>>> b/drivers/gpu/drm/scheduler/Makefile
>>>> index 53863621829f..46dfdca0758a 100644
>>>> --- a/drivers/gpu/drm/scheduler/Makefile
>>>> +++ b/drivers/gpu/drm/scheduler/Makefile
>>>> @@ -23,3 +23,4 @@
>>>>    gpu-sched-y := sched_main.o sched_fence.o sched_entity.o
>>>>    
>>>>    obj-$(CONFIG_DRM_SCHED) += gpu-sched.o
>>>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/
>>>> diff --git a/drivers/gpu/drm/scheduler/tests/Makefile
>>>> b/drivers/gpu/drm/scheduler/tests/Makefile
>>>> new file mode 100644
>>>> index 000000000000..d69eab6a2e9b
>>>> --- /dev/null
>>>> +++ b/drivers/gpu/drm/scheduler/tests/Makefile
>>>> @@ -0,0 +1,4 @@
>>>> +
>>>> +obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \
>>>> +        drm_mock_scheduler.o \
>>>> +        drm_sched_tests_basic.o
>>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>>>> b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>>>> new file mode 100644
>>>> index 000000000000..c9205f9ff524
>>>> --- /dev/null
>>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
>>>> @@ -0,0 +1,29 @@
>>>> +
>>>> +#include "drm_sched_tests.h"
>>>> +
>>>> +struct drm_mock_sched_entity *
>>>> +drm_mock_sched_entity_new(struct kunit *test,
>>>> +			  enum drm_sched_priority priority,
>>>> +			  struct drm_mock_scheduler *sched)
>>>
>>>
>>> I personally do like this function head style and
>>>
>>>
>>>> +{
>>>> +	struct drm_sched_mock_entity *entity;
>>>> +	int ret;
>>>> +
>>>> +	entity = kunit_kmalloc(test, sizeof(*entity),
>>>> GFP_KERNEL);
>>>> +	KUNIT_ASSERT_NOT_NULL(test, entity);
>>>> +
>>>> +	ret = drm_sched_entity_init(&entity->base,
>>>> +				    priority,
>>>> +				    &sched->base, 1,
>>>> +				    NULL);
>>>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>>>> +
>>>> +	entity->test = test;
>>>> +
>>>> +	return entity;
>>>> +}
>>>> +
>>>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>>>> *entity)
>>>
>>> do believe it should then be used consistently everywhere,
>>> regardless
>>> of length.
>>>
>>>
>>>> +{
>>>> +	drm_sched_entity_fini(&entity->base);
>>>> +}
>>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>>>> b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>>>> new file mode 100644
>>>> index 000000000000..d94568cf3da9
>>>> --- /dev/null
>>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
>>>> @@ -0,0 +1,3 @@
>>>> +
>>>> +#include "drm_sched_tests.h"
>>>> +
>>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>>>> b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>>>> new file mode 100644
>>>> index 000000000000..f1985900a6ba
>>>> --- /dev/null
>>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
>>>> @@ -0,0 +1,254 @@
>>>> +
>>>> +#include "drm_sched_tests.h"
>>>> +
>>>> +struct drm_mock_sched_entity *
>>>> +drm_mock_new_sched_entity(struct kunit *test,
>>>> +			  enum drm_sched_priority priority,
>>>> +			  struct drm_mock_scheduler *sched)
>>>> +{
>>>> +	struct drm_mock_sched_entity *entity;
>>>> +	struct drm_gpu_scheduler *drm_sched;
>>>> +	int ret;
>>>> +
>>>> +	entity = kunit_kzalloc(test, sizeof(*entity),
>>>> GFP_KERNEL);
>>>> +	KUNIT_ASSERT_NOT_NULL(test, entity);
>>>> +
>>>> +	drm_sched = &sched->base;
>>>> +	ret = drm_sched_entity_init(&entity->base,
>>>> +				    priority,
>>>> +				    &drm_sched, 1,
>>>> +				    NULL);
>>>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>>>> +
>>>> +	entity->test = test;
>>>> +
>>>> +	return entity;
>>>> +}
>>>> +
>>>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>>>> *entity)
>>>> +{
>>>> +	drm_sched_entity_destroy(&entity->base);
>>>> +}
>>>> +
>>>> +static enum hrtimer_restart
>>>> +drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer)
>>>
>>> I think we should get up with a consistent naming convention. Some
>>> things are called drm_mock_sched_, some others drm_sched_mock_.
>>> As far as I saw, drm_mock_* does not yet exist.
>>> So do you want to introduce drm_mock as something generic for drm?
>>> or
>>> just for drm/sched?
>>
>> This one is literally the only when where I tranposed the two. But it
>> is
>> also local and I am not too bothered about naming conventions of a
>> local functions. If it were to me I would favour brevity and emit the
>> long drm_.. prefixes which IMO do not help readability.
> 
> I tentatively agree with that. I just think whatever the style is, it
> should be consistent.
> 
> Same applies for new files' names
> 
> 
>>   If anything,
>> seeing something with a long drm_.. prefix can only be confusing
>> since
>> one can assume it is some sort of exported interface.
>>
>> I will change this instance.
>>
>>>> +{
>>>> +	struct drm_mock_sched_job *upto =
>>>> +		container_of(hrtimer, typeof(*upto), timer);
>>>> +	struct drm_mock_scheduler *sched =
>>>> +		drm_sched_to_mock_sched(upto->base.sched);
>>>> +	struct drm_mock_sched_job *job, *next;
>>>> +	ktime_t now = ktime_get();
>>>> +	unsigned long flags;
>>>> +	LIST_HEAD(signal);
>>>> +
>>>> +	spin_lock_irqsave(&sched->lock, flags);
>>>> +	list_for_each_entry_safe(job, next, &sched->job_list,
>>>> link)
>>>> {
>>>> +		if (!job->duration_us)
>>>> +			break;
>>>> +
>>>> +		if (ktime_before(now, job->finish_at))
>>>> +			break;
>>>> +
>>>> +		list_move_tail(&job->link, &signal);
>>>> +		sched->seqno = job->hw_fence.seqno;
>>>> +	}
>>>> +	spin_unlock_irqrestore(&sched->lock, flags);
>>>> +
>>>> +	list_for_each_entry(job, &signal, link) {
>>>> +		dma_fence_signal(&job->hw_fence);
>>>> +		dma_fence_put(&job->hw_fence);
>>>> +	}
>>>> +
>>>> +	return HRTIMER_NORESTART;
>>>> +}
>>>> +
>>>> +struct drm_mock_sched_job *
>>>> +drm_mock_new_sched_job(struct kunit *test,
>>>> +		       struct drm_mock_sched_entity *entity)
>>>> +{
>>>> +	struct drm_mock_sched_job *job;
>>>> +	int ret;
>>>> +
>>>> +	job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL);
>>>> +	KUNIT_ASSERT_NOT_NULL(test, job);
>>>> +
>>>> +	ret = drm_sched_job_init(&job->base,
>>>> +				 &entity->base,
>>>> +				 1,
>>>> +				 NULL);
>>>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>>>> +
>>>> +	job->test = test;
>>>> +
>>>> +	spin_lock_init(&job->lock);
>>>> +	INIT_LIST_HEAD(&job->link);
>>>> +	hrtimer_init(&job->timer, CLOCK_MONOTONIC,
>>>> HRTIMER_MODE_ABS);
>>>> +	job->timer.function = drm_sched_mock_job_signal_timer;
>>>> +
>>>> +	return job;
>>>> +}
>>>> +
>>>> +static const char *drm_mock_sched_hw_fence_driver_name(struct
>>>> dma_fence *fence)
>>>> +{
>>>> +	return "drm_mock_sched";
>>>> +}
>>>> +
>>>> +static const char *
>>>> +drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence)
>>>> +{
>>>> +	struct drm_mock_sched_job *job =
>>>> +		container_of(fence, typeof(*job), hw_fence);
>>>> +
>>>> +	return (const char *)job->base.sched->name;
>>>> +}
>>>> +static void drm_mock_sched_hw_fence_release(struct dma_fence
>>>> *fence)
>>>> +{
>>>> +	struct drm_mock_sched_job *job =
>>>> +		container_of(fence, typeof(*job), hw_fence);
>>>> +
>>>> +	hrtimer_cancel(&job->timer);
>>>> +
>>>> +	/* Freed by the kunit framework */
>>>> +}
>>>> +
>>>> +static const struct dma_fence_ops drm_mock_sched_hw_fence_ops =
>>>> {
>>>> +	.get_driver_name = drm_mock_sched_hw_fence_driver_name,
>>>> +	.get_timeline_name =
>>>> drm_mock_sched_hw_fence_timeline_name,
>>>> +	.release = drm_mock_sched_hw_fence_release,
>>>> +};
>>>> +
>>>> +static struct dma_fence *mock_sched_run_job(struct drm_sched_job
>>>> *sched_job)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched =
>>>> +		drm_sched_to_mock_sched(sched_job->sched);
>>>> +	struct drm_mock_sched_job *job =
>>>> drm_sched_job_to_mock_job(sched_job);
>>>> +
>>>> +	dma_fence_init(&job->hw_fence,
>>>> +		       &drm_mock_sched_hw_fence_ops,
>>>> +		       &job->lock,
>>>> +		       sched->hw_fence.context,
>>>> +		       atomic_inc_return(&sched-
>>>>> hw_fence.seqno));
>>>> +
>>>> +	dma_fence_get(&job->hw_fence); /* Reference for the
>>>> job_list
>>>> */
>>>> +
>>>> +	spin_lock_irq(&sched->lock);
>>>> +	if (job->duration_us) {
>>>> +		ktime_t prev_finish_at = 0;
>>>> +
>>>> +		if (!list_empty(&sched->job_list)) {
>>>> +			struct drm_mock_sched_job *prev =
>>>> +				list_last_entry(&sched-
>>>>> job_list,
>>>> typeof(*prev),
>>>> +						link);
>>>> +
>>>> +			prev_finish_at = prev->finish_at;
>>>> +		}
>>>> +
>>>> +		if (!prev_finish_at)
>>>> +			prev_finish_at = ktime_get();
>>>> +
>>>> +		job->finish_at = ktime_add_us(prev_finish_at,
>>>> job-
>>>>> duration_us);
>>>> +	}
>>>> +	list_add_tail(&job->link, &sched->job_list);
>>>> +	if (job->finish_at)
>>>> +		hrtimer_start(&job->timer, job->finish_at,
>>>> HRTIMER_MODE_ABS);
>>>> +	spin_unlock_irq(&sched->lock);
>>>> +
>>>> +	return &job->hw_fence;
>>>> +}
>>>> +
>>>> +static enum drm_gpu_sched_stat
>>>> +mock_sched_timedout_job(struct drm_sched_job *sched_job)
>>>> +{
>>>> +	return DRM_GPU_SCHED_STAT_ENODEV;
>>>> +}
>>>> +
>>>> +static void mock_sched_free_job(struct drm_sched_job *sched_job)
>>>> +{
>>>> +	drm_sched_job_cleanup(sched_job);
>>>> +}
>>>> +
>>>> +static const struct drm_sched_backend_ops drm_mock_scheduler_ops
>>>> = {
>>>> +	.run_job = mock_sched_run_job,
>>>> +	.timedout_job = mock_sched_timedout_job,
>>>> +	.free_job = mock_sched_free_job
>>>> +};
>>>> +
>>>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
>>>> *test)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched;
>>>> +	int ret;
>>>> +
>>>> +	sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL);
>>>> +	KUNIT_ASSERT_NOT_NULL(test, sched);
>>>> +
>>>> +	ret = drm_sched_init(&sched->base,
>>>> +			     &drm_mock_scheduler_ops,
>>>> +			     NULL, /* wq */
>>>> +			     DRM_SCHED_PRIORITY_COUNT,
>>>> +			     U32_MAX, /* max credits */
>>>> +			     UINT_MAX, /* hang limit */
>>>> +			     MAX_SCHEDULE_TIMEOUT, /* timeout */
>>>> +			     NULL, /* timeout wq */
>>>> +			     NULL, /* score */
>>>> +			     "drm-mock-scheduler",
>>>> +			     NULL /* dev */);
>>>> +	KUNIT_ASSERT_EQ(test, ret, 0);
>>>> +
>>>> +	sched->test = test;
>>>> +	sched->hw_fence.context = dma_fence_context_alloc(1);
>>>> +	atomic_set(&sched->hw_fence.seqno, 0);
>>>> +	INIT_LIST_HEAD(&sched->job_list);
>>>> +	spin_lock_init(&sched->lock);
>>>> +
>>>> +	return sched;
>>>> +}
>>>> +
>>>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched)
>>>> +{
>>>> +	struct drm_mock_sched_job *job, *next;
>>>> +	unsigned long flags;
>>>> +	LIST_HEAD(signal);
>>>> +
>>>> +	spin_lock_irqsave(&sched->lock, flags);
>>>> +	list_for_each_entry_safe(job, next, &sched->job_list,
>>>> link)
>>>> +		list_move_tail(&job->link, &signal);
>>>> +	spin_unlock_irqrestore(&sched->lock, flags);
>>>
>>> So maybe you can help me get up to speed here a bit. What is the
>>> purpose behind job->link? Is the general idea documented somewhere?
>>
>> List versus link suffix convention I use to distinguish struct
>> list_head
>> which is a list versus struct list_head which is used to put
>> something
>> on the list.
>>
>> In this case job->link is for the mock GPU driver to keep track of
>> jobs
>> which have been submitted to the "hardware" for "execution".
>>
>> I will of course document these things once the high level design is
>> settled.
>>
>>>> +
>>>> +	list_for_each_entry(job, &signal, link) {
>>>> +		hrtimer_cancel(&job->timer);
>>>> +		dma_fence_put(&job->hw_fence);
>>>> +	}
>>>> +
>>>> +	drm_sched_fini(&sched->base);
>>>> +}
>>>> +
>>>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
>>>> *sched,
>>>> +				    unsigned int num)
>>>> +{
>>>> +	struct drm_mock_sched_job *job, *next;
>>>> +	unsigned int found = 0;
>>>> +	unsigned long flags;
>>>> +	LIST_HEAD(signal);
>>>> +
>>>> +	spin_lock_irqsave(&sched->lock, flags);
>>>> +	if (WARN_ON_ONCE(sched->seqno + num < sched->seqno))
>>>> +		goto unlock;
>>>> +	sched->seqno += num;
>>>> +	list_for_each_entry_safe(job, next, &sched->job_list,
>>>> link)
>>>> {
>>>> +		if (sched->seqno < job->hw_fence.seqno)
>>>> +			break;
>>>> +
>>>> +		list_move_tail(&job->link, &signal);
>>>> +		found++;
>>>> +	}
>>>> +unlock:
>>>> +	spin_unlock_irqrestore(&sched->lock, flags);
>>>> +
>>>> +	list_for_each_entry(job, &signal, link) {
>>>> +		dma_fence_signal(&job->hw_fence);
>>>> +		dma_fence_put(&job->hw_fence);
>>>> +	}
>>>> +
>>>> +	return found;
>>>> +}
>>>> diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>>>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>>>> new file mode 100644
>>>> index 000000000000..421ee2712985
>>>> --- /dev/null
>>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
>>>> @@ -0,0 +1,124 @@
>>>> +
>>>> +#include <kunit/test.h>
>>>> +#include <linux/atomic.h>
>>>> +#include <linux/dma-fence.h>
>>>> +#include <linux/hrtimer.h>
>>>> +#include <linux/ktime.h>
>>>> +#include <linux/list.h>
>>>> +#include <linux/atomic.h>
>>>> +#include <linux/mutex.h>
>>>> +#include <linux/types.h>
>>>> +
>>>> +#include <drm/gpu_scheduler.h>
>>>> +
>>>> +struct drm_mock_scheduler {
>>>> +	struct drm_gpu_scheduler base;
>>>> +
>>>> +	struct kunit		*test;
>>>> +
>>>> +	struct {
>>>> +		u64		context;
>>>> +		atomic_t	seqno;
>>>> +	} hw_fence;
>>>
>>> Hm, well, so this is confusing. drm_mock_sched_job below contains
>>> an
>>> actual struct dma_fence that is also called hw_fence, whereas this
>>> here
>>> seems to be some pseudo-hw_fence?
>>>
>>> What is its purpose?
>>>
>>> I believe we agreed that "Hardware fence" should always mean a
>>> fence
>>> per job that is signaled by the hardware (driver interrupt) once
>>> the
>>> job is done.
>>>
>>> If this hw_fence here is the same, why is it per scheduler? That's
>>> confusing.
>>
>> Mock_job->hw_fence is what the mock GPU driver returns from the
>> sched->run_job().
>>
>> Mock_sched->hw_fence is what the mock GPU driver uses to track
>> execution
>> of jobs submitted to it for execution. If Irename this to hw_timeline
>> will it make it clearer?
> 
> How about "current_hw_fence"?

I prefer timeline. The current working version is:

struct drm_mock_scheduler {
	struct drm_gpu_scheduler base;

	struct kunit		*test;

	spinlock_t		lock;
	struct list_head	job_list; /* Protected by the lock */

	struct {
		u64		context;
		atomic_t	next_seqno;
		unsigned int	cur_seqno; /* Protected by the lock */
	} hw_timeline;
};

>>>> +
>>>> +	spinlock_t		lock;
>>>
>>> Maybe document the lock's purpose
>>>
>>>
>>>> +	unsigned int		seqno;
>>>> +	struct list_head	job_list;
>>>> +};
>>>> +
>>>> +struct drm_mock_sched_entity {
>>>> +	struct drm_sched_entity base;
>>>> +
>>>> +	struct kunit		*test;
>>>> +};
>>>> +
>>>> +struct drm_mock_sched_job {
>>>> +	struct drm_sched_job	base;
>>>> +
>>>> +	struct list_head	link;
>>>> +	struct hrtimer		timer;
>>>> +
>>>> +	unsigned int		duration_us;
>>>> +	ktime_t			finish_at;
>>>> +
>>>> +	spinlock_t		lock;
>>>
>>> Same
>>>
>>>> +	struct dma_fence	hw_fence;
>>>> +
>>>> +	struct kunit		*test;
>>>> +};
>>>> +
>>>> +static inline struct drm_mock_scheduler *
>>>> +drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched)
>>>> +{
>>>> +	return container_of(sched, struct drm_mock_scheduler,
>>>> base);
>>>> +};
>>>> +
>>>> +static inline struct drm_mock_sched_entity *
>>>> +drm_sched_entity_to_mock_entity(struct drm_sched_entity
>>>> *sched_entity)
>>>> +{
>>>> +	return container_of(sched_entity, struct
>>>> drm_mock_sched_entity, base);
>>>> +};
>>>> +
>>>> +static inline struct drm_mock_sched_job *
>>>> +drm_sched_job_to_mock_job(struct drm_sched_job *sched_job)
>>>> +{
>>>> +	return container_of(sched_job, struct
>>>> drm_mock_sched_job,
>>>> base);
>>>> +};
>>>> +
>>>> +struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit
>>>> *test);
>>>> +void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched);
>>>> +unsigned int drm_mock_sched_advance(struct drm_mock_scheduler
>>>> *sched,
>>>> +				    unsigned int num);
>>>> +
>>>> +struct drm_mock_sched_entity *
>>>> +drm_mock_new_sched_entity(struct kunit *test,
>>>> +			  enum drm_sched_priority priority,
>>>> +			  struct drm_mock_scheduler *sched);
>>>> +void drm_mock_sched_entity_free(struct drm_mock_sched_entity
>>>> *entity);
>>>> +
>>>> +struct drm_mock_sched_job *
>>>> +drm_mock_new_sched_job(struct kunit *test,
>>>> +		       struct drm_mock_sched_entity *entity);
>>>> +
>>>> +static inline void drm_mock_sched_job_submit(struct
>>>> drm_mock_sched_job *job)
>>>> +{
>>>> +	drm_sched_job_arm(&job->base);
>>>> +	drm_sched_entity_push_job(&job->base);
>>>> +}
>>>> +
>>>> +static inline void
>>>> +drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job
>>>> *job,
>>>> +				   unsigned int duration_us)
>>>> +{
>>>> +	job->duration_us = duration_us;
>>>> +}
>>>> +
>>>> +static inline bool
>>>> +drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job)
>>>> +{
>>>> +	return dma_fence_is_signaled(&job->base.s_fence-
>>>>> finished);
>>>> +}
>>>> +
>>>> +static inline bool
>>>> +drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job,
>>>> long timeout)
>>>> +{
>>>> +	long ret;
>>>> +
>>>> +	ret = dma_fence_wait_timeout(&job->base.s_fence-
>>>>> finished,
>>>> +				      false,
>>>> +				      timeout);
>>>> +
>>>> +	return ret != 0;
>>>> +}
>>>> +
>>>> +static inline long
>>>> +drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job
>>>> *job,
>>>> long timeout)
>>>> +{
>>>> +	long ret;
>>>> +
>>>> +	ret = dma_fence_wait_timeout(&job->base.s_fence-
>>>>> scheduled,
>>>> +				      false,
>>>> +				      timeout);
>>>> +
>>>> +	return ret != 0;
>>>> +}
>>>> diff --git
>>>> a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>>>> b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>>>> new file mode 100644
>>>> index 000000000000..6fd39bea95b1
>>>> --- /dev/null
>>>> +++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
>>>> @@ -0,0 +1,247 @@
>>>> +
>>>> +#include "drm_sched_tests.h"
>>>> +
>>>> +static int drm_sched_basic_init(struct kunit *test)
>>>> +{
>>>> +	test->priv = drm_mock_new_scheduler(test);
>>>> +
>>>> +	return 0;
>>>> +}
>>>> +
>>>> +static void drm_sched_basic_exit(struct kunit *test)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched = test->priv;
>>>> +
>>>> +	drm_mock_scheduler_fini(sched);
>>>> +}
>>>> +
>>>> +static void drm_sched_basic_submit(struct kunit *test)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched = test->priv;
>>>> +	struct drm_mock_sched_entity *entity;
>>>> +	struct drm_mock_sched_job *job;
>>>> +	unsigned int i;
>>>> +	bool done;
>>>> +
>>>> +	/*
>>>> +	 * Submit one job to the scheduler and verify that it
>>>> gets
>>>> scheduled
>>>> +	 * and completed only when the mock hw backend processes
>>>> it.
>>>> +	 */
>>>> +
>>>> +	entity = drm_mock_new_sched_entity(test,
>>>> +					
>>>> DRM_SCHED_PRIORITY_NORMAL,
>>>> +					   sched);
>>>> +	job = drm_mock_new_sched_job(test, entity);
>>>> +
>>>> +	drm_mock_sched_job_submit(job);
>>>> +
>>>> +	done = drm_mock_sched_job_wait_scheduled(job, HZ);
>>>> +	KUNIT_ASSERT_EQ(test, done, true);
>>>> +
>>>> +	done = drm_mock_sched_job_wait_finished(job, HZ / 2);
>>>> +	KUNIT_ASSERT_EQ(test, done, false);
>>>> +
>>>> +	i = drm_mock_sched_advance(sched, 1);
>>>> +	KUNIT_ASSERT_EQ(test, i, 1);
>>>> +
>>>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>>>> +	KUNIT_ASSERT_EQ(test, done, true);
>>>> +
>>>> +	drm_mock_sched_entity_free(entity);
>>>> +}
>>>> +
>>>> +static void drm_sched_basic_queue(struct kunit *test)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched = test->priv;
>>>> +	struct drm_mock_sched_entity *entity;
>>>> +	struct drm_mock_sched_job *job;
>>>> +	const unsigned int qd = 100;
>>>
>>> Not the best variable name – is this "queue depth"? Why is it 100?
>>> ->
>>> global define & document
>>
>> I wouldn't promote this to global and TBH if you look how small this
>> test function is it feels pretty self documenting.
> 
> I meant global in the sense of at the top of the file as a constant.
> 
> Wouldn't you agree that it would be good to have test parameters at a
> place where you can easily modify them, in case you want to compile the
> tests a bit differently?

I agree in principle, just don't want to go with one define for all tests.

I changed it locally to a parameterized test:

static const struct drm_sched_basic_params drm_sched_basic_cases[] = {
	{
		.description = "A queue of jobs in a single entity",
		.queue_depth = 100,
		.job_us = 1000,
		.num_entities = 1,
	},
	{
		.description = "A chain of dependent jobs across multiple entities",
		.queue_depth = 100,
		.job_us = 1000,
		.num_entities = 1,
		.dep_chain = true,
	},
	{
		.description = "Multiple independent job queues",
		.queue_depth = 100,
		.job_us = 1000,
		.num_entities = 4,
	},
	{
		.description = "Multiple inter-dependent job queues",
		.queue_depth = 100,
		.job_us = 1000,
		.num_entities = 4,
		.dep_chain = true,
	},
};

...
	KUNIT_CASE_PARAM(drm_sched_basic_test, drm_sched_basic_gen_params),
...

So now it is a single test body. Hopefully this satisfies both your 
desire to be able to modify easily, and my annoyance that there was too 
much of code duplication in v1.

Regards,

Tvrtko

> 
> 
> P.
> 
> 
>>
>>>
>>>
>>>> +	unsigned int i;
>>>> +	bool done;
>>>> +
>>>> +	/*
>>>> +	 * Submit a queue of jobs on the same entity, let them
>>>> be
>>>> completed by
>>>> +	 * the mock hw backend and check the status of the last
>>>> job.
>>>> +	 */
>>>> +
>>>> +	entity = drm_mock_new_sched_entity(test,
>>>> +					
>>>> DRM_SCHED_PRIORITY_NORMAL,
>>>> +					   sched);
>>>> +
>>>> +	for (i = 0; i < qd; i++) {
>>>> +		job = drm_mock_new_sched_job(test, entity);
>>>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>>>> +		drm_mock_sched_job_submit(job);
>>>> +	}
>>>> +
>>>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>>>> +	KUNIT_ASSERT_EQ(test, done, true);
>>>> +
>>>> +	drm_mock_sched_entity_free(entity);
>>>> +}
>>>> +
>>>> +static void drm_sched_basic_chain(struct kunit *test)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched = test->priv;
>>>> +	struct drm_mock_sched_job *job, *prev = NULL;
>>>> +	struct drm_mock_sched_entity *entity;
>>>> +	const unsigned int qd = 100;
>>>> +	unsigned int i;
>>>> +	bool done;
>>>> +
>>>> +	/*
>>>> +	 * Submit a queue of jobs on the same entity but with an
>>>> explicit
>>>> +	 * chain of dependencies between them. Let the jobs be
>>>> completed by
>>>> +	 * the mock hw backend and check the status of the last
>>>> job.
>>>> +	 */
>>>> +
>>>> +	entity = drm_mock_new_sched_entity(test,
>>>> +					
>>>> DRM_SCHED_PRIORITY_NORMAL,
>>>> +					   sched);
>>>> +
>>>> +	for (i = 0; i < qd; i++) {
>>>> +		job = drm_mock_new_sched_job(test, entity);
>>>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>>>> +		if (prev)
>>>> +			drm_sched_job_add_dependency(&job->base,
>>>> +						
>>>> dma_fence_get(&prev->base.s_fence->finished));
>>>> +		drm_mock_sched_job_submit(job);
>>>> +		prev = job;
>>>> +	}
>>>> +
>>>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>>>> +	KUNIT_ASSERT_EQ(test, done, true);
>>>> +
>>>> +	drm_mock_sched_entity_free(entity);
>>>> +}
>>>> +
>>>> +static void drm_sched_basic_entities(struct kunit *test)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched = test->priv;
>>>> +	struct drm_mock_sched_entity *entity[4];
>>>> +	struct drm_mock_sched_job *job;
>>>> +	const unsigned int qd = 100;
>>>> +	unsigned int i, cur_ent = 0;
>>>> +	bool done;
>>>> +
>>>> +	/*
>>>> +	 * Submit a queue of jobs across different entities, let
>>>> them be
>>>> +	 * completed by the mock hw backend and check the status
>>>> of
>>>> the last
>>>> +	 * job.
>>>> +	 */
>>>> +
>>>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>>>> +		entity[i] = drm_mock_new_sched_entity(test,
>>>> +						
>>>> DRM_SCHED_PRIORITY_NORMAL,
>>>> +						      sched);
>>>> +
>>>> +	for (i = 0; i < qd; i++) {
>>>> +		job = drm_mock_new_sched_job(test,
>>>> entity[cur_ent++]);
>>>> +		cur_ent %= ARRAY_SIZE(entity);
>>>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>>>> +		drm_mock_sched_job_submit(job);
>>>> +	}
>>>> +
>>>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>>>> +	KUNIT_ASSERT_EQ(test, done, true);
>>>> +
>>>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>>>> +		drm_mock_sched_entity_free(entity[i]);
>>>> +}
>>>> +
>>>> +static void drm_sched_basic_entities_chain(struct kunit *test)
>>>> +{
>>>> +	struct drm_mock_scheduler *sched = test->priv;
>>>> +	struct drm_mock_sched_job *job, *prev = NULL;
>>>> +	struct drm_mock_sched_entity *entity[4];
>>>> +	const unsigned int qd = 100;
>>>> +	unsigned int i, cur_ent = 0;
>>>> +	bool done;
>>>> +
>>>> +	/*
>>>> +	 * Submit a queue of jobs across different entities with
>>>> an
>>>> explicit
>>>> +	 * chain of dependencies between them. Let the jobs be
>>>> completed by
>>>> +	 * the mock hw backend and check the status of the last
>>>> job.
>>>> +	 */
>>>> +
>>>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>>>> +		entity[i] = drm_mock_new_sched_entity(test,
>>>> +						
>>>> DRM_SCHED_PRIORITY_NORMAL,
>>>> +						      sched);
>>>> +
>>>> +	for (i = 0; i < qd; i++) {
>>>> +		job = drm_mock_new_sched_job(test,
>>>> entity[cur_ent++]);
>>>> +		cur_ent %= ARRAY_SIZE(entity);
>>>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>>>> +		if (prev)
>>>> +			drm_sched_job_add_dependency(&job->base,
>>>> +						
>>>> dma_fence_get(&prev->base.s_fence->finished));
>>>> +		drm_mock_sched_job_submit(job);
>>>> +		prev = job;
>>>> +	}
>>>> +
>>>> +	done = drm_mock_sched_job_wait_finished(job, HZ);
>>>> +	KUNIT_ASSERT_EQ(test, done, true);
>>>> +
>>>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>>>> +		drm_mock_sched_entity_free(entity[i]);
>>>> +}
>>>> +
>>>> +static void drm_sched_basic_entity_cleanup(struct kunit *test)
>>>> +{
>>>> +	struct drm_mock_sched_job *job, *mid, *prev = NULL;
>>>> +	struct drm_mock_scheduler *sched = test->priv;
>>>> +	struct drm_mock_sched_entity *entity[4];
>>>
>>> 4 is used in several places, so could be defined globally. In case
>>> there's a special reason for why it's 4, that could be mentioned,
>>> too
>>
>> It's completely arbitrary. In this test we just need a bunch of
>> entities
>> to be active on the scheduler as we trigger a client exit in the
>> middle
>> of it. IMO the comment few lines below should be good enough to
>> explain
>> the idea. I fear that promoting this to a global define would just
>> give
>> it more weight that what it has. Then some test would want a
>> different
>> number etc.
>>
>> Regards,
>>
>> Tvrtko
>>
>>>> +	const unsigned int qd = 100;
>>>> +	unsigned int i, cur_ent = 0;
>>>> +	bool done;
>>>> +
>>>> +	/*
>>>> +	 * Submit a queue of jobs across different entities with
>>>> an
>>>> explicit
>>>> +	 * chain of dependencies between them and trigger entity
>>>> cleanup while
>>>> +	 * the queue is still being processed.
>>>> +	 */
>>>> +
>>>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>>>> +		entity[i] = drm_mock_new_sched_entity(test,
>>>> +						
>>>> DRM_SCHED_PRIORITY_NORMAL,
>>>> +						      sched);
>>>> +
>>>> +	for (i = 0; i < qd; i++) {
>>>> +		job = drm_mock_new_sched_job(test,
>>>> entity[cur_ent++]);
>>>> +		cur_ent %= ARRAY_SIZE(entity);
>>>> +		drm_mock_sched_job_set_duration_us(job, 1000);
>>>> +		if (prev)
>>>> +			drm_sched_job_add_dependency(&job->base,
>>>> +						
>>>> dma_fence_get(&prev->base.s_fence->finished));
>>>> +		drm_mock_sched_job_submit(job);
>>>> +		if (i == qd / 2)
>>>> +			mid = job;
>>>> +		prev = job;
>>>> +	}
>>>> +
>>>> +	done = drm_mock_sched_job_wait_finished(mid, HZ);
>>>> +	KUNIT_ASSERT_EQ(test, done, true);
>>>> +
>>>> +	/* Exit with half of the queue still pending to be
>>>> executed.
>>>> */
>>>> +	for (i = 0; i < ARRAY_SIZE(entity); i++)
>>>> +		drm_mock_sched_entity_free(entity[i]);}
>>>> +
>>>> +static struct kunit_case drm_sched_basic_tests[] = {
>>>> +	KUNIT_CASE(drm_sched_basic_submit),
>>>> +	KUNIT_CASE(drm_sched_basic_queue),
>>>> +	KUNIT_CASE(drm_sched_basic_chain),
>>>> +	KUNIT_CASE(drm_sched_basic_entities),
>>>> +	KUNIT_CASE(drm_sched_basic_entities_chain),
>>>> +	KUNIT_CASE(drm_sched_basic_entity_cleanup),
>>>> +	{}
>>>> +};
>>>> +
>>>> +static struct kunit_suite drm_sched_basic = {
>>>> +	.name = "drm_sched_basic_tests",
>>>> +	.init = drm_sched_basic_init,
>>>> +	.exit = drm_sched_basic_exit,
>>>> +	.test_cases = drm_sched_basic_tests,
>>>> +};
>>>> +
>>>> +kunit_test_suite(drm_sched_basic);
>>>
>
diff mbox series

Patch

diff --git a/drivers/gpu/drm/Kconfig.debug b/drivers/gpu/drm/Kconfig.debug
index a35d74171b7b..89f777f21e95 100644
--- a/drivers/gpu/drm/Kconfig.debug
+++ b/drivers/gpu/drm/Kconfig.debug
@@ -88,5 +88,17 @@  config DRM_TTM_KUNIT_TEST
 
 	  If in doubt, say "N".
 
+config DRM_SCHED_KUNIT_TEST
+	tristate "KUnit tests for the DRM scheduler" if !KUNIT_ALL_TESTS
+	select DRM_SCHED
+	depends on DRM && KUNIT
+	default KUNIT_ALL_TESTS
+	help
+	  Choose this option to build unit tests for the DRM scheduler.
+
+	  Recommended for driver developers only.
+
+	  If in doubt, say "N".
+
 config DRM_EXPORT_FOR_TESTS
 	bool
diff --git a/drivers/gpu/drm/scheduler/.kunitconfig b/drivers/gpu/drm/scheduler/.kunitconfig
new file mode 100644
index 000000000000..cece53609fcf
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/.kunitconfig
@@ -0,0 +1,12 @@ 
+CONFIG_KUNIT=y
+CONFIG_DRM=y
+CONFIG_DRM_SCHED_KUNIT_TEST=y
+CONFIG_EXPERT=y
+CONFIG_DEBUG_SPINLOCK=y
+CONFIG_DEBUG_MUTEXES=y
+CONFIG_DEBUG_ATOMIC_SLEEP=y
+CONFIG_LOCK_DEBUGGING_SUPPORT=y
+CONFIG_PROVE_LOCKING=y
+CONFIG_LOCKDEP=y
+CONFIG_DEBUG_LOCKDEP=y
+CONFIG_DEBUG_LIST=y
diff --git a/drivers/gpu/drm/scheduler/Makefile b/drivers/gpu/drm/scheduler/Makefile
index 53863621829f..46dfdca0758a 100644
--- a/drivers/gpu/drm/scheduler/Makefile
+++ b/drivers/gpu/drm/scheduler/Makefile
@@ -23,3 +23,4 @@ 
 gpu-sched-y := sched_main.o sched_fence.o sched_entity.o
 
 obj-$(CONFIG_DRM_SCHED) += gpu-sched.o
+obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += tests/
diff --git a/drivers/gpu/drm/scheduler/tests/Makefile b/drivers/gpu/drm/scheduler/tests/Makefile
new file mode 100644
index 000000000000..d69eab6a2e9b
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/Makefile
@@ -0,0 +1,4 @@ 
+
+obj-$(CONFIG_DRM_SCHED_KUNIT_TEST) += \
+        drm_mock_scheduler.o \
+        drm_sched_tests_basic.o
diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
new file mode 100644
index 000000000000..c9205f9ff524
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_mock_entity.c
@@ -0,0 +1,29 @@ 
+
+#include "drm_sched_tests.h"
+
+struct drm_mock_sched_entity *
+drm_mock_sched_entity_new(struct kunit *test,
+			  enum drm_sched_priority priority,
+			  struct drm_mock_scheduler *sched)
+{
+	struct drm_sched_mock_entity *entity;
+	int ret;
+
+	entity = kunit_kmalloc(test, sizeof(*entity), GFP_KERNEL);
+	KUNIT_ASSERT_NOT_NULL(test, entity);
+
+	ret = drm_sched_entity_init(&entity->base,
+				    priority,
+				    &sched->base, 1,
+				    NULL);
+	KUNIT_ASSERT_EQ(test, ret, 0);
+
+	entity->test = test;
+
+	return entity;
+}
+
+void drm_mock_sched_entity_free(struct drm_mock_sched_entity *entity)
+{
+	drm_sched_entity_fini(&entity->base);
+}
diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_job.c b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
new file mode 100644
index 000000000000..d94568cf3da9
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_mock_job.c
@@ -0,0 +1,3 @@ 
+
+#include "drm_sched_tests.h"
+
diff --git a/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
new file mode 100644
index 000000000000..f1985900a6ba
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_mock_scheduler.c
@@ -0,0 +1,254 @@ 
+
+#include "drm_sched_tests.h"
+
+struct drm_mock_sched_entity *
+drm_mock_new_sched_entity(struct kunit *test,
+			  enum drm_sched_priority priority,
+			  struct drm_mock_scheduler *sched)
+{
+	struct drm_mock_sched_entity *entity;
+	struct drm_gpu_scheduler *drm_sched;
+	int ret;
+
+	entity = kunit_kzalloc(test, sizeof(*entity), GFP_KERNEL);
+	KUNIT_ASSERT_NOT_NULL(test, entity);
+
+	drm_sched = &sched->base;
+	ret = drm_sched_entity_init(&entity->base,
+				    priority,
+				    &drm_sched, 1,
+				    NULL);
+	KUNIT_ASSERT_EQ(test, ret, 0);
+
+	entity->test = test;
+
+	return entity;
+}
+
+void drm_mock_sched_entity_free(struct drm_mock_sched_entity *entity)
+{
+	drm_sched_entity_destroy(&entity->base);
+}
+
+static enum hrtimer_restart
+drm_sched_mock_job_signal_timer(struct hrtimer *hrtimer)
+{
+	struct drm_mock_sched_job *upto =
+		container_of(hrtimer, typeof(*upto), timer);
+	struct drm_mock_scheduler *sched =
+		drm_sched_to_mock_sched(upto->base.sched);
+	struct drm_mock_sched_job *job, *next;
+	ktime_t now = ktime_get();
+	unsigned long flags;
+	LIST_HEAD(signal);
+
+	spin_lock_irqsave(&sched->lock, flags);
+	list_for_each_entry_safe(job, next, &sched->job_list, link) {
+		if (!job->duration_us)
+			break;
+
+		if (ktime_before(now, job->finish_at))
+			break;
+
+		list_move_tail(&job->link, &signal);
+		sched->seqno = job->hw_fence.seqno;
+	}
+	spin_unlock_irqrestore(&sched->lock, flags);
+
+	list_for_each_entry(job, &signal, link) {
+		dma_fence_signal(&job->hw_fence);
+		dma_fence_put(&job->hw_fence);
+	}
+
+	return HRTIMER_NORESTART;
+}
+
+struct drm_mock_sched_job *
+drm_mock_new_sched_job(struct kunit *test,
+		       struct drm_mock_sched_entity *entity)
+{
+	struct drm_mock_sched_job *job;
+	int ret;
+
+	job = kunit_kzalloc(test, sizeof(*job), GFP_KERNEL);
+	KUNIT_ASSERT_NOT_NULL(test, job);
+
+	ret = drm_sched_job_init(&job->base,
+				 &entity->base,
+				 1,
+				 NULL);
+	KUNIT_ASSERT_EQ(test, ret, 0);
+
+	job->test = test;
+
+	spin_lock_init(&job->lock);
+	INIT_LIST_HEAD(&job->link);
+	hrtimer_init(&job->timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS);
+	job->timer.function = drm_sched_mock_job_signal_timer;
+
+	return job;
+}
+
+static const char *drm_mock_sched_hw_fence_driver_name(struct dma_fence *fence)
+{
+	return "drm_mock_sched";
+}
+
+static const char *
+drm_mock_sched_hw_fence_timeline_name(struct dma_fence *fence)
+{
+	struct drm_mock_sched_job *job =
+		container_of(fence, typeof(*job), hw_fence);
+
+	return (const char *)job->base.sched->name;
+}
+static void drm_mock_sched_hw_fence_release(struct dma_fence *fence)
+{
+	struct drm_mock_sched_job *job =
+		container_of(fence, typeof(*job), hw_fence);
+
+	hrtimer_cancel(&job->timer);
+
+	/* Freed by the kunit framework */
+}
+
+static const struct dma_fence_ops drm_mock_sched_hw_fence_ops = {
+	.get_driver_name = drm_mock_sched_hw_fence_driver_name,
+	.get_timeline_name = drm_mock_sched_hw_fence_timeline_name,
+	.release = drm_mock_sched_hw_fence_release,
+};
+
+static struct dma_fence *mock_sched_run_job(struct drm_sched_job *sched_job)
+{
+	struct drm_mock_scheduler *sched =
+		drm_sched_to_mock_sched(sched_job->sched);
+	struct drm_mock_sched_job *job = drm_sched_job_to_mock_job(sched_job);
+
+	dma_fence_init(&job->hw_fence,
+		       &drm_mock_sched_hw_fence_ops,
+		       &job->lock,
+		       sched->hw_fence.context,
+		       atomic_inc_return(&sched->hw_fence.seqno));
+
+	dma_fence_get(&job->hw_fence); /* Reference for the job_list */
+
+	spin_lock_irq(&sched->lock);
+	if (job->duration_us) {
+		ktime_t prev_finish_at = 0;
+
+		if (!list_empty(&sched->job_list)) {
+			struct drm_mock_sched_job *prev =
+				list_last_entry(&sched->job_list, typeof(*prev),
+						link);
+
+			prev_finish_at = prev->finish_at;
+		}
+
+		if (!prev_finish_at)
+			prev_finish_at = ktime_get();
+
+		job->finish_at = ktime_add_us(prev_finish_at, job->duration_us);
+	}
+	list_add_tail(&job->link, &sched->job_list);
+	if (job->finish_at)
+		hrtimer_start(&job->timer, job->finish_at, HRTIMER_MODE_ABS);
+	spin_unlock_irq(&sched->lock);
+
+	return &job->hw_fence;
+}
+
+static enum drm_gpu_sched_stat
+mock_sched_timedout_job(struct drm_sched_job *sched_job)
+{
+	return DRM_GPU_SCHED_STAT_ENODEV;
+}
+
+static void mock_sched_free_job(struct drm_sched_job *sched_job)
+{
+	drm_sched_job_cleanup(sched_job);
+}
+
+static const struct drm_sched_backend_ops drm_mock_scheduler_ops = {
+	.run_job = mock_sched_run_job,
+	.timedout_job = mock_sched_timedout_job,
+	.free_job = mock_sched_free_job
+};
+
+struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit *test)
+{
+	struct drm_mock_scheduler *sched;
+	int ret;
+
+	sched = kunit_kzalloc(test, sizeof(*sched), GFP_KERNEL);
+	KUNIT_ASSERT_NOT_NULL(test, sched);
+
+	ret = drm_sched_init(&sched->base,
+			     &drm_mock_scheduler_ops,
+			     NULL, /* wq */
+			     DRM_SCHED_PRIORITY_COUNT,
+			     U32_MAX, /* max credits */
+			     UINT_MAX, /* hang limit */
+			     MAX_SCHEDULE_TIMEOUT, /* timeout */
+			     NULL, /* timeout wq */
+			     NULL, /* score */
+			     "drm-mock-scheduler",
+			     NULL /* dev */);
+	KUNIT_ASSERT_EQ(test, ret, 0);
+
+	sched->test = test;
+	sched->hw_fence.context = dma_fence_context_alloc(1);
+	atomic_set(&sched->hw_fence.seqno, 0);
+	INIT_LIST_HEAD(&sched->job_list);
+	spin_lock_init(&sched->lock);
+
+	return sched;
+}
+
+void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched)
+{
+	struct drm_mock_sched_job *job, *next;
+	unsigned long flags;
+	LIST_HEAD(signal);
+
+	spin_lock_irqsave(&sched->lock, flags);
+	list_for_each_entry_safe(job, next, &sched->job_list, link)
+		list_move_tail(&job->link, &signal);
+	spin_unlock_irqrestore(&sched->lock, flags);
+
+	list_for_each_entry(job, &signal, link) {
+		hrtimer_cancel(&job->timer);
+		dma_fence_put(&job->hw_fence);
+	}
+
+	drm_sched_fini(&sched->base);
+}
+
+unsigned int drm_mock_sched_advance(struct drm_mock_scheduler *sched,
+				    unsigned int num)
+{
+	struct drm_mock_sched_job *job, *next;
+	unsigned int found = 0;
+	unsigned long flags;
+	LIST_HEAD(signal);
+
+	spin_lock_irqsave(&sched->lock, flags);
+	if (WARN_ON_ONCE(sched->seqno + num < sched->seqno))
+		goto unlock;
+	sched->seqno += num;
+	list_for_each_entry_safe(job, next, &sched->job_list, link) {
+		if (sched->seqno < job->hw_fence.seqno)
+			break;
+
+		list_move_tail(&job->link, &signal);
+		found++;
+	}
+unlock:
+	spin_unlock_irqrestore(&sched->lock, flags);
+
+	list_for_each_entry(job, &signal, link) {
+		dma_fence_signal(&job->hw_fence);
+		dma_fence_put(&job->hw_fence);
+	}
+
+	return found;
+}
diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
new file mode 100644
index 000000000000..421ee2712985
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests.h
@@ -0,0 +1,124 @@ 
+
+#include <kunit/test.h>
+#include <linux/atomic.h>
+#include <linux/dma-fence.h>
+#include <linux/hrtimer.h>
+#include <linux/ktime.h>
+#include <linux/list.h>
+#include <linux/atomic.h>
+#include <linux/mutex.h>
+#include <linux/types.h>
+
+#include <drm/gpu_scheduler.h>
+
+struct drm_mock_scheduler {
+	struct drm_gpu_scheduler base;
+
+	struct kunit		*test;
+
+	struct {
+		u64		context;
+		atomic_t	seqno;
+	} hw_fence;
+
+	spinlock_t		lock;
+	unsigned int		seqno;
+	struct list_head	job_list;
+};
+
+struct drm_mock_sched_entity {
+	struct drm_sched_entity base;
+
+	struct kunit		*test;
+};
+
+struct drm_mock_sched_job {
+	struct drm_sched_job	base;
+
+	struct list_head	link;
+	struct hrtimer		timer;
+
+	unsigned int		duration_us;
+	ktime_t			finish_at;
+
+	spinlock_t		lock;
+	struct dma_fence	hw_fence;
+
+	struct kunit		*test;
+};
+
+static inline struct drm_mock_scheduler *
+drm_sched_to_mock_sched(struct drm_gpu_scheduler *sched)
+{
+	return container_of(sched, struct drm_mock_scheduler, base);
+};
+
+static inline struct drm_mock_sched_entity *
+drm_sched_entity_to_mock_entity(struct drm_sched_entity *sched_entity)
+{
+	return container_of(sched_entity, struct drm_mock_sched_entity, base);
+};
+
+static inline struct drm_mock_sched_job *
+drm_sched_job_to_mock_job(struct drm_sched_job *sched_job)
+{
+	return container_of(sched_job, struct drm_mock_sched_job, base);
+};
+
+struct drm_mock_scheduler *drm_mock_new_scheduler(struct kunit *test);
+void drm_mock_scheduler_fini(struct drm_mock_scheduler *sched);
+unsigned int drm_mock_sched_advance(struct drm_mock_scheduler *sched,
+				    unsigned int num);
+
+struct drm_mock_sched_entity *
+drm_mock_new_sched_entity(struct kunit *test,
+			  enum drm_sched_priority priority,
+			  struct drm_mock_scheduler *sched);
+void drm_mock_sched_entity_free(struct drm_mock_sched_entity *entity);
+
+struct drm_mock_sched_job *
+drm_mock_new_sched_job(struct kunit *test,
+		       struct drm_mock_sched_entity *entity);
+
+static inline void drm_mock_sched_job_submit(struct drm_mock_sched_job *job)
+{
+	drm_sched_job_arm(&job->base);
+	drm_sched_entity_push_job(&job->base);
+}
+
+static inline void
+drm_mock_sched_job_set_duration_us(struct drm_mock_sched_job *job,
+				   unsigned int duration_us)
+{
+	job->duration_us = duration_us;
+}
+
+static inline bool
+drm_mock_sched_job_is_finished(struct drm_mock_sched_job *job)
+{
+	return dma_fence_is_signaled(&job->base.s_fence->finished);
+}
+
+static inline bool
+drm_mock_sched_job_wait_finished(struct drm_mock_sched_job *job, long timeout)
+{
+	long ret;
+
+	ret = dma_fence_wait_timeout(&job->base.s_fence->finished,
+				      false,
+				      timeout);
+
+	return ret != 0;
+}
+
+static inline long
+drm_mock_sched_job_wait_scheduled(struct drm_mock_sched_job *job, long timeout)
+{
+	long ret;
+
+	ret = dma_fence_wait_timeout(&job->base.s_fence->scheduled,
+				      false,
+				      timeout);
+
+	return ret != 0;
+}
diff --git a/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
new file mode 100644
index 000000000000..6fd39bea95b1
--- /dev/null
+++ b/drivers/gpu/drm/scheduler/tests/drm_sched_tests_basic.c
@@ -0,0 +1,247 @@ 
+
+#include "drm_sched_tests.h"
+
+static int drm_sched_basic_init(struct kunit *test)
+{
+	test->priv = drm_mock_new_scheduler(test);
+
+	return 0;
+}
+
+static void drm_sched_basic_exit(struct kunit *test)
+{
+	struct drm_mock_scheduler *sched = test->priv;
+
+	drm_mock_scheduler_fini(sched);
+}
+
+static void drm_sched_basic_submit(struct kunit *test)
+{
+	struct drm_mock_scheduler *sched = test->priv;
+	struct drm_mock_sched_entity *entity;
+	struct drm_mock_sched_job *job;
+	unsigned int i;
+	bool done;
+
+	/*
+	 * Submit one job to the scheduler and verify that it gets scheduled
+	 * and completed only when the mock hw backend processes it.
+	 */
+
+	entity = drm_mock_new_sched_entity(test,
+					   DRM_SCHED_PRIORITY_NORMAL,
+					   sched);
+	job = drm_mock_new_sched_job(test, entity);
+
+	drm_mock_sched_job_submit(job);
+
+	done = drm_mock_sched_job_wait_scheduled(job, HZ);
+	KUNIT_ASSERT_EQ(test, done, true);
+
+	done = drm_mock_sched_job_wait_finished(job, HZ / 2);
+	KUNIT_ASSERT_EQ(test, done, false);
+
+	i = drm_mock_sched_advance(sched, 1);
+	KUNIT_ASSERT_EQ(test, i, 1);
+
+	done = drm_mock_sched_job_wait_finished(job, HZ);
+	KUNIT_ASSERT_EQ(test, done, true);
+
+	drm_mock_sched_entity_free(entity);
+}
+
+static void drm_sched_basic_queue(struct kunit *test)
+{
+	struct drm_mock_scheduler *sched = test->priv;
+	struct drm_mock_sched_entity *entity;
+	struct drm_mock_sched_job *job;
+	const unsigned int qd = 100;
+	unsigned int i;
+	bool done;
+
+	/*
+	 * Submit a queue of jobs on the same entity, let them be completed by
+	 * the mock hw backend and check the status of the last job.
+	 */
+
+	entity = drm_mock_new_sched_entity(test,
+					   DRM_SCHED_PRIORITY_NORMAL,
+					   sched);
+
+	for (i = 0; i < qd; i++) {
+		job = drm_mock_new_sched_job(test, entity);
+		drm_mock_sched_job_set_duration_us(job, 1000);
+		drm_mock_sched_job_submit(job);
+	}
+
+	done = drm_mock_sched_job_wait_finished(job, HZ);
+	KUNIT_ASSERT_EQ(test, done, true);
+
+	drm_mock_sched_entity_free(entity);
+}
+
+static void drm_sched_basic_chain(struct kunit *test)
+{
+	struct drm_mock_scheduler *sched = test->priv;
+	struct drm_mock_sched_job *job, *prev = NULL;
+	struct drm_mock_sched_entity *entity;
+	const unsigned int qd = 100;
+	unsigned int i;
+	bool done;
+
+	/*
+	 * Submit a queue of jobs on the same entity but with an explicit
+	 * chain of dependencies between them. Let the jobs be completed by
+	 * the mock hw backend and check the status of the last job.
+	 */
+
+	entity = drm_mock_new_sched_entity(test,
+					   DRM_SCHED_PRIORITY_NORMAL,
+					   sched);
+
+	for (i = 0; i < qd; i++) {
+		job = drm_mock_new_sched_job(test, entity);
+		drm_mock_sched_job_set_duration_us(job, 1000);
+		if (prev)
+			drm_sched_job_add_dependency(&job->base,
+						     dma_fence_get(&prev->base.s_fence->finished));
+		drm_mock_sched_job_submit(job);
+		prev = job;
+	}
+
+	done = drm_mock_sched_job_wait_finished(job, HZ);
+	KUNIT_ASSERT_EQ(test, done, true);
+
+	drm_mock_sched_entity_free(entity);
+}
+
+static void drm_sched_basic_entities(struct kunit *test)
+{
+	struct drm_mock_scheduler *sched = test->priv;
+	struct drm_mock_sched_entity *entity[4];
+	struct drm_mock_sched_job *job;
+	const unsigned int qd = 100;
+	unsigned int i, cur_ent = 0;
+	bool done;
+
+	/*
+	 * Submit a queue of jobs across different entities, let them be
+	 * completed by the mock hw backend and check the status of the last
+	 * job.
+	 */
+
+	for (i = 0; i < ARRAY_SIZE(entity); i++)
+		entity[i] = drm_mock_new_sched_entity(test,
+						      DRM_SCHED_PRIORITY_NORMAL,
+						      sched);
+
+	for (i = 0; i < qd; i++) {
+		job = drm_mock_new_sched_job(test, entity[cur_ent++]);
+		cur_ent %= ARRAY_SIZE(entity);
+		drm_mock_sched_job_set_duration_us(job, 1000);
+		drm_mock_sched_job_submit(job);
+	}
+
+	done = drm_mock_sched_job_wait_finished(job, HZ);
+	KUNIT_ASSERT_EQ(test, done, true);
+
+	for (i = 0; i < ARRAY_SIZE(entity); i++)
+		drm_mock_sched_entity_free(entity[i]);
+}
+
+static void drm_sched_basic_entities_chain(struct kunit *test)
+{
+	struct drm_mock_scheduler *sched = test->priv;
+	struct drm_mock_sched_job *job, *prev = NULL;
+	struct drm_mock_sched_entity *entity[4];
+	const unsigned int qd = 100;
+	unsigned int i, cur_ent = 0;
+	bool done;
+
+	/*
+	 * Submit a queue of jobs across different entities with an explicit
+	 * chain of dependencies between them. Let the jobs be completed by
+	 * the mock hw backend and check the status of the last job.
+	 */
+
+	for (i = 0; i < ARRAY_SIZE(entity); i++)
+		entity[i] = drm_mock_new_sched_entity(test,
+						      DRM_SCHED_PRIORITY_NORMAL,
+						      sched);
+
+	for (i = 0; i < qd; i++) {
+		job = drm_mock_new_sched_job(test, entity[cur_ent++]);
+		cur_ent %= ARRAY_SIZE(entity);
+		drm_mock_sched_job_set_duration_us(job, 1000);
+		if (prev)
+			drm_sched_job_add_dependency(&job->base,
+						     dma_fence_get(&prev->base.s_fence->finished));
+		drm_mock_sched_job_submit(job);
+		prev = job;
+	}
+
+	done = drm_mock_sched_job_wait_finished(job, HZ);
+	KUNIT_ASSERT_EQ(test, done, true);
+
+	for (i = 0; i < ARRAY_SIZE(entity); i++)
+		drm_mock_sched_entity_free(entity[i]);
+}
+
+static void drm_sched_basic_entity_cleanup(struct kunit *test)
+{
+	struct drm_mock_sched_job *job, *mid, *prev = NULL;
+	struct drm_mock_scheduler *sched = test->priv;
+	struct drm_mock_sched_entity *entity[4];
+	const unsigned int qd = 100;
+	unsigned int i, cur_ent = 0;
+	bool done;
+
+	/*
+	 * Submit a queue of jobs across different entities with an explicit
+	 * chain of dependencies between them and trigger entity cleanup while
+	 * the queue is still being processed.
+	 */
+
+	for (i = 0; i < ARRAY_SIZE(entity); i++)
+		entity[i] = drm_mock_new_sched_entity(test,
+						      DRM_SCHED_PRIORITY_NORMAL,
+						      sched);
+
+	for (i = 0; i < qd; i++) {
+		job = drm_mock_new_sched_job(test, entity[cur_ent++]);
+		cur_ent %= ARRAY_SIZE(entity);
+		drm_mock_sched_job_set_duration_us(job, 1000);
+		if (prev)
+			drm_sched_job_add_dependency(&job->base,
+						     dma_fence_get(&prev->base.s_fence->finished));
+		drm_mock_sched_job_submit(job);
+		if (i == qd / 2)
+			mid = job;
+		prev = job;
+	}
+
+	done = drm_mock_sched_job_wait_finished(mid, HZ);
+	KUNIT_ASSERT_EQ(test, done, true);
+
+	/* Exit with half of the queue still pending to be executed. */
+	for (i = 0; i < ARRAY_SIZE(entity); i++)
+		drm_mock_sched_entity_free(entity[i]);}
+
+static struct kunit_case drm_sched_basic_tests[] = {
+	KUNIT_CASE(drm_sched_basic_submit),
+	KUNIT_CASE(drm_sched_basic_queue),
+	KUNIT_CASE(drm_sched_basic_chain),
+	KUNIT_CASE(drm_sched_basic_entities),
+	KUNIT_CASE(drm_sched_basic_entities_chain),
+	KUNIT_CASE(drm_sched_basic_entity_cleanup),
+	{}
+};
+
+static struct kunit_suite drm_sched_basic = {
+	.name = "drm_sched_basic_tests",
+	.init = drm_sched_basic_init,
+	.exit = drm_sched_basic_exit,
+	.test_cases = drm_sched_basic_tests,
+};
+
+kunit_test_suite(drm_sched_basic);