diff mbox series

[v3] drm/scheduler: Improve documentation

Message ID 20240917144732.2758572-1-shuicheng.lin@intel.com (mailing list archive)
State New, archived
Headers show
Series [v3] drm/scheduler: Improve documentation | expand

Commit Message

Shuicheng Lin Sept. 17, 2024, 2:47 p.m. UTC
Function drm_sched_entity_push_job() doesn't have a return value,
remove the return value description for it.
Correct several other typo errors.

v2 (Philipp):
- more correction with related comments.

Signed-off-by: Shuicheng Lin <shuicheng.lin@intel.com>
Reviewed-by: Philipp Stanner <pstanner@redhat.com>
---
 drivers/gpu/drm/scheduler/sched_entity.c | 10 ++++------
 drivers/gpu/drm/scheduler/sched_main.c   |  4 ++--
 include/drm/gpu_scheduler.h              | 12 ++++++------
 include/linux/dma-resv.h                 |  6 +++---
 4 files changed, 15 insertions(+), 17 deletions(-)

Comments

Shuicheng Lin Sept. 22, 2024, 5:29 p.m. UTC | #1
Hi all,
I am not familiar with the process yet. To get it merged, should I add more mail-list or how to notify the maintainers?
Thanks in advance for your guide.

Best Regards
Shuicheng

> -----Original Message-----
> From: Lin, Shuicheng <shuicheng.lin@intel.com>
> Sent: Tuesday, September 17, 2024 7:48 AM
> To: dri-devel@lists.freedesktop.org
> Cc: Lin, Shuicheng <shuicheng.lin@intel.com>; Philipp Stanner
> <pstanner@redhat.com>
> Subject: [PATCH v3] drm/scheduler: Improve documentation
> 
> Function drm_sched_entity_push_job() doesn't have a return value, remove the
> return value description for it.
> Correct several other typo errors.
> 
> v2 (Philipp):
> - more correction with related comments.
> 
> Signed-off-by: Shuicheng Lin <shuicheng.lin@intel.com>
> Reviewed-by: Philipp Stanner <pstanner@redhat.com>
> ---
>  drivers/gpu/drm/scheduler/sched_entity.c | 10 ++++------
>  drivers/gpu/drm/scheduler/sched_main.c   |  4 ++--
>  include/drm/gpu_scheduler.h              | 12 ++++++------
>  include/linux/dma-resv.h                 |  6 +++---
>  4 files changed, 15 insertions(+), 17 deletions(-)
> 
> diff --git a/drivers/gpu/drm/scheduler/sched_entity.c
> b/drivers/gpu/drm/scheduler/sched_entity.c
> index 58c8161289fe..ffa3e765f5db 100644
> --- a/drivers/gpu/drm/scheduler/sched_entity.c
> +++ b/drivers/gpu/drm/scheduler/sched_entity.c
> @@ -51,7 +51,7 @@
>   * drm_sched_entity_set_priority(). For changing the set of schedulers
>   * @sched_list at runtime see drm_sched_entity_modify_sched().
>   *
> - * An entity is cleaned up by callind drm_sched_entity_fini(). See also
> + * An entity is cleaned up by calling drm_sched_entity_fini(). See also
>   * drm_sched_entity_destroy().
>   *
>   * Returns 0 on success or a negative error code on failure.
> @@ -370,8 +370,8 @@ static void drm_sched_entity_clear_dep(struct
> dma_fence *f,  }
> 
>  /*
> - * drm_sched_entity_clear_dep - callback to clear the entities dependency and
> - * wake up scheduler
> + * drm_sched_entity_wakeup - callback to clear the entity's dependency
> + and
> + * wake up the scheduler
>   */
>  static void drm_sched_entity_wakeup(struct dma_fence *f,
>  				    struct dma_fence_cb *cb)
> @@ -389,7 +389,7 @@ static void drm_sched_entity_wakeup(struct dma_fence
> *f,
>   * @entity: scheduler entity
>   * @priority: scheduler priority
>   *
> - * Update the priority of runqueus used for the entity.
> + * Update the priority of runqueues used for the entity.
>   */
>  void drm_sched_entity_set_priority(struct drm_sched_entity *entity,
>  				   enum drm_sched_priority priority) @@ -574,8
> +574,6 @@ void drm_sched_entity_select_rq(struct drm_sched_entity *entity)
>   * fence sequence number this function should be called with
> drm_sched_job_arm()
>   * under common lock for the struct drm_sched_entity that was set up for
>   * @sched_job in drm_sched_job_init().
> - *
> - * Returns 0 for success, negative error code otherwise.
>   */
>  void drm_sched_entity_push_job(struct drm_sched_job *sched_job)  { diff --git
> a/drivers/gpu/drm/scheduler/sched_main.c
> b/drivers/gpu/drm/scheduler/sched_main.c
> index ab53ab486fe6..cadf1662bc01 100644
> --- a/drivers/gpu/drm/scheduler/sched_main.c
> +++ b/drivers/gpu/drm/scheduler/sched_main.c
> @@ -41,7 +41,7 @@
>   * 4. Entities themselves maintain a queue of jobs that will be scheduled on
>   *    the hardware.
>   *
> - * The jobs in a entity are always scheduled in the order that they were pushed.
> + * The jobs in an entity are always scheduled in the order in which they were
> pushed.
>   *
>   * Note that once a job was taken from the entities queue and pushed to the
>   * hardware, i.e. the pending queue, the entity must not be referenced anymore
> @@ -1339,7 +1339,7 @@ void drm_sched_fini(struct drm_gpu_scheduler
> *sched)
>  		list_for_each_entry(s_entity, &rq->entities, list)
>  			/*
>  			 * Prevents reinsertion and marks job_queue as idle,
> -			 * it will removed from rq in drm_sched_entity_fini
> +			 * it will be removed from the rq in
> drm_sched_entity_fini()
>  			 * eventually
>  			 */
>  			s_entity->stopped = true;
> diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h index
> fe8edb917360..ef23113451e4 100644
> --- a/include/drm/gpu_scheduler.h
> +++ b/include/drm/gpu_scheduler.h
> @@ -33,11 +33,11 @@
>  #define MAX_WAIT_SCHED_ENTITY_Q_EMPTY msecs_to_jiffies(1000)
> 
>  /**
> - * DRM_SCHED_FENCE_DONT_PIPELINE - Prefent dependency pipelining
> + * DRM_SCHED_FENCE_DONT_PIPELINE - Prevent dependency pipelining
>   *
>   * Setting this flag on a scheduler fence prevents pipelining of jobs depending
>   * on this fence. In other words we always insert a full CPU round trip before
> - * dependen jobs are pushed to the hw queue.
> + * dependent jobs are pushed to the hw queue.
>   */
>  #define DRM_SCHED_FENCE_DONT_PIPELINE	DMA_FENCE_FLAG_USER_BITS
> 
> @@ -71,7 +71,7 @@ enum drm_sched_priority {
>  	DRM_SCHED_PRIORITY_COUNT
>  };
> 
> -/* Used to chose between FIFO and RR jobs scheduling */
> +/* Used to choose between FIFO and RR job-scheduling */
>  extern int drm_sched_policy;
> 
>  #define DRM_SCHED_POLICY_RR    0
> @@ -198,7 +198,7 @@ struct drm_sched_entity {
>  	 *
>  	 * Points to the finished fence of the last scheduled job. Only written
>  	 * by the scheduler thread, can be accessed locklessly from
> -	 * drm_sched_job_arm() iff the queue is empty.
> +	 * drm_sched_job_arm() if the queue is empty.
>  	 */
>  	struct dma_fence __rcu		*last_scheduled;
> 
> @@ -247,7 +247,7 @@ struct drm_sched_entity {
>   * @sched: the scheduler to which this rq belongs to.
>   * @entities: list of the entities to be scheduled.
>   * @current_entity: the entity which is to be scheduled.
> - * @rb_tree_root: root of time based priory queue of entities for FIFO scheduling
> + * @rb_tree_root: root of time based priority queue of entities for
> + FIFO scheduling
>   *
>   * Run queue is a set of entities scheduling command submissions for
>   * one specific ring. It implements the scheduling policy that selects @@ -321,7
> +321,7 @@ struct drm_sched_fence *to_drm_sched_fence(struct dma_fence
> *f);
>   * @s_fence: contains the fences for the scheduling of job.
>   * @finish_cb: the callback for the finished fence.
>   * @credits: the number of credits this job contributes to the scheduler
> - * @work: Helper to reschdeule job kill to different context.
> + * @work: Helper to reschedule job kill to different context.
>   * @id: a unique id assigned to each job scheduled on the scheduler.
>   * @karma: increment on every hang caused by this job. If this exceeds the hang
>   *         limit of the scheduler then the job is marked guilty and will not
> diff --git a/include/linux/dma-resv.h b/include/linux/dma-resv.h index
> 8d0e34dad446..c5ab6fd9ebe8 100644
> --- a/include/linux/dma-resv.h
> +++ b/include/linux/dma-resv.h
> @@ -105,10 +105,10 @@ enum dma_resv_usage {
>  	 * This should be used by submissions which don't want to participate in
>  	 * any implicit synchronization.
>  	 *
> -	 * The most common case are preemption fences, page table updates,
> TLB
> -	 * flushes as well as explicit synced user submissions.
> +	 * The most common cases are preemption fences, page table updates,
> TLB
> +	 * flushes as well as explicitly synced user submissions.
>  	 *
> -	 * Explicit synced user user submissions can be promoted to
> +	 * Explicitly synced user submissions can be promoted to
>  	 * DMA_RESV_USAGE_READ or DMA_RESV_USAGE_WRITE as needed
> using
>  	 * dma_buf_import_sync_file() when implicit synchronization should
>  	 * become necessary after initial adding of the fence.
> --
> 2.25.1
Simona Vetter Sept. 24, 2024, 10:03 a.m. UTC | #2
On Sun, Sep 22, 2024 at 05:29:36PM +0000, Lin, Shuicheng wrote:
> Hi all,
> I am not familiar with the process yet. To get it merged, should I add more mail-list or how to notify the maintainers?
> Thanks in advance for your guide.

drm/sched is a bit undermaintained, things unfortunately fall through
cracks. I've picked this up and merged it to drm-misc-next, thanks a lot.
-Sima

> 
> Best Regards
> Shuicheng
> 
> > -----Original Message-----
> > From: Lin, Shuicheng <shuicheng.lin@intel.com>
> > Sent: Tuesday, September 17, 2024 7:48 AM
> > To: dri-devel@lists.freedesktop.org
> > Cc: Lin, Shuicheng <shuicheng.lin@intel.com>; Philipp Stanner
> > <pstanner@redhat.com>
> > Subject: [PATCH v3] drm/scheduler: Improve documentation
> > 
> > Function drm_sched_entity_push_job() doesn't have a return value, remove the
> > return value description for it.
> > Correct several other typo errors.
> > 
> > v2 (Philipp):
> > - more correction with related comments.
> > 
> > Signed-off-by: Shuicheng Lin <shuicheng.lin@intel.com>
> > Reviewed-by: Philipp Stanner <pstanner@redhat.com>
> > ---
> >  drivers/gpu/drm/scheduler/sched_entity.c | 10 ++++------
> >  drivers/gpu/drm/scheduler/sched_main.c   |  4 ++--
> >  include/drm/gpu_scheduler.h              | 12 ++++++------
> >  include/linux/dma-resv.h                 |  6 +++---
> >  4 files changed, 15 insertions(+), 17 deletions(-)
> > 
> > diff --git a/drivers/gpu/drm/scheduler/sched_entity.c
> > b/drivers/gpu/drm/scheduler/sched_entity.c
> > index 58c8161289fe..ffa3e765f5db 100644
> > --- a/drivers/gpu/drm/scheduler/sched_entity.c
> > +++ b/drivers/gpu/drm/scheduler/sched_entity.c
> > @@ -51,7 +51,7 @@
> >   * drm_sched_entity_set_priority(). For changing the set of schedulers
> >   * @sched_list at runtime see drm_sched_entity_modify_sched().
> >   *
> > - * An entity is cleaned up by callind drm_sched_entity_fini(). See also
> > + * An entity is cleaned up by calling drm_sched_entity_fini(). See also
> >   * drm_sched_entity_destroy().
> >   *
> >   * Returns 0 on success or a negative error code on failure.
> > @@ -370,8 +370,8 @@ static void drm_sched_entity_clear_dep(struct
> > dma_fence *f,  }
> > 
> >  /*
> > - * drm_sched_entity_clear_dep - callback to clear the entities dependency and
> > - * wake up scheduler
> > + * drm_sched_entity_wakeup - callback to clear the entity's dependency
> > + and
> > + * wake up the scheduler
> >   */
> >  static void drm_sched_entity_wakeup(struct dma_fence *f,
> >  				    struct dma_fence_cb *cb)
> > @@ -389,7 +389,7 @@ static void drm_sched_entity_wakeup(struct dma_fence
> > *f,
> >   * @entity: scheduler entity
> >   * @priority: scheduler priority
> >   *
> > - * Update the priority of runqueus used for the entity.
> > + * Update the priority of runqueues used for the entity.
> >   */
> >  void drm_sched_entity_set_priority(struct drm_sched_entity *entity,
> >  				   enum drm_sched_priority priority) @@ -574,8
> > +574,6 @@ void drm_sched_entity_select_rq(struct drm_sched_entity *entity)
> >   * fence sequence number this function should be called with
> > drm_sched_job_arm()
> >   * under common lock for the struct drm_sched_entity that was set up for
> >   * @sched_job in drm_sched_job_init().
> > - *
> > - * Returns 0 for success, negative error code otherwise.
> >   */
> >  void drm_sched_entity_push_job(struct drm_sched_job *sched_job)  { diff --git
> > a/drivers/gpu/drm/scheduler/sched_main.c
> > b/drivers/gpu/drm/scheduler/sched_main.c
> > index ab53ab486fe6..cadf1662bc01 100644
> > --- a/drivers/gpu/drm/scheduler/sched_main.c
> > +++ b/drivers/gpu/drm/scheduler/sched_main.c
> > @@ -41,7 +41,7 @@
> >   * 4. Entities themselves maintain a queue of jobs that will be scheduled on
> >   *    the hardware.
> >   *
> > - * The jobs in a entity are always scheduled in the order that they were pushed.
> > + * The jobs in an entity are always scheduled in the order in which they were
> > pushed.
> >   *
> >   * Note that once a job was taken from the entities queue and pushed to the
> >   * hardware, i.e. the pending queue, the entity must not be referenced anymore
> > @@ -1339,7 +1339,7 @@ void drm_sched_fini(struct drm_gpu_scheduler
> > *sched)
> >  		list_for_each_entry(s_entity, &rq->entities, list)
> >  			/*
> >  			 * Prevents reinsertion and marks job_queue as idle,
> > -			 * it will removed from rq in drm_sched_entity_fini
> > +			 * it will be removed from the rq in
> > drm_sched_entity_fini()
> >  			 * eventually
> >  			 */
> >  			s_entity->stopped = true;
> > diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h index
> > fe8edb917360..ef23113451e4 100644
> > --- a/include/drm/gpu_scheduler.h
> > +++ b/include/drm/gpu_scheduler.h
> > @@ -33,11 +33,11 @@
> >  #define MAX_WAIT_SCHED_ENTITY_Q_EMPTY msecs_to_jiffies(1000)
> > 
> >  /**
> > - * DRM_SCHED_FENCE_DONT_PIPELINE - Prefent dependency pipelining
> > + * DRM_SCHED_FENCE_DONT_PIPELINE - Prevent dependency pipelining
> >   *
> >   * Setting this flag on a scheduler fence prevents pipelining of jobs depending
> >   * on this fence. In other words we always insert a full CPU round trip before
> > - * dependen jobs are pushed to the hw queue.
> > + * dependent jobs are pushed to the hw queue.
> >   */
> >  #define DRM_SCHED_FENCE_DONT_PIPELINE	DMA_FENCE_FLAG_USER_BITS
> > 
> > @@ -71,7 +71,7 @@ enum drm_sched_priority {
> >  	DRM_SCHED_PRIORITY_COUNT
> >  };
> > 
> > -/* Used to chose between FIFO and RR jobs scheduling */
> > +/* Used to choose between FIFO and RR job-scheduling */
> >  extern int drm_sched_policy;
> > 
> >  #define DRM_SCHED_POLICY_RR    0
> > @@ -198,7 +198,7 @@ struct drm_sched_entity {
> >  	 *
> >  	 * Points to the finished fence of the last scheduled job. Only written
> >  	 * by the scheduler thread, can be accessed locklessly from
> > -	 * drm_sched_job_arm() iff the queue is empty.
> > +	 * drm_sched_job_arm() if the queue is empty.
> >  	 */
> >  	struct dma_fence __rcu		*last_scheduled;
> > 
> > @@ -247,7 +247,7 @@ struct drm_sched_entity {
> >   * @sched: the scheduler to which this rq belongs to.
> >   * @entities: list of the entities to be scheduled.
> >   * @current_entity: the entity which is to be scheduled.
> > - * @rb_tree_root: root of time based priory queue of entities for FIFO scheduling
> > + * @rb_tree_root: root of time based priority queue of entities for
> > + FIFO scheduling
> >   *
> >   * Run queue is a set of entities scheduling command submissions for
> >   * one specific ring. It implements the scheduling policy that selects @@ -321,7
> > +321,7 @@ struct drm_sched_fence *to_drm_sched_fence(struct dma_fence
> > *f);
> >   * @s_fence: contains the fences for the scheduling of job.
> >   * @finish_cb: the callback for the finished fence.
> >   * @credits: the number of credits this job contributes to the scheduler
> > - * @work: Helper to reschdeule job kill to different context.
> > + * @work: Helper to reschedule job kill to different context.
> >   * @id: a unique id assigned to each job scheduled on the scheduler.
> >   * @karma: increment on every hang caused by this job. If this exceeds the hang
> >   *         limit of the scheduler then the job is marked guilty and will not
> > diff --git a/include/linux/dma-resv.h b/include/linux/dma-resv.h index
> > 8d0e34dad446..c5ab6fd9ebe8 100644
> > --- a/include/linux/dma-resv.h
> > +++ b/include/linux/dma-resv.h
> > @@ -105,10 +105,10 @@ enum dma_resv_usage {
> >  	 * This should be used by submissions which don't want to participate in
> >  	 * any implicit synchronization.
> >  	 *
> > -	 * The most common case are preemption fences, page table updates,
> > TLB
> > -	 * flushes as well as explicit synced user submissions.
> > +	 * The most common cases are preemption fences, page table updates,
> > TLB
> > +	 * flushes as well as explicitly synced user submissions.
> >  	 *
> > -	 * Explicit synced user user submissions can be promoted to
> > +	 * Explicitly synced user submissions can be promoted to
> >  	 * DMA_RESV_USAGE_READ or DMA_RESV_USAGE_WRITE as needed
> > using
> >  	 * dma_buf_import_sync_file() when implicit synchronization should
> >  	 * become necessary after initial adding of the fence.
> > --
> > 2.25.1
>
Philipp Stanner Sept. 26, 2024, 8:20 a.m. UTC | #3
On Tue, 2024-09-24 at 12:03 +0200, Simona Vetter wrote:
> On Sun, Sep 22, 2024 at 05:29:36PM +0000, Lin, Shuicheng wrote:
> > Hi all,
> > I am not familiar with the process yet. To get it merged, should I
> > add more mail-list or how to notify the maintainers?
> > Thanks in advance for your guide.
> 
> drm/sched is a bit undermaintained, things unfortunately fall through
> cracks. I've picked this up and merged it to drm-misc-next, thanks a
> lot.
> -Sima

Thx!

Feel free to ping Danilo and me in the future. We might be unavailable
at times individually, but in generally will take care of that.

P.

> 
> > 
> > Best Regards
> > Shuicheng
> > 
> > > -----Original Message-----
> > > From: Lin, Shuicheng <shuicheng.lin@intel.com>
> > > Sent: Tuesday, September 17, 2024 7:48 AM
> > > To: dri-devel@lists.freedesktop.org
> > > Cc: Lin, Shuicheng <shuicheng.lin@intel.com>; Philipp Stanner
> > > <pstanner@redhat.com>
> > > Subject: [PATCH v3] drm/scheduler: Improve documentation
> > > 
> > > Function drm_sched_entity_push_job() doesn't have a return value,
> > > remove the
> > > return value description for it.
> > > Correct several other typo errors.
> > > 
> > > v2 (Philipp):
> > > - more correction with related comments.
> > > 
> > > Signed-off-by: Shuicheng Lin <shuicheng.lin@intel.com>
> > > Reviewed-by: Philipp Stanner <pstanner@redhat.com>
> > > ---
> > >  drivers/gpu/drm/scheduler/sched_entity.c | 10 ++++------
> > >  drivers/gpu/drm/scheduler/sched_main.c   |  4 ++--
> > >  include/drm/gpu_scheduler.h              | 12 ++++++------
> > >  include/linux/dma-resv.h                 |  6 +++---
> > >  4 files changed, 15 insertions(+), 17 deletions(-)
> > > 
> > > diff --git a/drivers/gpu/drm/scheduler/sched_entity.c
> > > b/drivers/gpu/drm/scheduler/sched_entity.c
> > > index 58c8161289fe..ffa3e765f5db 100644
> > > --- a/drivers/gpu/drm/scheduler/sched_entity.c
> > > +++ b/drivers/gpu/drm/scheduler/sched_entity.c
> > > @@ -51,7 +51,7 @@
> > >   * drm_sched_entity_set_priority(). For changing the set of
> > > schedulers
> > >   * @sched_list at runtime see drm_sched_entity_modify_sched().
> > >   *
> > > - * An entity is cleaned up by callind drm_sched_entity_fini().
> > > See also
> > > + * An entity is cleaned up by calling drm_sched_entity_fini().
> > > See also
> > >   * drm_sched_entity_destroy().
> > >   *
> > >   * Returns 0 on success or a negative error code on failure.
> > > @@ -370,8 +370,8 @@ static void drm_sched_entity_clear_dep(struct
> > > dma_fence *f,  }
> > > 
> > >  /*
> > > - * drm_sched_entity_clear_dep - callback to clear the entities
> > > dependency and
> > > - * wake up scheduler
> > > + * drm_sched_entity_wakeup - callback to clear the entity's
> > > dependency
> > > + and
> > > + * wake up the scheduler
> > >   */
> > >  static void drm_sched_entity_wakeup(struct dma_fence *f,
> > >  				    struct dma_fence_cb *cb)
> > > @@ -389,7 +389,7 @@ static void drm_sched_entity_wakeup(struct
> > > dma_fence
> > > *f,
> > >   * @entity: scheduler entity
> > >   * @priority: scheduler priority
> > >   *
> > > - * Update the priority of runqueus used for the entity.
> > > + * Update the priority of runqueues used for the entity.
> > >   */
> > >  void drm_sched_entity_set_priority(struct drm_sched_entity
> > > *entity,
> > >  				   enum drm_sched_priority
> > > priority) @@ -574,8
> > > +574,6 @@ void drm_sched_entity_select_rq(struct drm_sched_entity
> > > *entity)
> > >   * fence sequence number this function should be called with
> > > drm_sched_job_arm()
> > >   * under common lock for the struct drm_sched_entity that was
> > > set up for
> > >   * @sched_job in drm_sched_job_init().
> > > - *
> > > - * Returns 0 for success, negative error code otherwise.
> > >   */
> > >  void drm_sched_entity_push_job(struct drm_sched_job *sched_job) 
> > > { diff --git
> > > a/drivers/gpu/drm/scheduler/sched_main.c
> > > b/drivers/gpu/drm/scheduler/sched_main.c
> > > index ab53ab486fe6..cadf1662bc01 100644
> > > --- a/drivers/gpu/drm/scheduler/sched_main.c
> > > +++ b/drivers/gpu/drm/scheduler/sched_main.c
> > > @@ -41,7 +41,7 @@
> > >   * 4. Entities themselves maintain a queue of jobs that will be
> > > scheduled on
> > >   *    the hardware.
> > >   *
> > > - * The jobs in a entity are always scheduled in the order that
> > > they were pushed.
> > > + * The jobs in an entity are always scheduled in the order in
> > > which they were
> > > pushed.
> > >   *
> > >   * Note that once a job was taken from the entities queue and
> > > pushed to the
> > >   * hardware, i.e. the pending queue, the entity must not be
> > > referenced anymore
> > > @@ -1339,7 +1339,7 @@ void drm_sched_fini(struct
> > > drm_gpu_scheduler
> > > *sched)
> > >  		list_for_each_entry(s_entity, &rq->entities,
> > > list)
> > >  			/*
> > >  			 * Prevents reinsertion and marks
> > > job_queue as idle,
> > > -			 * it will removed from rq in
> > > drm_sched_entity_fini
> > > +			 * it will be removed from the rq in
> > > drm_sched_entity_fini()
> > >  			 * eventually
> > >  			 */
> > >  			s_entity->stopped = true;
> > > diff --git a/include/drm/gpu_scheduler.h
> > > b/include/drm/gpu_scheduler.h index
> > > fe8edb917360..ef23113451e4 100644
> > > --- a/include/drm/gpu_scheduler.h
> > > +++ b/include/drm/gpu_scheduler.h
> > > @@ -33,11 +33,11 @@
> > >  #define MAX_WAIT_SCHED_ENTITY_Q_EMPTY msecs_to_jiffies(1000)
> > > 
> > >  /**
> > > - * DRM_SCHED_FENCE_DONT_PIPELINE - Prefent dependency pipelining
> > > + * DRM_SCHED_FENCE_DONT_PIPELINE - Prevent dependency pipelining
> > >   *
> > >   * Setting this flag on a scheduler fence prevents pipelining of
> > > jobs depending
> > >   * on this fence. In other words we always insert a full CPU
> > > round trip before
> > > - * dependen jobs are pushed to the hw queue.
> > > + * dependent jobs are pushed to the hw queue.
> > >   */
> > >  #define DRM_SCHED_FENCE_DONT_PIPELINE	DMA_FENCE_FLAG_USER_BITS
> > > 
> > > @@ -71,7 +71,7 @@ enum drm_sched_priority {
> > >  	DRM_SCHED_PRIORITY_COUNT
> > >  };
> > > 
> > > -/* Used to chose between FIFO and RR jobs scheduling */
> > > +/* Used to choose between FIFO and RR job-scheduling */
> > >  extern int drm_sched_policy;
> > > 
> > >  #define DRM_SCHED_POLICY_RR    0
> > > @@ -198,7 +198,7 @@ struct drm_sched_entity {
> > >  	 *
> > >  	 * Points to the finished fence of the last scheduled
> > > job. Only written
> > >  	 * by the scheduler thread, can be accessed locklessly
> > > from
> > > -	 * drm_sched_job_arm() iff the queue is empty.
> > > +	 * drm_sched_job_arm() if the queue is empty.
> > >  	 */
> > >  	struct dma_fence __rcu		*last_scheduled;
> > > 
> > > @@ -247,7 +247,7 @@ struct drm_sched_entity {
> > >   * @sched: the scheduler to which this rq belongs to.
> > >   * @entities: list of the entities to be scheduled.
> > >   * @current_entity: the entity which is to be scheduled.
> > > - * @rb_tree_root: root of time based priory queue of entities
> > > for FIFO scheduling
> > > + * @rb_tree_root: root of time based priority queue of entities
> > > for
> > > + FIFO scheduling
> > >   *
> > >   * Run queue is a set of entities scheduling command submissions
> > > for
> > >   * one specific ring. It implements the scheduling policy that
> > > selects @@ -321,7
> > > +321,7 @@ struct drm_sched_fence *to_drm_sched_fence(struct
> > > dma_fence
> > > *f);
> > >   * @s_fence: contains the fences for the scheduling of job.
> > >   * @finish_cb: the callback for the finished fence.
> > >   * @credits: the number of credits this job contributes to the
> > > scheduler
> > > - * @work: Helper to reschdeule job kill to different context.
> > > + * @work: Helper to reschedule job kill to different context.
> > >   * @id: a unique id assigned to each job scheduled on the
> > > scheduler.
> > >   * @karma: increment on every hang caused by this job. If this
> > > exceeds the hang
> > >   *         limit of the scheduler then the job is marked guilty
> > > and will not
> > > diff --git a/include/linux/dma-resv.h b/include/linux/dma-resv.h
> > > index
> > > 8d0e34dad446..c5ab6fd9ebe8 100644
> > > --- a/include/linux/dma-resv.h
> > > +++ b/include/linux/dma-resv.h
> > > @@ -105,10 +105,10 @@ enum dma_resv_usage {
> > >  	 * This should be used by submissions which don't want
> > > to participate in
> > >  	 * any implicit synchronization.
> > >  	 *
> > > -	 * The most common case are preemption fences, page
> > > table updates,
> > > TLB
> > > -	 * flushes as well as explicit synced user submissions.
> > > +	 * The most common cases are preemption fences, page
> > > table updates,
> > > TLB
> > > +	 * flushes as well as explicitly synced user
> > > submissions.
> > >  	 *
> > > -	 * Explicit synced user user submissions can be promoted
> > > to
> > > +	 * Explicitly synced user submissions can be promoted to
> > >  	 * DMA_RESV_USAGE_READ or DMA_RESV_USAGE_WRITE as needed
> > > using
> > >  	 * dma_buf_import_sync_file() when implicit
> > > synchronization should
> > >  	 * become necessary after initial adding of the fence.
> > > --
> > > 2.25.1
> > 
>
diff mbox series

Patch

diff --git a/drivers/gpu/drm/scheduler/sched_entity.c b/drivers/gpu/drm/scheduler/sched_entity.c
index 58c8161289fe..ffa3e765f5db 100644
--- a/drivers/gpu/drm/scheduler/sched_entity.c
+++ b/drivers/gpu/drm/scheduler/sched_entity.c
@@ -51,7 +51,7 @@ 
  * drm_sched_entity_set_priority(). For changing the set of schedulers
  * @sched_list at runtime see drm_sched_entity_modify_sched().
  *
- * An entity is cleaned up by callind drm_sched_entity_fini(). See also
+ * An entity is cleaned up by calling drm_sched_entity_fini(). See also
  * drm_sched_entity_destroy().
  *
  * Returns 0 on success or a negative error code on failure.
@@ -370,8 +370,8 @@  static void drm_sched_entity_clear_dep(struct dma_fence *f,
 }
 
 /*
- * drm_sched_entity_clear_dep - callback to clear the entities dependency and
- * wake up scheduler
+ * drm_sched_entity_wakeup - callback to clear the entity's dependency and
+ * wake up the scheduler
  */
 static void drm_sched_entity_wakeup(struct dma_fence *f,
 				    struct dma_fence_cb *cb)
@@ -389,7 +389,7 @@  static void drm_sched_entity_wakeup(struct dma_fence *f,
  * @entity: scheduler entity
  * @priority: scheduler priority
  *
- * Update the priority of runqueus used for the entity.
+ * Update the priority of runqueues used for the entity.
  */
 void drm_sched_entity_set_priority(struct drm_sched_entity *entity,
 				   enum drm_sched_priority priority)
@@ -574,8 +574,6 @@  void drm_sched_entity_select_rq(struct drm_sched_entity *entity)
  * fence sequence number this function should be called with drm_sched_job_arm()
  * under common lock for the struct drm_sched_entity that was set up for
  * @sched_job in drm_sched_job_init().
- *
- * Returns 0 for success, negative error code otherwise.
  */
 void drm_sched_entity_push_job(struct drm_sched_job *sched_job)
 {
diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c
index ab53ab486fe6..cadf1662bc01 100644
--- a/drivers/gpu/drm/scheduler/sched_main.c
+++ b/drivers/gpu/drm/scheduler/sched_main.c
@@ -41,7 +41,7 @@ 
  * 4. Entities themselves maintain a queue of jobs that will be scheduled on
  *    the hardware.
  *
- * The jobs in a entity are always scheduled in the order that they were pushed.
+ * The jobs in an entity are always scheduled in the order in which they were pushed.
  *
  * Note that once a job was taken from the entities queue and pushed to the
  * hardware, i.e. the pending queue, the entity must not be referenced anymore
@@ -1339,7 +1339,7 @@  void drm_sched_fini(struct drm_gpu_scheduler *sched)
 		list_for_each_entry(s_entity, &rq->entities, list)
 			/*
 			 * Prevents reinsertion and marks job_queue as idle,
-			 * it will removed from rq in drm_sched_entity_fini
+			 * it will be removed from the rq in drm_sched_entity_fini()
 			 * eventually
 			 */
 			s_entity->stopped = true;
diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h
index fe8edb917360..ef23113451e4 100644
--- a/include/drm/gpu_scheduler.h
+++ b/include/drm/gpu_scheduler.h
@@ -33,11 +33,11 @@ 
 #define MAX_WAIT_SCHED_ENTITY_Q_EMPTY msecs_to_jiffies(1000)
 
 /**
- * DRM_SCHED_FENCE_DONT_PIPELINE - Prefent dependency pipelining
+ * DRM_SCHED_FENCE_DONT_PIPELINE - Prevent dependency pipelining
  *
  * Setting this flag on a scheduler fence prevents pipelining of jobs depending
  * on this fence. In other words we always insert a full CPU round trip before
- * dependen jobs are pushed to the hw queue.
+ * dependent jobs are pushed to the hw queue.
  */
 #define DRM_SCHED_FENCE_DONT_PIPELINE	DMA_FENCE_FLAG_USER_BITS
 
@@ -71,7 +71,7 @@  enum drm_sched_priority {
 	DRM_SCHED_PRIORITY_COUNT
 };
 
-/* Used to chose between FIFO and RR jobs scheduling */
+/* Used to choose between FIFO and RR job-scheduling */
 extern int drm_sched_policy;
 
 #define DRM_SCHED_POLICY_RR    0
@@ -198,7 +198,7 @@  struct drm_sched_entity {
 	 *
 	 * Points to the finished fence of the last scheduled job. Only written
 	 * by the scheduler thread, can be accessed locklessly from
-	 * drm_sched_job_arm() iff the queue is empty.
+	 * drm_sched_job_arm() if the queue is empty.
 	 */
 	struct dma_fence __rcu		*last_scheduled;
 
@@ -247,7 +247,7 @@  struct drm_sched_entity {
  * @sched: the scheduler to which this rq belongs to.
  * @entities: list of the entities to be scheduled.
  * @current_entity: the entity which is to be scheduled.
- * @rb_tree_root: root of time based priory queue of entities for FIFO scheduling
+ * @rb_tree_root: root of time based priority queue of entities for FIFO scheduling
  *
  * Run queue is a set of entities scheduling command submissions for
  * one specific ring. It implements the scheduling policy that selects
@@ -321,7 +321,7 @@  struct drm_sched_fence *to_drm_sched_fence(struct dma_fence *f);
  * @s_fence: contains the fences for the scheduling of job.
  * @finish_cb: the callback for the finished fence.
  * @credits: the number of credits this job contributes to the scheduler
- * @work: Helper to reschdeule job kill to different context.
+ * @work: Helper to reschedule job kill to different context.
  * @id: a unique id assigned to each job scheduled on the scheduler.
  * @karma: increment on every hang caused by this job. If this exceeds the hang
  *         limit of the scheduler then the job is marked guilty and will not
diff --git a/include/linux/dma-resv.h b/include/linux/dma-resv.h
index 8d0e34dad446..c5ab6fd9ebe8 100644
--- a/include/linux/dma-resv.h
+++ b/include/linux/dma-resv.h
@@ -105,10 +105,10 @@  enum dma_resv_usage {
 	 * This should be used by submissions which don't want to participate in
 	 * any implicit synchronization.
 	 *
-	 * The most common case are preemption fences, page table updates, TLB
-	 * flushes as well as explicit synced user submissions.
+	 * The most common cases are preemption fences, page table updates, TLB
+	 * flushes as well as explicitly synced user submissions.
 	 *
-	 * Explicit synced user user submissions can be promoted to
+	 * Explicitly synced user submissions can be promoted to
 	 * DMA_RESV_USAGE_READ or DMA_RESV_USAGE_WRITE as needed using
 	 * dma_buf_import_sync_file() when implicit synchronization should
 	 * become necessary after initial adding of the fence.