Message ID | 20230920144343.64830-4-dakr@redhat.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | DRM GPUVA Manager GPU-VM features | expand |
On Thu, 21 Sept 2023 at 00:44, Danilo Krummrich <dakr@redhat.com> wrote: > > Rename struct drm_gpuvm within struct nouveau_uvmm from 'umgr' to base. > > Signed-off-by: Danilo Krummrich <dakr@redhat.com> Reviewed-by: Dave Airlie <airlied@redhat.com> > --- > drivers/gpu/drm/nouveau/nouveau_debugfs.c | 2 +- > drivers/gpu/drm/nouveau/nouveau_exec.c | 4 +-- > drivers/gpu/drm/nouveau/nouveau_uvmm.c | 32 +++++++++++------------ > drivers/gpu/drm/nouveau/nouveau_uvmm.h | 6 ++--- > 4 files changed, 22 insertions(+), 22 deletions(-) > > diff --git a/drivers/gpu/drm/nouveau/nouveau_debugfs.c b/drivers/gpu/drm/nouveau/nouveau_debugfs.c > index 053f703f2f68..e83db051e851 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_debugfs.c > +++ b/drivers/gpu/drm/nouveau/nouveau_debugfs.c > @@ -231,7 +231,7 @@ nouveau_debugfs_gpuva(struct seq_file *m, void *data) > continue; > > nouveau_uvmm_lock(uvmm); > - drm_debugfs_gpuva_info(m, &uvmm->umgr); > + drm_debugfs_gpuva_info(m, &uvmm->base); > seq_puts(m, "\n"); > nouveau_debugfs_gpuva_regions(m, uvmm); > nouveau_uvmm_unlock(uvmm); > diff --git a/drivers/gpu/drm/nouveau/nouveau_exec.c b/drivers/gpu/drm/nouveau/nouveau_exec.c > index c001952cd678..b4239af29e5a 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_exec.c > +++ b/drivers/gpu/drm/nouveau/nouveau_exec.c > @@ -106,8 +106,8 @@ nouveau_exec_job_submit(struct nouveau_job *job) > drm_exec_until_all_locked(exec) { > struct drm_gpuva *va; > > - drm_gpuvm_for_each_va(va, &uvmm->umgr) { > - if (unlikely(va == &uvmm->umgr.kernel_alloc_node)) > + drm_gpuvm_for_each_va(va, &uvmm->base) { > + if (unlikely(va == &uvmm->base.kernel_alloc_node)) > continue; > > ret = drm_exec_prepare_obj(exec, va->gem.obj, 1); > diff --git a/drivers/gpu/drm/nouveau/nouveau_uvmm.c b/drivers/gpu/drm/nouveau/nouveau_uvmm.c > index c750072cb268..6c86b64273c3 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_uvmm.c > +++ b/drivers/gpu/drm/nouveau/nouveau_uvmm.c > @@ -329,7 +329,7 @@ nouveau_uvma_region_create(struct nouveau_uvmm *uvmm, > struct nouveau_uvma_region *reg; > int ret; > > - if (!drm_gpuva_interval_empty(&uvmm->umgr, addr, range)) > + if (!drm_gpuva_interval_empty(&uvmm->base, addr, range)) > return -ENOSPC; > > ret = nouveau_uvma_region_alloc(®); > @@ -384,7 +384,7 @@ nouveau_uvma_region_empty(struct nouveau_uvma_region *reg) > { > struct nouveau_uvmm *uvmm = reg->uvmm; > > - return drm_gpuva_interval_empty(&uvmm->umgr, > + return drm_gpuva_interval_empty(&uvmm->base, > reg->va.addr, > reg->va.range); > } > @@ -589,7 +589,7 @@ op_map_prepare(struct nouveau_uvmm *uvmm, > uvma->region = args->region; > uvma->kind = args->kind; > > - drm_gpuva_map(&uvmm->umgr, &uvma->va, op); > + drm_gpuva_map(&uvmm->base, &uvma->va, op); > > /* Keep a reference until this uvma is destroyed. */ > nouveau_uvma_gem_get(uvma); > @@ -1194,7 +1194,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) > goto unwind_continue; > } > > - op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->umgr, > + op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->base, > op->va.addr, > op->va.range); > if (IS_ERR(op->ops)) { > @@ -1205,7 +1205,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) > ret = nouveau_uvmm_sm_unmap_prepare(uvmm, &op->new, > op->ops); > if (ret) { > - drm_gpuva_ops_free(&uvmm->umgr, op->ops); > + drm_gpuva_ops_free(&uvmm->base, op->ops); > op->ops = NULL; > op->reg = NULL; > goto unwind_continue; > @@ -1240,7 +1240,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) > } > } > > - op->ops = drm_gpuvm_sm_map_ops_create(&uvmm->umgr, > + op->ops = drm_gpuvm_sm_map_ops_create(&uvmm->base, > op->va.addr, > op->va.range, > op->gem.obj, > @@ -1256,7 +1256,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) > op->va.range, > op->flags & 0xff); > if (ret) { > - drm_gpuva_ops_free(&uvmm->umgr, op->ops); > + drm_gpuva_ops_free(&uvmm->base, op->ops); > op->ops = NULL; > goto unwind_continue; > } > @@ -1264,7 +1264,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) > break; > } > case OP_UNMAP: > - op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->umgr, > + op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->base, > op->va.addr, > op->va.range); > if (IS_ERR(op->ops)) { > @@ -1275,7 +1275,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) > ret = nouveau_uvmm_sm_unmap_prepare(uvmm, &op->new, > op->ops); > if (ret) { > - drm_gpuva_ops_free(&uvmm->umgr, op->ops); > + drm_gpuva_ops_free(&uvmm->base, op->ops); > op->ops = NULL; > goto unwind_continue; > } > @@ -1404,7 +1404,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) > break; > } > > - drm_gpuva_ops_free(&uvmm->umgr, op->ops); > + drm_gpuva_ops_free(&uvmm->base, op->ops); > op->ops = NULL; > op->reg = NULL; > } > @@ -1509,7 +1509,7 @@ nouveau_uvmm_bind_job_free_work_fn(struct work_struct *work) > } > > if (!IS_ERR_OR_NULL(op->ops)) > - drm_gpuva_ops_free(&uvmm->umgr, op->ops); > + drm_gpuva_ops_free(&uvmm->base, op->ops); > > if (obj) > drm_gem_object_put(obj); > @@ -1836,7 +1836,7 @@ nouveau_uvmm_init(struct nouveau_uvmm *uvmm, struct nouveau_cli *cli, > uvmm->kernel_managed_addr = kernel_managed_addr; > uvmm->kernel_managed_size = kernel_managed_size; > > - drm_gpuvm_init(&uvmm->umgr, cli->name, > + drm_gpuvm_init(&uvmm->base, cli->name, > NOUVEAU_VA_SPACE_START, > NOUVEAU_VA_SPACE_END, > kernel_managed_addr, kernel_managed_size, > @@ -1855,7 +1855,7 @@ nouveau_uvmm_init(struct nouveau_uvmm *uvmm, struct nouveau_cli *cli, > return 0; > > out_free_gpuva_mgr: > - drm_gpuvm_destroy(&uvmm->umgr); > + drm_gpuvm_destroy(&uvmm->base); > out_unlock: > mutex_unlock(&cli->mutex); > return ret; > @@ -1877,11 +1877,11 @@ nouveau_uvmm_fini(struct nouveau_uvmm *uvmm) > wait_event(entity->job.wq, list_empty(&entity->job.list.head)); > > nouveau_uvmm_lock(uvmm); > - drm_gpuvm_for_each_va_safe(va, next, &uvmm->umgr) { > + drm_gpuvm_for_each_va_safe(va, next, &uvmm->base) { > struct nouveau_uvma *uvma = uvma_from_va(va); > struct drm_gem_object *obj = va->gem.obj; > > - if (unlikely(va == &uvmm->umgr.kernel_alloc_node)) > + if (unlikely(va == &uvmm->base.kernel_alloc_node)) > continue; > > drm_gpuva_remove(va); > @@ -1910,7 +1910,7 @@ nouveau_uvmm_fini(struct nouveau_uvmm *uvmm) > > mutex_lock(&cli->mutex); > nouveau_vmm_fini(&uvmm->vmm); > - drm_gpuvm_destroy(&uvmm->umgr); > + drm_gpuvm_destroy(&uvmm->base); > mutex_unlock(&cli->mutex); > > dma_resv_fini(&uvmm->resv); > diff --git a/drivers/gpu/drm/nouveau/nouveau_uvmm.h b/drivers/gpu/drm/nouveau/nouveau_uvmm.h > index e96c9919d1bd..a308c59760a5 100644 > --- a/drivers/gpu/drm/nouveau/nouveau_uvmm.h > +++ b/drivers/gpu/drm/nouveau/nouveau_uvmm.h > @@ -8,8 +8,8 @@ > #include "nouveau_drv.h" > > struct nouveau_uvmm { > + struct drm_gpuvm base; > struct nouveau_vmm vmm; > - struct drm_gpuvm umgr; > struct maple_tree region_mt; > struct mutex mutex; > struct dma_resv resv; > @@ -41,10 +41,10 @@ struct nouveau_uvma { > u8 kind; > }; > > -#define uvmm_from_mgr(x) container_of((x), struct nouveau_uvmm, umgr) > +#define uvmm_from_gpuvm(x) container_of((x), struct nouveau_uvmm, base) > #define uvma_from_va(x) container_of((x), struct nouveau_uvma, va) > > -#define to_uvmm(x) uvmm_from_mgr((x)->va.vm) > +#define to_uvmm(x) uvmm_from_gpuvm((x)->va.vm) > > struct nouveau_uvmm_bind_job { > struct nouveau_job base; > -- > 2.41.0 >
diff --git a/drivers/gpu/drm/nouveau/nouveau_debugfs.c b/drivers/gpu/drm/nouveau/nouveau_debugfs.c index 053f703f2f68..e83db051e851 100644 --- a/drivers/gpu/drm/nouveau/nouveau_debugfs.c +++ b/drivers/gpu/drm/nouveau/nouveau_debugfs.c @@ -231,7 +231,7 @@ nouveau_debugfs_gpuva(struct seq_file *m, void *data) continue; nouveau_uvmm_lock(uvmm); - drm_debugfs_gpuva_info(m, &uvmm->umgr); + drm_debugfs_gpuva_info(m, &uvmm->base); seq_puts(m, "\n"); nouveau_debugfs_gpuva_regions(m, uvmm); nouveau_uvmm_unlock(uvmm); diff --git a/drivers/gpu/drm/nouveau/nouveau_exec.c b/drivers/gpu/drm/nouveau/nouveau_exec.c index c001952cd678..b4239af29e5a 100644 --- a/drivers/gpu/drm/nouveau/nouveau_exec.c +++ b/drivers/gpu/drm/nouveau/nouveau_exec.c @@ -106,8 +106,8 @@ nouveau_exec_job_submit(struct nouveau_job *job) drm_exec_until_all_locked(exec) { struct drm_gpuva *va; - drm_gpuvm_for_each_va(va, &uvmm->umgr) { - if (unlikely(va == &uvmm->umgr.kernel_alloc_node)) + drm_gpuvm_for_each_va(va, &uvmm->base) { + if (unlikely(va == &uvmm->base.kernel_alloc_node)) continue; ret = drm_exec_prepare_obj(exec, va->gem.obj, 1); diff --git a/drivers/gpu/drm/nouveau/nouveau_uvmm.c b/drivers/gpu/drm/nouveau/nouveau_uvmm.c index c750072cb268..6c86b64273c3 100644 --- a/drivers/gpu/drm/nouveau/nouveau_uvmm.c +++ b/drivers/gpu/drm/nouveau/nouveau_uvmm.c @@ -329,7 +329,7 @@ nouveau_uvma_region_create(struct nouveau_uvmm *uvmm, struct nouveau_uvma_region *reg; int ret; - if (!drm_gpuva_interval_empty(&uvmm->umgr, addr, range)) + if (!drm_gpuva_interval_empty(&uvmm->base, addr, range)) return -ENOSPC; ret = nouveau_uvma_region_alloc(®); @@ -384,7 +384,7 @@ nouveau_uvma_region_empty(struct nouveau_uvma_region *reg) { struct nouveau_uvmm *uvmm = reg->uvmm; - return drm_gpuva_interval_empty(&uvmm->umgr, + return drm_gpuva_interval_empty(&uvmm->base, reg->va.addr, reg->va.range); } @@ -589,7 +589,7 @@ op_map_prepare(struct nouveau_uvmm *uvmm, uvma->region = args->region; uvma->kind = args->kind; - drm_gpuva_map(&uvmm->umgr, &uvma->va, op); + drm_gpuva_map(&uvmm->base, &uvma->va, op); /* Keep a reference until this uvma is destroyed. */ nouveau_uvma_gem_get(uvma); @@ -1194,7 +1194,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) goto unwind_continue; } - op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->umgr, + op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->base, op->va.addr, op->va.range); if (IS_ERR(op->ops)) { @@ -1205,7 +1205,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) ret = nouveau_uvmm_sm_unmap_prepare(uvmm, &op->new, op->ops); if (ret) { - drm_gpuva_ops_free(&uvmm->umgr, op->ops); + drm_gpuva_ops_free(&uvmm->base, op->ops); op->ops = NULL; op->reg = NULL; goto unwind_continue; @@ -1240,7 +1240,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) } } - op->ops = drm_gpuvm_sm_map_ops_create(&uvmm->umgr, + op->ops = drm_gpuvm_sm_map_ops_create(&uvmm->base, op->va.addr, op->va.range, op->gem.obj, @@ -1256,7 +1256,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) op->va.range, op->flags & 0xff); if (ret) { - drm_gpuva_ops_free(&uvmm->umgr, op->ops); + drm_gpuva_ops_free(&uvmm->base, op->ops); op->ops = NULL; goto unwind_continue; } @@ -1264,7 +1264,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) break; } case OP_UNMAP: - op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->umgr, + op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->base, op->va.addr, op->va.range); if (IS_ERR(op->ops)) { @@ -1275,7 +1275,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) ret = nouveau_uvmm_sm_unmap_prepare(uvmm, &op->new, op->ops); if (ret) { - drm_gpuva_ops_free(&uvmm->umgr, op->ops); + drm_gpuva_ops_free(&uvmm->base, op->ops); op->ops = NULL; goto unwind_continue; } @@ -1404,7 +1404,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) break; } - drm_gpuva_ops_free(&uvmm->umgr, op->ops); + drm_gpuva_ops_free(&uvmm->base, op->ops); op->ops = NULL; op->reg = NULL; } @@ -1509,7 +1509,7 @@ nouveau_uvmm_bind_job_free_work_fn(struct work_struct *work) } if (!IS_ERR_OR_NULL(op->ops)) - drm_gpuva_ops_free(&uvmm->umgr, op->ops); + drm_gpuva_ops_free(&uvmm->base, op->ops); if (obj) drm_gem_object_put(obj); @@ -1836,7 +1836,7 @@ nouveau_uvmm_init(struct nouveau_uvmm *uvmm, struct nouveau_cli *cli, uvmm->kernel_managed_addr = kernel_managed_addr; uvmm->kernel_managed_size = kernel_managed_size; - drm_gpuvm_init(&uvmm->umgr, cli->name, + drm_gpuvm_init(&uvmm->base, cli->name, NOUVEAU_VA_SPACE_START, NOUVEAU_VA_SPACE_END, kernel_managed_addr, kernel_managed_size, @@ -1855,7 +1855,7 @@ nouveau_uvmm_init(struct nouveau_uvmm *uvmm, struct nouveau_cli *cli, return 0; out_free_gpuva_mgr: - drm_gpuvm_destroy(&uvmm->umgr); + drm_gpuvm_destroy(&uvmm->base); out_unlock: mutex_unlock(&cli->mutex); return ret; @@ -1877,11 +1877,11 @@ nouveau_uvmm_fini(struct nouveau_uvmm *uvmm) wait_event(entity->job.wq, list_empty(&entity->job.list.head)); nouveau_uvmm_lock(uvmm); - drm_gpuvm_for_each_va_safe(va, next, &uvmm->umgr) { + drm_gpuvm_for_each_va_safe(va, next, &uvmm->base) { struct nouveau_uvma *uvma = uvma_from_va(va); struct drm_gem_object *obj = va->gem.obj; - if (unlikely(va == &uvmm->umgr.kernel_alloc_node)) + if (unlikely(va == &uvmm->base.kernel_alloc_node)) continue; drm_gpuva_remove(va); @@ -1910,7 +1910,7 @@ nouveau_uvmm_fini(struct nouveau_uvmm *uvmm) mutex_lock(&cli->mutex); nouveau_vmm_fini(&uvmm->vmm); - drm_gpuvm_destroy(&uvmm->umgr); + drm_gpuvm_destroy(&uvmm->base); mutex_unlock(&cli->mutex); dma_resv_fini(&uvmm->resv); diff --git a/drivers/gpu/drm/nouveau/nouveau_uvmm.h b/drivers/gpu/drm/nouveau/nouveau_uvmm.h index e96c9919d1bd..a308c59760a5 100644 --- a/drivers/gpu/drm/nouveau/nouveau_uvmm.h +++ b/drivers/gpu/drm/nouveau/nouveau_uvmm.h @@ -8,8 +8,8 @@ #include "nouveau_drv.h" struct nouveau_uvmm { + struct drm_gpuvm base; struct nouveau_vmm vmm; - struct drm_gpuvm umgr; struct maple_tree region_mt; struct mutex mutex; struct dma_resv resv; @@ -41,10 +41,10 @@ struct nouveau_uvma { u8 kind; }; -#define uvmm_from_mgr(x) container_of((x), struct nouveau_uvmm, umgr) +#define uvmm_from_gpuvm(x) container_of((x), struct nouveau_uvmm, base) #define uvma_from_va(x) container_of((x), struct nouveau_uvma, va) -#define to_uvmm(x) uvmm_from_mgr((x)->va.vm) +#define to_uvmm(x) uvmm_from_gpuvm((x)->va.vm) struct nouveau_uvmm_bind_job { struct nouveau_job base;
Rename struct drm_gpuvm within struct nouveau_uvmm from 'umgr' to base. Signed-off-by: Danilo Krummrich <dakr@redhat.com> --- drivers/gpu/drm/nouveau/nouveau_debugfs.c | 2 +- drivers/gpu/drm/nouveau/nouveau_exec.c | 4 +-- drivers/gpu/drm/nouveau/nouveau_uvmm.c | 32 +++++++++++------------ drivers/gpu/drm/nouveau/nouveau_uvmm.h | 6 ++--- 4 files changed, 22 insertions(+), 22 deletions(-)