Message ID | 20211219212500.61432-3-andi.shyti@linux.intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | More preparation for multi gt patches | expand |
On Sun, Dec 19, 2021 at 11:24:56PM +0200, Andi Shyti wrote: > From: Michał Winiarski <michal.winiarski@intel.com> > > GGTT is currently available both through i915->ggtt and gt->ggtt, and we > eventually want to get rid of the i915->ggtt one. > Use to_gt() for all i915->ggtt accesses to help with the future > refactoring. > > Signed-off-by: Michał Winiarski <michal.winiarski@intel.com> > Cc: Michal Wajdeczko <michal.wajdeczko@intel.com> > Signed-off-by: Andi Shyti <andi.shyti@linux.intel.com> > --- > drivers/gpu/drm/i915/gvt/dmabuf.c | 2 +- > drivers/gpu/drm/i915/i915_debugfs.c | 4 ++-- > drivers/gpu/drm/i915/i915_driver.c | 4 ++-- > drivers/gpu/drm/i915/i915_drv.h | 2 +- > drivers/gpu/drm/i915/i915_gem.c | 23 ++++++++++++----------- > drivers/gpu/drm/i915/i915_gem_gtt.c | 6 +++--- > drivers/gpu/drm/i915/i915_getparam.c | 2 +- > drivers/gpu/drm/i915/i915_perf.c | 4 ++-- > 8 files changed, 24 insertions(+), 23 deletions(-) > > diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c > index 8e65cd8258b9..94c3eb1586b0 100644 > --- a/drivers/gpu/drm/i915/gvt/dmabuf.c > +++ b/drivers/gpu/drm/i915/gvt/dmabuf.c > @@ -84,7 +84,7 @@ static int vgpu_gem_get_pages( > kfree(st); > return ret; > } > - gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + > + gtt_entries = (gen8_pte_t __iomem *)to_gt(dev_priv)->ggtt->gsm + > (fb_info->start >> PAGE_SHIFT); > for_each_sg(st->sgl, sg, page_num, i) { > dma_addr_t dma_addr = > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c > index e0e052cdf8b8..6966fe08df92 100644 > --- a/drivers/gpu/drm/i915/i915_debugfs.c > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > @@ -390,9 +390,9 @@ static int i915_swizzle_info(struct seq_file *m, void *data) > intel_wakeref_t wakeref; > > seq_printf(m, "bit6 swizzle for X-tiling = %s\n", > - swizzle_string(dev_priv->ggtt.bit_6_swizzle_x)); > + swizzle_string(to_gt(dev_priv)->ggtt->bit_6_swizzle_x)); > seq_printf(m, "bit6 swizzle for Y-tiling = %s\n", > - swizzle_string(dev_priv->ggtt.bit_6_swizzle_y)); > + swizzle_string(to_gt(dev_priv)->ggtt->bit_6_swizzle_y)); > > if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES) > seq_puts(m, "L-shaped memory detected\n"); > diff --git a/drivers/gpu/drm/i915/i915_driver.c b/drivers/gpu/drm/i915/i915_driver.c > index 60f8cbf24de7..3c984553d86f 100644 > --- a/drivers/gpu/drm/i915/i915_driver.c > +++ b/drivers/gpu/drm/i915/i915_driver.c > @@ -1146,7 +1146,7 @@ static int i915_drm_suspend(struct drm_device *dev) > > /* Must be called before GGTT is suspended. */ > intel_dpt_suspend(dev_priv); > - i915_ggtt_suspend(&dev_priv->ggtt); > + i915_ggtt_suspend(to_gt(dev_priv)->ggtt); > > i915_save_display(dev_priv); > > @@ -1270,7 +1270,7 @@ static int i915_drm_resume(struct drm_device *dev) > if (ret) > drm_err(&dev_priv->drm, "failed to re-enable GGTT\n"); > > - i915_ggtt_resume(&dev_priv->ggtt); > + i915_ggtt_resume(to_gt(dev_priv)->ggtt); > /* Must be called after GGTT is resumed. */ > intel_dpt_resume(dev_priv); > > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > index 471be2716abe..524025790fe0 100644 > --- a/drivers/gpu/drm/i915/i915_drv.h > +++ b/drivers/gpu/drm/i915/i915_drv.h > @@ -1749,7 +1749,7 @@ static inline bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_objec > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > > - return i915->ggtt.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && > + return to_gt(i915)->ggtt->bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && > i915_gem_object_is_tiled(obj); > } > > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > index 8ba2119092f2..45e3b4c540a1 100644 > --- a/drivers/gpu/drm/i915/i915_gem.c > +++ b/drivers/gpu/drm/i915/i915_gem.c > @@ -88,7 +88,8 @@ int > i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, > struct drm_file *file) > { > - struct i915_ggtt *ggtt = &to_i915(dev)->ggtt; > + struct drm_i915_private *i915 = to_i915(dev); > + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; > struct drm_i915_gem_get_aperture *args = data; > struct i915_vma *vma; > u64 pinned; > @@ -289,7 +290,7 @@ static struct i915_vma *i915_gem_gtt_prepare(struct drm_i915_gem_object *obj, > bool write) > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > - struct i915_ggtt *ggtt = &i915->ggtt; > + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; > struct i915_vma *vma; > struct i915_gem_ww_ctx ww; > int ret; > @@ -350,7 +351,7 @@ static void i915_gem_gtt_cleanup(struct drm_i915_gem_object *obj, > struct i915_vma *vma) > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > - struct i915_ggtt *ggtt = &i915->ggtt; > + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; > > i915_gem_object_unpin_pages(obj); > if (drm_mm_node_allocated(node)) { > @@ -366,7 +367,7 @@ i915_gem_gtt_pread(struct drm_i915_gem_object *obj, > const struct drm_i915_gem_pread *args) > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > - struct i915_ggtt *ggtt = &i915->ggtt; > + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; > intel_wakeref_t wakeref; > struct drm_mm_node node; > void __user *user_data; > @@ -522,7 +523,7 @@ i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj, > const struct drm_i915_gem_pwrite *args) > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > - struct i915_ggtt *ggtt = &i915->ggtt; > + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; > struct intel_runtime_pm *rpm = &i915->runtime_pm; > intel_wakeref_t wakeref; > struct drm_mm_node node; > @@ -823,7 +824,7 @@ void i915_gem_runtime_suspend(struct drm_i915_private *i915) > */ > > list_for_each_entry_safe(obj, on, > - &i915->ggtt.userfault_list, userfault_link) > + &to_gt(i915)->ggtt->userfault_list, userfault_link) > __i915_gem_object_release_mmap_gtt(obj); > > /* > @@ -831,8 +832,8 @@ void i915_gem_runtime_suspend(struct drm_i915_private *i915) > * in use by hardware (i.e. they are pinned), we should not be powering > * down! All other fences will be reacquired by the user upon waking. > */ > - for (i = 0; i < i915->ggtt.num_fences; i++) { > - struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i]; > + for (i = 0; i < to_gt(i915)->ggtt->num_fences; i++) { > + struct i915_fence_reg *reg = &to_gt(i915)->ggtt->fence_regs[i]; > > /* > * Ideally we want to assert that the fence register is not > @@ -873,7 +874,7 @@ i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj, > u64 size, u64 alignment, u64 flags) > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > - struct i915_ggtt *ggtt = &i915->ggtt; > + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; > struct i915_vma *vma; > int ret; > > @@ -1101,7 +1102,7 @@ int i915_gem_init(struct drm_i915_private *dev_priv) > > /* Minimal basic recovery for KMS */ > ret = i915_ggtt_enable_hw(dev_priv); > - i915_ggtt_resume(&dev_priv->ggtt); > + i915_ggtt_resume(to_gt(dev_priv)->ggtt); > intel_init_clock_gating(dev_priv); > } > > @@ -1124,7 +1125,7 @@ void i915_gem_driver_unregister(struct drm_i915_private *i915) > > void i915_gem_driver_remove(struct drm_i915_private *dev_priv) > { > - intel_wakeref_auto_fini(&dev_priv->ggtt.userfault_wakeref); > + intel_wakeref_auto_fini(&to_gt(dev_priv)->ggtt->userfault_wakeref); > > i915_gem_suspend_late(dev_priv); > intel_gt_driver_remove(to_gt(dev_priv)); > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c > index cd5f2348a187..2f2ba7a2955d 100644 > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c > @@ -56,7 +56,7 @@ void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj, > struct sg_table *pages) > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > - struct i915_ggtt *ggtt = &i915->ggtt; > + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; > > /* XXX This does not prevent more requests being submitted! */ > if (unlikely(ggtt->do_idle_maps)) > @@ -103,7 +103,7 @@ int i915_gem_gtt_reserve(struct i915_address_space *vm, > GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); > GEM_BUG_ON(!IS_ALIGNED(offset, I915_GTT_MIN_ALIGNMENT)); > GEM_BUG_ON(range_overflows(offset, size, vm->total)); > - GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm); > + GEM_BUG_ON(vm == &to_gt(vm->i915)->ggtt->alias->vm); > GEM_BUG_ON(drm_mm_node_allocated(node)); > > node->size = size; > @@ -201,7 +201,7 @@ int i915_gem_gtt_insert(struct i915_address_space *vm, > GEM_BUG_ON(start >= end); > GEM_BUG_ON(start > 0 && !IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); > GEM_BUG_ON(end < U64_MAX && !IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); > - GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm); > + GEM_BUG_ON(vm == &to_gt(vm->i915)->ggtt->alias->vm); > GEM_BUG_ON(drm_mm_node_allocated(node)); > > if (unlikely(range_overflows(start, size, end))) > diff --git a/drivers/gpu/drm/i915/i915_getparam.c b/drivers/gpu/drm/i915/i915_getparam.c > index 7f80ad247bc8..5b8a2157d797 100644 > --- a/drivers/gpu/drm/i915/i915_getparam.c > +++ b/drivers/gpu/drm/i915/i915_getparam.c > @@ -31,7 +31,7 @@ int i915_getparam_ioctl(struct drm_device *dev, void *data, > value = pdev->revision; > break; > case I915_PARAM_NUM_FENCES_AVAIL: > - value = i915->ggtt.num_fences; > + value = to_gt(i915)->ggtt->num_fences; > break; > case I915_PARAM_HAS_OVERLAY: > value = !!i915->overlay; > diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c > index 170bba913c30..128315aec517 100644 > --- a/drivers/gpu/drm/i915/i915_perf.c > +++ b/drivers/gpu/drm/i915/i915_perf.c > @@ -1630,7 +1630,7 @@ static int alloc_noa_wait(struct i915_perf_stream *stream) > struct drm_i915_gem_object *bo; > struct i915_vma *vma; > const u64 delay_ticks = 0xffffffffffffffff - > - intel_gt_ns_to_clock_interval(stream->perf->i915->ggtt.vm.gt, > + intel_gt_ns_to_clock_interval(to_gt(stream->perf->i915)->ggtt->vm.gt, I'm not too familiar with the perf code, but this looks a bit roundabout since we're ultimately trying to get to a GT...do we even need to go through the ggtt structure here or can we just pass "to_gt(stream->perf->i915)" as the first parameter? > atomic64_read(&stream->perf->noa_programming_delay)); > const u32 base = stream->engine->mmio_base; > #define CS_GPR(x) GEN8_RING_CS_GPR(base, x) > @@ -3542,7 +3542,7 @@ i915_perf_open_ioctl_locked(struct i915_perf *perf, > > static u64 oa_exponent_to_ns(struct i915_perf *perf, int exponent) > { > - return intel_gt_clock_interval_to_ns(perf->i915->ggtt.vm.gt, > + return intel_gt_clock_interval_to_ns(to_gt(perf->i915)->ggtt->vm.gt, Ditto; this looks like "to_gt(perf->i915)" might be all we need? Matt > 2ULL << exponent); > } > > -- > 2.34.1 >
Hi Matt, > > diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c > > index 170bba913c30..128315aec517 100644 > > --- a/drivers/gpu/drm/i915/i915_perf.c > > +++ b/drivers/gpu/drm/i915/i915_perf.c > > @@ -1630,7 +1630,7 @@ static int alloc_noa_wait(struct i915_perf_stream *stream) > > struct drm_i915_gem_object *bo; > > struct i915_vma *vma; > > const u64 delay_ticks = 0xffffffffffffffff - > > - intel_gt_ns_to_clock_interval(stream->perf->i915->ggtt.vm.gt, > > + intel_gt_ns_to_clock_interval(to_gt(stream->perf->i915)->ggtt->vm.gt, > > I'm not too familiar with the perf code, but this looks a bit roundabout > since we're ultimately trying to get to a GT...do we even need to go > through the ggtt structure here or can we just pass > "to_gt(stream->perf->i915)" as the first parameter? > > > atomic64_read(&stream->perf->noa_programming_delay)); > > const u32 base = stream->engine->mmio_base; > > #define CS_GPR(x) GEN8_RING_CS_GPR(base, x) > > @@ -3542,7 +3542,7 @@ i915_perf_open_ioctl_locked(struct i915_perf *perf, > > > > static u64 oa_exponent_to_ns(struct i915_perf *perf, int exponent) > > { > > - return intel_gt_clock_interval_to_ns(perf->i915->ggtt.vm.gt, > > + return intel_gt_clock_interval_to_ns(to_gt(perf->i915)->ggtt->vm.gt, > > Ditto; this looks like "to_gt(perf->i915)" might be all we need? I think this function is looking for the GT coming from the VM, otherwise originally it could have taken it from &i915->gt. In my first version I proposed a wrapper around this but it was rejected by Lucas. Besides, as we discussed earlier when I was proposed the static allocation, the ggtt might not always be linked to the same gt, so that I assumed that sometimes: to_gt(perf->i915)->ggtt->vm.gt != to_gt(perf->i915) if two GTs are sharing the same ggtt, what would the ggtt->vm.gt link be? Thanks, Andi
On Tue, Dec 21, 2021 at 09:46:29PM +0200, Andi Shyti wrote: > Hi Matt, > > > > diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c > > > index 170bba913c30..128315aec517 100644 > > > --- a/drivers/gpu/drm/i915/i915_perf.c > > > +++ b/drivers/gpu/drm/i915/i915_perf.c > > > @@ -1630,7 +1630,7 @@ static int alloc_noa_wait(struct i915_perf_stream *stream) > > > struct drm_i915_gem_object *bo; > > > struct i915_vma *vma; > > > const u64 delay_ticks = 0xffffffffffffffff - > > > - intel_gt_ns_to_clock_interval(stream->perf->i915->ggtt.vm.gt, > > > + intel_gt_ns_to_clock_interval(to_gt(stream->perf->i915)->ggtt->vm.gt, > > > > I'm not too familiar with the perf code, but this looks a bit roundabout > > since we're ultimately trying to get to a GT...do we even need to go > > through the ggtt structure here or can we just pass > > "to_gt(stream->perf->i915)" as the first parameter? > > > > > atomic64_read(&stream->perf->noa_programming_delay)); > > > const u32 base = stream->engine->mmio_base; > > > #define CS_GPR(x) GEN8_RING_CS_GPR(base, x) > > > @@ -3542,7 +3542,7 @@ i915_perf_open_ioctl_locked(struct i915_perf *perf, > > > > > > static u64 oa_exponent_to_ns(struct i915_perf *perf, int exponent) > > > { > > > - return intel_gt_clock_interval_to_ns(perf->i915->ggtt.vm.gt, > > > + return intel_gt_clock_interval_to_ns(to_gt(perf->i915)->ggtt->vm.gt, > > > > Ditto; this looks like "to_gt(perf->i915)" might be all we need? > > I think this function is looking for the GT coming from the VM, > otherwise originally it could have taken it from &i915->gt. In my > first version I proposed a wrapper around this but it was > rejected by Lucas. > > Besides, as we discussed earlier when I was proposed the static > allocation, the ggtt might not always be linked to the same gt, > so that I assumed that sometimes: > > to_gt(perf->i915)->ggtt->vm.gt != to_gt(perf->i915) > > if two GTs are sharing the same ggtt, what would the ggtt->vm.gt > link be? From the git history, it doesn't look like this really needs to care about the GGTT at all; I think it was just unintentionally written in a roundabout manner when intel_gt was first being introduced in the code. The reference here first showed up in commit f170523a7b8e ("drm/i915/gt: Consolidate the CS timestamp clocks"). Actually the most correct thing to do is probably to use 'stream->engine->gt' to ensure we grab the GT actually associated with the stream's engine. Matt > > Thanks, > Andi
On Mon, Jan 03, 2022 at 01:17:10PM -0800, Matt Roper wrote: >On Tue, Dec 21, 2021 at 09:46:29PM +0200, Andi Shyti wrote: >> Hi Matt, >> >> > > diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c >> > > index 170bba913c30..128315aec517 100644 >> > > --- a/drivers/gpu/drm/i915/i915_perf.c >> > > +++ b/drivers/gpu/drm/i915/i915_perf.c >> > > @@ -1630,7 +1630,7 @@ static int alloc_noa_wait(struct i915_perf_stream *stream) >> > > struct drm_i915_gem_object *bo; >> > > struct i915_vma *vma; >> > > const u64 delay_ticks = 0xffffffffffffffff - >> > > - intel_gt_ns_to_clock_interval(stream->perf->i915->ggtt.vm.gt, >> > > + intel_gt_ns_to_clock_interval(to_gt(stream->perf->i915)->ggtt->vm.gt, >> > >> > I'm not too familiar with the perf code, but this looks a bit roundabout >> > since we're ultimately trying to get to a GT...do we even need to go >> > through the ggtt structure here or can we just pass >> > "to_gt(stream->perf->i915)" as the first parameter? >> > >> > > atomic64_read(&stream->perf->noa_programming_delay)); >> > > const u32 base = stream->engine->mmio_base; >> > > #define CS_GPR(x) GEN8_RING_CS_GPR(base, x) >> > > @@ -3542,7 +3542,7 @@ i915_perf_open_ioctl_locked(struct i915_perf *perf, >> > > >> > > static u64 oa_exponent_to_ns(struct i915_perf *perf, int exponent) >> > > { >> > > - return intel_gt_clock_interval_to_ns(perf->i915->ggtt.vm.gt, >> > > + return intel_gt_clock_interval_to_ns(to_gt(perf->i915)->ggtt->vm.gt, >> > >> > Ditto; this looks like "to_gt(perf->i915)" might be all we need? >> >> I think this function is looking for the GT coming from the VM, >> otherwise originally it could have taken it from &i915->gt. In my >> first version I proposed a wrapper around this but it was >> rejected by Lucas. >> >> Besides, as we discussed earlier when I was proposed the static >> allocation, the ggtt might not always be linked to the same gt, >> so that I assumed that sometimes: >> >> to_gt(perf->i915)->ggtt->vm.gt != to_gt(perf->i915) >> >> if two GTs are sharing the same ggtt, what would the ggtt->vm.gt >> link be? > >From the git history, it doesn't look like this really needs to care >about the GGTT at all; I think it was just unintentionally written in a >roundabout manner when intel_gt was first being introduced in the code. >The reference here first showed up in commit f170523a7b8e ("drm/i915/gt: >Consolidate the CS timestamp clocks"). > >Actually the most correct thing to do is probably to use >'stream->engine->gt' to ensure we grab the GT actually associated with >the stream's engine. > stream is not yet created at this point, so I would do this: pass intel_gt to the helper instead of perf: static u64 oa_exponent_to_ns(struct intel_gt *gt, int exponent) { return intel_gt_clock_interval_to_ns(gt, 2ULL << exponent); } caller would then be: oa_period = oa_exponent_to_ns(props->engine->gt, value); Thanks, Umesh > >Matt > > >> >> Thanks, >> Andi > >-- >Matt Roper >Graphics Software Engineer >VTT-OSGC Platform Enablement >Intel Corporation >(916) 356-2795
Hi guys, > > > > > struct drm_i915_gem_object *bo; > > > > > struct i915_vma *vma; > > > > > const u64 delay_ticks = 0xffffffffffffffff - > > > > > - intel_gt_ns_to_clock_interval(stream->perf->i915->ggtt.vm.gt, > > > > > + intel_gt_ns_to_clock_interval(to_gt(stream->perf->i915)->ggtt->vm.gt, > > > > > > > > I'm not too familiar with the perf code, but this looks a bit roundabout > > > > since we're ultimately trying to get to a GT...do we even need to go > > > > through the ggtt structure here or can we just pass > > > > "to_gt(stream->perf->i915)" as the first parameter? > > > > > > > > > atomic64_read(&stream->perf->noa_programming_delay)); > > > > > const u32 base = stream->engine->mmio_base; > > > > > #define CS_GPR(x) GEN8_RING_CS_GPR(base, x) > > > > > @@ -3542,7 +3542,7 @@ i915_perf_open_ioctl_locked(struct i915_perf *perf, > > > > > > > > > > static u64 oa_exponent_to_ns(struct i915_perf *perf, int exponent) > > > > > { > > > > > - return intel_gt_clock_interval_to_ns(perf->i915->ggtt.vm.gt, > > > > > + return intel_gt_clock_interval_to_ns(to_gt(perf->i915)->ggtt->vm.gt, > > > > > > > > Ditto; this looks like "to_gt(perf->i915)" might be all we need? > > > > > > I think this function is looking for the GT coming from the VM, > > > otherwise originally it could have taken it from &i915->gt. In my > > > first version I proposed a wrapper around this but it was > > > rejected by Lucas. > > > > > > Besides, as we discussed earlier when I was proposed the static > > > allocation, the ggtt might not always be linked to the same gt, > > > so that I assumed that sometimes: > > > > > > to_gt(perf->i915)->ggtt->vm.gt != to_gt(perf->i915) > > > > > > if two GTs are sharing the same ggtt, what would the ggtt->vm.gt > > > link be? > > > > From the git history, it doesn't look like this really needs to care > > about the GGTT at all; I think it was just unintentionally written in a > > roundabout manner when intel_gt was first being introduced in the code. > > The reference here first showed up in commit f170523a7b8e ("drm/i915/gt: > > Consolidate the CS timestamp clocks"). > > > > Actually the most correct thing to do is probably to use > > 'stream->engine->gt' to ensure we grab the GT actually associated with > > the stream's engine. > > > > stream is not yet created at this point, so I would do this: > > pass intel_gt to the helper instead of perf: > static u64 oa_exponent_to_ns(struct intel_gt *gt, int exponent) > { > return intel_gt_clock_interval_to_ns(gt, 2ULL << exponent); > } > > caller would then be: > oa_period = oa_exponent_to_ns(props->engine->gt, value); thanks for the suggestions, but this is out of the scope of this patch... I did propose a wrapper but it was rejected because it was, indeed, out of scope. I'm going to use to_gt(perf->i915) as Matt suggested originally, patch is ready. Thanks, Andi
diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c index 8e65cd8258b9..94c3eb1586b0 100644 --- a/drivers/gpu/drm/i915/gvt/dmabuf.c +++ b/drivers/gpu/drm/i915/gvt/dmabuf.c @@ -84,7 +84,7 @@ static int vgpu_gem_get_pages( kfree(st); return ret; } - gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + + gtt_entries = (gen8_pte_t __iomem *)to_gt(dev_priv)->ggtt->gsm + (fb_info->start >> PAGE_SHIFT); for_each_sg(st->sgl, sg, page_num, i) { dma_addr_t dma_addr = diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index e0e052cdf8b8..6966fe08df92 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c @@ -390,9 +390,9 @@ static int i915_swizzle_info(struct seq_file *m, void *data) intel_wakeref_t wakeref; seq_printf(m, "bit6 swizzle for X-tiling = %s\n", - swizzle_string(dev_priv->ggtt.bit_6_swizzle_x)); + swizzle_string(to_gt(dev_priv)->ggtt->bit_6_swizzle_x)); seq_printf(m, "bit6 swizzle for Y-tiling = %s\n", - swizzle_string(dev_priv->ggtt.bit_6_swizzle_y)); + swizzle_string(to_gt(dev_priv)->ggtt->bit_6_swizzle_y)); if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES) seq_puts(m, "L-shaped memory detected\n"); diff --git a/drivers/gpu/drm/i915/i915_driver.c b/drivers/gpu/drm/i915/i915_driver.c index 60f8cbf24de7..3c984553d86f 100644 --- a/drivers/gpu/drm/i915/i915_driver.c +++ b/drivers/gpu/drm/i915/i915_driver.c @@ -1146,7 +1146,7 @@ static int i915_drm_suspend(struct drm_device *dev) /* Must be called before GGTT is suspended. */ intel_dpt_suspend(dev_priv); - i915_ggtt_suspend(&dev_priv->ggtt); + i915_ggtt_suspend(to_gt(dev_priv)->ggtt); i915_save_display(dev_priv); @@ -1270,7 +1270,7 @@ static int i915_drm_resume(struct drm_device *dev) if (ret) drm_err(&dev_priv->drm, "failed to re-enable GGTT\n"); - i915_ggtt_resume(&dev_priv->ggtt); + i915_ggtt_resume(to_gt(dev_priv)->ggtt); /* Must be called after GGTT is resumed. */ intel_dpt_resume(dev_priv); diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index 471be2716abe..524025790fe0 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h @@ -1749,7 +1749,7 @@ static inline bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_objec { struct drm_i915_private *i915 = to_i915(obj->base.dev); - return i915->ggtt.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && + return to_gt(i915)->ggtt->bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && i915_gem_object_is_tiled(obj); } diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 8ba2119092f2..45e3b4c540a1 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c @@ -88,7 +88,8 @@ int i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, struct drm_file *file) { - struct i915_ggtt *ggtt = &to_i915(dev)->ggtt; + struct drm_i915_private *i915 = to_i915(dev); + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; struct drm_i915_gem_get_aperture *args = data; struct i915_vma *vma; u64 pinned; @@ -289,7 +290,7 @@ static struct i915_vma *i915_gem_gtt_prepare(struct drm_i915_gem_object *obj, bool write) { struct drm_i915_private *i915 = to_i915(obj->base.dev); - struct i915_ggtt *ggtt = &i915->ggtt; + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; struct i915_vma *vma; struct i915_gem_ww_ctx ww; int ret; @@ -350,7 +351,7 @@ static void i915_gem_gtt_cleanup(struct drm_i915_gem_object *obj, struct i915_vma *vma) { struct drm_i915_private *i915 = to_i915(obj->base.dev); - struct i915_ggtt *ggtt = &i915->ggtt; + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; i915_gem_object_unpin_pages(obj); if (drm_mm_node_allocated(node)) { @@ -366,7 +367,7 @@ i915_gem_gtt_pread(struct drm_i915_gem_object *obj, const struct drm_i915_gem_pread *args) { struct drm_i915_private *i915 = to_i915(obj->base.dev); - struct i915_ggtt *ggtt = &i915->ggtt; + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; intel_wakeref_t wakeref; struct drm_mm_node node; void __user *user_data; @@ -522,7 +523,7 @@ i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj, const struct drm_i915_gem_pwrite *args) { struct drm_i915_private *i915 = to_i915(obj->base.dev); - struct i915_ggtt *ggtt = &i915->ggtt; + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; struct intel_runtime_pm *rpm = &i915->runtime_pm; intel_wakeref_t wakeref; struct drm_mm_node node; @@ -823,7 +824,7 @@ void i915_gem_runtime_suspend(struct drm_i915_private *i915) */ list_for_each_entry_safe(obj, on, - &i915->ggtt.userfault_list, userfault_link) + &to_gt(i915)->ggtt->userfault_list, userfault_link) __i915_gem_object_release_mmap_gtt(obj); /* @@ -831,8 +832,8 @@ void i915_gem_runtime_suspend(struct drm_i915_private *i915) * in use by hardware (i.e. they are pinned), we should not be powering * down! All other fences will be reacquired by the user upon waking. */ - for (i = 0; i < i915->ggtt.num_fences; i++) { - struct i915_fence_reg *reg = &i915->ggtt.fence_regs[i]; + for (i = 0; i < to_gt(i915)->ggtt->num_fences; i++) { + struct i915_fence_reg *reg = &to_gt(i915)->ggtt->fence_regs[i]; /* * Ideally we want to assert that the fence register is not @@ -873,7 +874,7 @@ i915_gem_object_ggtt_pin_ww(struct drm_i915_gem_object *obj, u64 size, u64 alignment, u64 flags) { struct drm_i915_private *i915 = to_i915(obj->base.dev); - struct i915_ggtt *ggtt = &i915->ggtt; + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; struct i915_vma *vma; int ret; @@ -1101,7 +1102,7 @@ int i915_gem_init(struct drm_i915_private *dev_priv) /* Minimal basic recovery for KMS */ ret = i915_ggtt_enable_hw(dev_priv); - i915_ggtt_resume(&dev_priv->ggtt); + i915_ggtt_resume(to_gt(dev_priv)->ggtt); intel_init_clock_gating(dev_priv); } @@ -1124,7 +1125,7 @@ void i915_gem_driver_unregister(struct drm_i915_private *i915) void i915_gem_driver_remove(struct drm_i915_private *dev_priv) { - intel_wakeref_auto_fini(&dev_priv->ggtt.userfault_wakeref); + intel_wakeref_auto_fini(&to_gt(dev_priv)->ggtt->userfault_wakeref); i915_gem_suspend_late(dev_priv); intel_gt_driver_remove(to_gt(dev_priv)); diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c index cd5f2348a187..2f2ba7a2955d 100644 --- a/drivers/gpu/drm/i915/i915_gem_gtt.c +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c @@ -56,7 +56,7 @@ void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) { struct drm_i915_private *i915 = to_i915(obj->base.dev); - struct i915_ggtt *ggtt = &i915->ggtt; + struct i915_ggtt *ggtt = to_gt(i915)->ggtt; /* XXX This does not prevent more requests being submitted! */ if (unlikely(ggtt->do_idle_maps)) @@ -103,7 +103,7 @@ int i915_gem_gtt_reserve(struct i915_address_space *vm, GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); GEM_BUG_ON(!IS_ALIGNED(offset, I915_GTT_MIN_ALIGNMENT)); GEM_BUG_ON(range_overflows(offset, size, vm->total)); - GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm); + GEM_BUG_ON(vm == &to_gt(vm->i915)->ggtt->alias->vm); GEM_BUG_ON(drm_mm_node_allocated(node)); node->size = size; @@ -201,7 +201,7 @@ int i915_gem_gtt_insert(struct i915_address_space *vm, GEM_BUG_ON(start >= end); GEM_BUG_ON(start > 0 && !IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); GEM_BUG_ON(end < U64_MAX && !IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); - GEM_BUG_ON(vm == &vm->i915->ggtt.alias->vm); + GEM_BUG_ON(vm == &to_gt(vm->i915)->ggtt->alias->vm); GEM_BUG_ON(drm_mm_node_allocated(node)); if (unlikely(range_overflows(start, size, end))) diff --git a/drivers/gpu/drm/i915/i915_getparam.c b/drivers/gpu/drm/i915/i915_getparam.c index 7f80ad247bc8..5b8a2157d797 100644 --- a/drivers/gpu/drm/i915/i915_getparam.c +++ b/drivers/gpu/drm/i915/i915_getparam.c @@ -31,7 +31,7 @@ int i915_getparam_ioctl(struct drm_device *dev, void *data, value = pdev->revision; break; case I915_PARAM_NUM_FENCES_AVAIL: - value = i915->ggtt.num_fences; + value = to_gt(i915)->ggtt->num_fences; break; case I915_PARAM_HAS_OVERLAY: value = !!i915->overlay; diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c index 170bba913c30..128315aec517 100644 --- a/drivers/gpu/drm/i915/i915_perf.c +++ b/drivers/gpu/drm/i915/i915_perf.c @@ -1630,7 +1630,7 @@ static int alloc_noa_wait(struct i915_perf_stream *stream) struct drm_i915_gem_object *bo; struct i915_vma *vma; const u64 delay_ticks = 0xffffffffffffffff - - intel_gt_ns_to_clock_interval(stream->perf->i915->ggtt.vm.gt, + intel_gt_ns_to_clock_interval(to_gt(stream->perf->i915)->ggtt->vm.gt, atomic64_read(&stream->perf->noa_programming_delay)); const u32 base = stream->engine->mmio_base; #define CS_GPR(x) GEN8_RING_CS_GPR(base, x) @@ -3542,7 +3542,7 @@ i915_perf_open_ioctl_locked(struct i915_perf *perf, static u64 oa_exponent_to_ns(struct i915_perf *perf, int exponent) { - return intel_gt_clock_interval_to_ns(perf->i915->ggtt.vm.gt, + return intel_gt_clock_interval_to_ns(to_gt(perf->i915)->ggtt->vm.gt, 2ULL << exponent); }