Message ID | 20230628161023.2944972-1-tvrtko.ursulin@linux.intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [RFC] drm/i915: Refactor PAT/cache handling | expand |
> From: Tvrtko Ursulin <tvrtko.ursulin@intel.com> > > Informal commit message for now. > > I got a bit impatient and curious to see if the idea we discussed would > work so sketched something out. I think it is what I was describing back > then.. Oops, you beat me on this, shame on me. > So high level idea is to teach the driver what caching modes are hidden > behind PAT indices. Given you already had that in static tables, if we > just turn the tables a bit around and add a driver abstraction of caching > modes this is what happens: > > * We can lose the ugly runtime i915_gem_get_pat_index. > > * We can have a smarter i915_gem_object_has_cache_level, which now can > use the above mentioned table to understand the caching modes and so > does not have to pessimistically return true for _any_ input when user > has set the PAT index. This may improve things even for MTL. > > * We can simplify the debugfs printout to be platform agnostic. > > * We are perhaps opening the door to un-regress the dodgy addition > made to i915_gem_object_can_bypass_llc? See QQQ/FIXME in the patch. > > I hope I did not forget anything, but anyway, please have a read and see > what you think. I think it has potential. > > Proper commit message can come later. > > Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> > Cc: Fei Yang <fei.yang@intel.com> > --- > drivers/gpu/drm/i915/Makefile | 1 + > drivers/gpu/drm/i915/gem/i915_gem_domain.c | 34 ++--- > .../gpu/drm/i915/gem/i915_gem_execbuffer.c | 13 +- > drivers/gpu/drm/i915/gem/i915_gem_mman.c | 10 +- > drivers/gpu/drm/i915/gem/i915_gem_object.c | 78 ++++------- > drivers/gpu/drm/i915/gem/i915_gem_object.h | 18 ++- > .../gpu/drm/i915/gem/i915_gem_object_types.h | 99 +------------- > drivers/gpu/drm/i915/gem/i915_gem_stolen.c | 7 +- > drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c | 26 ++-- > .../gpu/drm/i915/gem/selftests/huge_pages.c | 2 +- > drivers/gpu/drm/i915/gt/gen6_ppgtt.c | 4 +- > drivers/gpu/drm/i915/gt/gen8_ppgtt.c | 13 +- > drivers/gpu/drm/i915/gt/intel_ggtt.c | 9 +- > drivers/gpu/drm/i915/gt/intel_migrate.c | 11 +- > drivers/gpu/drm/i915/gt/selftest_migrate.c | 9 +- > drivers/gpu/drm/i915/gt/selftest_reset.c | 14 +- > drivers/gpu/drm/i915/gt/selftest_timeline.c | 5 +- > drivers/gpu/drm/i915/gt/selftest_tlb.c | 5 +- > drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c | 8 +- > drivers/gpu/drm/i915/i915_cache.c | 59 ++++++++ > drivers/gpu/drm/i915/i915_cache.h | 129 ++++++++++++++++++ > drivers/gpu/drm/i915/i915_debugfs.c | 83 ++++++----- > drivers/gpu/drm/i915/i915_driver.c | 3 + > drivers/gpu/drm/i915/i915_drv.h | 3 + > drivers/gpu/drm/i915/i915_gem.c | 21 +-- > drivers/gpu/drm/i915/i915_gpu_error.c | 7 +- > drivers/gpu/drm/i915/i915_pci.c | 83 +++++------ > drivers/gpu/drm/i915/intel_device_info.h | 6 +- > drivers/gpu/drm/i915/selftests/i915_gem.c | 5 +- > .../gpu/drm/i915/selftests/i915_gem_evict.c | 4 +- > drivers/gpu/drm/i915/selftests/i915_gem_gtt.c | 13 +- > .../drm/i915/selftests/intel_memory_region.c | 4 +- > .../gpu/drm/i915/selftests/mock_gem_device.c | 10 +- > 33 files changed, 415 insertions(+), 381 deletions(-) > create mode 100644 drivers/gpu/drm/i915/i915_cache.c > create mode 100644 drivers/gpu/drm/i915/i915_cache.h > > diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile > index 2be9dd960540..2c3da8f0c78e 100644 > --- a/drivers/gpu/drm/i915/Makefile > +++ b/drivers/gpu/drm/i915/Makefile > @@ -30,6 +30,7 @@ subdir-ccflags-y += -I$(srctree)/$(src) > # core driver code > i915-y += i915_driver.o \ > i915_drm_client.o \ > + i915_cache.o \ > i915_config.o \ > i915_getparam.o \ > i915_ioctl.o \ > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_domain.c b/drivers/gpu/drm/i915/gem/i915_gem_domain.c > index dfaaa8b66ac3..49bfae45390f 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_domain.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_domain.c > @@ -8,6 +8,7 @@ > #include "display/intel_frontbuffer.h" > #include "gt/intel_gt.h" > > +#include "i915_cache.h" > #include "i915_drv.h" > #include "i915_gem_clflush.h" > #include "i915_gem_domain.h" > @@ -27,15 +28,8 @@ static bool gpu_write_needs_clflush(struct drm_i915_gem_object *obj) > if (IS_DGFX(i915)) > return false; > > - /* > - * For objects created by userspace through GEM_CREATE with pat_index > - * set by set_pat extension, i915_gem_object_has_cache_level() will > - * always return true, because the coherency of such object is managed > - * by userspace. Othereise the call here would fall back to checking > - * whether the object is un-cached or write-through. > - */ > - return !(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) || > - i915_gem_object_has_cache_level(obj, I915_CACHE_WT)); > + return i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1 && > + i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WT) != 1; Why is it necessary to define I915_CACHE_MODE's while there is already i915_cache_level? I thought we wanted to get rid of such abstractions instead of adding more. This patch also introduced INTEL_INFO(i915)->cache_modes, why don't we directly add the platform specific PAT there? For example, add the following for MTL, INTEL_INFO(i915)->pat[] = { [0] = MTL_PPAT_L4_0_WB, \ [1] = MTL_PPAT_L4_1_WT, \ [2] = MTL_PPAT_L4_3_UC, \ [3] = MTL_PPAT_L4_0_WB | MTL_2_COH_1W, \ [4] = MTL_PPAT_L4_0_WB | MTL_3_COH_2W, \ } Everything here has already been defined, no need to introduce new macros. This can also be used to initialize the PAT index registers, like in xelpg_setup_private_ppat() and xelpmp_setup_private_ppat(). > } > > bool i915_gem_cpu_write_needs_clflush(struct drm_i915_gem_object *obj) > @@ -272,15 +266,18 @@ i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write) > int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, > enum i915_cache_level cache_level) s/enum i915_cache_level cache_level/unsigned int pat_index This function is for KMD objects only, I don't think we even need to keep the i915_cache_level, simply passing in INTEL_INFO(i915)->pat_uc/wb/wt is good enough. > { > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > + i915_cache_t mode; > int ret; > > - /* > - * For objects created by userspace through GEM_CREATE with pat_index > - * set by set_pat extension, simply return 0 here without touching > - * the cache setting, because such objects should have an immutable > - * cache setting by desgin and always managed by userspace. > - */ > - if (i915_gem_object_has_cache_level(obj, cache_level)) > + if (obj->pat_set_by_user) > + return -EINVAL; I don't think this condition would ever be true, but okay to keep it. > + ret = i915_cache_level_to_pat_and_mode(i915, cache_level, &mode); > + if (ret < 0) > + return -EINVAL; > + > + if (mode == obj->cache_mode) The above lines can be just one line, if (pat_index == obj->pat_index) > return 0; > > ret = i915_gem_object_wait(obj, > @@ -326,10 +323,9 @@ int i915_gem_get_caching_ioctl(struct drm_device *dev, void *data, > goto out; > } > > - if (i915_gem_object_has_cache_level(obj, I915_CACHE_LLC) || > - i915_gem_object_has_cache_level(obj, I915_CACHE_L3_LLC)) > + if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WB)) > args->caching = I915_CACHING_CACHED; > - else if (i915_gem_object_has_cache_level(obj, I915_CACHE_WT)) > + else if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WT)) > args->caching = I915_CACHING_DISPLAY; > else > args->caching = I915_CACHING_NONE; > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > index d3208a325614..ee85221fa6eb 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > @@ -640,15 +640,9 @@ static inline int use_cpu_reloc(const struct reloc_cache *cache, > if (DBG_FORCE_RELOC == FORCE_GTT_RELOC) > return false; > > - /* > - * For objects created by userspace through GEM_CREATE with pat_index > - * set by set_pat extension, i915_gem_object_has_cache_level() always > - * return true, otherwise the call would fall back to checking whether > - * the object is un-cached. > - */ > return (cache->has_llc || > obj->cache_dirty || > - !i915_gem_object_has_cache_level(obj, I915_CACHE_NONE)); > + i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1); > } > > static int eb_reserve_vma(struct i915_execbuffer *eb, > @@ -1329,10 +1323,7 @@ static void *reloc_iomap(struct i915_vma *batch, > if (drm_mm_node_allocated(&cache->node)) { > ggtt->vm.insert_page(&ggtt->vm, > i915_gem_object_get_dma_address(obj, page), > - offset, > - i915_gem_get_pat_index(ggtt->vm.i915, > - I915_CACHE_NONE), > - 0); > + offset, eb->i915->pat_uc, 0); > } else { > offset += page << PAGE_SHIFT; > } > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_mman.c b/drivers/gpu/drm/i915/gem/i915_gem_mman.c > index aa4d842d4c5a..5e21aedb02d2 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_mman.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_mman.c > @@ -386,13 +386,11 @@ static vm_fault_t vm_fault_gtt(struct vm_fault *vmf) > /* > * For objects created by userspace through GEM_CREATE with pat_index > * set by set_pat extension, coherency is managed by userspace, make > - * sure we don't fail handling the vm fault by calling > - * i915_gem_object_has_cache_level() which always return true for such > - * objects. Otherwise this helper function would fall back to checking > - * whether the object is un-cached. > + * sure we don't fail handling the vm fault by making sure that we > + * know the object is uncached or that we have LLC. > */ > - if (!(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) || > - HAS_LLC(i915))) { > + if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1 && > + !HAS_LLC(i915)) { > ret = -EFAULT; > goto err_unpin; > } > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.c b/drivers/gpu/drm/i915/gem/i915_gem_object.c > index 0004d5fa7cc2..52c6c5f09bdd 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_object.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.c > @@ -45,33 +45,6 @@ static struct kmem_cache *slab_objects; > > static const struct drm_gem_object_funcs i915_gem_object_funcs; > > -unsigned int i915_gem_get_pat_index(struct drm_i915_private *i915, > - enum i915_cache_level level) > -{ > - if (drm_WARN_ON(&i915->drm, level >= I915_MAX_CACHE_LEVEL)) > - return 0; > - > - return INTEL_INFO(i915)->cachelevel_to_pat[level]; > -} Yes, this can be removed. INTEL_INFO(i915)->pat_uc/wb/wt should be sufficient, > - > -bool i915_gem_object_has_cache_level(const struct drm_i915_gem_object *obj, > - enum i915_cache_level lvl) > -{ If we had INTEL_INFO(i915)->pat[] setup, it would be easier just to keep this function, because we can simply check the cache policy bit field in INTEL_INFO(i915)->pat[obj->pat_index] to see whether it is cached, uncached, or write-through. See Bspec https://gfxspecs.intel.com/Predator/Home/Index/44235 For MTL check bit[3:2], for other gen12 platforms check bit[1:0] > - /* > - * In case the pat_index is set by user space, this kernel mode > - * driver should leave the coherency to be managed by user space, > - * simply return true here. > - */ > - if (obj->pat_set_by_user) > - return true; > - > - /* > - * Otherwise the pat_index should have been converted from cache_level > - * so that the following comparison is valid. > - */ > - return obj->pat_index == i915_gem_get_pat_index(obj_to_i915(obj), lvl); > -} > - > struct drm_i915_gem_object *i915_gem_object_alloc(void) > { > struct drm_i915_gem_object *obj; > @@ -144,6 +117,24 @@ void __i915_gem_object_fini(struct drm_i915_gem_object *obj) > dma_resv_fini(&obj->base._resv); > } > > +void __i915_gem_object_update_coherency(struct drm_i915_gem_object *obj) > +{ > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > + const unsigned int mode = I915_CACHE_MODE(obj->cache_mode); obj->cache_mode seems to be redundant if we have INTEL_INFO(i915)->pat[] and obj->pat_index. > + > + if (!(mode == I915_CACHE_MODE_UNKNOWN || mode == I915_CACHE_MODE_UC)) > + obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | > + I915_BO_CACHE_COHERENT_FOR_WRITE); > + else if (mode != I915_CACHE_MODE_UNKNOWN && HAS_LLC(i915)) > + obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; > + else > + obj->cache_coherent = 0; > + > + obj->cache_dirty = > + !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && > + !IS_DGFX(i915); > +} > + > /** > * i915_gem_object_set_cache_coherency - Mark up the object's coherency levels > * for a given cache_level > @@ -154,20 +145,15 @@ void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj, > unsigned int cache_level) > { > struct drm_i915_private *i915 = to_i915(obj->base.dev); > + i915_cache_t mode; > + int found; > > - obj->pat_index = i915_gem_get_pat_index(i915, cache_level); > + found = i915_cache_level_to_pat_and_mode(i915, cache_level, &mode); > + GEM_WARN_ON(found < 0); > + obj->pat_index = found; > + obj->cache_mode = mode; > > - if (cache_level != I915_CACHE_NONE) > - obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | > - I915_BO_CACHE_COHERENT_FOR_WRITE); > - else if (HAS_LLC(i915)) > - obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; > - else > - obj->cache_coherent = 0; > - > - obj->cache_dirty = > - !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && > - !IS_DGFX(i915); > + __i915_gem_object_update_coherency(obj); > } > > /** > @@ -187,18 +173,9 @@ void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj, > return; > > obj->pat_index = pat_index; > + obj->cache_mode = INTEL_INFO(i915)->cache_modes[pat_index]; > > - if (pat_index != i915_gem_get_pat_index(i915, I915_CACHE_NONE)) > - obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | > - I915_BO_CACHE_COHERENT_FOR_WRITE); > - else if (HAS_LLC(i915)) > - obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; > - else > - obj->cache_coherent = 0; > - > - obj->cache_dirty = > - !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && > - !IS_DGFX(i915); > + __i915_gem_object_update_coherency(obj); > } > > bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj) > @@ -215,6 +192,7 @@ bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj) > /* > * Always flush cache for UMD objects at creation time. > */ > + /* QQQ/FIXME why? avoidable performance penalty? */ > if (obj->pat_set_by_user) > return true; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.h b/drivers/gpu/drm/i915/gem/i915_gem_object.h > index 884a17275b3a..f84f41e9f81f 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_object.h > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.h > @@ -13,6 +13,7 @@ > > #include "display/intel_frontbuffer.h" > #include "intel_memory_region.h" > +#include "i915_cache.h" > #include "i915_gem_object_types.h" > #include "i915_gem_gtt.h" > #include "i915_gem_ww.h" > @@ -32,10 +33,18 @@ static inline bool i915_gem_object_size_2big(u64 size) > return false; > } > > -unsigned int i915_gem_get_pat_index(struct drm_i915_private *i915, > - enum i915_cache_level level); > -bool i915_gem_object_has_cache_level(const struct drm_i915_gem_object *obj, > - enum i915_cache_level lvl); > +static inline int > +i915_gem_object_has_cache_mode(const struct drm_i915_gem_object *obj, > + unsigned int mode) > +{ > + if (I915_CACHE_MODE(obj->cache_mode) == mode) > + return 1; > + else if (obj->pat_set_by_user) > + return -1; /* Unknown, callers should assume no. */ > + else > + return 0; > +} > + > void i915_gem_init__objects(struct drm_i915_private *i915); > > void i915_objects_module_exit(void); > @@ -764,6 +773,7 @@ int i915_gem_object_wait_moving_fence(struct drm_i915_gem_object *obj, > bool intr); > bool i915_gem_object_has_unknown_state(struct drm_i915_gem_object *obj); > > +void __i915_gem_object_update_coherency(struct drm_i915_gem_object *obj); > void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj, > unsigned int cache_level); > void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj, > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > index 8de2b91b3edf..1f9fa28d07df 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > @@ -14,6 +14,7 @@ > #include <uapi/drm/i915_drm.h> > > #include "i915_active.h" > +#include "i915_cache.h" > #include "i915_selftest.h" > #include "i915_vma_resource.h" > > @@ -116,93 +117,6 @@ struct drm_i915_gem_object_ops { > const char *name; /* friendly name for debug, e.g. lockdep classes */ > }; > > -/** > - * enum i915_cache_level - The supported GTT caching values for system memory > - * pages. > - * > - * These translate to some special GTT PTE bits when binding pages into some > - * address space. It also determines whether an object, or rather its pages are > - * coherent with the GPU, when also reading or writing through the CPU cache > - * with those pages. > - * > - * Userspace can also control this through struct drm_i915_gem_caching. > - */ > -enum i915_cache_level { > - /** > - * @I915_CACHE_NONE: > - * > - * GPU access is not coherent with the CPU cache. If the cache is dirty > - * and we need the underlying pages to be coherent with some later GPU > - * access then we need to manually flush the pages. > - * > - * On shared LLC platforms reads and writes through the CPU cache are > - * still coherent even with this setting. See also > - * &drm_i915_gem_object.cache_coherent for more details. Due to this we > - * should only ever use uncached for scanout surfaces, otherwise we end > - * up over-flushing in some places. > - * > - * This is the default on non-LLC platforms. > - */ > - I915_CACHE_NONE = 0, > - /** > - * @I915_CACHE_LLC: > - * > - * GPU access is coherent with the CPU cache. If the cache is dirty, > - * then the GPU will ensure that access remains coherent, when both > - * reading and writing through the CPU cache. GPU writes can dirty the > - * CPU cache. > - * > - * Not used for scanout surfaces. > - * > - * Applies to both platforms with shared LLC(HAS_LLC), and snooping > - * based platforms(HAS_SNOOP). > - * > - * This is the default on shared LLC platforms. The only exception is > - * scanout objects, where the display engine is not coherent with the > - * CPU cache. For such objects I915_CACHE_NONE or I915_CACHE_WT is > - * automatically applied by the kernel in pin_for_display, if userspace > - * has not done so already. > - */ > - I915_CACHE_LLC, > - /** > - * @I915_CACHE_L3_LLC: > - * > - * Explicitly enable the Gfx L3 cache, with coherent LLC. > - * > - * The Gfx L3 sits between the domain specific caches, e.g > - * sampler/render caches, and the larger LLC. LLC is coherent with the > - * GPU, but L3 is only visible to the GPU, so likely needs to be flushed > - * when the workload completes. > - * > - * Not used for scanout surfaces. > - * > - * Only exposed on some gen7 + GGTT. More recent hardware has dropped > - * this explicit setting, where it should now be enabled by default. > - */ > - I915_CACHE_L3_LLC, > - /** > - * @I915_CACHE_WT: > - * > - * Write-through. Used for scanout surfaces. > - * > - * The GPU can utilise the caches, while still having the display engine > - * be coherent with GPU writes, as a result we don't need to flush the > - * CPU caches when moving out of the render domain. This is the default > - * setting chosen by the kernel, if supported by the HW, otherwise we > - * fallback to I915_CACHE_NONE. On the CPU side writes through the CPU > - * cache still need to be flushed, to remain coherent with the display > - * engine. > - */ > - I915_CACHE_WT, > - /** > - * @I915_MAX_CACHE_LEVEL: > - * > - * Mark the last entry in the enum. Used for defining cachelevel_to_pat > - * array for cache_level to pat translation table. > - */ > - I915_MAX_CACHE_LEVEL, > -}; > - > enum i915_map_type { > I915_MAP_WB = 0, > I915_MAP_WC, > @@ -375,6 +289,9 @@ struct drm_i915_gem_object { > unsigned int mem_flags; > #define I915_BO_FLAG_STRUCT_PAGE BIT(0) /* Object backed by struct pages */ > #define I915_BO_FLAG_IOMEM BIT(1) /* Object backed by IO memory */ > + > + i915_cache_t cache_mode; > + > /** > * @pat_index: The desired PAT index. > * > @@ -409,9 +326,7 @@ struct drm_i915_gem_object { > * Check for @pat_set_by_user to find out if an object has pat index set > * by userspace. The ioctl's to change cache settings have also been > * disabled for the objects with pat index set by userspace. Please don't > - * assume @cache_coherent having the flags set as describe here. A helper > - * function i915_gem_object_has_cache_level() provides one way to bypass > - * the use of this field. > + * assume @cache_coherent having the flags set as describe here. > * > * Track whether the pages are coherent with the GPU if reading or > * writing through the CPU caches. The largely depends on the > @@ -492,9 +407,7 @@ struct drm_i915_gem_object { > * Check for @pat_set_by_user to find out if an object has pat index set > * by userspace. The ioctl's to change cache settings have also been > * disabled for the objects with pat_index set by userspace. Please don't > - * assume @cache_dirty is set as describe here. Also see helper function > - * i915_gem_object_has_cache_level() for possible ways to bypass the use > - * of this field. > + * assume @cache_dirty is set as describe here. > * > * Track if we are we dirty with writes through the CPU cache for this > * object. As a result reading directly from main memory might yield > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > index 3b094d36a0b0..a7012f1a9c70 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > @@ -563,11 +563,8 @@ static void dbg_poison(struct i915_ggtt *ggtt, > while (size) { > void __iomem *s; > > - ggtt->vm.insert_page(&ggtt->vm, addr, > - ggtt->error_capture.start, > - i915_gem_get_pat_index(ggtt->vm.i915, > - I915_CACHE_NONE), > - 0); > + ggtt->vm.insert_page(&ggtt->vm, addr, ggtt->error_capture.start, > + ggtt->vm.i915->pat_uc, 0); > mb(); > > s = io_mapping_map_wc(&ggtt->iomap, > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c > index 7078af2f8f79..e794bd2a7ccb 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c > +++ b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c > @@ -58,6 +58,16 @@ i915_ttm_cache_level(struct drm_i915_private *i915, struct ttm_resource *res, > I915_CACHE_NONE; > } > > +static unsigned int > +i915_ttm_cache_pat(struct drm_i915_private *i915, struct ttm_resource *res, > + struct ttm_tt *ttm) > +{ > + return ((HAS_LLC(i915) || HAS_SNOOP(i915)) && > + !i915_ttm_gtt_binds_lmem(res) && > + ttm->caching == ttm_cached) ? i915->pat_wb : > + i915->pat_uc; > +} > + > static struct intel_memory_region * > i915_ttm_region(struct ttm_device *bdev, int ttm_mem_type) > { > @@ -196,7 +206,7 @@ static struct dma_fence *i915_ttm_accel_move(struct ttm_buffer_object *bo, > struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo); > struct i915_request *rq; > struct ttm_tt *src_ttm = bo->ttm; > - enum i915_cache_level src_level, dst_level; > + unsigned int src_pat, dst_pat; > int ret; > > if (!to_gt(i915)->migrate.context || intel_gt_is_wedged(to_gt(i915))) > @@ -206,16 +216,15 @@ static struct dma_fence *i915_ttm_accel_move(struct ttm_buffer_object *bo, > if (I915_SELFTEST_ONLY(fail_gpu_migration)) > clear = true; > > - dst_level = i915_ttm_cache_level(i915, dst_mem, dst_ttm); > + dst_pat = i915_ttm_cache_pat(i915, dst_mem, dst_ttm); > if (clear) { > if (bo->type == ttm_bo_type_kernel && > !I915_SELFTEST_ONLY(fail_gpu_migration)) > return ERR_PTR(-EINVAL); > > intel_engine_pm_get(to_gt(i915)->migrate.context->engine); > - ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, > - dst_st->sgl, > - i915_gem_get_pat_index(i915, dst_level), > + ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, > + deps, dst_st->sgl, dst_pat, > i915_ttm_gtt_binds_lmem(dst_mem), > 0, &rq); > } else { > @@ -225,14 +234,13 @@ static struct dma_fence *i915_ttm_accel_move(struct ttm_buffer_object *bo, > if (IS_ERR(src_rsgt)) > return ERR_CAST(src_rsgt); > > - src_level = i915_ttm_cache_level(i915, bo->resource, src_ttm); > + src_pat = i915_ttm_cache_pat(i915, bo->resource, src_ttm); > intel_engine_pm_get(to_gt(i915)->migrate.context->engine); > ret = intel_context_migrate_copy(to_gt(i915)->migrate.context, > deps, src_rsgt->table.sgl, > - i915_gem_get_pat_index(i915, src_level), > + src_pat, > i915_ttm_gtt_binds_lmem(bo->resource), > - dst_st->sgl, > - i915_gem_get_pat_index(i915, dst_level), > + dst_st->sgl, dst_pat, > i915_ttm_gtt_binds_lmem(dst_mem), > &rq); > > diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > index df6c9a84252c..c8925918784e 100644 > --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > @@ -354,7 +354,7 @@ fake_huge_pages_object(struct drm_i915_private *i915, u64 size, bool single) > > obj->write_domain = I915_GEM_DOMAIN_CPU; > obj->read_domains = I915_GEM_DOMAIN_CPU; > - obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE); > + i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE); > > return obj; > } > diff --git a/drivers/gpu/drm/i915/gt/gen6_ppgtt.c b/drivers/gpu/drm/i915/gt/gen6_ppgtt.c > index c2bdc133c89a..fb69f667652a 100644 > --- a/drivers/gpu/drm/i915/gt/gen6_ppgtt.c > +++ b/drivers/gpu/drm/i915/gt/gen6_ppgtt.c > @@ -226,9 +226,7 @@ static int gen6_ppgtt_init_scratch(struct gen6_ppgtt *ppgtt) > return ret; > > vm->scratch[0]->encode = > - vm->pte_encode(px_dma(vm->scratch[0]), > - i915_gem_get_pat_index(vm->i915, > - I915_CACHE_NONE), > + vm->pte_encode(px_dma(vm->scratch[0]), vm->i915->pat_uc, > PTE_READ_ONLY); > > vm->scratch[1] = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); > diff --git a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c > index f948d33e5ec5..a6692ea1a91e 100644 > --- a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c > +++ b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c > @@ -40,16 +40,11 @@ static u64 gen8_pte_encode(dma_addr_t addr, > if (flags & PTE_LM) > pte |= GEN12_PPGTT_PTE_LM; > > - /* > - * For pre-gen12 platforms pat_index is the same as enum > - * i915_cache_level, so the switch-case here is still valid. > - * See translation table defined by LEGACY_CACHELEVEL. > - */ > switch (pat_index) { > - case I915_CACHE_NONE: > + case 0: > pte |= PPAT_UNCACHED; > break; > - case I915_CACHE_WT: > + case 3: > pte |= PPAT_DISPLAY_ELLC; > break; > default: > @@ -853,9 +848,7 @@ static int gen8_init_scratch(struct i915_address_space *vm) > pte_flags |= PTE_LM; > > vm->scratch[0]->encode = > - vm->pte_encode(px_dma(vm->scratch[0]), > - i915_gem_get_pat_index(vm->i915, > - I915_CACHE_NONE), > + vm->pte_encode(px_dma(vm->scratch[0]), vm->i915->pat_uc, > pte_flags); > > for (i = 1; i <= vm->top; i++) { > diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c > index dd0ed941441a..c97379cf8241 100644 > --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c > +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c > @@ -921,9 +921,7 @@ static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size) > pte_flags |= PTE_LM; > > ggtt->vm.scratch[0]->encode = > - ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]), > - i915_gem_get_pat_index(i915, > - I915_CACHE_NONE), > + ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]), i915->pat_uc, > pte_flags); > > return 0; > @@ -1297,10 +1295,7 @@ bool i915_ggtt_resume_vm(struct i915_address_space *vm) > * ptes to be repopulated. > */ > vma->resource->bound_flags = 0; > - vma->ops->bind_vma(vm, NULL, vma->resource, > - obj ? obj->pat_index : > - i915_gem_get_pat_index(vm->i915, > - I915_CACHE_NONE), > + vma->ops->bind_vma(vm, NULL, vma->resource, obj->cache_mode, > was_bound); > > if (obj) { /* only used during resume => exclusive access */ > diff --git a/drivers/gpu/drm/i915/gt/intel_migrate.c b/drivers/gpu/drm/i915/gt/intel_migrate.c > index 6023288b0e2d..81f7834cc2db 100644 > --- a/drivers/gpu/drm/i915/gt/intel_migrate.c > +++ b/drivers/gpu/drm/i915/gt/intel_migrate.c > @@ -45,9 +45,7 @@ static void xehpsdv_toggle_pdes(struct i915_address_space *vm, > * Insert a dummy PTE into every PT that will map to LMEM to ensure > * we have a correctly setup PDE structure for later use. > */ > - vm->insert_page(vm, 0, d->offset, > - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), > - PTE_LM); > + vm->insert_page(vm, 0, d->offset, vm->i915->pat_uc, PTE_LM); > GEM_BUG_ON(!pt->is_compact); > d->offset += SZ_2M; > } > @@ -65,9 +63,7 @@ static void xehpsdv_insert_pte(struct i915_address_space *vm, > * alignment is 64K underneath for the pt, and we are careful > * not to access the space in the void. > */ > - vm->insert_page(vm, px_dma(pt), d->offset, > - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), > - PTE_LM); > + vm->insert_page(vm, px_dma(pt), d->offset, vm->i915->pat_uc, PTE_LM); > d->offset += SZ_64K; > } > > @@ -77,8 +73,7 @@ static void insert_pte(struct i915_address_space *vm, > { > struct insert_pte_data *d = data; > > - vm->insert_page(vm, px_dma(pt), d->offset, > - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), > + vm->insert_page(vm, px_dma(pt), d->offset, vm->i915->pat_uc, > i915_gem_object_is_lmem(pt->base) ? PTE_LM : 0); > d->offset += PAGE_SIZE; > } > diff --git a/drivers/gpu/drm/i915/gt/selftest_migrate.c b/drivers/gpu/drm/i915/gt/selftest_migrate.c > index 3def5ca72dec..a67ede65d816 100644 > --- a/drivers/gpu/drm/i915/gt/selftest_migrate.c > +++ b/drivers/gpu/drm/i915/gt/selftest_migrate.c > @@ -904,8 +904,7 @@ static int perf_clear_blt(void *arg) > > err = __perf_clear_blt(gt->migrate.context, > dst->mm.pages->sgl, > - i915_gem_get_pat_index(gt->i915, > - I915_CACHE_NONE), > + gt->i915->pat_uc, > i915_gem_object_is_lmem(dst), > sizes[i]); > > @@ -995,12 +994,10 @@ static int perf_copy_blt(void *arg) > > err = __perf_copy_blt(gt->migrate.context, > src->mm.pages->sgl, > - i915_gem_get_pat_index(gt->i915, > - I915_CACHE_NONE), > + gt->i915->pat_uc, > i915_gem_object_is_lmem(src), > dst->mm.pages->sgl, > - i915_gem_get_pat_index(gt->i915, > - I915_CACHE_NONE), > + gt->i915->pat_uc, > i915_gem_object_is_lmem(dst), > sz); > > diff --git a/drivers/gpu/drm/i915/gt/selftest_reset.c b/drivers/gpu/drm/i915/gt/selftest_reset.c > index 79aa6ac66ad2..327dc9294e0f 100644 > --- a/drivers/gpu/drm/i915/gt/selftest_reset.c > +++ b/drivers/gpu/drm/i915/gt/selftest_reset.c > @@ -84,11 +84,8 @@ __igt_reset_stolen(struct intel_gt *gt, > void __iomem *s; > void *in; > > - ggtt->vm.insert_page(&ggtt->vm, dma, > - ggtt->error_capture.start, > - i915_gem_get_pat_index(gt->i915, > - I915_CACHE_NONE), > - 0); > + ggtt->vm.insert_page(&ggtt->vm, dma, ggtt->error_capture.start, > + gt->i915->pat_uc, 0); > mb(); > > s = io_mapping_map_wc(&ggtt->iomap, > @@ -127,11 +124,8 @@ __igt_reset_stolen(struct intel_gt *gt, > void *in; > u32 x; > > - ggtt->vm.insert_page(&ggtt->vm, dma, > - ggtt->error_capture.start, > - i915_gem_get_pat_index(gt->i915, > - I915_CACHE_NONE), > - 0); > + ggtt->vm.insert_page(&ggtt->vm, dma, ggtt->error_capture.start, > + gt->i915->pat_uc, 0); > mb(); > > s = io_mapping_map_wc(&ggtt->iomap, > diff --git a/drivers/gpu/drm/i915/gt/selftest_timeline.c b/drivers/gpu/drm/i915/gt/selftest_timeline.c > index 39c3ec12df1a..db64dc7d3fce 100644 > --- a/drivers/gpu/drm/i915/gt/selftest_timeline.c > +++ b/drivers/gpu/drm/i915/gt/selftest_timeline.c > @@ -836,7 +836,10 @@ static int setup_watcher(struct hwsp_watcher *w, struct intel_gt *gt, > return PTR_ERR(obj); > > /* keep the same cache settings as timeline */ > - i915_gem_object_set_pat_index(obj, tl->hwsp_ggtt->obj->pat_index); > + obj->pat_index = tl->hwsp_ggtt->obj->pat_index; > + obj->cache_mode = tl->hwsp_ggtt->obj->cache_mode; > + __i915_gem_object_update_coherency(obj); > + > w->map = i915_gem_object_pin_map_unlocked(obj, > page_unmask_bits(tl->hwsp_ggtt->obj->mm.mapping)); > if (IS_ERR(w->map)) { > diff --git a/drivers/gpu/drm/i915/gt/selftest_tlb.c b/drivers/gpu/drm/i915/gt/selftest_tlb.c > index 3bd6b540257b..6049f01be219 100644 > --- a/drivers/gpu/drm/i915/gt/selftest_tlb.c > +++ b/drivers/gpu/drm/i915/gt/selftest_tlb.c > @@ -36,8 +36,6 @@ pte_tlbinv(struct intel_context *ce, > u64 length, > struct rnd_state *prng) > { > - const unsigned int pat_index = > - i915_gem_get_pat_index(ce->vm->i915, I915_CACHE_NONE); > struct drm_i915_gem_object *batch; > struct drm_mm_node vb_node; > struct i915_request *rq; > @@ -157,7 +155,8 @@ pte_tlbinv(struct intel_context *ce, > /* Flip the PTE between A and B */ > if (i915_gem_object_is_lmem(vb->obj)) > pte_flags |= PTE_LM; > - ce->vm->insert_entries(ce->vm, &vb_res, pat_index, pte_flags); > + ce->vm->insert_entries(ce->vm, &vb_res, ce->vm->i915->pat_uc, > + pte_flags); > > /* Flush the PTE update to concurrent HW */ > tlbinv(ce->vm, addr & -length, length); > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c b/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c > index d408856ae4c0..e099414d624d 100644 > --- a/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c > +++ b/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c > @@ -991,14 +991,10 @@ static void uc_fw_bind_ggtt(struct intel_uc_fw *uc_fw) > > if (ggtt->vm.raw_insert_entries) > ggtt->vm.raw_insert_entries(&ggtt->vm, vma_res, > - i915_gem_get_pat_index(ggtt->vm.i915, > - I915_CACHE_NONE), > - pte_flags); > + ggtt->vm.i915->pat_uc, pte_flags); > else > ggtt->vm.insert_entries(&ggtt->vm, vma_res, > - i915_gem_get_pat_index(ggtt->vm.i915, > - I915_CACHE_NONE), > - pte_flags); > + ggtt->vm.i915->pat_uc, pte_flags); > } > > static void uc_fw_unbind_ggtt(struct intel_uc_fw *uc_fw) > diff --git a/drivers/gpu/drm/i915/i915_cache.c b/drivers/gpu/drm/i915/i915_cache.c > new file mode 100644 > index 000000000000..7a8002ebd2ec > --- /dev/null > +++ b/drivers/gpu/drm/i915/i915_cache.c > @@ -0,0 +1,59 @@ > +/* > + * SPDX-License-Identifier: MIT > + * > + * Copyright © 2023 Intel Corporation > + */ > + > +#include "i915_cache.h" > +#include "i915_drv.h" > + > +static int find_pat(const struct intel_device_info *info, i915_cache_t mode) > +{ > + int i; > + > + for (i = 0; i < ARRAY_SIZE(info->cache_modes); i++) { > + if (info->cache_modes[i] == mode) > + return i; > + } > + > + return -1; > +} > + > +void i915_cache_init(struct drm_i915_private *i915) > +{ > + int ret; > + > + ret = find_pat(INTEL_INFO(i915), I915_CACHE(UC)); > + WARN_ON(ret < 0); > + drm_info(&i915->drm, "Using PAT index %u for uncached access\n", ret); > + i915->pat_uc = ret; > + > + ret = find_pat(INTEL_INFO(i915), I915_CACHE(WB)); > + WARN_ON(ret < 0); > + drm_info(&i915->drm, "Using PAT index %u for write-back access\n", ret); > + i915->pat_wb = ret; > +} I don't think we need the above two functions. Why don't we just hard code pat_uc and pat_wb in device_info? plus pat_wt too? These are predetermined, and used by KMD only. > +int i915_cache_level_to_pat_and_mode(struct drm_i915_private *i915, > + unsigned int cache_level, > + i915_cache_t *mode) > +{ > + const struct intel_device_info *info = INTEL_INFO(i915); > + i915_cache_t level_to_mode[] = { > + [I915_CACHE_NONE] = I915_CACHE(UC), > + [I915_CACHE_WT] = I915_CACHE(WT), > + [I915_CACHE_LLC] = I915_CACHE(WB), > + }; > + int ret; > + > + if (GRAPHICS_VER(i915) >= 12) > + level_to_mode[I915_CACHE_L3_LLC] = I915_CACHE(WB); > + else > + level_to_mode[I915_CACHE_L3_LLC] = _I915_CACHE(WB, LLC); > + > + ret = find_pat(info, level_to_mode[cache_level]); > + if (ret >= 0 && mode) > + *mode = info->cache_modes[ret]; > + > + return ret; > +} > diff --git a/drivers/gpu/drm/i915/i915_cache.h b/drivers/gpu/drm/i915/i915_cache.h > new file mode 100644 > index 000000000000..0df03f1f01ef > --- /dev/null > +++ b/drivers/gpu/drm/i915/i915_cache.h > @@ -0,0 +1,129 @@ > +/* SPDX-License-Identifier: MIT */ > +/* > + * Copyright © 2023 Intel Corporation > + */ > + > +#ifndef __I915_CACHE_H__ > +#define __I915_CACHE_H__ > + > +#include <linux/types.h> > + > +struct drm_i915_private; > + > +typedef u16 i915_cache_t; > + > +#define I915_CACHE(mode) \ > + (i915_cache_t)(I915_CACHE_MODE_##mode) > + > +#define _I915_CACHE(mode, flag) \ > + (i915_cache_t)((I915_CACHE_MODE_##mode) | ( BIT(8 + I915_CACHE_##flag))) > + > +#define I915_CACHE_MODE(cache) \ > + (unsigned int)(((i915_cache_t)(cache)) & 0xff) > +#define I915_CACHE_FLAGS(cache) \ > + (unsigned int)((((i915_cache_t)(cache) & 0xff00)) >> 16) > + > +/* Cache mode values */ > +#define I915_CACHE_MODE_UNKNOWN (0) > +#define I915_CACHE_MODE_UC (1) > +#define I915_CACHE_MODE_WB (2) > +#define I915_CACHE_MODE_WT (3) > +#define I915_CACHE_MODE_WC (4) Why do you need these CACHE_MODE's? Aren't they the same as i915_cache_level, which need some sort of translation? > + > +/* Mode flag bits */ > +#define I915_CACHE_L3 (0) > +#define I915_CACHE_COH1W (1) > +#define I915_CACHE_COH2W (2) > +#define I915_CACHE_CLOS1 (3) > +#define I915_CACHE_CLOS2 (4) These had been defined in drivers/gpu/drm/i915/gt/intel_gtt.h already, why add new ones? The CLOS ones are not needed in upstream unless we want to support PVC here. > + > +void i915_cache_init(struct drm_i915_private *i915); > +int i915_cache_level_to_pat_and_mode(struct drm_i915_private *i915, > + unsigned int cache_level, > + i915_cache_t *mode); > + > +/* > + * Legacy/kernel internal interface below: > + */ > + > +/** > + * enum i915_cache_level - The supported GTT caching values for system memory > + * pages. > + * > + * These translate to some special GTT PTE bits when binding pages into some > + * address space. It also determines whether an object, or rather its pages are > + * coherent with the GPU, when also reading or writing through the CPU cache > + * with those pages. > + * > + * Userspace can also control this through struct drm_i915_gem_caching. > + */ > +enum i915_cache_level { Shouldn't we completely get rid of this enum now? It should be replaced by INTEL_INFO(i915)->pat_uc/wb/wt. > + /** > + * @I915_CACHE_NONE: > + * > + * GPU access is not coherent with the CPU cache. If the cache is dirty > + * and we need the underlying pages to be coherent with some later GPU > + * access then we need to manually flush the pages. > + * > + * On shared LLC platforms reads and writes through the CPU cache are > + * still coherent even with this setting. See also > + * &drm_i915_gem_object.cache_coherent for more details. Due to this we > + * should only ever use uncached for scanout surfaces, otherwise we end > + * up over-flushing in some places. > + * > + * This is the default on non-LLC platforms. > + */ > + I915_CACHE_NONE = 0, > + /** > + * @I915_CACHE_LLC: > + * > + * GPU access is coherent with the CPU cache. If the cache is dirty, > + * then the GPU will ensure that access remains coherent, when both > + * reading and writing through the CPU cache. GPU writes can dirty the > + * CPU cache. > + * > + * Not used for scanout surfaces. > + * > + * Applies to both platforms with shared LLC(HAS_LLC), and snooping > + * based platforms(HAS_SNOOP). > + * > + * This is the default on shared LLC platforms. The only exception is > + * scanout objects, where the display engine is not coherent with the > + * CPU cache. For such objects I915_CACHE_NONE or I915_CACHE_WT is > + * automatically applied by the kernel in pin_for_display, if userspace > + * has not done so already. > + */ > + I915_CACHE_LLC, > + /** > + * @I915_CACHE_L3_LLC: > + * > + * Explicitly enable the Gfx L3 cache, with coherent LLC. > + * > + * The Gfx L3 sits between the domain specific caches, e.g > + * sampler/render caches, and the larger LLC. LLC is coherent with the > + * GPU, but L3 is only visible to the GPU, so likely needs to be flushed > + * when the workload completes. > + * > + * Not used for scanout surfaces. > + * > + * Only exposed on some gen7 + GGTT. More recent hardware has dropped > + * this explicit setting, where it should now be enabled by default. > + */ > + I915_CACHE_L3_LLC, > + /** > + * @I915_CACHE_WT: > + * > + * Write-through. Used for scanout surfaces. > + * > + * The GPU can utilise the caches, while still having the display engine > + * be coherent with GPU writes, as a result we don't need to flush the > + * CPU caches when moving out of the render domain. This is the default > + * setting chosen by the kernel, if supported by the HW, otherwise we > + * fallback to I915_CACHE_NONE. On the CPU side writes through the CPU > + * cache still need to be flushed, to remain coherent with the display > + * engine. > + */ > + I915_CACHE_WT, > +}; > + > +#endif /* __I915_CACHE_H__ */ > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c > index 76ccd4e03e31..e2da57397770 100644 > --- a/drivers/gpu/drm/i915/i915_debugfs.c > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > @@ -139,48 +139,37 @@ static const char *stringify_vma_type(const struct i915_vma *vma) > return "ppgtt"; > } > > -static const char *i915_cache_level_str(struct drm_i915_gem_object *obj) > +static void obj_cache_str(struct seq_file *m, struct drm_i915_gem_object *obj) > { > - struct drm_i915_private *i915 = obj_to_i915(obj); > + const i915_cache_t cache = obj->cache_mode; > + const unsigned int mode = I915_CACHE_MODE(cache); > + const unsigned long flags = I915_CACHE_FLAGS(cache); > + static const char *mode_str[] = { > + [I915_CACHE_MODE_UC] = "UC", > + [I915_CACHE_MODE_WB] = "WB", > + [I915_CACHE_MODE_WT] = "WT", > + [I915_CACHE_MODE_WC] = "WC", > + }; > + static const char *flag_str[] = { > + [I915_CACHE_L3] = "L3", > + [I915_CACHE_COH1W] = "1-Way-Coherent", > + [I915_CACHE_COH2W] = "2-Way-Coherent", > + [I915_CACHE_CLOS1] = "CLOS1", > + [I915_CACHE_CLOS2] = "CLOS2", > + }; > > - if (IS_METEORLAKE(i915)) { > - switch (obj->pat_index) { > - case 0: return " WB"; > - case 1: return " WT"; > - case 2: return " UC"; > - case 3: return " WB (1-Way Coh)"; > - case 4: return " WB (2-Way Coh)"; > - default: return " not defined"; > - } > - } else if (IS_PONTEVECCHIO(i915)) { > - switch (obj->pat_index) { > - case 0: return " UC"; > - case 1: return " WC"; > - case 2: return " WT"; > - case 3: return " WB"; > - case 4: return " WT (CLOS1)"; > - case 5: return " WB (CLOS1)"; > - case 6: return " WT (CLOS2)"; > - case 7: return " WT (CLOS2)"; > - default: return " not defined"; > - } > - } else if (GRAPHICS_VER(i915) >= 12) { > - switch (obj->pat_index) { > - case 0: return " WB"; > - case 1: return " WC"; > - case 2: return " WT"; > - case 3: return " UC"; > - default: return " not defined"; > - } > + if (mode == I915_CACHE_MODE_UNKNOWN || mode > ARRAY_SIZE(mode_str)) { > + if (obj->pat_set_by_user) > + seq_printf(m, " PAT-%u", obj->pat_index); > + else > + seq_printf(m, " PAT-%u-???", obj->pat_index); > } else { > - switch (obj->pat_index) { > - case 0: return " UC"; > - case 1: return HAS_LLC(i915) ? > - " LLC" : " snooped"; > - case 2: return " L3+LLC"; > - case 3: return " WT"; > - default: return " not defined"; > - } > + unsigned long bit; > + > + seq_printf(m, " %s", mode_str[mode]); > + > + for_each_set_bit(bit, &flags, sizeof(i915_cache_t)) > + seq_printf(m, "-%s", flag_str[bit]); > } > } > > @@ -190,17 +179,23 @@ i915_debugfs_describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) > struct i915_vma *vma; > int pin_count = 0; > > - seq_printf(m, "%pK: %c%c%c %8zdKiB %02x %02x %s%s%s", > + seq_printf(m, "%pK: %c%c%c %8zdKiB %02x %02x", > &obj->base, > get_tiling_flag(obj), > get_global_flag(obj), > get_pin_mapped_flag(obj), > obj->base.size / 1024, > obj->read_domains, > - obj->write_domain, > - i915_cache_level_str(obj), > - obj->mm.dirty ? " dirty" : "", > - obj->mm.madv == I915_MADV_DONTNEED ? " purgeable" : ""); > + obj->write_domain); > + > + obj_cache_str(m, obj); > + > + if (obj->mm.dirty) > + seq_puts(m, " dirty"); > + > + if (obj->mm.madv == I915_MADV_DONTNEED) > + seq_puts(m, " purgeable"); > + > if (obj->base.name) > seq_printf(m, " (name: %d)", obj->base.name); > > diff --git a/drivers/gpu/drm/i915/i915_driver.c b/drivers/gpu/drm/i915/i915_driver.c > index 222d0a1f3b55..deab26752ba4 100644 > --- a/drivers/gpu/drm/i915/i915_driver.c > +++ b/drivers/gpu/drm/i915/i915_driver.c > @@ -80,6 +80,7 @@ > #include "soc/intel_dram.h" > #include "soc/intel_gmch.h" > > +#include "i915_cache.h" > #include "i915_debugfs.h" > #include "i915_driver.h" > #include "i915_drm_client.h" > @@ -267,6 +268,8 @@ static int i915_driver_early_probe(struct drm_i915_private *dev_priv) > > intel_detect_preproduction_hw(dev_priv); > > + i915_cache_init(dev_priv); > + > return 0; > > err_rootgt: > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > index b4cf6f0f636d..cb1c0c9d98ef 100644 > --- a/drivers/gpu/drm/i915/i915_drv.h > +++ b/drivers/gpu/drm/i915/i915_drv.h > @@ -251,6 +251,9 @@ struct drm_i915_private { > unsigned int hpll_freq; > unsigned int czclk_freq; > > + unsigned int pat_uc; > + unsigned int pat_wb; How about making these part of INTEL_INFO(i915)? They are predetermined, no need to be dynamic. > + > /** > * wq - Driver workqueue for GEM. > * > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > index 7ae42f746cc2..9aae75862e6f 100644 > --- a/drivers/gpu/drm/i915/i915_gem.c > +++ b/drivers/gpu/drm/i915/i915_gem.c > @@ -422,9 +422,7 @@ i915_gem_gtt_pread(struct drm_i915_gem_object *obj, > ggtt->vm.insert_page(&ggtt->vm, > i915_gem_object_get_dma_address(obj, > offset >> PAGE_SHIFT), > - node.start, > - i915_gem_get_pat_index(i915, > - I915_CACHE_NONE), 0); > + node.start, i915->pat_uc, 0); > } else { > page_base += offset & PAGE_MASK; > } > @@ -603,9 +601,7 @@ i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj, > ggtt->vm.insert_page(&ggtt->vm, > i915_gem_object_get_dma_address(obj, > offset >> PAGE_SHIFT), > - node.start, > - i915_gem_get_pat_index(i915, > - I915_CACHE_NONE), 0); > + node.start, i915->pat_uc, 0); > wmb(); /* flush modifications to the GGTT (insert_page) */ > } else { > page_base += offset & PAGE_MASK; > @@ -1148,19 +1144,6 @@ int i915_gem_init(struct drm_i915_private *dev_priv) > unsigned int i; > int ret; > > - /* > - * In the proccess of replacing cache_level with pat_index a tricky > - * dependency is created on the definition of the enum i915_cache_level. > - * in case this enum is changed, PTE encode would be broken. > - * Add a WARNING here. And remove when we completely quit using this > - * enum > - */ > - BUILD_BUG_ON(I915_CACHE_NONE != 0 || > - I915_CACHE_LLC != 1 || > - I915_CACHE_L3_LLC != 2 || > - I915_CACHE_WT != 3 || > - I915_MAX_CACHE_LEVEL != 4); > - > /* We need to fallback to 4K pages if host doesn't support huge gtt. */ > if (intel_vgpu_active(dev_priv) && !intel_vgpu_has_huge_gtt(dev_priv)) > RUNTIME_INFO(dev_priv)->page_sizes = I915_GTT_PAGE_SIZE_4K; > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c > index 4749f99e6320..fad336a45699 100644 > --- a/drivers/gpu/drm/i915/i915_gpu_error.c > +++ b/drivers/gpu/drm/i915/i915_gpu_error.c > @@ -1122,14 +1122,11 @@ i915_vma_coredump_create(const struct intel_gt *gt, > mutex_lock(&ggtt->error_mutex); > if (ggtt->vm.raw_insert_page) > ggtt->vm.raw_insert_page(&ggtt->vm, dma, slot, > - i915_gem_get_pat_index(gt->i915, > - I915_CACHE_NONE), > + ggtt->vm.i915->pat_uc, > 0); > else > ggtt->vm.insert_page(&ggtt->vm, dma, slot, > - i915_gem_get_pat_index(gt->i915, > - I915_CACHE_NONE), > - 0); > + ggtt->vm.i915->pat_uc, 0); > mb(); > > s = io_mapping_map_wc(&ggtt->iomap, slot, PAGE_SIZE); > diff --git a/drivers/gpu/drm/i915/i915_pci.c b/drivers/gpu/drm/i915/i915_pci.c > index 3d7a5db9833b..fbdce31afeb1 100644 > --- a/drivers/gpu/drm/i915/i915_pci.c > +++ b/drivers/gpu/drm/i915/i915_pci.c > @@ -32,6 +32,7 @@ > #include "gt/intel_sa_media.h" > #include "gem/i915_gem_object_types.h" > > +#include "i915_cache.h" > #include "i915_driver.h" > #include "i915_drv.h" > #include "i915_pci.h" > @@ -46,36 +47,42 @@ __diag_ignore_all("-Woverride-init", "Allow overriding inherited members"); > .__runtime.graphics.ip.ver = (x), \ > .__runtime.media.ip.ver = (x) > > -#define LEGACY_CACHELEVEL \ > - .cachelevel_to_pat = { \ > - [I915_CACHE_NONE] = 0, \ > - [I915_CACHE_LLC] = 1, \ > - [I915_CACHE_L3_LLC] = 2, \ > - [I915_CACHE_WT] = 3, \ > +/* TODO/QQQ index 1 & 2 */ > +#define LEGACY_CACHE_MODES \ I was thinking to just put the PAT settings here, instead of cache_modes, simply .pat = {\ GEN8_PPAT_WB, \ GEN8_PPAT_WC, \ GEN8_PPAT_WT, \ GEN8_PPAT_UC, } > + .cache_modes = { \ > + [0] = I915_CACHE(UC), \ > + [1] = I915_CACHE(WB), \ > + [2] = _I915_CACHE(WB, L3), \ > + [3] = I915_CACHE(WT), \ > } > > -#define TGL_CACHELEVEL \ > - .cachelevel_to_pat = { \ > - [I915_CACHE_NONE] = 3, \ > - [I915_CACHE_LLC] = 0, \ > - [I915_CACHE_L3_LLC] = 0, \ > - [I915_CACHE_WT] = 2, \ > +#define GEN12_CACHE_MODES \ > + .cache_modes = { \ > + [0] = I915_CACHE(WB), \ > + [1] = I915_CACHE(WC), \ > + [2] = I915_CACHE(WT), \ > + [3] = I915_CACHE(UC), \ > } > > -#define PVC_CACHELEVEL \ > - .cachelevel_to_pat = { \ > - [I915_CACHE_NONE] = 0, \ > - [I915_CACHE_LLC] = 3, \ > - [I915_CACHE_L3_LLC] = 3, \ > - [I915_CACHE_WT] = 2, \ > +#define PVC_CACHE_MODES \ > + .cache_modes = { \ > + [0] = I915_CACHE(UC), \ > + [1] = I915_CACHE(WC), \ > + [2] = I915_CACHE(WT), \ > + [3] = I915_CACHE(WB), \ > + [4] = _I915_CACHE(WT, CLOS1), \ > + [5] = _I915_CACHE(WB, CLOS1), \ > + [6] = _I915_CACHE(WT, CLOS2), \ > + [7] = _I915_CACHE(WB, CLOS2), \ > } .pat = {\ GEN8_PPAT_UC, \ GEN8_PPAT_WC, \ GEN8_PPAT_WT, \ GEN8_PPAT_WB, \ GEN12_PPAT_CLOS(1) | GEN8_PPAT_WT, \ GEN12_PPAT_CLOS(1) | GEN8_PPAT_WB, \ GEN12_PPAT_CLOS(2) | GEN8_PPAT_WT, \ GEN12_PPAT_CLOS(2) | GEN8_PPAT_WB, \ } > > -#define MTL_CACHELEVEL \ > - .cachelevel_to_pat = { \ > - [I915_CACHE_NONE] = 2, \ > - [I915_CACHE_LLC] = 3, \ > - [I915_CACHE_L3_LLC] = 3, \ > - [I915_CACHE_WT] = 1, \ > +#define MTL_CACHE_MODES \ > + .cache_modes = { \ > + [0] = I915_CACHE(WB), \ > + [1] = I915_CACHE(WT), \ > + [2] = I915_CACHE(UC), \ > + [3] = _I915_CACHE(WB, COH1W), \ > + [4] = _I915_CACHE(WB, COH2W), \ > } .pat = {\ MTL_PPAT_L4_0_WB, \ MTL_PPAT_L4_1_WT, \ MTL_PPAT_L4_3_UC, \ MTL_PPAT_L4_0_WB | MTL_2_COH_1W, \ MTL_PPAT_L4_0_WB | MTL_3_COH_2W, \ } > /* Keep in gen based order, and chronological order within a gen */ > @@ -100,7 +107,7 @@ __diag_ignore_all("-Woverride-init", "Allow overriding inherited members"); > .max_pat_index = 3, \ > GEN_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > #define I845_FEATURES \ > GEN(2), \ > @@ -115,7 +122,7 @@ __diag_ignore_all("-Woverride-init", "Allow overriding inherited members"); > .max_pat_index = 3, \ > GEN_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > static const struct intel_device_info i830_info = { > I830_FEATURES, > @@ -148,7 +155,7 @@ static const struct intel_device_info i865g_info = { > .max_pat_index = 3, \ > GEN_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > static const struct intel_device_info i915g_info = { > GEN3_FEATURES, > @@ -211,7 +218,7 @@ static const struct intel_device_info pnv_m_info = { > .max_pat_index = 3, \ > GEN_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > static const struct intel_device_info i965g_info = { > GEN4_FEATURES, > @@ -255,7 +262,7 @@ static const struct intel_device_info gm45_info = { > .max_pat_index = 3, \ > GEN_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > static const struct intel_device_info ilk_d_info = { > GEN5_FEATURES, > @@ -285,7 +292,7 @@ static const struct intel_device_info ilk_m_info = { > .__runtime.ppgtt_size = 31, \ > GEN_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > #define SNB_D_PLATFORM \ > GEN6_FEATURES, \ > @@ -333,7 +340,7 @@ static const struct intel_device_info snb_m_gt2_info = { > .__runtime.ppgtt_size = 31, \ > GEN_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > #define IVB_D_PLATFORM \ > GEN7_FEATURES, \ > @@ -390,7 +397,7 @@ static const struct intel_device_info vlv_info = { > .__runtime.platform_engine_mask = BIT(RCS0) | BIT(VCS0) | BIT(BCS0), > GEN_DEFAULT_PAGE_SIZES, > GEN_DEFAULT_REGIONS, > - LEGACY_CACHELEVEL, > + LEGACY_CACHE_MODES > }; > > #define G75_FEATURES \ > @@ -476,7 +483,7 @@ static const struct intel_device_info chv_info = { > .has_coherent_ggtt = false, > GEN_DEFAULT_PAGE_SIZES, > GEN_DEFAULT_REGIONS, > - LEGACY_CACHELEVEL, > + LEGACY_CACHE_MODES > }; > > #define GEN9_DEFAULT_PAGE_SIZES \ > @@ -539,7 +546,7 @@ static const struct intel_device_info skl_gt4_info = { > .max_pat_index = 3, \ > GEN9_DEFAULT_PAGE_SIZES, \ > GEN_DEFAULT_REGIONS, \ > - LEGACY_CACHELEVEL > + LEGACY_CACHE_MODES > > static const struct intel_device_info bxt_info = { > GEN9_LP_FEATURES, > @@ -643,7 +650,7 @@ static const struct intel_device_info jsl_info = { > #define GEN12_FEATURES \ > GEN11_FEATURES, \ > GEN(12), \ > - TGL_CACHELEVEL, \ > + GEN12_CACHE_MODES, \ > .has_global_mocs = 1, \ > .has_pxp = 1, \ > .max_pat_index = 3 > @@ -711,7 +718,7 @@ static const struct intel_device_info adl_p_info = { > .__runtime.graphics.ip.ver = 12, \ > .__runtime.graphics.ip.rel = 50, \ > XE_HP_PAGE_SIZES, \ > - TGL_CACHELEVEL, \ > + GEN12_CACHE_MODES, \ > .dma_mask_size = 46, \ > .has_3d_pipeline = 1, \ > .has_64bit_reloc = 1, \ > @@ -806,7 +813,7 @@ static const struct intel_device_info pvc_info = { > BIT(VCS0) | > BIT(CCS0) | BIT(CCS1) | BIT(CCS2) | BIT(CCS3), > .require_force_probe = 1, > - PVC_CACHELEVEL, > + PVC_CACHE_MODES > }; > > static const struct intel_gt_definition xelpmp_extra_gt[] = { > @@ -841,7 +848,7 @@ static const struct intel_device_info mtl_info = { > .__runtime.memory_regions = REGION_SMEM | REGION_STOLEN_LMEM, > .__runtime.platform_engine_mask = BIT(RCS0) | BIT(BCS0) | BIT(CCS0), > .require_force_probe = 1, > - MTL_CACHELEVEL, > + MTL_CACHE_MODES > }; > > #undef PLATFORM > diff --git a/drivers/gpu/drm/i915/intel_device_info.h b/drivers/gpu/drm/i915/intel_device_info.h > index 069291b3bd37..5cbae7c2ee30 100644 > --- a/drivers/gpu/drm/i915/intel_device_info.h > +++ b/drivers/gpu/drm/i915/intel_device_info.h > @@ -27,6 +27,8 @@ > > #include <uapi/drm/i915_drm.h> > > +#include "i915_cache.h" > + > #include "intel_step.h" > > #include "display/intel_display_device.h" > @@ -248,8 +250,8 @@ struct intel_device_info { > */ > const struct intel_runtime_info __runtime; > > - u32 cachelevel_to_pat[I915_MAX_CACHE_LEVEL]; > - u32 max_pat_index; > + i915_cache_t cache_modes[9]; u32 pat[16]; See https://gfxspecs.intel.com/Predator/Home/Index/63019, there are PAT[3..0] -Fei > + unsigned int max_pat_index; > }; > > struct intel_driver_caps { > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem.c b/drivers/gpu/drm/i915/selftests/i915_gem.c > index 61da4ed9d521..e620f73793a5 100644 > --- a/drivers/gpu/drm/i915/selftests/i915_gem.c > +++ b/drivers/gpu/drm/i915/selftests/i915_gem.c > @@ -57,10 +57,7 @@ static void trash_stolen(struct drm_i915_private *i915) > u32 __iomem *s; > int x; > > - ggtt->vm.insert_page(&ggtt->vm, dma, slot, > - i915_gem_get_pat_index(i915, > - I915_CACHE_NONE), > - 0); > + ggtt->vm.insert_page(&ggtt->vm, dma, slot, i915->pat_uc, 0); > > s = io_mapping_map_atomic_wc(&ggtt->iomap, slot); > for (x = 0; x < PAGE_SIZE / sizeof(u32); x++) { > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_evict.c b/drivers/gpu/drm/i915/selftests/i915_gem_evict.c > index f8fe3681c3dc..658a5b59545e 100644 > --- a/drivers/gpu/drm/i915/selftests/i915_gem_evict.c > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_evict.c > @@ -246,7 +246,7 @@ static int igt_evict_for_cache_color(void *arg) > struct drm_mm_node target = { > .start = I915_GTT_PAGE_SIZE * 2, > .size = I915_GTT_PAGE_SIZE, > - .color = i915_gem_get_pat_index(gt->i915, I915_CACHE_LLC), > + .color = I915_CACHE(WB), > }; > struct drm_i915_gem_object *obj; > struct i915_vma *vma; > @@ -309,7 +309,7 @@ static int igt_evict_for_cache_color(void *arg) > /* Attempt to remove the first *pinned* vma, by removing the (empty) > * neighbour -- this should fail. > */ > - target.color = i915_gem_get_pat_index(gt->i915, I915_CACHE_L3_LLC); > + target.color = _I915_CACHE(WB, LLC); > > mutex_lock(&ggtt->vm.mutex); > err = i915_gem_evict_for_node(&ggtt->vm, NULL, &target, 0); > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > index 5c397a2df70e..a24585784f75 100644 > --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > @@ -135,7 +135,7 @@ fake_dma_object(struct drm_i915_private *i915, u64 size) > > obj->write_domain = I915_GEM_DOMAIN_CPU; > obj->read_domains = I915_GEM_DOMAIN_CPU; > - obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE); > + obj->pat_index = i915->pat_uc; > > /* Preallocate the "backing storage" */ > if (i915_gem_object_pin_pages_unlocked(obj)) > @@ -358,10 +358,8 @@ static int lowlevel_hole(struct i915_address_space *vm, > mock_vma_res->start = addr; > > with_intel_runtime_pm(vm->gt->uncore->rpm, wakeref) > - vm->insert_entries(vm, mock_vma_res, > - i915_gem_get_pat_index(vm->i915, > - I915_CACHE_NONE), > - 0); > + vm->insert_entries(vm, mock_vma_res, > + vm->i915->pat_uc, 0); > } > count = n; > > @@ -1379,10 +1377,7 @@ static int igt_ggtt_page(void *arg) > > ggtt->vm.insert_page(&ggtt->vm, > i915_gem_object_get_dma_address(obj, 0), > - offset, > - i915_gem_get_pat_index(i915, > - I915_CACHE_NONE), > - 0); > + offset, ggtt->vm.i915->pat_uc, 0); > } > > order = i915_random_order(count, &prng); > diff --git a/drivers/gpu/drm/i915/selftests/intel_memory_region.c b/drivers/gpu/drm/i915/selftests/intel_memory_region.c > index d985d9bae2e8..b82fe0ef8cd7 100644 > --- a/drivers/gpu/drm/i915/selftests/intel_memory_region.c > +++ b/drivers/gpu/drm/i915/selftests/intel_memory_region.c > @@ -1070,9 +1070,7 @@ static int igt_lmem_write_cpu(void *arg) > /* Put the pages into a known state -- from the gpu for added fun */ > intel_engine_pm_get(engine); > err = intel_context_migrate_clear(engine->gt->migrate.context, NULL, > - obj->mm.pages->sgl, > - i915_gem_get_pat_index(i915, > - I915_CACHE_NONE), > + obj->mm.pages->sgl, i915->pat_uc, > true, 0xdeadbeaf, &rq); > if (rq) { > dma_resv_add_fence(obj->base.resv, &rq->fence, > diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > index 09d4bbcdcdbf..ad778842cba2 100644 > --- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c > +++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > @@ -126,7 +126,12 @@ struct drm_i915_private *mock_gem_device(void) > struct drm_i915_private *i915; > struct intel_device_info *i915_info; > struct pci_dev *pdev; > - unsigned int i; > + static const i915_cache_t legacy_cache_modes[] = { > + [0] = I915_CACHE(UC), > + [1] = I915_CACHE(WB), > + [2] = _I915_CACHE(WB, L3), > + [3] = I915_CACHE(WT), > + }; > int ret; > > pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); > @@ -187,8 +192,7 @@ struct drm_i915_private *mock_gem_device(void) > /* simply use legacy cache level for mock device */ > i915_info = (struct intel_device_info *)INTEL_INFO(i915); > i915_info->max_pat_index = 3; > - for (i = 0; i < I915_MAX_CACHE_LEVEL; i++) > - i915_info->cachelevel_to_pat[i] = i; > + memcpy(i915_info->cache_modes, legacy_cache_modes, sizeof(legacy_cache_modes)); > > intel_memory_regions_hw_probe(i915); > > -- > 2.39.2
On 29/06/2023 06:39, Yang, Fei wrote: > > From: Tvrtko Ursulin <tvrtko.ursulin@intel.com> > > > > Informal commit message for now. > > > > I got a bit impatient and curious to see if the idea we discussed would > > work so sketched something out. I think it is what I was describing back > > then.. > > Oops, you beat me on this, shame on me. Not at all, thanks for a detailed look into my attempt! You were right with most of the observations and I think I've done most of them and now it even passes BAT (trybot). Because I rushed the v1 to sketch out the idea, indeed the double caching abstraction and obj->cache_mode made no sense. It's all ripped out now. Only thing I disagree with is to use hardware bits in the lookup table. Maybe in a different table, one which could be used for PAT index to hw bits lookup in the PTE encode functions. But for the higher level in the driver I need the abstracted view so driver can understand the modes and ask questions about them. Let me go through the individual comments below to see if there is something more to cover. > > So high level idea is to teach the driver what caching modes are hidden > > behind PAT indices. Given you already had that in static tables, if we > > just turn the tables a bit around and add a driver abstraction of caching > > modes this is what happens: > > > > * We can lose the ugly runtime i915_gem_get_pat_index. > > > > * We can have a smarter i915_gem_object_has_cache_level, which now can > > use the above mentioned table to understand the caching modes and so > > does not have to pessimistically return true for _any_ input when user > > has set the PAT index. This may improve things even for MTL. > > > > * We can simplify the debugfs printout to be platform agnostic. > > > > * We are perhaps opening the door to un-regress the dodgy addition > > made to i915_gem_object_can_bypass_llc? See QQQ/FIXME in the patch. > > > > I hope I did not forget anything, but anyway, please have a read and see > > what you think. I think it has potential. > > > > Proper commit message can come later. > > > > Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> > > Cc: Fei Yang <fei.yang@intel.com> > > --- > > drivers/gpu/drm/i915/Makefile | 1 + > > drivers/gpu/drm/i915/gem/i915_gem_domain.c | 34 ++--- > > .../gpu/drm/i915/gem/i915_gem_execbuffer.c | 13 +- > > drivers/gpu/drm/i915/gem/i915_gem_mman.c | 10 +- > > drivers/gpu/drm/i915/gem/i915_gem_object.c | 78 ++++------- > > drivers/gpu/drm/i915/gem/i915_gem_object.h | 18 ++- > > .../gpu/drm/i915/gem/i915_gem_object_types.h | 99 +------------- > > drivers/gpu/drm/i915/gem/i915_gem_stolen.c | 7 +- > > drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c | 26 ++-- > > .../gpu/drm/i915/gem/selftests/huge_pages.c | 2 +- > > drivers/gpu/drm/i915/gt/gen6_ppgtt.c | 4 +- > > drivers/gpu/drm/i915/gt/gen8_ppgtt.c | 13 +- > > drivers/gpu/drm/i915/gt/intel_ggtt.c | 9 +- > > drivers/gpu/drm/i915/gt/intel_migrate.c | 11 +- > > drivers/gpu/drm/i915/gt/selftest_migrate.c | 9 +- > > drivers/gpu/drm/i915/gt/selftest_reset.c | 14 +- > > drivers/gpu/drm/i915/gt/selftest_timeline.c | 5 +- > > drivers/gpu/drm/i915/gt/selftest_tlb.c | 5 +- > > drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c | 8 +- > > drivers/gpu/drm/i915/i915_cache.c | 59 ++++++++ > > drivers/gpu/drm/i915/i915_cache.h | 129 ++++++++++++++++++ > > drivers/gpu/drm/i915/i915_debugfs.c | 83 ++++++----- > > drivers/gpu/drm/i915/i915_driver.c | 3 + > > drivers/gpu/drm/i915/i915_drv.h | 3 + > > drivers/gpu/drm/i915/i915_gem.c | 21 +-- > > drivers/gpu/drm/i915/i915_gpu_error.c | 7 +- > > drivers/gpu/drm/i915/i915_pci.c | 83 +++++------ > > drivers/gpu/drm/i915/intel_device_info.h | 6 +- > > drivers/gpu/drm/i915/selftests/i915_gem.c | 5 +- > > .../gpu/drm/i915/selftests/i915_gem_evict.c | 4 +- > > drivers/gpu/drm/i915/selftests/i915_gem_gtt.c | 13 +- > > .../drm/i915/selftests/intel_memory_region.c | 4 +- > > .../gpu/drm/i915/selftests/mock_gem_device.c | 10 +- > > 33 files changed, 415 insertions(+), 381 deletions(-) > > create mode 100644 drivers/gpu/drm/i915/i915_cache.c > > create mode 100644 drivers/gpu/drm/i915/i915_cache.h > > > > diff --git a/drivers/gpu/drm/i915/Makefile > b/drivers/gpu/drm/i915/Makefile > > index 2be9dd960540..2c3da8f0c78e 100644 > > --- a/drivers/gpu/drm/i915/Makefile > > +++ b/drivers/gpu/drm/i915/Makefile > > @@ -30,6 +30,7 @@ subdir-ccflags-y += -I$(srctree)/$(src) > > # core driver code > > i915-y += i915_driver.o \ > > i915_drm_client.o \ > > + i915_cache.o \ > > i915_config.o \ > > i915_getparam.o \ > > i915_ioctl.o \ > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_domain.c > b/drivers/gpu/drm/i915/gem/i915_gem_domain.c > > index dfaaa8b66ac3..49bfae45390f 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_domain.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_domain.c > > @@ -8,6 +8,7 @@ > > #include "display/intel_frontbuffer.h" > > #include "gt/intel_gt.h" > > > > +#include "i915_cache.h" > > #include "i915_drv.h" > > #include "i915_gem_clflush.h" > > #include "i915_gem_domain.h" > > @@ -27,15 +28,8 @@ static bool gpu_write_needs_clflush(struct > drm_i915_gem_object *obj) > > if (IS_DGFX(i915)) > > return false; > > > > - /* > > - * For objects created by userspace through GEM_CREATE with pat_index > > - * set by set_pat extension, i915_gem_object_has_cache_level() will > > - * always return true, because the coherency of such object is > managed > > - * by userspace. Othereise the call here would fall back to checking > > - * whether the object is un-cached or write-through. > > - */ > > - return !(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) || > > - i915_gem_object_has_cache_level(obj, I915_CACHE_WT)); > > + return i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != > 1 && > > + i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WT) != 1; > > Why is it necessary to define I915_CACHE_MODE's while there is already > i915_cache_level? > I thought we wanted to get rid of such abstractions instead of adding more. Indeed. Dropped. > > This patch also introduced INTEL_INFO(i915)->cache_modes, why don't we > directly add the > platform specific PAT there? For example, add the following for MTL, > > INTEL_INFO(i915)->pat[] = { > [0] = MTL_PPAT_L4_0_WB, \ > [1] = MTL_PPAT_L4_1_WT, \ > [2] = MTL_PPAT_L4_3_UC, \ > [3] = MTL_PPAT_L4_0_WB | MTL_2_COH_1W, \ > [4] = MTL_PPAT_L4_0_WB | MTL_3_COH_2W, \ > } > > Everything here has already been defined, no need to introduce new macros. > > This can also be used to initialize the PAT index registers, like in > xelpg_setup_private_ppat() and xelpmp_setup_private_ppat(). As exaplained above, I think a 2nd table could be used for this. > > } > > > > bool i915_gem_cpu_write_needs_clflush(struct drm_i915_gem_object *obj) > > @@ -272,15 +266,18 @@ i915_gem_object_set_to_gtt_domain(struct > drm_i915_gem_object *obj, bool write) > > int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, > > enum i915_cache_level cache_level) > > s/enum i915_cache_level cache_level/unsigned int pat_index > > This function is for KMD objects only, I don't think we even need to keep > the i915_cache_level, simply passing in INTEL_INFO(i915)->pat_uc/wb/wt is > good enough. Yep, it works on the new i915_cache_t in v2. > > { > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > + i915_cache_t mode; > > int ret; > > > > - /* > > - * For objects created by userspace through GEM_CREATE with pat_index > > - * set by set_pat extension, simply return 0 here without touching > > - * the cache setting, because such objects should have an immutable > > - * cache setting by desgin and always managed by userspace. > > - */ > > - if (i915_gem_object_has_cache_level(obj, cache_level)) > > + if (obj->pat_set_by_user) > > + return -EINVAL; > > I don't think this condition would ever be true, but okay to keep it. Could be that I got confused. Or it would be best to add a drm_err or drm_WARN_ON_ONCE just to guard against any surprises. > > + ret = i915_cache_level_to_pat_and_mode(i915, cache_level, &mode); > > + if (ret < 0) > > + return -EINVAL; > > + > > + if (mode == obj->cache_mode) > > The above lines can be just one line, > if (pat_index == obj->pat_index) Yep done. > > return 0; > > > > ret = i915_gem_object_wait(obj, > > @@ -326,10 +323,9 @@ int i915_gem_get_caching_ioctl(struct drm_device > *dev, void *data, > > goto out; > > } > > > > - if (i915_gem_object_has_cache_level(obj, I915_CACHE_LLC) || > > - i915_gem_object_has_cache_level(obj, I915_CACHE_L3_LLC)) > > + if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WB)) > > args->caching = I915_CACHING_CACHED; > > - else if (i915_gem_object_has_cache_level(obj, I915_CACHE_WT)) > > + else if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WT)) > > args->caching = I915_CACHING_DISPLAY; > > else > > args->caching = I915_CACHING_NONE; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > index d3208a325614..ee85221fa6eb 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > @@ -640,15 +640,9 @@ static inline int use_cpu_reloc(const struct > reloc_cache *cache, > > if (DBG_FORCE_RELOC == FORCE_GTT_RELOC) > > return false; > > > > - /* > > - * For objects created by userspace through GEM_CREATE with pat_index > > - * set by set_pat extension, i915_gem_object_has_cache_level() always > > - * return true, otherwise the call would fall back to checking > whether > > - * the object is un-cached. > > - */ > > return (cache->has_llc || > > obj->cache_dirty || > > - !i915_gem_object_has_cache_level(obj, I915_CACHE_NONE)); > > + i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1); > > } > > > > static int eb_reserve_vma(struct i915_execbuffer *eb, > > @@ -1329,10 +1323,7 @@ static void *reloc_iomap(struct i915_vma *batch, > > if (drm_mm_node_allocated(&cache->node)) { > > ggtt->vm.insert_page(&ggtt->vm, > > i915_gem_object_get_dma_address(obj, page), > > - offset, > > - i915_gem_get_pat_index(ggtt->vm.i915, > > - I915_CACHE_NONE), > > - 0); > > + offset, eb->i915->pat_uc, 0); > > } else { > > offset += page << PAGE_SHIFT; > > } > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_mman.c > b/drivers/gpu/drm/i915/gem/i915_gem_mman.c > > index aa4d842d4c5a..5e21aedb02d2 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_mman.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_mman.c > > @@ -386,13 +386,11 @@ static vm_fault_t vm_fault_gtt(struct vm_fault > *vmf) > > /* > > * For objects created by userspace through GEM_CREATE with pat_index > > * set by set_pat extension, coherency is managed by userspace, make > > - * sure we don't fail handling the vm fault by calling > > - * i915_gem_object_has_cache_level() which always return true for > such > > - * objects. Otherwise this helper function would fall back to > checking > > - * whether the object is un-cached. > > + * sure we don't fail handling the vm fault by making sure that we > > + * know the object is uncached or that we have LLC. > > */ > > - if (!(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) || > > - HAS_LLC(i915))) { > > + if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1 && > > + !HAS_LLC(i915)) { > > ret = -EFAULT; > > goto err_unpin; > > } > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.c > b/drivers/gpu/drm/i915/gem/i915_gem_object.c > > index 0004d5fa7cc2..52c6c5f09bdd 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_object.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.c > > @@ -45,33 +45,6 @@ static struct kmem_cache *slab_objects; > > > > static const struct drm_gem_object_funcs i915_gem_object_funcs; > > > > -unsigned int i915_gem_get_pat_index(struct drm_i915_private *i915, > > - enum i915_cache_level level) > > -{ > > - if (drm_WARN_ON(&i915->drm, level >= I915_MAX_CACHE_LEVEL)) > > - return 0; > > - > > - return INTEL_INFO(i915)->cachelevel_to_pat[level]; > > -} > > Yes, this can be removed. INTEL_INFO(i915)->pat_uc/wb/wt should be > sufficient, > > > - > > -bool i915_gem_object_has_cache_level(const struct > drm_i915_gem_object *obj, > > - enum i915_cache_level lvl) > > -{ > > If we had INTEL_INFO(i915)->pat[] setup, it would be easier just to keep > this > function, because we can simply check the cache policy bit field in > INTEL_INFO(i915)->pat[obj->pat_index] to see whether it is cached, uncached, > or write-through. Yep, also done. > See Bspec https://gfxspecs.intel.com/Predator/Home/Index/44235 > For MTL check bit[3:2], for other gen12 platforms check bit[1:0] > > > - /* > > - * In case the pat_index is set by user space, this kernel mode > > - * driver should leave the coherency to be managed by user space, > > - * simply return true here. > > - */ > > - if (obj->pat_set_by_user) > > - return true; > > - > > - /* > > - * Otherwise the pat_index should have been converted from > cache_level > > - * so that the following comparison is valid. > > - */ > > - return obj->pat_index == i915_gem_get_pat_index(obj_to_i915(obj), > lvl); > > -} > > - > > struct drm_i915_gem_object *i915_gem_object_alloc(void) > > { > > struct drm_i915_gem_object *obj; > > @@ -144,6 +117,24 @@ void __i915_gem_object_fini(struct > drm_i915_gem_object *obj) > > dma_resv_fini(&obj->base._resv); > > } > > > > +void __i915_gem_object_update_coherency(struct drm_i915_gem_object *obj) > > +{ > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > + const unsigned int mode = I915_CACHE_MODE(obj->cache_mode); > > obj->cache_mode seems to be redundant if we have INTEL_INFO(i915)->pat[] and > obj->pat_index. Ditto. > > > + > > + if (!(mode == I915_CACHE_MODE_UNKNOWN || mode == I915_CACHE_MODE_UC)) > > + obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | > > + I915_BO_CACHE_COHERENT_FOR_WRITE); > > + else if (mode != I915_CACHE_MODE_UNKNOWN && HAS_LLC(i915)) > > + obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; > > + else > > + obj->cache_coherent = 0; > > + > > + obj->cache_dirty = > > + !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && > > + !IS_DGFX(i915); > > +} > > + > > /** > > * i915_gem_object_set_cache_coherency - Mark up the object's > coherency levels > > * for a given cache_level > > @@ -154,20 +145,15 @@ void i915_gem_object_set_cache_coherency(struct > drm_i915_gem_object *obj, > > unsigned int cache_level) > > { > > struct drm_i915_private *i915 = to_i915(obj->base.dev); > > + i915_cache_t mode; > > + int found; > > > > - obj->pat_index = i915_gem_get_pat_index(i915, cache_level); > > + found = i915_cache_level_to_pat_and_mode(i915, cache_level, &mode); > > + GEM_WARN_ON(found < 0); > > + obj->pat_index = found; > > + obj->cache_mode = mode; > > > > - if (cache_level != I915_CACHE_NONE) > > - obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | > > - I915_BO_CACHE_COHERENT_FOR_WRITE); > > - else if (HAS_LLC(i915)) > > - obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; > > - else > > - obj->cache_coherent = 0; > > - > > - obj->cache_dirty = > > - !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && > > - !IS_DGFX(i915); > > + __i915_gem_object_update_coherency(obj); > > } > > > > /** > > @@ -187,18 +173,9 @@ void i915_gem_object_set_pat_index(struct > drm_i915_gem_object *obj, > > return; > > > > obj->pat_index = pat_index; > > + obj->cache_mode = INTEL_INFO(i915)->cache_modes[pat_index]; > > > > - if (pat_index != i915_gem_get_pat_index(i915, I915_CACHE_NONE)) > > - obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | > > - I915_BO_CACHE_COHERENT_FOR_WRITE); > > - else if (HAS_LLC(i915)) > > - obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; > > - else > > - obj->cache_coherent = 0; > > - > > - obj->cache_dirty = > > - !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && > > - !IS_DGFX(i915); > > + __i915_gem_object_update_coherency(obj); > > } > > > > bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj) > > @@ -215,6 +192,7 @@ bool i915_gem_object_can_bypass_llc(struct > drm_i915_gem_object *obj) > > /* > > * Always flush cache for UMD objects at creation time. > > */ > > + /* QQQ/FIXME why? avoidable performance penalty? */ > > if (obj->pat_set_by_user) > > return true; > > > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.h > b/drivers/gpu/drm/i915/gem/i915_gem_object.h > > index 884a17275b3a..f84f41e9f81f 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_object.h > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.h > > @@ -13,6 +13,7 @@ > > > > #include "display/intel_frontbuffer.h" > > #include "intel_memory_region.h" > > +#include "i915_cache.h" > > #include "i915_gem_object_types.h" > > #include "i915_gem_gtt.h" > > #include "i915_gem_ww.h" > > @@ -32,10 +33,18 @@ static inline bool i915_gem_object_size_2big(u64 > size) > > return false; > > } > > > > -unsigned int i915_gem_get_pat_index(struct drm_i915_private *i915, > > - enum i915_cache_level level); > > -bool i915_gem_object_has_cache_level(const struct > drm_i915_gem_object *obj, > > - enum i915_cache_level lvl); > > +static inline int > > +i915_gem_object_has_cache_mode(const struct drm_i915_gem_object *obj, > > + unsigned int mode) > > +{ > > + if (I915_CACHE_MODE(obj->cache_mode) == mode) > > + return 1; > > + else if (obj->pat_set_by_user) > > + return -1; /* Unknown, callers should assume no. */ > > + else > > + return 0; > > +} > > + > > void i915_gem_init__objects(struct drm_i915_private *i915); > > > > void i915_objects_module_exit(void); > > @@ -764,6 +773,7 @@ int i915_gem_object_wait_moving_fence(struct > drm_i915_gem_object *obj, > > bool intr); > > bool i915_gem_object_has_unknown_state(struct drm_i915_gem_object *obj); > > > > +void __i915_gem_object_update_coherency(struct drm_i915_gem_object > *obj); > > void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object > *obj, > > unsigned int cache_level); > > void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj, > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > > index 8de2b91b3edf..1f9fa28d07df 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > > @@ -14,6 +14,7 @@ > > #include <uapi/drm/i915_drm.h> > > > > #include "i915_active.h" > > +#include "i915_cache.h" > > #include "i915_selftest.h" > > #include "i915_vma_resource.h" > > > > @@ -116,93 +117,6 @@ struct drm_i915_gem_object_ops { > > const char *name; /* friendly name for debug, e.g. lockdep classes */ > > }; > > > > -/** > > - * enum i915_cache_level - The supported GTT caching values for > system memory > > - * pages. > > - * > > - * These translate to some special GTT PTE bits when binding pages > into some > > - * address space. It also determines whether an object, or rather > its pages are > > - * coherent with the GPU, when also reading or writing through the > CPU cache > > - * with those pages. > > - * > > - * Userspace can also control this through struct drm_i915_gem_caching. > > - */ > > -enum i915_cache_level { > > - /** > > - * @I915_CACHE_NONE: > > - * > > - * GPU access is not coherent with the CPU cache. If the cache is > dirty > > - * and we need the underlying pages to be coherent with some > later GPU > > - * access then we need to manually flush the pages. > > - * > > - * On shared LLC platforms reads and writes through the CPU cache are > > - * still coherent even with this setting. See also > > - * &drm_i915_gem_object.cache_coherent for more details. Due to > this we > > - * should only ever use uncached for scanout surfaces, otherwise > we end > > - * up over-flushing in some places. > > - * > > - * This is the default on non-LLC platforms. > > - */ > > - I915_CACHE_NONE = 0, > > - /** > > - * @I915_CACHE_LLC: > > - * > > - * GPU access is coherent with the CPU cache. If the cache is dirty, > > - * then the GPU will ensure that access remains coherent, when both > > - * reading and writing through the CPU cache. GPU writes can > dirty the > > - * CPU cache. > > - * > > - * Not used for scanout surfaces. > > - * > > - * Applies to both platforms with shared LLC(HAS_LLC), and snooping > > - * based platforms(HAS_SNOOP). > > - * > > - * This is the default on shared LLC platforms. The only > exception is > > - * scanout objects, where the display engine is not coherent with the > > - * CPU cache. For such objects I915_CACHE_NONE or I915_CACHE_WT is > > - * automatically applied by the kernel in pin_for_display, if > userspace > > - * has not done so already. > > - */ > > - I915_CACHE_LLC, > > - /** > > - * @I915_CACHE_L3_LLC: > > - * > > - * Explicitly enable the Gfx L3 cache, with coherent LLC. > > - * > > - * The Gfx L3 sits between the domain specific caches, e.g > > - * sampler/render caches, and the larger LLC. LLC is coherent > with the > > - * GPU, but L3 is only visible to the GPU, so likely needs to be > flushed > > - * when the workload completes. > > - * > > - * Not used for scanout surfaces. > > - * > > - * Only exposed on some gen7 + GGTT. More recent hardware has dropped > > - * this explicit setting, where it should now be enabled by default. > > - */ > > - I915_CACHE_L3_LLC, > > - /** > > - * @I915_CACHE_WT: > > - * > > - * Write-through. Used for scanout surfaces. > > - * > > - * The GPU can utilise the caches, while still having the display > engine > > - * be coherent with GPU writes, as a result we don't need to > flush the > > - * CPU caches when moving out of the render domain. This is the > default > > - * setting chosen by the kernel, if supported by the HW, otherwise we > > - * fallback to I915_CACHE_NONE. On the CPU side writes through > the CPU > > - * cache still need to be flushed, to remain coherent with the > display > > - * engine. > > - */ > > - I915_CACHE_WT, > > - /** > > - * @I915_MAX_CACHE_LEVEL: > > - * > > - * Mark the last entry in the enum. Used for defining > cachelevel_to_pat > > - * array for cache_level to pat translation table. > > - */ > > - I915_MAX_CACHE_LEVEL, > > -}; > > - > > enum i915_map_type { > > I915_MAP_WB = 0, > > I915_MAP_WC, > > @@ -375,6 +289,9 @@ struct drm_i915_gem_object { > > unsigned int mem_flags; > > #define I915_BO_FLAG_STRUCT_PAGE BIT(0) /* Object backed by struct > pages */ > > #define I915_BO_FLAG_IOMEM BIT(1) /* Object backed by IO memory */ > > + > > + i915_cache_t cache_mode; > > + > > /** > > * @pat_index: The desired PAT index. > > * > > @@ -409,9 +326,7 @@ struct drm_i915_gem_object { > > * Check for @pat_set_by_user to find out if an object has pat > index set > > * by userspace. The ioctl's to change cache settings have also been > > * disabled for the objects with pat index set by userspace. > Please don't > > - * assume @cache_coherent having the flags set as describe here. > A helper > > - * function i915_gem_object_has_cache_level() provides one way to > bypass > > - * the use of this field. > > + * assume @cache_coherent having the flags set as describe here. > > * > > * Track whether the pages are coherent with the GPU if reading or > > * writing through the CPU caches. The largely depends on the > > @@ -492,9 +407,7 @@ struct drm_i915_gem_object { > > * Check for @pat_set_by_user to find out if an object has pat > index set > > * by userspace. The ioctl's to change cache settings have also been > > * disabled for the objects with pat_index set by userspace. > Please don't > > - * assume @cache_dirty is set as describe here. Also see helper > function > > - * i915_gem_object_has_cache_level() for possible ways to bypass > the use > > - * of this field. > > + * assume @cache_dirty is set as describe here. > > * > > * Track if we are we dirty with writes through the CPU cache for > this > > * object. As a result reading directly from main memory might yield > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > index 3b094d36a0b0..a7012f1a9c70 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > @@ -563,11 +563,8 @@ static void dbg_poison(struct i915_ggtt *ggtt, > > while (size) { > > void __iomem *s; > > > > - ggtt->vm.insert_page(&ggtt->vm, addr, > > - ggtt->error_capture.start, > > - i915_gem_get_pat_index(ggtt->vm.i915, > > - I915_CACHE_NONE), > > - 0); > > + ggtt->vm.insert_page(&ggtt->vm, addr, > ggtt->error_capture.start, > > + ggtt->vm.i915->pat_uc, 0); > > mb(); > > > > s = io_mapping_map_wc(&ggtt->iomap, > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c > b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c > > index 7078af2f8f79..e794bd2a7ccb 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c > > @@ -58,6 +58,16 @@ i915_ttm_cache_level(struct drm_i915_private > *i915, struct ttm_resource *res, > > I915_CACHE_NONE; > > } > > > > +static unsigned int > > +i915_ttm_cache_pat(struct drm_i915_private *i915, struct > ttm_resource *res, > > + struct ttm_tt *ttm) > > +{ > > + return ((HAS_LLC(i915) || HAS_SNOOP(i915)) && > > + !i915_ttm_gtt_binds_lmem(res) && > > + ttm->caching == ttm_cached) ? i915->pat_wb : > > + i915->pat_uc; > > +} > > + > > static struct intel_memory_region * > > i915_ttm_region(struct ttm_device *bdev, int ttm_mem_type) > > { > > @@ -196,7 +206,7 @@ static struct dma_fence > *i915_ttm_accel_move(struct ttm_buffer_object *bo, > > struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo); > > struct i915_request *rq; > > struct ttm_tt *src_ttm = bo->ttm; > > - enum i915_cache_level src_level, dst_level; > > + unsigned int src_pat, dst_pat; > > int ret; > > > > if (!to_gt(i915)->migrate.context || intel_gt_is_wedged(to_gt(i915))) > > @@ -206,16 +216,15 @@ static struct dma_fence > *i915_ttm_accel_move(struct ttm_buffer_object *bo, > > if (I915_SELFTEST_ONLY(fail_gpu_migration)) > > clear = true; > > > > - dst_level = i915_ttm_cache_level(i915, dst_mem, dst_ttm); > > + dst_pat = i915_ttm_cache_pat(i915, dst_mem, dst_ttm); > > if (clear) { > > if (bo->type == ttm_bo_type_kernel && > > !I915_SELFTEST_ONLY(fail_gpu_migration)) > > return ERR_PTR(-EINVAL); > > > > intel_engine_pm_get(to_gt(i915)->migrate.context->engine); > > - ret = > intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, > > - dst_st->sgl, > > - i915_gem_get_pat_index(i915, > dst_level), > > + ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, > > + deps, dst_st->sgl, dst_pat, > > i915_ttm_gtt_binds_lmem(dst_mem), > > 0, &rq); > > } else { > > @@ -225,14 +234,13 @@ static struct dma_fence > *i915_ttm_accel_move(struct ttm_buffer_object *bo, > > if (IS_ERR(src_rsgt)) > > return ERR_CAST(src_rsgt); > > > > - src_level = i915_ttm_cache_level(i915, bo->resource, src_ttm); > > + src_pat = i915_ttm_cache_pat(i915, bo->resource, src_ttm); > > intel_engine_pm_get(to_gt(i915)->migrate.context->engine); > > ret = intel_context_migrate_copy(to_gt(i915)->migrate.context, > > deps, src_rsgt->table.sgl, > > - i915_gem_get_pat_index(i915, > src_level), > > + src_pat, > > i915_ttm_gtt_binds_lmem(bo->resource), > > - dst_st->sgl, > > - i915_gem_get_pat_index(i915, > dst_level), > > + dst_st->sgl, dst_pat, > > i915_ttm_gtt_binds_lmem(dst_mem), > > &rq); > > > > diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > index df6c9a84252c..c8925918784e 100644 > > --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > @@ -354,7 +354,7 @@ fake_huge_pages_object(struct drm_i915_private > *i915, u64 size, bool single) > > > > obj->write_domain = I915_GEM_DOMAIN_CPU; > > obj->read_domains = I915_GEM_DOMAIN_CPU; > > - obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE); > > + i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE); > > > > return obj; > > } > > diff --git a/drivers/gpu/drm/i915/gt/gen6_ppgtt.c > b/drivers/gpu/drm/i915/gt/gen6_ppgtt.c > > index c2bdc133c89a..fb69f667652a 100644 > > --- a/drivers/gpu/drm/i915/gt/gen6_ppgtt.c > > +++ b/drivers/gpu/drm/i915/gt/gen6_ppgtt.c > > @@ -226,9 +226,7 @@ static int gen6_ppgtt_init_scratch(struct > gen6_ppgtt *ppgtt) > > return ret; > > > > vm->scratch[0]->encode = > > - vm->pte_encode(px_dma(vm->scratch[0]), > > - i915_gem_get_pat_index(vm->i915, > > - I915_CACHE_NONE), > > + vm->pte_encode(px_dma(vm->scratch[0]), vm->i915->pat_uc, > > PTE_READ_ONLY); > > > > vm->scratch[1] = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); > > diff --git a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c > b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c > > index f948d33e5ec5..a6692ea1a91e 100644 > > --- a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c > > +++ b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c > > @@ -40,16 +40,11 @@ static u64 gen8_pte_encode(dma_addr_t addr, > > if (flags & PTE_LM) > > pte |= GEN12_PPGTT_PTE_LM; > > > > - /* > > - * For pre-gen12 platforms pat_index is the same as enum > > - * i915_cache_level, so the switch-case here is still valid. > > - * See translation table defined by LEGACY_CACHELEVEL. > > - */ > > switch (pat_index) { > > - case I915_CACHE_NONE: > > + case 0: > > pte |= PPAT_UNCACHED; > > break; > > - case I915_CACHE_WT: > > + case 3: > > pte |= PPAT_DISPLAY_ELLC; > > break; > > default: > > @@ -853,9 +848,7 @@ static int gen8_init_scratch(struct > i915_address_space *vm) > > pte_flags |= PTE_LM; > > > > vm->scratch[0]->encode = > > - vm->pte_encode(px_dma(vm->scratch[0]), > > - i915_gem_get_pat_index(vm->i915, > > - I915_CACHE_NONE), > > + vm->pte_encode(px_dma(vm->scratch[0]), vm->i915->pat_uc, > > pte_flags); > > > > for (i = 1; i <= vm->top; i++) { > > diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c > b/drivers/gpu/drm/i915/gt/intel_ggtt.c > > index dd0ed941441a..c97379cf8241 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c > > +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c > > @@ -921,9 +921,7 @@ static int ggtt_probe_common(struct i915_ggtt > *ggtt, u64 size) > > pte_flags |= PTE_LM; > > > > ggtt->vm.scratch[0]->encode = > > - ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]), > > - i915_gem_get_pat_index(i915, > > - I915_CACHE_NONE), > > + ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]), i915->pat_uc, > > pte_flags); > > > > return 0; > > @@ -1297,10 +1295,7 @@ bool i915_ggtt_resume_vm(struct > i915_address_space *vm) > > * ptes to be repopulated. > > */ > > vma->resource->bound_flags = 0; > > - vma->ops->bind_vma(vm, NULL, vma->resource, > > - obj ? obj->pat_index : > > - i915_gem_get_pat_index(vm->i915, > > - I915_CACHE_NONE), > > + vma->ops->bind_vma(vm, NULL, vma->resource, obj->cache_mode, > > was_bound); > > > > if (obj) { /* only used during resume => exclusive access */ > > diff --git a/drivers/gpu/drm/i915/gt/intel_migrate.c > b/drivers/gpu/drm/i915/gt/intel_migrate.c > > index 6023288b0e2d..81f7834cc2db 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_migrate.c > > +++ b/drivers/gpu/drm/i915/gt/intel_migrate.c > > @@ -45,9 +45,7 @@ static void xehpsdv_toggle_pdes(struct > i915_address_space *vm, > > * Insert a dummy PTE into every PT that will map to LMEM to ensure > > * we have a correctly setup PDE structure for later use. > > */ > > - vm->insert_page(vm, 0, d->offset, > > - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), > > - PTE_LM); > > + vm->insert_page(vm, 0, d->offset, vm->i915->pat_uc, PTE_LM); > > GEM_BUG_ON(!pt->is_compact); > > d->offset += SZ_2M; > > } > > @@ -65,9 +63,7 @@ static void xehpsdv_insert_pte(struct > i915_address_space *vm, > > * alignment is 64K underneath for the pt, and we are careful > > * not to access the space in the void. > > */ > > - vm->insert_page(vm, px_dma(pt), d->offset, > > - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), > > - PTE_LM); > > + vm->insert_page(vm, px_dma(pt), d->offset, vm->i915->pat_uc, PTE_LM); > > d->offset += SZ_64K; > > } > > > > @@ -77,8 +73,7 @@ static void insert_pte(struct i915_address_space *vm, > > { > > struct insert_pte_data *d = data; > > > > - vm->insert_page(vm, px_dma(pt), d->offset, > > - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), > > + vm->insert_page(vm, px_dma(pt), d->offset, vm->i915->pat_uc, > > i915_gem_object_is_lmem(pt->base) ? PTE_LM : 0); > > d->offset += PAGE_SIZE; > > } > > diff --git a/drivers/gpu/drm/i915/gt/selftest_migrate.c > b/drivers/gpu/drm/i915/gt/selftest_migrate.c > > index 3def5ca72dec..a67ede65d816 100644 > > --- a/drivers/gpu/drm/i915/gt/selftest_migrate.c > > +++ b/drivers/gpu/drm/i915/gt/selftest_migrate.c > > @@ -904,8 +904,7 @@ static int perf_clear_blt(void *arg) > > > > err = __perf_clear_blt(gt->migrate.context, > > dst->mm.pages->sgl, > > - i915_gem_get_pat_index(gt->i915, > > - I915_CACHE_NONE), > > + gt->i915->pat_uc, > > i915_gem_object_is_lmem(dst), > > sizes[i]); > > > > @@ -995,12 +994,10 @@ static int perf_copy_blt(void *arg) > > > > err = __perf_copy_blt(gt->migrate.context, > > src->mm.pages->sgl, > > - i915_gem_get_pat_index(gt->i915, > > - I915_CACHE_NONE), > > + gt->i915->pat_uc, > > i915_gem_object_is_lmem(src), > > dst->mm.pages->sgl, > > - i915_gem_get_pat_index(gt->i915, > > - I915_CACHE_NONE), > > + gt->i915->pat_uc, > > i915_gem_object_is_lmem(dst), > > sz); > > > > diff --git a/drivers/gpu/drm/i915/gt/selftest_reset.c > b/drivers/gpu/drm/i915/gt/selftest_reset.c > > index 79aa6ac66ad2..327dc9294e0f 100644 > > --- a/drivers/gpu/drm/i915/gt/selftest_reset.c > > +++ b/drivers/gpu/drm/i915/gt/selftest_reset.c > > @@ -84,11 +84,8 @@ __igt_reset_stolen(struct intel_gt *gt, > > void __iomem *s; > > void *in; > > > > - ggtt->vm.insert_page(&ggtt->vm, dma, > > - ggtt->error_capture.start, > > - i915_gem_get_pat_index(gt->i915, > > - I915_CACHE_NONE), > > - 0); > > + ggtt->vm.insert_page(&ggtt->vm, dma, ggtt->error_capture.start, > > + gt->i915->pat_uc, 0); > > mb(); > > > > s = io_mapping_map_wc(&ggtt->iomap, > > @@ -127,11 +124,8 @@ __igt_reset_stolen(struct intel_gt *gt, > > void *in; > > u32 x; > > > > - ggtt->vm.insert_page(&ggtt->vm, dma, > > - ggtt->error_capture.start, > > - i915_gem_get_pat_index(gt->i915, > > - I915_CACHE_NONE), > > - 0); > > + ggtt->vm.insert_page(&ggtt->vm, dma, ggtt->error_capture.start, > > + gt->i915->pat_uc, 0); > > mb(); > > > > s = io_mapping_map_wc(&ggtt->iomap, > > diff --git a/drivers/gpu/drm/i915/gt/selftest_timeline.c > b/drivers/gpu/drm/i915/gt/selftest_timeline.c > > index 39c3ec12df1a..db64dc7d3fce 100644 > > --- a/drivers/gpu/drm/i915/gt/selftest_timeline.c > > +++ b/drivers/gpu/drm/i915/gt/selftest_timeline.c > > @@ -836,7 +836,10 @@ static int setup_watcher(struct hwsp_watcher *w, > struct intel_gt *gt, > > return PTR_ERR(obj); > > > > /* keep the same cache settings as timeline */ > > - i915_gem_object_set_pat_index(obj, tl->hwsp_ggtt->obj->pat_index); > > + obj->pat_index = tl->hwsp_ggtt->obj->pat_index; > > + obj->cache_mode = tl->hwsp_ggtt->obj->cache_mode; > > + __i915_gem_object_update_coherency(obj); > > + > > w->map = i915_gem_object_pin_map_unlocked(obj, > > > page_unmask_bits(tl->hwsp_ggtt->obj->mm.mapping)); > > if (IS_ERR(w->map)) { > > diff --git a/drivers/gpu/drm/i915/gt/selftest_tlb.c > b/drivers/gpu/drm/i915/gt/selftest_tlb.c > > index 3bd6b540257b..6049f01be219 100644 > > --- a/drivers/gpu/drm/i915/gt/selftest_tlb.c > > +++ b/drivers/gpu/drm/i915/gt/selftest_tlb.c > > @@ -36,8 +36,6 @@ pte_tlbinv(struct intel_context *ce, > > u64 length, > > struct rnd_state *prng) > > { > > - const unsigned int pat_index = > > - i915_gem_get_pat_index(ce->vm->i915, I915_CACHE_NONE); > > struct drm_i915_gem_object *batch; > > struct drm_mm_node vb_node; > > struct i915_request *rq; > > @@ -157,7 +155,8 @@ pte_tlbinv(struct intel_context *ce, > > /* Flip the PTE between A and B */ > > if (i915_gem_object_is_lmem(vb->obj)) > > pte_flags |= PTE_LM; > > - ce->vm->insert_entries(ce->vm, &vb_res, pat_index, pte_flags); > > + ce->vm->insert_entries(ce->vm, &vb_res, ce->vm->i915->pat_uc, > > + pte_flags); > > > > /* Flush the PTE update to concurrent HW */ > > tlbinv(ce->vm, addr & -length, length); > > diff --git a/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c > b/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c > > index d408856ae4c0..e099414d624d 100644 > > --- a/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c > > +++ b/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c > > @@ -991,14 +991,10 @@ static void uc_fw_bind_ggtt(struct intel_uc_fw > *uc_fw) > > > > if (ggtt->vm.raw_insert_entries) > > ggtt->vm.raw_insert_entries(&ggtt->vm, vma_res, > > - i915_gem_get_pat_index(ggtt->vm.i915, > > - I915_CACHE_NONE), > > - pte_flags); > > + ggtt->vm.i915->pat_uc, pte_flags); > > else > > ggtt->vm.insert_entries(&ggtt->vm, vma_res, > > - i915_gem_get_pat_index(ggtt->vm.i915, > > - I915_CACHE_NONE), > > - pte_flags); > > + ggtt->vm.i915->pat_uc, pte_flags); > > } > > > > static void uc_fw_unbind_ggtt(struct intel_uc_fw *uc_fw) > > diff --git a/drivers/gpu/drm/i915/i915_cache.c > b/drivers/gpu/drm/i915/i915_cache.c > > new file mode 100644 > > index 000000000000..7a8002ebd2ec > > --- /dev/null > > +++ b/drivers/gpu/drm/i915/i915_cache.c > > @@ -0,0 +1,59 @@ > > +/* > > + * SPDX-License-Identifier: MIT > > + * > > + * Copyright © 2023 Intel Corporation > > + */ > > + > > +#include "i915_cache.h" > > +#include "i915_drv.h" > > + > > +static int find_pat(const struct intel_device_info *info, > i915_cache_t mode) > > +{ > > + int i; > > + > > + for (i = 0; i < ARRAY_SIZE(info->cache_modes); i++) { > > + if (info->cache_modes[i] == mode) > > + return i; > > + } > > + > > + return -1; > > +} > > + > > +void i915_cache_init(struct drm_i915_private *i915) > > +{ > > + int ret; > > + > > + ret = find_pat(INTEL_INFO(i915), I915_CACHE(UC)); > > + WARN_ON(ret < 0); > > + drm_info(&i915->drm, "Using PAT index %u for uncached access\n", > ret); > > + i915->pat_uc = ret; > > + > > + ret = find_pat(INTEL_INFO(i915), I915_CACHE(WB)); > > + WARN_ON(ret < 0); > > + drm_info(&i915->drm, "Using PAT index %u for write-back > access\n", ret); > > + i915->pat_wb = ret; > > +} > > I don't think we need the above two functions. Why don't we just hard code > pat_uc and pat_wb in device_info? plus pat_wt too? These are predetermined, > and used by KMD only. We could add a table of enum i915_cache_mode to pat index but since we only use two from within the driver I was a bit reluctant to add a set of tables with more entries. Just a slight preference at the moment, I guess I could be convinced we really should go the 2nd table route. > > +int i915_cache_level_to_pat_and_mode(struct drm_i915_private *i915, > > + unsigned int cache_level, > > + i915_cache_t *mode) > > +{ > > + const struct intel_device_info *info = INTEL_INFO(i915); > > + i915_cache_t level_to_mode[] = { > > + [I915_CACHE_NONE] = I915_CACHE(UC), > > + [I915_CACHE_WT] = I915_CACHE(WT), > > + [I915_CACHE_LLC] = I915_CACHE(WB), > > + }; > > + int ret; > > + > > + if (GRAPHICS_VER(i915) >= 12) > > + level_to_mode[I915_CACHE_L3_LLC] = I915_CACHE(WB); > > + else > > + level_to_mode[I915_CACHE_L3_LLC] = _I915_CACHE(WB, LLC); > > + > > + ret = find_pat(info, level_to_mode[cache_level]); > > + if (ret >= 0 && mode) > > + *mode = info->cache_modes[ret]; > > + > > + return ret; > > +} > > diff --git a/drivers/gpu/drm/i915/i915_cache.h > b/drivers/gpu/drm/i915/i915_cache.h > > new file mode 100644 > > index 000000000000..0df03f1f01ef > > --- /dev/null > > +++ b/drivers/gpu/drm/i915/i915_cache.h > > @@ -0,0 +1,129 @@ > > +/* SPDX-License-Identifier: MIT */ > > +/* > > + * Copyright © 2023 Intel Corporation > > + */ > > + > > +#ifndef __I915_CACHE_H__ > > +#define __I915_CACHE_H__ > > + > > +#include <linux/types.h> > > + > > +struct drm_i915_private; > > + > > +typedef u16 i915_cache_t; > > + > > +#define I915_CACHE(mode) \ > > + (i915_cache_t)(I915_CACHE_MODE_##mode) > > + > > +#define _I915_CACHE(mode, flag) \ > > + (i915_cache_t)((I915_CACHE_MODE_##mode) | ( BIT(8 + > I915_CACHE_##flag))) > > + > > +#define I915_CACHE_MODE(cache) \ > > + (unsigned int)(((i915_cache_t)(cache)) & 0xff) > > +#define I915_CACHE_FLAGS(cache) \ > > + (unsigned int)((((i915_cache_t)(cache) & 0xff00)) >> 16) > > + > > +/* Cache mode values */ > > +#define I915_CACHE_MODE_UNKNOWN (0) > > +#define I915_CACHE_MODE_UC (1) > > +#define I915_CACHE_MODE_WB (2) > > +#define I915_CACHE_MODE_WT (3) > > +#define I915_CACHE_MODE_WC (4) > > Why do you need these CACHE_MODE's? Aren't they the same as > i915_cache_level, which > need some sort of translation? Yes, in v2 I dropped i915_cache_level instead. > > + > > +/* Mode flag bits */ > > +#define I915_CACHE_L3 (0) > > +#define I915_CACHE_COH1W (1) > > +#define I915_CACHE_COH2W (2) > > +#define I915_CACHE_CLOS1 (3) > > +#define I915_CACHE_CLOS2 (4) > > These had been defined in drivers/gpu/drm/i915/gt/intel_gtt.h already, > why add new ones? > The CLOS ones are not needed in upstream unless we want to support PVC here. I simply added entries for everything I could find currently in the driver. In parallel I am checking if we should/could remove all the PVC bits and pieces from the driver. > > +void i915_cache_init(struct drm_i915_private *i915); > > +int i915_cache_level_to_pat_and_mode(struct drm_i915_private *i915, > > + unsigned int cache_level, > > + i915_cache_t *mode); > > + > > +/* > > + * Legacy/kernel internal interface below: > > + */ > > + > > +/** > > + * enum i915_cache_level - The supported GTT caching values for > system memory > > + * pages. > > + * > > + * These translate to some special GTT PTE bits when binding pages > into some > > + * address space. It also determines whether an object, or rather > its pages are > > + * coherent with the GPU, when also reading or writing through the > CPU cache > > + * with those pages. > > + * > > + * Userspace can also control this through struct drm_i915_gem_caching. > > + */ > > +enum i915_cache_level { > > Shouldn't we completely get rid of this enum now? It should be replaced by > INTEL_INFO(i915)->pat_uc/wb/wt. Yep, done. > > + /** > > + * @I915_CACHE_NONE: > > + * > > + * GPU access is not coherent with the CPU cache. If the cache is > dirty > > + * and we need the underlying pages to be coherent with some > later GPU > > + * access then we need to manually flush the pages. > > + * > > + * On shared LLC platforms reads and writes through the CPU cache are > > + * still coherent even with this setting. See also > > + * &drm_i915_gem_object.cache_coherent for more details. Due to > this we > > + * should only ever use uncached for scanout surfaces, otherwise > we end > > + * up over-flushing in some places. > > + * > > + * This is the default on non-LLC platforms. > > + */ > > + I915_CACHE_NONE = 0, > > + /** > > + * @I915_CACHE_LLC: > > + * > > + * GPU access is coherent with the CPU cache. If the cache is dirty, > > + * then the GPU will ensure that access remains coherent, when both > > + * reading and writing through the CPU cache. GPU writes can > dirty the > > + * CPU cache. > > + * > > + * Not used for scanout surfaces. > > + * > > + * Applies to both platforms with shared LLC(HAS_LLC), and snooping > > + * based platforms(HAS_SNOOP). > > + * > > + * This is the default on shared LLC platforms. The only > exception is > > + * scanout objects, where the display engine is not coherent with the > > + * CPU cache. For such objects I915_CACHE_NONE or I915_CACHE_WT is > > + * automatically applied by the kernel in pin_for_display, if > userspace > > + * has not done so already. > > + */ > > + I915_CACHE_LLC, > > + /** > > + * @I915_CACHE_L3_LLC: > > + * > > + * Explicitly enable the Gfx L3 cache, with coherent LLC. > > + * > > + * The Gfx L3 sits between the domain specific caches, e.g > > + * sampler/render caches, and the larger LLC. LLC is coherent > with the > > + * GPU, but L3 is only visible to the GPU, so likely needs to be > flushed > > + * when the workload completes. > > + * > > + * Not used for scanout surfaces. > > + * > > + * Only exposed on some gen7 + GGTT. More recent hardware has dropped > > + * this explicit setting, where it should now be enabled by default. > > + */ > > + I915_CACHE_L3_LLC, > > + /** > > + * @I915_CACHE_WT: > > + * > > + * Write-through. Used for scanout surfaces. > > + * > > + * The GPU can utilise the caches, while still having the display > engine > > + * be coherent with GPU writes, as a result we don't need to > flush the > > + * CPU caches when moving out of the render domain. This is the > default > > + * setting chosen by the kernel, if supported by the HW, otherwise we > > + * fallback to I915_CACHE_NONE. On the CPU side writes through > the CPU > > + * cache still need to be flushed, to remain coherent with the > display > > + * engine. > > + */ > > + I915_CACHE_WT, > > +}; > > + > > +#endif /* __I915_CACHE_H__ */ > > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c > b/drivers/gpu/drm/i915/i915_debugfs.c > > index 76ccd4e03e31..e2da57397770 100644 > > --- a/drivers/gpu/drm/i915/i915_debugfs.c > > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > > @@ -139,48 +139,37 @@ static const char *stringify_vma_type(const > struct i915_vma *vma) > > return "ppgtt"; > > } > > > > -static const char *i915_cache_level_str(struct drm_i915_gem_object *obj) > > +static void obj_cache_str(struct seq_file *m, struct > drm_i915_gem_object *obj) > > { > > - struct drm_i915_private *i915 = obj_to_i915(obj); > > + const i915_cache_t cache = obj->cache_mode; > > + const unsigned int mode = I915_CACHE_MODE(cache); > > + const unsigned long flags = I915_CACHE_FLAGS(cache); > > + static const char *mode_str[] = { > > + [I915_CACHE_MODE_UC] = "UC", > > + [I915_CACHE_MODE_WB] = "WB", > > + [I915_CACHE_MODE_WT] = "WT", > > + [I915_CACHE_MODE_WC] = "WC", > > + }; > > + static const char *flag_str[] = { > > + [I915_CACHE_L3] = "L3", > > + [I915_CACHE_COH1W] = "1-Way-Coherent", > > + [I915_CACHE_COH2W] = "2-Way-Coherent", > > + [I915_CACHE_CLOS1] = "CLOS1", > > + [I915_CACHE_CLOS2] = "CLOS2", > > + }; > > > > - if (IS_METEORLAKE(i915)) { > > - switch (obj->pat_index) { > > - case 0: return " WB"; > > - case 1: return " WT"; > > - case 2: return " UC"; > > - case 3: return " WB (1-Way Coh)"; > > - case 4: return " WB (2-Way Coh)"; > > - default: return " not defined"; > > - } > > - } else if (IS_PONTEVECCHIO(i915)) { > > - switch (obj->pat_index) { > > - case 0: return " UC"; > > - case 1: return " WC"; > > - case 2: return " WT"; > > - case 3: return " WB"; > > - case 4: return " WT (CLOS1)"; > > - case 5: return " WB (CLOS1)"; > > - case 6: return " WT (CLOS2)"; > > - case 7: return " WT (CLOS2)"; > > - default: return " not defined"; > > - } > > - } else if (GRAPHICS_VER(i915) >= 12) { > > - switch (obj->pat_index) { > > - case 0: return " WB"; > > - case 1: return " WC"; > > - case 2: return " WT"; > > - case 3: return " UC"; > > - default: return " not defined"; > > - } > > + if (mode == I915_CACHE_MODE_UNKNOWN || mode > ARRAY_SIZE(mode_str)) { > > + if (obj->pat_set_by_user) > > + seq_printf(m, " PAT-%u", obj->pat_index); > > + else > > + seq_printf(m, " PAT-%u-???", obj->pat_index); > > } else { > > - switch (obj->pat_index) { > > - case 0: return " UC"; > > - case 1: return HAS_LLC(i915) ? > > - " LLC" : " snooped"; > > - case 2: return " L3+LLC"; > > - case 3: return " WT"; > > - default: return " not defined"; > > - } > > + unsigned long bit; > > + > > + seq_printf(m, " %s", mode_str[mode]); > > + > > + for_each_set_bit(bit, &flags, sizeof(i915_cache_t)) > > + seq_printf(m, "-%s", flag_str[bit]); > > } > > } > > > > @@ -190,17 +179,23 @@ i915_debugfs_describe_obj(struct seq_file *m, > struct drm_i915_gem_object *obj) > > struct i915_vma *vma; > > int pin_count = 0; > > > > - seq_printf(m, "%pK: %c%c%c %8zdKiB %02x %02x %s%s%s", > > + seq_printf(m, "%pK: %c%c%c %8zdKiB %02x %02x", > > &obj->base, > > get_tiling_flag(obj), > > get_global_flag(obj), > > get_pin_mapped_flag(obj), > > obj->base.size / 1024, > > obj->read_domains, > > - obj->write_domain, > > - i915_cache_level_str(obj), > > - obj->mm.dirty ? " dirty" : "", > > - obj->mm.madv == I915_MADV_DONTNEED ? " purgeable" : ""); > > + obj->write_domain); > > + > > + obj_cache_str(m, obj); > > + > > + if (obj->mm.dirty) > > + seq_puts(m, " dirty"); > > + > > + if (obj->mm.madv == I915_MADV_DONTNEED) > > + seq_puts(m, " purgeable"); > > + > > if (obj->base.name) > > seq_printf(m, " (name: %d)", obj->base.name); > > > > diff --git a/drivers/gpu/drm/i915/i915_driver.c > b/drivers/gpu/drm/i915/i915_driver.c > > index 222d0a1f3b55..deab26752ba4 100644 > > --- a/drivers/gpu/drm/i915/i915_driver.c > > +++ b/drivers/gpu/drm/i915/i915_driver.c > > @@ -80,6 +80,7 @@ > > #include "soc/intel_dram.h" > > #include "soc/intel_gmch.h" > > > > +#include "i915_cache.h" > > #include "i915_debugfs.h" > > #include "i915_driver.h" > > #include "i915_drm_client.h" > > @@ -267,6 +268,8 @@ static int i915_driver_early_probe(struct > drm_i915_private *dev_priv) > > > > intel_detect_preproduction_hw(dev_priv); > > > > + i915_cache_init(dev_priv); > > + > > return 0; > > > > err_rootgt: > > diff --git a/drivers/gpu/drm/i915/i915_drv.h > b/drivers/gpu/drm/i915/i915_drv.h > > index b4cf6f0f636d..cb1c0c9d98ef 100644 > > --- a/drivers/gpu/drm/i915/i915_drv.h > > +++ b/drivers/gpu/drm/i915/i915_drv.h > > @@ -251,6 +251,9 @@ struct drm_i915_private { > > unsigned int hpll_freq; > > unsigned int czclk_freq; > > > > + unsigned int pat_uc; > > + unsigned int pat_wb; > > How about making these part of INTEL_INFO(i915)? They are predetermined, > no need to be > dynamic. Discussed somewhere above. > > > + > > /** > > * wq - Driver workqueue for GEM. > > * > > diff --git a/drivers/gpu/drm/i915/i915_gem.c > b/drivers/gpu/drm/i915/i915_gem.c > > index 7ae42f746cc2..9aae75862e6f 100644 > > --- a/drivers/gpu/drm/i915/i915_gem.c > > +++ b/drivers/gpu/drm/i915/i915_gem.c > > @@ -422,9 +422,7 @@ i915_gem_gtt_pread(struct drm_i915_gem_object *obj, > > ggtt->vm.insert_page(&ggtt->vm, > > i915_gem_object_get_dma_address(obj, > > offset >> > PAGE_SHIFT), > > - node.start, > > - i915_gem_get_pat_index(i915, > > - I915_CACHE_NONE), 0); > > + node.start, i915->pat_uc, 0); > > } else { > > page_base += offset & PAGE_MASK; > > } > > @@ -603,9 +601,7 @@ i915_gem_gtt_pwrite_fast(struct > drm_i915_gem_object *obj, > > ggtt->vm.insert_page(&ggtt->vm, > > i915_gem_object_get_dma_address(obj, > > offset >> > PAGE_SHIFT), > > - node.start, > > - i915_gem_get_pat_index(i915, > > - I915_CACHE_NONE), 0); > > + node.start, i915->pat_uc, 0); > > wmb(); /* flush modifications to the GGTT > (insert_page) */ > > } else { > > page_base += offset & PAGE_MASK; > > @@ -1148,19 +1144,6 @@ int i915_gem_init(struct drm_i915_private > *dev_priv) > > unsigned int i; > > int ret; > > > > - /* > > - * In the proccess of replacing cache_level with pat_index a tricky > > - * dependency is created on the definition of the enum > i915_cache_level. > > - * in case this enum is changed, PTE encode would be broken. > > - * Add a WARNING here. And remove when we completely quit using this > > - * enum > > - */ > > - BUILD_BUG_ON(I915_CACHE_NONE != 0 || > > - I915_CACHE_LLC != 1 || > > - I915_CACHE_L3_LLC != 2 || > > - I915_CACHE_WT != 3 || > > - I915_MAX_CACHE_LEVEL != 4); > > - > > /* We need to fallback to 4K pages if host doesn't support huge > gtt. */ > > if (intel_vgpu_active(dev_priv) && > !intel_vgpu_has_huge_gtt(dev_priv)) > > RUNTIME_INFO(dev_priv)->page_sizes = I915_GTT_PAGE_SIZE_4K; > > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c > b/drivers/gpu/drm/i915/i915_gpu_error.c > > index 4749f99e6320..fad336a45699 100644 > > --- a/drivers/gpu/drm/i915/i915_gpu_error.c > > +++ b/drivers/gpu/drm/i915/i915_gpu_error.c > > @@ -1122,14 +1122,11 @@ i915_vma_coredump_create(const struct > intel_gt *gt, > > mutex_lock(&ggtt->error_mutex); > > if (ggtt->vm.raw_insert_page) > > ggtt->vm.raw_insert_page(&ggtt->vm, dma, slot, > > - i915_gem_get_pat_index(gt->i915, > > - > I915_CACHE_NONE), > > + ggtt->vm.i915->pat_uc, > > 0); > > else > > ggtt->vm.insert_page(&ggtt->vm, dma, slot, > > - i915_gem_get_pat_index(gt->i915, > > - I915_CACHE_NONE), > > - 0); > > + ggtt->vm.i915->pat_uc, 0); > > mb(); > > > > s = io_mapping_map_wc(&ggtt->iomap, slot, PAGE_SIZE); > > diff --git a/drivers/gpu/drm/i915/i915_pci.c > b/drivers/gpu/drm/i915/i915_pci.c > > index 3d7a5db9833b..fbdce31afeb1 100644 > > --- a/drivers/gpu/drm/i915/i915_pci.c > > +++ b/drivers/gpu/drm/i915/i915_pci.c > > @@ -32,6 +32,7 @@ > > #include "gt/intel_sa_media.h" > > #include "gem/i915_gem_object_types.h" > > > > +#include "i915_cache.h" > > #include "i915_driver.h" > > #include "i915_drv.h" > > #include "i915_pci.h" > > @@ -46,36 +47,42 @@ __diag_ignore_all("-Woverride-init", "Allow > overriding inherited members"); > > .__runtime.graphics.ip.ver = (x), \ > > .__runtime.media.ip.ver = (x) > > > > -#define LEGACY_CACHELEVEL \ > > - .cachelevel_to_pat = { \ > > - [I915_CACHE_NONE] = 0, \ > > - [I915_CACHE_LLC] = 1, \ > > - [I915_CACHE_L3_LLC] = 2, \ > > - [I915_CACHE_WT] = 3, \ > > +/* TODO/QQQ index 1 & 2 */ > > +#define LEGACY_CACHE_MODES \ > > I was thinking to just put the PAT settings here, instead of > cache_modes, simply > .pat = {\ > GEN8_PPAT_WB, \ > GEN8_PPAT_WC, \ > GEN8_PPAT_WT, \ > GEN8_PPAT_UC, > } Also discussed above - I want to be able to have a functioning platform agnostic "has cache mode" helper and I think this would not help. > > > + .cache_modes = { \ > > + [0] = I915_CACHE(UC), \ > > + [1] = I915_CACHE(WB), \ > > + [2] = _I915_CACHE(WB, L3), \ > > + [3] = I915_CACHE(WT), \ > > } > > > > -#define TGL_CACHELEVEL \ > > - .cachelevel_to_pat = { \ > > - [I915_CACHE_NONE] = 3, \ > > - [I915_CACHE_LLC] = 0, \ > > - [I915_CACHE_L3_LLC] = 0, \ > > - [I915_CACHE_WT] = 2, \ > > +#define GEN12_CACHE_MODES \ > > + .cache_modes = { \ > > + [0] = I915_CACHE(WB), \ > > + [1] = I915_CACHE(WC), \ > > + [2] = I915_CACHE(WT), \ > > + [3] = I915_CACHE(UC), \ > > } > > > > -#define PVC_CACHELEVEL \ > > - .cachelevel_to_pat = { \ > > - [I915_CACHE_NONE] = 0, \ > > - [I915_CACHE_LLC] = 3, \ > > - [I915_CACHE_L3_LLC] = 3, \ > > - [I915_CACHE_WT] = 2, \ > > +#define PVC_CACHE_MODES \ > > + .cache_modes = { \ > > + [0] = I915_CACHE(UC), \ > > + [1] = I915_CACHE(WC), \ > > + [2] = I915_CACHE(WT), \ > > + [3] = I915_CACHE(WB), \ > > + [4] = _I915_CACHE(WT, CLOS1), \ > > + [5] = _I915_CACHE(WB, CLOS1), \ > > + [6] = _I915_CACHE(WT, CLOS2), \ > > + [7] = _I915_CACHE(WB, CLOS2), \ > > } > > .pat = {\ > GEN8_PPAT_UC, \ > GEN8_PPAT_WC, \ > GEN8_PPAT_WT, \ > GEN8_PPAT_WB, \ > GEN12_PPAT_CLOS(1) | GEN8_PPAT_WT, \ > GEN12_PPAT_CLOS(1) | GEN8_PPAT_WB, \ > GEN12_PPAT_CLOS(2) | GEN8_PPAT_WT, \ > GEN12_PPAT_CLOS(2) | GEN8_PPAT_WB, \ > } > > > > > -#define MTL_CACHELEVEL \ > > - .cachelevel_to_pat = { \ > > - [I915_CACHE_NONE] = 2, \ > > - [I915_CACHE_LLC] = 3, \ > > - [I915_CACHE_L3_LLC] = 3, \ > > - [I915_CACHE_WT] = 1, \ > > +#define MTL_CACHE_MODES \ > > + .cache_modes = { \ > > + [0] = I915_CACHE(WB), \ > > + [1] = I915_CACHE(WT), \ > > + [2] = I915_CACHE(UC), \ > > + [3] = _I915_CACHE(WB, COH1W), \ > > + [4] = _I915_CACHE(WB, COH2W), \ > > } > > .pat = {\ > MTL_PPAT_L4_0_WB, \ > MTL_PPAT_L4_1_WT, \ > MTL_PPAT_L4_3_UC, \ > MTL_PPAT_L4_0_WB | MTL_2_COH_1W, \ > MTL_PPAT_L4_0_WB | MTL_3_COH_2W, \ > } > > > > /* Keep in gen based order, and chronological order within a gen */ > > @@ -100,7 +107,7 @@ __diag_ignore_all("-Woverride-init", "Allow > overriding inherited members"); > > .max_pat_index = 3, \ > > GEN_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > #define I845_FEATURES \ > > GEN(2), \ > > @@ -115,7 +122,7 @@ __diag_ignore_all("-Woverride-init", "Allow > overriding inherited members"); > > .max_pat_index = 3, \ > > GEN_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > static const struct intel_device_info i830_info = { > > I830_FEATURES, > > @@ -148,7 +155,7 @@ static const struct intel_device_info i865g_info = { > > .max_pat_index = 3, \ > > GEN_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > static const struct intel_device_info i915g_info = { > > GEN3_FEATURES, > > @@ -211,7 +218,7 @@ static const struct intel_device_info pnv_m_info = { > > .max_pat_index = 3, \ > > GEN_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > static const struct intel_device_info i965g_info = { > > GEN4_FEATURES, > > @@ -255,7 +262,7 @@ static const struct intel_device_info gm45_info = { > > .max_pat_index = 3, \ > > GEN_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > static const struct intel_device_info ilk_d_info = { > > GEN5_FEATURES, > > @@ -285,7 +292,7 @@ static const struct intel_device_info ilk_m_info = { > > .__runtime.ppgtt_size = 31, \ > > GEN_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > #define SNB_D_PLATFORM \ > > GEN6_FEATURES, \ > > @@ -333,7 +340,7 @@ static const struct intel_device_info > snb_m_gt2_info = { > > .__runtime.ppgtt_size = 31, \ > > GEN_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > #define IVB_D_PLATFORM \ > > GEN7_FEATURES, \ > > @@ -390,7 +397,7 @@ static const struct intel_device_info vlv_info = { > > .__runtime.platform_engine_mask = BIT(RCS0) | BIT(VCS0) | BIT(BCS0), > > GEN_DEFAULT_PAGE_SIZES, > > GEN_DEFAULT_REGIONS, > > - LEGACY_CACHELEVEL, > > + LEGACY_CACHE_MODES > > }; > > > > #define G75_FEATURES \ > > @@ -476,7 +483,7 @@ static const struct intel_device_info chv_info = { > > .has_coherent_ggtt = false, > > GEN_DEFAULT_PAGE_SIZES, > > GEN_DEFAULT_REGIONS, > > - LEGACY_CACHELEVEL, > > + LEGACY_CACHE_MODES > > }; > > > > #define GEN9_DEFAULT_PAGE_SIZES \ > > @@ -539,7 +546,7 @@ static const struct intel_device_info > skl_gt4_info = { > > .max_pat_index = 3, \ > > GEN9_DEFAULT_PAGE_SIZES, \ > > GEN_DEFAULT_REGIONS, \ > > - LEGACY_CACHELEVEL > > + LEGACY_CACHE_MODES > > > > static const struct intel_device_info bxt_info = { > > GEN9_LP_FEATURES, > > @@ -643,7 +650,7 @@ static const struct intel_device_info jsl_info = { > > #define GEN12_FEATURES \ > > GEN11_FEATURES, \ > > GEN(12), \ > > - TGL_CACHELEVEL, \ > > + GEN12_CACHE_MODES, \ > > .has_global_mocs = 1, \ > > .has_pxp = 1, \ > > .max_pat_index = 3 > > @@ -711,7 +718,7 @@ static const struct intel_device_info adl_p_info = { > > .__runtime.graphics.ip.ver = 12, \ > > .__runtime.graphics.ip.rel = 50, \ > > XE_HP_PAGE_SIZES, \ > > - TGL_CACHELEVEL, \ > > + GEN12_CACHE_MODES, \ > > .dma_mask_size = 46, \ > > .has_3d_pipeline = 1, \ > > .has_64bit_reloc = 1, \ > > @@ -806,7 +813,7 @@ static const struct intel_device_info pvc_info = { > > BIT(VCS0) | > > BIT(CCS0) | BIT(CCS1) | BIT(CCS2) | BIT(CCS3), > > .require_force_probe = 1, > > - PVC_CACHELEVEL, > > + PVC_CACHE_MODES > > }; > > > > static const struct intel_gt_definition xelpmp_extra_gt[] = { > > @@ -841,7 +848,7 @@ static const struct intel_device_info mtl_info = { > > .__runtime.memory_regions = REGION_SMEM | REGION_STOLEN_LMEM, > > .__runtime.platform_engine_mask = BIT(RCS0) | BIT(BCS0) | BIT(CCS0), > > .require_force_probe = 1, > > - MTL_CACHELEVEL, > > + MTL_CACHE_MODES > > }; > > > > #undef PLATFORM > > diff --git a/drivers/gpu/drm/i915/intel_device_info.h > b/drivers/gpu/drm/i915/intel_device_info.h > > index 069291b3bd37..5cbae7c2ee30 100644 > > --- a/drivers/gpu/drm/i915/intel_device_info.h > > +++ b/drivers/gpu/drm/i915/intel_device_info.h > > @@ -27,6 +27,8 @@ > > > > #include <uapi/drm/i915_drm.h> > > > > +#include "i915_cache.h" > > + > > #include "intel_step.h" > > > > #include "display/intel_display_device.h" > > @@ -248,8 +250,8 @@ struct intel_device_info { > > */ > > const struct intel_runtime_info __runtime; > > > > - u32 cachelevel_to_pat[I915_MAX_CACHE_LEVEL]; > > - u32 max_pat_index; > > + i915_cache_t cache_modes[9]; > > u32 pat[16]; > > See https://gfxspecs.intel.com/Predator/Home/Index/63019, there are > PAT[3..0] What do you mean by this? Anyway I am waiting for Jani to merge the mock_gem_device cleanup patch before I rebase and send v2 to intel-gfx. Until then you can have a peek on trybot. Regards, Tvrtko > > -Fei > > > + unsigned int max_pat_index; > > }; > > > > struct intel_driver_caps { > > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem.c > b/drivers/gpu/drm/i915/selftests/i915_gem.c > > index 61da4ed9d521..e620f73793a5 100644 > > --- a/drivers/gpu/drm/i915/selftests/i915_gem.c > > +++ b/drivers/gpu/drm/i915/selftests/i915_gem.c > > @@ -57,10 +57,7 @@ static void trash_stolen(struct drm_i915_private > *i915) > > u32 __iomem *s; > > int x; > > > > - ggtt->vm.insert_page(&ggtt->vm, dma, slot, > > - i915_gem_get_pat_index(i915, > > - I915_CACHE_NONE), > > - 0); > > + ggtt->vm.insert_page(&ggtt->vm, dma, slot, i915->pat_uc, 0); > > > > s = io_mapping_map_atomic_wc(&ggtt->iomap, slot); > > for (x = 0; x < PAGE_SIZE / sizeof(u32); x++) { > > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_evict.c > b/drivers/gpu/drm/i915/selftests/i915_gem_evict.c > > index f8fe3681c3dc..658a5b59545e 100644 > > --- a/drivers/gpu/drm/i915/selftests/i915_gem_evict.c > > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_evict.c > > @@ -246,7 +246,7 @@ static int igt_evict_for_cache_color(void *arg) > > struct drm_mm_node target = { > > .start = I915_GTT_PAGE_SIZE * 2, > > .size = I915_GTT_PAGE_SIZE, > > - .color = i915_gem_get_pat_index(gt->i915, I915_CACHE_LLC), > > + .color = I915_CACHE(WB), > > }; > > struct drm_i915_gem_object *obj; > > struct i915_vma *vma; > > @@ -309,7 +309,7 @@ static int igt_evict_for_cache_color(void *arg) > > /* Attempt to remove the first *pinned* vma, by removing the (empty) > > * neighbour -- this should fail. > > */ > > - target.color = i915_gem_get_pat_index(gt->i915, I915_CACHE_L3_LLC); > > + target.color = _I915_CACHE(WB, LLC); > > > > mutex_lock(&ggtt->vm.mutex); > > err = i915_gem_evict_for_node(&ggtt->vm, NULL, &target, 0); > > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > index 5c397a2df70e..a24585784f75 100644 > > --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > @@ -135,7 +135,7 @@ fake_dma_object(struct drm_i915_private *i915, > u64 size) > > > > obj->write_domain = I915_GEM_DOMAIN_CPU; > > obj->read_domains = I915_GEM_DOMAIN_CPU; > > - obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE); > > + obj->pat_index = i915->pat_uc; > > > > /* Preallocate the "backing storage" */ > > if (i915_gem_object_pin_pages_unlocked(obj)) > > @@ -358,10 +358,8 @@ static int lowlevel_hole(struct > i915_address_space *vm, > > mock_vma_res->start = addr; > > > > with_intel_runtime_pm(vm->gt->uncore->rpm, wakeref) > > - vm->insert_entries(vm, mock_vma_res, > > - i915_gem_get_pat_index(vm->i915, > > - I915_CACHE_NONE), > > - 0); > > + vm->insert_entries(vm, mock_vma_res, > > + vm->i915->pat_uc, 0); > > } > > count = n; > > > > @@ -1379,10 +1377,7 @@ static int igt_ggtt_page(void *arg) > > > > ggtt->vm.insert_page(&ggtt->vm, > > i915_gem_object_get_dma_address(obj, 0), > > - offset, > > - i915_gem_get_pat_index(i915, > > - I915_CACHE_NONE), > > - 0); > > + offset, ggtt->vm.i915->pat_uc, 0); > > } > > > > order = i915_random_order(count, &prng); > > diff --git a/drivers/gpu/drm/i915/selftests/intel_memory_region.c > b/drivers/gpu/drm/i915/selftests/intel_memory_region.c > > index d985d9bae2e8..b82fe0ef8cd7 100644 > > --- a/drivers/gpu/drm/i915/selftests/intel_memory_region.c > > +++ b/drivers/gpu/drm/i915/selftests/intel_memory_region.c > > @@ -1070,9 +1070,7 @@ static int igt_lmem_write_cpu(void *arg) > > /* Put the pages into a known state -- from the gpu for added fun */ > > intel_engine_pm_get(engine); > > err = intel_context_migrate_clear(engine->gt->migrate.context, NULL, > > - obj->mm.pages->sgl, > > - i915_gem_get_pat_index(i915, > > - I915_CACHE_NONE), > > + obj->mm.pages->sgl, i915->pat_uc, > > true, 0xdeadbeaf, &rq); > > if (rq) { > > dma_resv_add_fence(obj->base.resv, &rq->fence, > > diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c > b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > > index 09d4bbcdcdbf..ad778842cba2 100644 > > --- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c > > +++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c > > @@ -126,7 +126,12 @@ struct drm_i915_private *mock_gem_device(void) > > struct drm_i915_private *i915; > > struct intel_device_info *i915_info; > > struct pci_dev *pdev; > > - unsigned int i; > > + static const i915_cache_t legacy_cache_modes[] = { > > + [0] = I915_CACHE(UC), > > + [1] = I915_CACHE(WB), > > + [2] = _I915_CACHE(WB, L3), > > + [3] = I915_CACHE(WT), > > + }; > > int ret; > > > > pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); > > @@ -187,8 +192,7 @@ struct drm_i915_private *mock_gem_device(void) > > /* simply use legacy cache level for mock device */ > > i915_info = (struct intel_device_info *)INTEL_INFO(i915); > > i915_info->max_pat_index = 3; > > - for (i = 0; i < I915_MAX_CACHE_LEVEL; i++) > > - i915_info->cachelevel_to_pat[i] = i; > > + memcpy(i915_info->cache_modes, legacy_cache_modes, > sizeof(legacy_cache_modes)); > > > > intel_memory_regions_hw_probe(i915); > > > > -- > > 2.39.2 >
diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile index 2be9dd960540..2c3da8f0c78e 100644 --- a/drivers/gpu/drm/i915/Makefile +++ b/drivers/gpu/drm/i915/Makefile @@ -30,6 +30,7 @@ subdir-ccflags-y += -I$(srctree)/$(src) # core driver code i915-y += i915_driver.o \ i915_drm_client.o \ + i915_cache.o \ i915_config.o \ i915_getparam.o \ i915_ioctl.o \ diff --git a/drivers/gpu/drm/i915/gem/i915_gem_domain.c b/drivers/gpu/drm/i915/gem/i915_gem_domain.c index dfaaa8b66ac3..49bfae45390f 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_domain.c +++ b/drivers/gpu/drm/i915/gem/i915_gem_domain.c @@ -8,6 +8,7 @@ #include "display/intel_frontbuffer.h" #include "gt/intel_gt.h" +#include "i915_cache.h" #include "i915_drv.h" #include "i915_gem_clflush.h" #include "i915_gem_domain.h" @@ -27,15 +28,8 @@ static bool gpu_write_needs_clflush(struct drm_i915_gem_object *obj) if (IS_DGFX(i915)) return false; - /* - * For objects created by userspace through GEM_CREATE with pat_index - * set by set_pat extension, i915_gem_object_has_cache_level() will - * always return true, because the coherency of such object is managed - * by userspace. Othereise the call here would fall back to checking - * whether the object is un-cached or write-through. - */ - return !(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) || - i915_gem_object_has_cache_level(obj, I915_CACHE_WT)); + return i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1 && + i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WT) != 1; } bool i915_gem_cpu_write_needs_clflush(struct drm_i915_gem_object *obj) @@ -272,15 +266,18 @@ i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write) int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, enum i915_cache_level cache_level) { + struct drm_i915_private *i915 = to_i915(obj->base.dev); + i915_cache_t mode; int ret; - /* - * For objects created by userspace through GEM_CREATE with pat_index - * set by set_pat extension, simply return 0 here without touching - * the cache setting, because such objects should have an immutable - * cache setting by desgin and always managed by userspace. - */ - if (i915_gem_object_has_cache_level(obj, cache_level)) + if (obj->pat_set_by_user) + return -EINVAL; + + ret = i915_cache_level_to_pat_and_mode(i915, cache_level, &mode); + if (ret < 0) + return -EINVAL; + + if (mode == obj->cache_mode) return 0; ret = i915_gem_object_wait(obj, @@ -326,10 +323,9 @@ int i915_gem_get_caching_ioctl(struct drm_device *dev, void *data, goto out; } - if (i915_gem_object_has_cache_level(obj, I915_CACHE_LLC) || - i915_gem_object_has_cache_level(obj, I915_CACHE_L3_LLC)) + if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WB)) args->caching = I915_CACHING_CACHED; - else if (i915_gem_object_has_cache_level(obj, I915_CACHE_WT)) + else if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_WT)) args->caching = I915_CACHING_DISPLAY; else args->caching = I915_CACHING_NONE; diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c index d3208a325614..ee85221fa6eb 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c @@ -640,15 +640,9 @@ static inline int use_cpu_reloc(const struct reloc_cache *cache, if (DBG_FORCE_RELOC == FORCE_GTT_RELOC) return false; - /* - * For objects created by userspace through GEM_CREATE with pat_index - * set by set_pat extension, i915_gem_object_has_cache_level() always - * return true, otherwise the call would fall back to checking whether - * the object is un-cached. - */ return (cache->has_llc || obj->cache_dirty || - !i915_gem_object_has_cache_level(obj, I915_CACHE_NONE)); + i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1); } static int eb_reserve_vma(struct i915_execbuffer *eb, @@ -1329,10 +1323,7 @@ static void *reloc_iomap(struct i915_vma *batch, if (drm_mm_node_allocated(&cache->node)) { ggtt->vm.insert_page(&ggtt->vm, i915_gem_object_get_dma_address(obj, page), - offset, - i915_gem_get_pat_index(ggtt->vm.i915, - I915_CACHE_NONE), - 0); + offset, eb->i915->pat_uc, 0); } else { offset += page << PAGE_SHIFT; } diff --git a/drivers/gpu/drm/i915/gem/i915_gem_mman.c b/drivers/gpu/drm/i915/gem/i915_gem_mman.c index aa4d842d4c5a..5e21aedb02d2 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_mman.c +++ b/drivers/gpu/drm/i915/gem/i915_gem_mman.c @@ -386,13 +386,11 @@ static vm_fault_t vm_fault_gtt(struct vm_fault *vmf) /* * For objects created by userspace through GEM_CREATE with pat_index * set by set_pat extension, coherency is managed by userspace, make - * sure we don't fail handling the vm fault by calling - * i915_gem_object_has_cache_level() which always return true for such - * objects. Otherwise this helper function would fall back to checking - * whether the object is un-cached. + * sure we don't fail handling the vm fault by making sure that we + * know the object is uncached or that we have LLC. */ - if (!(i915_gem_object_has_cache_level(obj, I915_CACHE_NONE) || - HAS_LLC(i915))) { + if (i915_gem_object_has_cache_mode(obj, I915_CACHE_MODE_UC) != 1 && + !HAS_LLC(i915)) { ret = -EFAULT; goto err_unpin; } diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.c b/drivers/gpu/drm/i915/gem/i915_gem_object.c index 0004d5fa7cc2..52c6c5f09bdd 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_object.c +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.c @@ -45,33 +45,6 @@ static struct kmem_cache *slab_objects; static const struct drm_gem_object_funcs i915_gem_object_funcs; -unsigned int i915_gem_get_pat_index(struct drm_i915_private *i915, - enum i915_cache_level level) -{ - if (drm_WARN_ON(&i915->drm, level >= I915_MAX_CACHE_LEVEL)) - return 0; - - return INTEL_INFO(i915)->cachelevel_to_pat[level]; -} - -bool i915_gem_object_has_cache_level(const struct drm_i915_gem_object *obj, - enum i915_cache_level lvl) -{ - /* - * In case the pat_index is set by user space, this kernel mode - * driver should leave the coherency to be managed by user space, - * simply return true here. - */ - if (obj->pat_set_by_user) - return true; - - /* - * Otherwise the pat_index should have been converted from cache_level - * so that the following comparison is valid. - */ - return obj->pat_index == i915_gem_get_pat_index(obj_to_i915(obj), lvl); -} - struct drm_i915_gem_object *i915_gem_object_alloc(void) { struct drm_i915_gem_object *obj; @@ -144,6 +117,24 @@ void __i915_gem_object_fini(struct drm_i915_gem_object *obj) dma_resv_fini(&obj->base._resv); } +void __i915_gem_object_update_coherency(struct drm_i915_gem_object *obj) +{ + struct drm_i915_private *i915 = to_i915(obj->base.dev); + const unsigned int mode = I915_CACHE_MODE(obj->cache_mode); + + if (!(mode == I915_CACHE_MODE_UNKNOWN || mode == I915_CACHE_MODE_UC)) + obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | + I915_BO_CACHE_COHERENT_FOR_WRITE); + else if (mode != I915_CACHE_MODE_UNKNOWN && HAS_LLC(i915)) + obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; + else + obj->cache_coherent = 0; + + obj->cache_dirty = + !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && + !IS_DGFX(i915); +} + /** * i915_gem_object_set_cache_coherency - Mark up the object's coherency levels * for a given cache_level @@ -154,20 +145,15 @@ void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj, unsigned int cache_level) { struct drm_i915_private *i915 = to_i915(obj->base.dev); + i915_cache_t mode; + int found; - obj->pat_index = i915_gem_get_pat_index(i915, cache_level); + found = i915_cache_level_to_pat_and_mode(i915, cache_level, &mode); + GEM_WARN_ON(found < 0); + obj->pat_index = found; + obj->cache_mode = mode; - if (cache_level != I915_CACHE_NONE) - obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | - I915_BO_CACHE_COHERENT_FOR_WRITE); - else if (HAS_LLC(i915)) - obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; - else - obj->cache_coherent = 0; - - obj->cache_dirty = - !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && - !IS_DGFX(i915); + __i915_gem_object_update_coherency(obj); } /** @@ -187,18 +173,9 @@ void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj, return; obj->pat_index = pat_index; + obj->cache_mode = INTEL_INFO(i915)->cache_modes[pat_index]; - if (pat_index != i915_gem_get_pat_index(i915, I915_CACHE_NONE)) - obj->cache_coherent = (I915_BO_CACHE_COHERENT_FOR_READ | - I915_BO_CACHE_COHERENT_FOR_WRITE); - else if (HAS_LLC(i915)) - obj->cache_coherent = I915_BO_CACHE_COHERENT_FOR_READ; - else - obj->cache_coherent = 0; - - obj->cache_dirty = - !(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE) && - !IS_DGFX(i915); + __i915_gem_object_update_coherency(obj); } bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj) @@ -215,6 +192,7 @@ bool i915_gem_object_can_bypass_llc(struct drm_i915_gem_object *obj) /* * Always flush cache for UMD objects at creation time. */ + /* QQQ/FIXME why? avoidable performance penalty? */ if (obj->pat_set_by_user) return true; diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.h b/drivers/gpu/drm/i915/gem/i915_gem_object.h index 884a17275b3a..f84f41e9f81f 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_object.h +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.h @@ -13,6 +13,7 @@ #include "display/intel_frontbuffer.h" #include "intel_memory_region.h" +#include "i915_cache.h" #include "i915_gem_object_types.h" #include "i915_gem_gtt.h" #include "i915_gem_ww.h" @@ -32,10 +33,18 @@ static inline bool i915_gem_object_size_2big(u64 size) return false; } -unsigned int i915_gem_get_pat_index(struct drm_i915_private *i915, - enum i915_cache_level level); -bool i915_gem_object_has_cache_level(const struct drm_i915_gem_object *obj, - enum i915_cache_level lvl); +static inline int +i915_gem_object_has_cache_mode(const struct drm_i915_gem_object *obj, + unsigned int mode) +{ + if (I915_CACHE_MODE(obj->cache_mode) == mode) + return 1; + else if (obj->pat_set_by_user) + return -1; /* Unknown, callers should assume no. */ + else + return 0; +} + void i915_gem_init__objects(struct drm_i915_private *i915); void i915_objects_module_exit(void); @@ -764,6 +773,7 @@ int i915_gem_object_wait_moving_fence(struct drm_i915_gem_object *obj, bool intr); bool i915_gem_object_has_unknown_state(struct drm_i915_gem_object *obj); +void __i915_gem_object_update_coherency(struct drm_i915_gem_object *obj); void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj, unsigned int cache_level); void i915_gem_object_set_pat_index(struct drm_i915_gem_object *obj, diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h index 8de2b91b3edf..1f9fa28d07df 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h +++ b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h @@ -14,6 +14,7 @@ #include <uapi/drm/i915_drm.h> #include "i915_active.h" +#include "i915_cache.h" #include "i915_selftest.h" #include "i915_vma_resource.h" @@ -116,93 +117,6 @@ struct drm_i915_gem_object_ops { const char *name; /* friendly name for debug, e.g. lockdep classes */ }; -/** - * enum i915_cache_level - The supported GTT caching values for system memory - * pages. - * - * These translate to some special GTT PTE bits when binding pages into some - * address space. It also determines whether an object, or rather its pages are - * coherent with the GPU, when also reading or writing through the CPU cache - * with those pages. - * - * Userspace can also control this through struct drm_i915_gem_caching. - */ -enum i915_cache_level { - /** - * @I915_CACHE_NONE: - * - * GPU access is not coherent with the CPU cache. If the cache is dirty - * and we need the underlying pages to be coherent with some later GPU - * access then we need to manually flush the pages. - * - * On shared LLC platforms reads and writes through the CPU cache are - * still coherent even with this setting. See also - * &drm_i915_gem_object.cache_coherent for more details. Due to this we - * should only ever use uncached for scanout surfaces, otherwise we end - * up over-flushing in some places. - * - * This is the default on non-LLC platforms. - */ - I915_CACHE_NONE = 0, - /** - * @I915_CACHE_LLC: - * - * GPU access is coherent with the CPU cache. If the cache is dirty, - * then the GPU will ensure that access remains coherent, when both - * reading and writing through the CPU cache. GPU writes can dirty the - * CPU cache. - * - * Not used for scanout surfaces. - * - * Applies to both platforms with shared LLC(HAS_LLC), and snooping - * based platforms(HAS_SNOOP). - * - * This is the default on shared LLC platforms. The only exception is - * scanout objects, where the display engine is not coherent with the - * CPU cache. For such objects I915_CACHE_NONE or I915_CACHE_WT is - * automatically applied by the kernel in pin_for_display, if userspace - * has not done so already. - */ - I915_CACHE_LLC, - /** - * @I915_CACHE_L3_LLC: - * - * Explicitly enable the Gfx L3 cache, with coherent LLC. - * - * The Gfx L3 sits between the domain specific caches, e.g - * sampler/render caches, and the larger LLC. LLC is coherent with the - * GPU, but L3 is only visible to the GPU, so likely needs to be flushed - * when the workload completes. - * - * Not used for scanout surfaces. - * - * Only exposed on some gen7 + GGTT. More recent hardware has dropped - * this explicit setting, where it should now be enabled by default. - */ - I915_CACHE_L3_LLC, - /** - * @I915_CACHE_WT: - * - * Write-through. Used for scanout surfaces. - * - * The GPU can utilise the caches, while still having the display engine - * be coherent with GPU writes, as a result we don't need to flush the - * CPU caches when moving out of the render domain. This is the default - * setting chosen by the kernel, if supported by the HW, otherwise we - * fallback to I915_CACHE_NONE. On the CPU side writes through the CPU - * cache still need to be flushed, to remain coherent with the display - * engine. - */ - I915_CACHE_WT, - /** - * @I915_MAX_CACHE_LEVEL: - * - * Mark the last entry in the enum. Used for defining cachelevel_to_pat - * array for cache_level to pat translation table. - */ - I915_MAX_CACHE_LEVEL, -}; - enum i915_map_type { I915_MAP_WB = 0, I915_MAP_WC, @@ -375,6 +289,9 @@ struct drm_i915_gem_object { unsigned int mem_flags; #define I915_BO_FLAG_STRUCT_PAGE BIT(0) /* Object backed by struct pages */ #define I915_BO_FLAG_IOMEM BIT(1) /* Object backed by IO memory */ + + i915_cache_t cache_mode; + /** * @pat_index: The desired PAT index. * @@ -409,9 +326,7 @@ struct drm_i915_gem_object { * Check for @pat_set_by_user to find out if an object has pat index set * by userspace. The ioctl's to change cache settings have also been * disabled for the objects with pat index set by userspace. Please don't - * assume @cache_coherent having the flags set as describe here. A helper - * function i915_gem_object_has_cache_level() provides one way to bypass - * the use of this field. + * assume @cache_coherent having the flags set as describe here. * * Track whether the pages are coherent with the GPU if reading or * writing through the CPU caches. The largely depends on the @@ -492,9 +407,7 @@ struct drm_i915_gem_object { * Check for @pat_set_by_user to find out if an object has pat index set * by userspace. The ioctl's to change cache settings have also been * disabled for the objects with pat_index set by userspace. Please don't - * assume @cache_dirty is set as describe here. Also see helper function - * i915_gem_object_has_cache_level() for possible ways to bypass the use - * of this field. + * assume @cache_dirty is set as describe here. * * Track if we are we dirty with writes through the CPU cache for this * object. As a result reading directly from main memory might yield diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c index 3b094d36a0b0..a7012f1a9c70 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c @@ -563,11 +563,8 @@ static void dbg_poison(struct i915_ggtt *ggtt, while (size) { void __iomem *s; - ggtt->vm.insert_page(&ggtt->vm, addr, - ggtt->error_capture.start, - i915_gem_get_pat_index(ggtt->vm.i915, - I915_CACHE_NONE), - 0); + ggtt->vm.insert_page(&ggtt->vm, addr, ggtt->error_capture.start, + ggtt->vm.i915->pat_uc, 0); mb(); s = io_mapping_map_wc(&ggtt->iomap, diff --git a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c index 7078af2f8f79..e794bd2a7ccb 100644 --- a/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c +++ b/drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c @@ -58,6 +58,16 @@ i915_ttm_cache_level(struct drm_i915_private *i915, struct ttm_resource *res, I915_CACHE_NONE; } +static unsigned int +i915_ttm_cache_pat(struct drm_i915_private *i915, struct ttm_resource *res, + struct ttm_tt *ttm) +{ + return ((HAS_LLC(i915) || HAS_SNOOP(i915)) && + !i915_ttm_gtt_binds_lmem(res) && + ttm->caching == ttm_cached) ? i915->pat_wb : + i915->pat_uc; +} + static struct intel_memory_region * i915_ttm_region(struct ttm_device *bdev, int ttm_mem_type) { @@ -196,7 +206,7 @@ static struct dma_fence *i915_ttm_accel_move(struct ttm_buffer_object *bo, struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo); struct i915_request *rq; struct ttm_tt *src_ttm = bo->ttm; - enum i915_cache_level src_level, dst_level; + unsigned int src_pat, dst_pat; int ret; if (!to_gt(i915)->migrate.context || intel_gt_is_wedged(to_gt(i915))) @@ -206,16 +216,15 @@ static struct dma_fence *i915_ttm_accel_move(struct ttm_buffer_object *bo, if (I915_SELFTEST_ONLY(fail_gpu_migration)) clear = true; - dst_level = i915_ttm_cache_level(i915, dst_mem, dst_ttm); + dst_pat = i915_ttm_cache_pat(i915, dst_mem, dst_ttm); if (clear) { if (bo->type == ttm_bo_type_kernel && !I915_SELFTEST_ONLY(fail_gpu_migration)) return ERR_PTR(-EINVAL); intel_engine_pm_get(to_gt(i915)->migrate.context->engine); - ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, - dst_st->sgl, - i915_gem_get_pat_index(i915, dst_level), + ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, + deps, dst_st->sgl, dst_pat, i915_ttm_gtt_binds_lmem(dst_mem), 0, &rq); } else { @@ -225,14 +234,13 @@ static struct dma_fence *i915_ttm_accel_move(struct ttm_buffer_object *bo, if (IS_ERR(src_rsgt)) return ERR_CAST(src_rsgt); - src_level = i915_ttm_cache_level(i915, bo->resource, src_ttm); + src_pat = i915_ttm_cache_pat(i915, bo->resource, src_ttm); intel_engine_pm_get(to_gt(i915)->migrate.context->engine); ret = intel_context_migrate_copy(to_gt(i915)->migrate.context, deps, src_rsgt->table.sgl, - i915_gem_get_pat_index(i915, src_level), + src_pat, i915_ttm_gtt_binds_lmem(bo->resource), - dst_st->sgl, - i915_gem_get_pat_index(i915, dst_level), + dst_st->sgl, dst_pat, i915_ttm_gtt_binds_lmem(dst_mem), &rq); diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c index df6c9a84252c..c8925918784e 100644 --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c @@ -354,7 +354,7 @@ fake_huge_pages_object(struct drm_i915_private *i915, u64 size, bool single) obj->write_domain = I915_GEM_DOMAIN_CPU; obj->read_domains = I915_GEM_DOMAIN_CPU; - obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE); + i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE); return obj; } diff --git a/drivers/gpu/drm/i915/gt/gen6_ppgtt.c b/drivers/gpu/drm/i915/gt/gen6_ppgtt.c index c2bdc133c89a..fb69f667652a 100644 --- a/drivers/gpu/drm/i915/gt/gen6_ppgtt.c +++ b/drivers/gpu/drm/i915/gt/gen6_ppgtt.c @@ -226,9 +226,7 @@ static int gen6_ppgtt_init_scratch(struct gen6_ppgtt *ppgtt) return ret; vm->scratch[0]->encode = - vm->pte_encode(px_dma(vm->scratch[0]), - i915_gem_get_pat_index(vm->i915, - I915_CACHE_NONE), + vm->pte_encode(px_dma(vm->scratch[0]), vm->i915->pat_uc, PTE_READ_ONLY); vm->scratch[1] = vm->alloc_pt_dma(vm, I915_GTT_PAGE_SIZE_4K); diff --git a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c index f948d33e5ec5..a6692ea1a91e 100644 --- a/drivers/gpu/drm/i915/gt/gen8_ppgtt.c +++ b/drivers/gpu/drm/i915/gt/gen8_ppgtt.c @@ -40,16 +40,11 @@ static u64 gen8_pte_encode(dma_addr_t addr, if (flags & PTE_LM) pte |= GEN12_PPGTT_PTE_LM; - /* - * For pre-gen12 platforms pat_index is the same as enum - * i915_cache_level, so the switch-case here is still valid. - * See translation table defined by LEGACY_CACHELEVEL. - */ switch (pat_index) { - case I915_CACHE_NONE: + case 0: pte |= PPAT_UNCACHED; break; - case I915_CACHE_WT: + case 3: pte |= PPAT_DISPLAY_ELLC; break; default: @@ -853,9 +848,7 @@ static int gen8_init_scratch(struct i915_address_space *vm) pte_flags |= PTE_LM; vm->scratch[0]->encode = - vm->pte_encode(px_dma(vm->scratch[0]), - i915_gem_get_pat_index(vm->i915, - I915_CACHE_NONE), + vm->pte_encode(px_dma(vm->scratch[0]), vm->i915->pat_uc, pte_flags); for (i = 1; i <= vm->top; i++) { diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt.c b/drivers/gpu/drm/i915/gt/intel_ggtt.c index dd0ed941441a..c97379cf8241 100644 --- a/drivers/gpu/drm/i915/gt/intel_ggtt.c +++ b/drivers/gpu/drm/i915/gt/intel_ggtt.c @@ -921,9 +921,7 @@ static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size) pte_flags |= PTE_LM; ggtt->vm.scratch[0]->encode = - ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]), - i915_gem_get_pat_index(i915, - I915_CACHE_NONE), + ggtt->vm.pte_encode(px_dma(ggtt->vm.scratch[0]), i915->pat_uc, pte_flags); return 0; @@ -1297,10 +1295,7 @@ bool i915_ggtt_resume_vm(struct i915_address_space *vm) * ptes to be repopulated. */ vma->resource->bound_flags = 0; - vma->ops->bind_vma(vm, NULL, vma->resource, - obj ? obj->pat_index : - i915_gem_get_pat_index(vm->i915, - I915_CACHE_NONE), + vma->ops->bind_vma(vm, NULL, vma->resource, obj->cache_mode, was_bound); if (obj) { /* only used during resume => exclusive access */ diff --git a/drivers/gpu/drm/i915/gt/intel_migrate.c b/drivers/gpu/drm/i915/gt/intel_migrate.c index 6023288b0e2d..81f7834cc2db 100644 --- a/drivers/gpu/drm/i915/gt/intel_migrate.c +++ b/drivers/gpu/drm/i915/gt/intel_migrate.c @@ -45,9 +45,7 @@ static void xehpsdv_toggle_pdes(struct i915_address_space *vm, * Insert a dummy PTE into every PT that will map to LMEM to ensure * we have a correctly setup PDE structure for later use. */ - vm->insert_page(vm, 0, d->offset, - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), - PTE_LM); + vm->insert_page(vm, 0, d->offset, vm->i915->pat_uc, PTE_LM); GEM_BUG_ON(!pt->is_compact); d->offset += SZ_2M; } @@ -65,9 +63,7 @@ static void xehpsdv_insert_pte(struct i915_address_space *vm, * alignment is 64K underneath for the pt, and we are careful * not to access the space in the void. */ - vm->insert_page(vm, px_dma(pt), d->offset, - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), - PTE_LM); + vm->insert_page(vm, px_dma(pt), d->offset, vm->i915->pat_uc, PTE_LM); d->offset += SZ_64K; } @@ -77,8 +73,7 @@ static void insert_pte(struct i915_address_space *vm, { struct insert_pte_data *d = data; - vm->insert_page(vm, px_dma(pt), d->offset, - i915_gem_get_pat_index(vm->i915, I915_CACHE_NONE), + vm->insert_page(vm, px_dma(pt), d->offset, vm->i915->pat_uc, i915_gem_object_is_lmem(pt->base) ? PTE_LM : 0); d->offset += PAGE_SIZE; } diff --git a/drivers/gpu/drm/i915/gt/selftest_migrate.c b/drivers/gpu/drm/i915/gt/selftest_migrate.c index 3def5ca72dec..a67ede65d816 100644 --- a/drivers/gpu/drm/i915/gt/selftest_migrate.c +++ b/drivers/gpu/drm/i915/gt/selftest_migrate.c @@ -904,8 +904,7 @@ static int perf_clear_blt(void *arg) err = __perf_clear_blt(gt->migrate.context, dst->mm.pages->sgl, - i915_gem_get_pat_index(gt->i915, - I915_CACHE_NONE), + gt->i915->pat_uc, i915_gem_object_is_lmem(dst), sizes[i]); @@ -995,12 +994,10 @@ static int perf_copy_blt(void *arg) err = __perf_copy_blt(gt->migrate.context, src->mm.pages->sgl, - i915_gem_get_pat_index(gt->i915, - I915_CACHE_NONE), + gt->i915->pat_uc, i915_gem_object_is_lmem(src), dst->mm.pages->sgl, - i915_gem_get_pat_index(gt->i915, - I915_CACHE_NONE), + gt->i915->pat_uc, i915_gem_object_is_lmem(dst), sz); diff --git a/drivers/gpu/drm/i915/gt/selftest_reset.c b/drivers/gpu/drm/i915/gt/selftest_reset.c index 79aa6ac66ad2..327dc9294e0f 100644 --- a/drivers/gpu/drm/i915/gt/selftest_reset.c +++ b/drivers/gpu/drm/i915/gt/selftest_reset.c @@ -84,11 +84,8 @@ __igt_reset_stolen(struct intel_gt *gt, void __iomem *s; void *in; - ggtt->vm.insert_page(&ggtt->vm, dma, - ggtt->error_capture.start, - i915_gem_get_pat_index(gt->i915, - I915_CACHE_NONE), - 0); + ggtt->vm.insert_page(&ggtt->vm, dma, ggtt->error_capture.start, + gt->i915->pat_uc, 0); mb(); s = io_mapping_map_wc(&ggtt->iomap, @@ -127,11 +124,8 @@ __igt_reset_stolen(struct intel_gt *gt, void *in; u32 x; - ggtt->vm.insert_page(&ggtt->vm, dma, - ggtt->error_capture.start, - i915_gem_get_pat_index(gt->i915, - I915_CACHE_NONE), - 0); + ggtt->vm.insert_page(&ggtt->vm, dma, ggtt->error_capture.start, + gt->i915->pat_uc, 0); mb(); s = io_mapping_map_wc(&ggtt->iomap, diff --git a/drivers/gpu/drm/i915/gt/selftest_timeline.c b/drivers/gpu/drm/i915/gt/selftest_timeline.c index 39c3ec12df1a..db64dc7d3fce 100644 --- a/drivers/gpu/drm/i915/gt/selftest_timeline.c +++ b/drivers/gpu/drm/i915/gt/selftest_timeline.c @@ -836,7 +836,10 @@ static int setup_watcher(struct hwsp_watcher *w, struct intel_gt *gt, return PTR_ERR(obj); /* keep the same cache settings as timeline */ - i915_gem_object_set_pat_index(obj, tl->hwsp_ggtt->obj->pat_index); + obj->pat_index = tl->hwsp_ggtt->obj->pat_index; + obj->cache_mode = tl->hwsp_ggtt->obj->cache_mode; + __i915_gem_object_update_coherency(obj); + w->map = i915_gem_object_pin_map_unlocked(obj, page_unmask_bits(tl->hwsp_ggtt->obj->mm.mapping)); if (IS_ERR(w->map)) { diff --git a/drivers/gpu/drm/i915/gt/selftest_tlb.c b/drivers/gpu/drm/i915/gt/selftest_tlb.c index 3bd6b540257b..6049f01be219 100644 --- a/drivers/gpu/drm/i915/gt/selftest_tlb.c +++ b/drivers/gpu/drm/i915/gt/selftest_tlb.c @@ -36,8 +36,6 @@ pte_tlbinv(struct intel_context *ce, u64 length, struct rnd_state *prng) { - const unsigned int pat_index = - i915_gem_get_pat_index(ce->vm->i915, I915_CACHE_NONE); struct drm_i915_gem_object *batch; struct drm_mm_node vb_node; struct i915_request *rq; @@ -157,7 +155,8 @@ pte_tlbinv(struct intel_context *ce, /* Flip the PTE between A and B */ if (i915_gem_object_is_lmem(vb->obj)) pte_flags |= PTE_LM; - ce->vm->insert_entries(ce->vm, &vb_res, pat_index, pte_flags); + ce->vm->insert_entries(ce->vm, &vb_res, ce->vm->i915->pat_uc, + pte_flags); /* Flush the PTE update to concurrent HW */ tlbinv(ce->vm, addr & -length, length); diff --git a/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c b/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c index d408856ae4c0..e099414d624d 100644 --- a/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c +++ b/drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c @@ -991,14 +991,10 @@ static void uc_fw_bind_ggtt(struct intel_uc_fw *uc_fw) if (ggtt->vm.raw_insert_entries) ggtt->vm.raw_insert_entries(&ggtt->vm, vma_res, - i915_gem_get_pat_index(ggtt->vm.i915, - I915_CACHE_NONE), - pte_flags); + ggtt->vm.i915->pat_uc, pte_flags); else ggtt->vm.insert_entries(&ggtt->vm, vma_res, - i915_gem_get_pat_index(ggtt->vm.i915, - I915_CACHE_NONE), - pte_flags); + ggtt->vm.i915->pat_uc, pte_flags); } static void uc_fw_unbind_ggtt(struct intel_uc_fw *uc_fw) diff --git a/drivers/gpu/drm/i915/i915_cache.c b/drivers/gpu/drm/i915/i915_cache.c new file mode 100644 index 000000000000..7a8002ebd2ec --- /dev/null +++ b/drivers/gpu/drm/i915/i915_cache.c @@ -0,0 +1,59 @@ +/* + * SPDX-License-Identifier: MIT + * + * Copyright © 2023 Intel Corporation + */ + +#include "i915_cache.h" +#include "i915_drv.h" + +static int find_pat(const struct intel_device_info *info, i915_cache_t mode) +{ + int i; + + for (i = 0; i < ARRAY_SIZE(info->cache_modes); i++) { + if (info->cache_modes[i] == mode) + return i; + } + + return -1; +} + +void i915_cache_init(struct drm_i915_private *i915) +{ + int ret; + + ret = find_pat(INTEL_INFO(i915), I915_CACHE(UC)); + WARN_ON(ret < 0); + drm_info(&i915->drm, "Using PAT index %u for uncached access\n", ret); + i915->pat_uc = ret; + + ret = find_pat(INTEL_INFO(i915), I915_CACHE(WB)); + WARN_ON(ret < 0); + drm_info(&i915->drm, "Using PAT index %u for write-back access\n", ret); + i915->pat_wb = ret; +} + +int i915_cache_level_to_pat_and_mode(struct drm_i915_private *i915, + unsigned int cache_level, + i915_cache_t *mode) +{ + const struct intel_device_info *info = INTEL_INFO(i915); + i915_cache_t level_to_mode[] = { + [I915_CACHE_NONE] = I915_CACHE(UC), + [I915_CACHE_WT] = I915_CACHE(WT), + [I915_CACHE_LLC] = I915_CACHE(WB), + }; + int ret; + + if (GRAPHICS_VER(i915) >= 12) + level_to_mode[I915_CACHE_L3_LLC] = I915_CACHE(WB); + else + level_to_mode[I915_CACHE_L3_LLC] = _I915_CACHE(WB, LLC); + + ret = find_pat(info, level_to_mode[cache_level]); + if (ret >= 0 && mode) + *mode = info->cache_modes[ret]; + + return ret; +} diff --git a/drivers/gpu/drm/i915/i915_cache.h b/drivers/gpu/drm/i915/i915_cache.h new file mode 100644 index 000000000000..0df03f1f01ef --- /dev/null +++ b/drivers/gpu/drm/i915/i915_cache.h @@ -0,0 +1,129 @@ +/* SPDX-License-Identifier: MIT */ +/* + * Copyright © 2023 Intel Corporation + */ + +#ifndef __I915_CACHE_H__ +#define __I915_CACHE_H__ + +#include <linux/types.h> + +struct drm_i915_private; + +typedef u16 i915_cache_t; + +#define I915_CACHE(mode) \ + (i915_cache_t)(I915_CACHE_MODE_##mode) + +#define _I915_CACHE(mode, flag) \ + (i915_cache_t)((I915_CACHE_MODE_##mode) | ( BIT(8 + I915_CACHE_##flag))) + +#define I915_CACHE_MODE(cache) \ + (unsigned int)(((i915_cache_t)(cache)) & 0xff) +#define I915_CACHE_FLAGS(cache) \ + (unsigned int)((((i915_cache_t)(cache) & 0xff00)) >> 16) + +/* Cache mode values */ +#define I915_CACHE_MODE_UNKNOWN (0) +#define I915_CACHE_MODE_UC (1) +#define I915_CACHE_MODE_WB (2) +#define I915_CACHE_MODE_WT (3) +#define I915_CACHE_MODE_WC (4) + +/* Mode flag bits */ +#define I915_CACHE_L3 (0) +#define I915_CACHE_COH1W (1) +#define I915_CACHE_COH2W (2) +#define I915_CACHE_CLOS1 (3) +#define I915_CACHE_CLOS2 (4) + +void i915_cache_init(struct drm_i915_private *i915); +int i915_cache_level_to_pat_and_mode(struct drm_i915_private *i915, + unsigned int cache_level, + i915_cache_t *mode); + +/* + * Legacy/kernel internal interface below: + */ + +/** + * enum i915_cache_level - The supported GTT caching values for system memory + * pages. + * + * These translate to some special GTT PTE bits when binding pages into some + * address space. It also determines whether an object, or rather its pages are + * coherent with the GPU, when also reading or writing through the CPU cache + * with those pages. + * + * Userspace can also control this through struct drm_i915_gem_caching. + */ +enum i915_cache_level { + /** + * @I915_CACHE_NONE: + * + * GPU access is not coherent with the CPU cache. If the cache is dirty + * and we need the underlying pages to be coherent with some later GPU + * access then we need to manually flush the pages. + * + * On shared LLC platforms reads and writes through the CPU cache are + * still coherent even with this setting. See also + * &drm_i915_gem_object.cache_coherent for more details. Due to this we + * should only ever use uncached for scanout surfaces, otherwise we end + * up over-flushing in some places. + * + * This is the default on non-LLC platforms. + */ + I915_CACHE_NONE = 0, + /** + * @I915_CACHE_LLC: + * + * GPU access is coherent with the CPU cache. If the cache is dirty, + * then the GPU will ensure that access remains coherent, when both + * reading and writing through the CPU cache. GPU writes can dirty the + * CPU cache. + * + * Not used for scanout surfaces. + * + * Applies to both platforms with shared LLC(HAS_LLC), and snooping + * based platforms(HAS_SNOOP). + * + * This is the default on shared LLC platforms. The only exception is + * scanout objects, where the display engine is not coherent with the + * CPU cache. For such objects I915_CACHE_NONE or I915_CACHE_WT is + * automatically applied by the kernel in pin_for_display, if userspace + * has not done so already. + */ + I915_CACHE_LLC, + /** + * @I915_CACHE_L3_LLC: + * + * Explicitly enable the Gfx L3 cache, with coherent LLC. + * + * The Gfx L3 sits between the domain specific caches, e.g + * sampler/render caches, and the larger LLC. LLC is coherent with the + * GPU, but L3 is only visible to the GPU, so likely needs to be flushed + * when the workload completes. + * + * Not used for scanout surfaces. + * + * Only exposed on some gen7 + GGTT. More recent hardware has dropped + * this explicit setting, where it should now be enabled by default. + */ + I915_CACHE_L3_LLC, + /** + * @I915_CACHE_WT: + * + * Write-through. Used for scanout surfaces. + * + * The GPU can utilise the caches, while still having the display engine + * be coherent with GPU writes, as a result we don't need to flush the + * CPU caches when moving out of the render domain. This is the default + * setting chosen by the kernel, if supported by the HW, otherwise we + * fallback to I915_CACHE_NONE. On the CPU side writes through the CPU + * cache still need to be flushed, to remain coherent with the display + * engine. + */ + I915_CACHE_WT, +}; + +#endif /* __I915_CACHE_H__ */ diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index 76ccd4e03e31..e2da57397770 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c @@ -139,48 +139,37 @@ static const char *stringify_vma_type(const struct i915_vma *vma) return "ppgtt"; } -static const char *i915_cache_level_str(struct drm_i915_gem_object *obj) +static void obj_cache_str(struct seq_file *m, struct drm_i915_gem_object *obj) { - struct drm_i915_private *i915 = obj_to_i915(obj); + const i915_cache_t cache = obj->cache_mode; + const unsigned int mode = I915_CACHE_MODE(cache); + const unsigned long flags = I915_CACHE_FLAGS(cache); + static const char *mode_str[] = { + [I915_CACHE_MODE_UC] = "UC", + [I915_CACHE_MODE_WB] = "WB", + [I915_CACHE_MODE_WT] = "WT", + [I915_CACHE_MODE_WC] = "WC", + }; + static const char *flag_str[] = { + [I915_CACHE_L3] = "L3", + [I915_CACHE_COH1W] = "1-Way-Coherent", + [I915_CACHE_COH2W] = "2-Way-Coherent", + [I915_CACHE_CLOS1] = "CLOS1", + [I915_CACHE_CLOS2] = "CLOS2", + }; - if (IS_METEORLAKE(i915)) { - switch (obj->pat_index) { - case 0: return " WB"; - case 1: return " WT"; - case 2: return " UC"; - case 3: return " WB (1-Way Coh)"; - case 4: return " WB (2-Way Coh)"; - default: return " not defined"; - } - } else if (IS_PONTEVECCHIO(i915)) { - switch (obj->pat_index) { - case 0: return " UC"; - case 1: return " WC"; - case 2: return " WT"; - case 3: return " WB"; - case 4: return " WT (CLOS1)"; - case 5: return " WB (CLOS1)"; - case 6: return " WT (CLOS2)"; - case 7: return " WT (CLOS2)"; - default: return " not defined"; - } - } else if (GRAPHICS_VER(i915) >= 12) { - switch (obj->pat_index) { - case 0: return " WB"; - case 1: return " WC"; - case 2: return " WT"; - case 3: return " UC"; - default: return " not defined"; - } + if (mode == I915_CACHE_MODE_UNKNOWN || mode > ARRAY_SIZE(mode_str)) { + if (obj->pat_set_by_user) + seq_printf(m, " PAT-%u", obj->pat_index); + else + seq_printf(m, " PAT-%u-???", obj->pat_index); } else { - switch (obj->pat_index) { - case 0: return " UC"; - case 1: return HAS_LLC(i915) ? - " LLC" : " snooped"; - case 2: return " L3+LLC"; - case 3: return " WT"; - default: return " not defined"; - } + unsigned long bit; + + seq_printf(m, " %s", mode_str[mode]); + + for_each_set_bit(bit, &flags, sizeof(i915_cache_t)) + seq_printf(m, "-%s", flag_str[bit]); } } @@ -190,17 +179,23 @@ i915_debugfs_describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) struct i915_vma *vma; int pin_count = 0; - seq_printf(m, "%pK: %c%c%c %8zdKiB %02x %02x %s%s%s", + seq_printf(m, "%pK: %c%c%c %8zdKiB %02x %02x", &obj->base, get_tiling_flag(obj), get_global_flag(obj), get_pin_mapped_flag(obj), obj->base.size / 1024, obj->read_domains, - obj->write_domain, - i915_cache_level_str(obj), - obj->mm.dirty ? " dirty" : "", - obj->mm.madv == I915_MADV_DONTNEED ? " purgeable" : ""); + obj->write_domain); + + obj_cache_str(m, obj); + + if (obj->mm.dirty) + seq_puts(m, " dirty"); + + if (obj->mm.madv == I915_MADV_DONTNEED) + seq_puts(m, " purgeable"); + if (obj->base.name) seq_printf(m, " (name: %d)", obj->base.name); diff --git a/drivers/gpu/drm/i915/i915_driver.c b/drivers/gpu/drm/i915/i915_driver.c index 222d0a1f3b55..deab26752ba4 100644 --- a/drivers/gpu/drm/i915/i915_driver.c +++ b/drivers/gpu/drm/i915/i915_driver.c @@ -80,6 +80,7 @@ #include "soc/intel_dram.h" #include "soc/intel_gmch.h" +#include "i915_cache.h" #include "i915_debugfs.h" #include "i915_driver.h" #include "i915_drm_client.h" @@ -267,6 +268,8 @@ static int i915_driver_early_probe(struct drm_i915_private *dev_priv) intel_detect_preproduction_hw(dev_priv); + i915_cache_init(dev_priv); + return 0; err_rootgt: diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index b4cf6f0f636d..cb1c0c9d98ef 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h @@ -251,6 +251,9 @@ struct drm_i915_private { unsigned int hpll_freq; unsigned int czclk_freq; + unsigned int pat_uc; + unsigned int pat_wb; + /** * wq - Driver workqueue for GEM. * diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 7ae42f746cc2..9aae75862e6f 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c @@ -422,9 +422,7 @@ i915_gem_gtt_pread(struct drm_i915_gem_object *obj, ggtt->vm.insert_page(&ggtt->vm, i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT), - node.start, - i915_gem_get_pat_index(i915, - I915_CACHE_NONE), 0); + node.start, i915->pat_uc, 0); } else { page_base += offset & PAGE_MASK; } @@ -603,9 +601,7 @@ i915_gem_gtt_pwrite_fast(struct drm_i915_gem_object *obj, ggtt->vm.insert_page(&ggtt->vm, i915_gem_object_get_dma_address(obj, offset >> PAGE_SHIFT), - node.start, - i915_gem_get_pat_index(i915, - I915_CACHE_NONE), 0); + node.start, i915->pat_uc, 0); wmb(); /* flush modifications to the GGTT (insert_page) */ } else { page_base += offset & PAGE_MASK; @@ -1148,19 +1144,6 @@ int i915_gem_init(struct drm_i915_private *dev_priv) unsigned int i; int ret; - /* - * In the proccess of replacing cache_level with pat_index a tricky - * dependency is created on the definition of the enum i915_cache_level. - * in case this enum is changed, PTE encode would be broken. - * Add a WARNING here. And remove when we completely quit using this - * enum - */ - BUILD_BUG_ON(I915_CACHE_NONE != 0 || - I915_CACHE_LLC != 1 || - I915_CACHE_L3_LLC != 2 || - I915_CACHE_WT != 3 || - I915_MAX_CACHE_LEVEL != 4); - /* We need to fallback to 4K pages if host doesn't support huge gtt. */ if (intel_vgpu_active(dev_priv) && !intel_vgpu_has_huge_gtt(dev_priv)) RUNTIME_INFO(dev_priv)->page_sizes = I915_GTT_PAGE_SIZE_4K; diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c index 4749f99e6320..fad336a45699 100644 --- a/drivers/gpu/drm/i915/i915_gpu_error.c +++ b/drivers/gpu/drm/i915/i915_gpu_error.c @@ -1122,14 +1122,11 @@ i915_vma_coredump_create(const struct intel_gt *gt, mutex_lock(&ggtt->error_mutex); if (ggtt->vm.raw_insert_page) ggtt->vm.raw_insert_page(&ggtt->vm, dma, slot, - i915_gem_get_pat_index(gt->i915, - I915_CACHE_NONE), + ggtt->vm.i915->pat_uc, 0); else ggtt->vm.insert_page(&ggtt->vm, dma, slot, - i915_gem_get_pat_index(gt->i915, - I915_CACHE_NONE), - 0); + ggtt->vm.i915->pat_uc, 0); mb(); s = io_mapping_map_wc(&ggtt->iomap, slot, PAGE_SIZE); diff --git a/drivers/gpu/drm/i915/i915_pci.c b/drivers/gpu/drm/i915/i915_pci.c index 3d7a5db9833b..fbdce31afeb1 100644 --- a/drivers/gpu/drm/i915/i915_pci.c +++ b/drivers/gpu/drm/i915/i915_pci.c @@ -32,6 +32,7 @@ #include "gt/intel_sa_media.h" #include "gem/i915_gem_object_types.h" +#include "i915_cache.h" #include "i915_driver.h" #include "i915_drv.h" #include "i915_pci.h" @@ -46,36 +47,42 @@ __diag_ignore_all("-Woverride-init", "Allow overriding inherited members"); .__runtime.graphics.ip.ver = (x), \ .__runtime.media.ip.ver = (x) -#define LEGACY_CACHELEVEL \ - .cachelevel_to_pat = { \ - [I915_CACHE_NONE] = 0, \ - [I915_CACHE_LLC] = 1, \ - [I915_CACHE_L3_LLC] = 2, \ - [I915_CACHE_WT] = 3, \ +/* TODO/QQQ index 1 & 2 */ +#define LEGACY_CACHE_MODES \ + .cache_modes = { \ + [0] = I915_CACHE(UC), \ + [1] = I915_CACHE(WB), \ + [2] = _I915_CACHE(WB, L3), \ + [3] = I915_CACHE(WT), \ } -#define TGL_CACHELEVEL \ - .cachelevel_to_pat = { \ - [I915_CACHE_NONE] = 3, \ - [I915_CACHE_LLC] = 0, \ - [I915_CACHE_L3_LLC] = 0, \ - [I915_CACHE_WT] = 2, \ +#define GEN12_CACHE_MODES \ + .cache_modes = { \ + [0] = I915_CACHE(WB), \ + [1] = I915_CACHE(WC), \ + [2] = I915_CACHE(WT), \ + [3] = I915_CACHE(UC), \ } -#define PVC_CACHELEVEL \ - .cachelevel_to_pat = { \ - [I915_CACHE_NONE] = 0, \ - [I915_CACHE_LLC] = 3, \ - [I915_CACHE_L3_LLC] = 3, \ - [I915_CACHE_WT] = 2, \ +#define PVC_CACHE_MODES \ + .cache_modes = { \ + [0] = I915_CACHE(UC), \ + [1] = I915_CACHE(WC), \ + [2] = I915_CACHE(WT), \ + [3] = I915_CACHE(WB), \ + [4] = _I915_CACHE(WT, CLOS1), \ + [5] = _I915_CACHE(WB, CLOS1), \ + [6] = _I915_CACHE(WT, CLOS2), \ + [7] = _I915_CACHE(WB, CLOS2), \ } -#define MTL_CACHELEVEL \ - .cachelevel_to_pat = { \ - [I915_CACHE_NONE] = 2, \ - [I915_CACHE_LLC] = 3, \ - [I915_CACHE_L3_LLC] = 3, \ - [I915_CACHE_WT] = 1, \ +#define MTL_CACHE_MODES \ + .cache_modes = { \ + [0] = I915_CACHE(WB), \ + [1] = I915_CACHE(WT), \ + [2] = I915_CACHE(UC), \ + [3] = _I915_CACHE(WB, COH1W), \ + [4] = _I915_CACHE(WB, COH2W), \ } /* Keep in gen based order, and chronological order within a gen */ @@ -100,7 +107,7 @@ __diag_ignore_all("-Woverride-init", "Allow overriding inherited members"); .max_pat_index = 3, \ GEN_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES #define I845_FEATURES \ GEN(2), \ @@ -115,7 +122,7 @@ __diag_ignore_all("-Woverride-init", "Allow overriding inherited members"); .max_pat_index = 3, \ GEN_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES static const struct intel_device_info i830_info = { I830_FEATURES, @@ -148,7 +155,7 @@ static const struct intel_device_info i865g_info = { .max_pat_index = 3, \ GEN_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES static const struct intel_device_info i915g_info = { GEN3_FEATURES, @@ -211,7 +218,7 @@ static const struct intel_device_info pnv_m_info = { .max_pat_index = 3, \ GEN_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES static const struct intel_device_info i965g_info = { GEN4_FEATURES, @@ -255,7 +262,7 @@ static const struct intel_device_info gm45_info = { .max_pat_index = 3, \ GEN_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES static const struct intel_device_info ilk_d_info = { GEN5_FEATURES, @@ -285,7 +292,7 @@ static const struct intel_device_info ilk_m_info = { .__runtime.ppgtt_size = 31, \ GEN_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES #define SNB_D_PLATFORM \ GEN6_FEATURES, \ @@ -333,7 +340,7 @@ static const struct intel_device_info snb_m_gt2_info = { .__runtime.ppgtt_size = 31, \ GEN_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES #define IVB_D_PLATFORM \ GEN7_FEATURES, \ @@ -390,7 +397,7 @@ static const struct intel_device_info vlv_info = { .__runtime.platform_engine_mask = BIT(RCS0) | BIT(VCS0) | BIT(BCS0), GEN_DEFAULT_PAGE_SIZES, GEN_DEFAULT_REGIONS, - LEGACY_CACHELEVEL, + LEGACY_CACHE_MODES }; #define G75_FEATURES \ @@ -476,7 +483,7 @@ static const struct intel_device_info chv_info = { .has_coherent_ggtt = false, GEN_DEFAULT_PAGE_SIZES, GEN_DEFAULT_REGIONS, - LEGACY_CACHELEVEL, + LEGACY_CACHE_MODES }; #define GEN9_DEFAULT_PAGE_SIZES \ @@ -539,7 +546,7 @@ static const struct intel_device_info skl_gt4_info = { .max_pat_index = 3, \ GEN9_DEFAULT_PAGE_SIZES, \ GEN_DEFAULT_REGIONS, \ - LEGACY_CACHELEVEL + LEGACY_CACHE_MODES static const struct intel_device_info bxt_info = { GEN9_LP_FEATURES, @@ -643,7 +650,7 @@ static const struct intel_device_info jsl_info = { #define GEN12_FEATURES \ GEN11_FEATURES, \ GEN(12), \ - TGL_CACHELEVEL, \ + GEN12_CACHE_MODES, \ .has_global_mocs = 1, \ .has_pxp = 1, \ .max_pat_index = 3 @@ -711,7 +718,7 @@ static const struct intel_device_info adl_p_info = { .__runtime.graphics.ip.ver = 12, \ .__runtime.graphics.ip.rel = 50, \ XE_HP_PAGE_SIZES, \ - TGL_CACHELEVEL, \ + GEN12_CACHE_MODES, \ .dma_mask_size = 46, \ .has_3d_pipeline = 1, \ .has_64bit_reloc = 1, \ @@ -806,7 +813,7 @@ static const struct intel_device_info pvc_info = { BIT(VCS0) | BIT(CCS0) | BIT(CCS1) | BIT(CCS2) | BIT(CCS3), .require_force_probe = 1, - PVC_CACHELEVEL, + PVC_CACHE_MODES }; static const struct intel_gt_definition xelpmp_extra_gt[] = { @@ -841,7 +848,7 @@ static const struct intel_device_info mtl_info = { .__runtime.memory_regions = REGION_SMEM | REGION_STOLEN_LMEM, .__runtime.platform_engine_mask = BIT(RCS0) | BIT(BCS0) | BIT(CCS0), .require_force_probe = 1, - MTL_CACHELEVEL, + MTL_CACHE_MODES }; #undef PLATFORM diff --git a/drivers/gpu/drm/i915/intel_device_info.h b/drivers/gpu/drm/i915/intel_device_info.h index 069291b3bd37..5cbae7c2ee30 100644 --- a/drivers/gpu/drm/i915/intel_device_info.h +++ b/drivers/gpu/drm/i915/intel_device_info.h @@ -27,6 +27,8 @@ #include <uapi/drm/i915_drm.h> +#include "i915_cache.h" + #include "intel_step.h" #include "display/intel_display_device.h" @@ -248,8 +250,8 @@ struct intel_device_info { */ const struct intel_runtime_info __runtime; - u32 cachelevel_to_pat[I915_MAX_CACHE_LEVEL]; - u32 max_pat_index; + i915_cache_t cache_modes[9]; + unsigned int max_pat_index; }; struct intel_driver_caps { diff --git a/drivers/gpu/drm/i915/selftests/i915_gem.c b/drivers/gpu/drm/i915/selftests/i915_gem.c index 61da4ed9d521..e620f73793a5 100644 --- a/drivers/gpu/drm/i915/selftests/i915_gem.c +++ b/drivers/gpu/drm/i915/selftests/i915_gem.c @@ -57,10 +57,7 @@ static void trash_stolen(struct drm_i915_private *i915) u32 __iomem *s; int x; - ggtt->vm.insert_page(&ggtt->vm, dma, slot, - i915_gem_get_pat_index(i915, - I915_CACHE_NONE), - 0); + ggtt->vm.insert_page(&ggtt->vm, dma, slot, i915->pat_uc, 0); s = io_mapping_map_atomic_wc(&ggtt->iomap, slot); for (x = 0; x < PAGE_SIZE / sizeof(u32); x++) { diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_evict.c b/drivers/gpu/drm/i915/selftests/i915_gem_evict.c index f8fe3681c3dc..658a5b59545e 100644 --- a/drivers/gpu/drm/i915/selftests/i915_gem_evict.c +++ b/drivers/gpu/drm/i915/selftests/i915_gem_evict.c @@ -246,7 +246,7 @@ static int igt_evict_for_cache_color(void *arg) struct drm_mm_node target = { .start = I915_GTT_PAGE_SIZE * 2, .size = I915_GTT_PAGE_SIZE, - .color = i915_gem_get_pat_index(gt->i915, I915_CACHE_LLC), + .color = I915_CACHE(WB), }; struct drm_i915_gem_object *obj; struct i915_vma *vma; @@ -309,7 +309,7 @@ static int igt_evict_for_cache_color(void *arg) /* Attempt to remove the first *pinned* vma, by removing the (empty) * neighbour -- this should fail. */ - target.color = i915_gem_get_pat_index(gt->i915, I915_CACHE_L3_LLC); + target.color = _I915_CACHE(WB, LLC); mutex_lock(&ggtt->vm.mutex); err = i915_gem_evict_for_node(&ggtt->vm, NULL, &target, 0); diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c index 5c397a2df70e..a24585784f75 100644 --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c @@ -135,7 +135,7 @@ fake_dma_object(struct drm_i915_private *i915, u64 size) obj->write_domain = I915_GEM_DOMAIN_CPU; obj->read_domains = I915_GEM_DOMAIN_CPU; - obj->pat_index = i915_gem_get_pat_index(i915, I915_CACHE_NONE); + obj->pat_index = i915->pat_uc; /* Preallocate the "backing storage" */ if (i915_gem_object_pin_pages_unlocked(obj)) @@ -358,10 +358,8 @@ static int lowlevel_hole(struct i915_address_space *vm, mock_vma_res->start = addr; with_intel_runtime_pm(vm->gt->uncore->rpm, wakeref) - vm->insert_entries(vm, mock_vma_res, - i915_gem_get_pat_index(vm->i915, - I915_CACHE_NONE), - 0); + vm->insert_entries(vm, mock_vma_res, + vm->i915->pat_uc, 0); } count = n; @@ -1379,10 +1377,7 @@ static int igt_ggtt_page(void *arg) ggtt->vm.insert_page(&ggtt->vm, i915_gem_object_get_dma_address(obj, 0), - offset, - i915_gem_get_pat_index(i915, - I915_CACHE_NONE), - 0); + offset, ggtt->vm.i915->pat_uc, 0); } order = i915_random_order(count, &prng); diff --git a/drivers/gpu/drm/i915/selftests/intel_memory_region.c b/drivers/gpu/drm/i915/selftests/intel_memory_region.c index d985d9bae2e8..b82fe0ef8cd7 100644 --- a/drivers/gpu/drm/i915/selftests/intel_memory_region.c +++ b/drivers/gpu/drm/i915/selftests/intel_memory_region.c @@ -1070,9 +1070,7 @@ static int igt_lmem_write_cpu(void *arg) /* Put the pages into a known state -- from the gpu for added fun */ intel_engine_pm_get(engine); err = intel_context_migrate_clear(engine->gt->migrate.context, NULL, - obj->mm.pages->sgl, - i915_gem_get_pat_index(i915, - I915_CACHE_NONE), + obj->mm.pages->sgl, i915->pat_uc, true, 0xdeadbeaf, &rq); if (rq) { dma_resv_add_fence(obj->base.resv, &rq->fence, diff --git a/drivers/gpu/drm/i915/selftests/mock_gem_device.c b/drivers/gpu/drm/i915/selftests/mock_gem_device.c index 09d4bbcdcdbf..ad778842cba2 100644 --- a/drivers/gpu/drm/i915/selftests/mock_gem_device.c +++ b/drivers/gpu/drm/i915/selftests/mock_gem_device.c @@ -126,7 +126,12 @@ struct drm_i915_private *mock_gem_device(void) struct drm_i915_private *i915; struct intel_device_info *i915_info; struct pci_dev *pdev; - unsigned int i; + static const i915_cache_t legacy_cache_modes[] = { + [0] = I915_CACHE(UC), + [1] = I915_CACHE(WB), + [2] = _I915_CACHE(WB, L3), + [3] = I915_CACHE(WT), + }; int ret; pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); @@ -187,8 +192,7 @@ struct drm_i915_private *mock_gem_device(void) /* simply use legacy cache level for mock device */ i915_info = (struct intel_device_info *)INTEL_INFO(i915); i915_info->max_pat_index = 3; - for (i = 0; i < I915_MAX_CACHE_LEVEL; i++) - i915_info->cachelevel_to_pat[i] = i; + memcpy(i915_info->cache_modes, legacy_cache_modes, sizeof(legacy_cache_modes)); intel_memory_regions_hw_probe(i915);