Message ID | 20190412160338.64994-10-thellstrom@vmware.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Emulated coherent graphics memory | expand |
Minor nits below Reviewed-by: Deepak Rawat <drawat@vmware.com> On Fri, 2019-04-12 at 16:04 +0000, Thomas Hellstrom wrote: > Add the callbacks necessary to implement emulated coherent memory for > surfaces. Add a flag to the gb_surface_create ioctl to indicate that > surface memory should be coherent. > Also bump the drm minor version to signal the availability of > coherent > surfaces. > > Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com> > --- > .../device_include/svga3d_surfacedefs.h | 209 +++++++++- > drivers/gpu/drm/vmwgfx/vmwgfx_drv.h | 4 +- > drivers/gpu/drm/vmwgfx/vmwgfx_surface.c | 390 > +++++++++++++++++- > include/uapi/drm/vmwgfx_drm.h | 4 +- > 4 files changed, 600 insertions(+), 7 deletions(-) > > diff --git > a/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h > b/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h > index f2bfd3d80598..d901206c04e3 100644 > --- a/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h > +++ b/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h > @@ -1280,7 +1280,6 @@ > svga3dsurface_get_pixel_offset(SVGA3dSurfaceFormat format, > return offset; > } > > - > static inline u32 > svga3dsurface_get_image_offset(SVGA3dSurfaceFormat format, > surf_size_struct baseLevelSize, > @@ -1375,4 +1374,212 @@ > svga3dsurface_is_screen_target_format(SVGA3dSurfaceFormat format) > return svga3dsurface_is_dx_screen_target_format(format); > } > > +/** > + * struct svga3dsurface_mip - Mimpmap level information > + * @bytes: Bytes required in the backing store of this mipmap level. > + * @img_stride: Byte stride per image. > + * @row_stride: Byte stride per block row. > + * @size: The size of the mipmap. > + */ > +struct svga3dsurface_mip { > + size_t bytes; > + size_t img_stride; > + size_t row_stride; > + struct drm_vmw_size size; > + > +}; > + > +/** > + * struct svga3dsurface_cache - Cached surface information > + * @desc: Pointer to the surface descriptor > + * @mip: Array of mipmap level information. Valid size is > @num_mip_levels. > + * @mip_chain_bytes: Bytes required in the backing store for the > whole chain > + * of mip levels. > + * @num_mip_levels: Valid size of the @mip array. Number of mipmap > levels in > + * a chain. > + * @num_layers: Number of slices in an array texture or number of > faces in > + * a cubemap texture. > + */ > +struct svga3dsurface_cache { > + const struct svga3d_surface_desc *desc; > + struct svga3dsurface_mip mip[DRM_VMW_MAX_MIP_LEVELS]; > + size_t mip_chain_bytes; > + u32 num_mip_levels; > + u32 num_layers; > +}; > + > +/** > + * struct svga3dsurface_loc - Surface location > + * @sub_resource: Surface subresource. Defined as layer * > num_mip_levels + > + * mip_level. > + * @x: X coordinate. > + * @y: Y coordinate. > + * @z: Z coordinate. > + */ > +struct svga3dsurface_loc { > + u32 sub_resource; > + u32 x, y, z; > +}; > + > +/** > + * svga3dsurface_subres - Compute the subresource from layer and > mipmap. > + * @cache: Surface layout data. > + * @mip_level: The mipmap level. > + * @layer: The surface layer (face or array slice). > + * > + * Return: The subresource. > + */ > +static inline u32 svga3dsurface_subres(const struct > svga3dsurface_cache *cache, > + u32 mip_level, u32 layer) > +{ > + return cache->num_mip_levels * layer + mip_level; > +} > + > +/** > + * svga3dsurface_setup_cache - Build a surface cache entry > + * @size: The surface base level dimensions. > + * @format: The surface format. > + * @num_mip_levels: Number of mipmap levels. > + * @num_layers: Number of layers. > + * @cache: Pointer to a struct svga3dsurface_cach object to be > filled in. > + */ > +static inline void svga3dsurface_setup_cache(const struct > drm_vmw_size *size, > + SVGA3dSurfaceFormat > format, > + u32 num_mip_levels, > + u32 num_layers, > + u32 num_samples, > + struct svga3dsurface_cache > *cache) > +{ > + const struct svga3d_surface_desc *desc; > + u32 i; > + > + memset(cache, 0, sizeof(*cache)); > + cache->desc = desc = svga3dsurface_get_desc(format); > + cache->num_mip_levels = num_mip_levels; > + cache->num_layers = num_layers; > + for (i = 0; i < cache->num_mip_levels; i++) { > + struct svga3dsurface_mip *mip = &cache->mip[i]; > + > + mip->size = svga3dsurface_get_mip_size(*size, i); > + mip->bytes = svga3dsurface_get_image_buffer_size > + (desc, &mip->size, 0) * num_samples; > + mip->row_stride = > + __KERNEL_DIV_ROUND_UP(mip->size.width, > + desc->block_size.width) * > + desc->bytes_per_block * num_samples; > + mip->img_stride = > + __KERNEL_DIV_ROUND_UP(mip->size.height, > + desc->block_size.height) > * > + mip->row_stride; > + cache->mip_chain_bytes += mip->bytes; > + } > +} > + > +/** > + * svga3dsurface_get_loc - Get a surface location from an offset > into the > + * backing store > + * @cache: Surface layout data. > + * @loc: Pointer to a struct svga3dsurface_loc to be filled in. > + * @offset: Offset into the surface backing store. > + */ > +static inline void > +svga3dsurface_get_loc(const struct svga3dsurface_cache *cache, > + struct svga3dsurface_loc *loc, > + size_t offset) > +{ > + u32 layer = offset / cache->mip_chain_bytes; > + const struct svga3dsurface_mip *mip = &cache->mip[0]; > + const struct svga3d_surface_desc *desc = cache->desc; > + int i; > + > + offset -= layer * cache->mip_chain_bytes; > + for (i = 0; i < cache->num_mip_levels; ++i, ++mip) { > + if (mip->bytes > offset) > + break; > + offset -= mip->bytes; > + } > + > + loc->sub_resource = svga3dsurface_subres(cache, i, layer); > + loc->z = offset / mip->img_stride; > + offset -= loc->z * mip->img_stride; > + loc->z *= desc->block_size.depth; > + loc->y = offset / mip->row_stride; > + offset -= loc->y * mip->row_stride; > + loc->y *= desc->block_size.height; > + loc->x = offset / desc->bytes_per_block; > + loc->x *= desc->block_size.width; > +} > + > +/** > + * svga3dsurface_inc_loc - Clamp increment a surface location with > one block > + * size > + * in each dimension. > + * @loc: Pointer to a struct svga3dsurface_loc to be incremented. > + * > + * When computing the size of a range as size = end - start, the > range does not > + * include the end element. However a location representing the last > byte > + * of a touched region in the backing store *is* included in the > range. > + * This function motifies such a location to match the end /s/motifies/modifies > definition > + * given as start + size which is the one used in a SVGA3dBox. > + */ > +static inline void > +svga3dsurface_inc_loc(const struct svga3dsurface_cache *cache, > + struct svga3dsurface_loc *loc) > +{ > + const struct svga3d_surface_desc *desc = cache->desc; > + u32 mip = loc->sub_resource % cache->num_mip_levels; > + const struct drm_vmw_size *size = &cache->mip[mip].size; > + > + loc->sub_resource++; > + loc->x += desc->block_size.width; > + if (loc->x > size->width) > + loc->x = size->width; > + loc->y += desc->block_size.height; > + if (loc->y > size->height) > + loc->y = size->height; > + loc->z += desc->block_size.depth; > + if (loc->z > size->depth) > + loc->z = size->depth; > +} > + > +/** > + * svga3dsurface_min_loc - The start location in a subresorce /s/subresorce/subresource > + * @cache: Surface layout data. > + * @sub_resource: The subresource. > + * @loc: Pointer to a struct svga3dsurface_loc to be filled in. > + */ > +static inline void > +svga3dsurface_min_loc(const struct svga3dsurface_cache *cache, > + u32 sub_resource, > + struct svga3dsurface_loc *loc) > +{ > + loc->sub_resource = sub_resource; > + loc->x = loc->y = loc->z = 0; > +} > + > +/** > + * svga3dsurface_min_loc - The end location in a subresorce svga3dsurface_max_loc. Also /s/subresorce/subresource > + * @cache: Surface layout data. > + * @sub_resource: The subresource. > + * @loc: Pointer to a struct svga3dsurface_loc to be filled in. > + * > + * Following the end definition given in svga3dsurface_inc_loc(), > + * Compute the end location of a surface subresource. > + */ > +static inline void > +svga3dsurface_max_loc(const struct svga3dsurface_cache *cache, > + u32 sub_resource, > + struct svga3dsurface_loc *loc) > +{ > + const struct drm_vmw_size *size; > + u32 mip; > + > + loc->sub_resource = sub_resource + 1; > + mip = sub_resource % cache->num_mip_levels; > + size = &cache->mip[mip].size; > + loc->x = size->width; > + loc->y = size->height; > + loc->z = size->depth; > +} > + > #endif /* _SVGA3D_SURFACEDEFS_H_ */ > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h > b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h > index 00794415335e..630a01d75a41 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h > @@ -44,9 +44,9 @@ > #include <linux/sync_file.h> > > #define VMWGFX_DRIVER_NAME "vmwgfx" > -#define VMWGFX_DRIVER_DATE "20180704" > +#define VMWGFX_DRIVER_DATE "20190328" > #define VMWGFX_DRIVER_MAJOR 2 > -#define VMWGFX_DRIVER_MINOR 15 > +#define VMWGFX_DRIVER_MINOR 16 > #define VMWGFX_DRIVER_PATCHLEVEL 0 > #define VMWGFX_FILE_PAGE_OFFSET 0x00100000 > #define VMWGFX_FIFO_STATIC_SIZE (1024*1024) > diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c > b/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c > index c40d44f4d9af..f56141529da5 100644 > --- a/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c > +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c > @@ -68,6 +68,20 @@ struct vmw_surface_offset { > uint32_t bo_offset; > }; > > +/** > + * vmw_surface_dirty - Surface dirty-tracker > + * @cache: Cached layout information of the surface. > + * @size: Accounting size for the struct vmw_surface_dirty. > + * @num_subres: Number of subresources. > + * @boxes: Array of SVGA3dBoxes indicating dirty regions. One per > subresource. > + */ > +struct vmw_surface_dirty { > + struct svga3dsurface_cache cache; > + size_t size; > + u32 num_subres; > + SVGA3dBox boxes[0]; > +}; > + > static void vmw_user_surface_free(struct vmw_resource *res); > static struct vmw_resource * > vmw_user_surface_base_to_res(struct ttm_base_object *base); > @@ -96,6 +110,13 @@ vmw_gb_surface_reference_internal(struct > drm_device *dev, > struct drm_vmw_gb_surface_ref_ext_rep > *rep, > struct drm_file *file_priv); > > +static void vmw_surface_dirty_free(struct vmw_resource *res); > +static int vmw_surface_dirty_alloc(struct vmw_resource *res); > +static int vmw_surface_dirty_sync(struct vmw_resource *res); > +static void vmw_surface_dirty_range_add(struct vmw_resource *res, > size_t start, > + size_t end); > +static int vmw_surface_clean(struct vmw_resource *res); > + > static const struct vmw_user_resource_conv user_surface_conv = { > .object_type = VMW_RES_SURFACE, > .base_obj_to_res = vmw_user_surface_base_to_res, > @@ -133,7 +154,12 @@ static const struct vmw_res_func > vmw_gb_surface_func = { > .create = vmw_gb_surface_create, > .destroy = vmw_gb_surface_destroy, > .bind = vmw_gb_surface_bind, > - .unbind = vmw_gb_surface_unbind > + .unbind = vmw_gb_surface_unbind, > + .dirty_alloc = vmw_surface_dirty_alloc, > + .dirty_free = vmw_surface_dirty_free, > + .dirty_sync = vmw_surface_dirty_sync, > + .dirty_range_add = vmw_surface_dirty_range_add, > + .clean = vmw_surface_clean, > }; > > /** > @@ -641,6 +667,7 @@ static void vmw_user_surface_free(struct > vmw_resource *res) > struct vmw_private *dev_priv = srf->res.dev_priv; > uint32_t size = user_srf->size; > > + WARN_ON_ONCE(res->dirty); > if (user_srf->master) > drm_master_put(&user_srf->master); > kfree(srf->offsets); > @@ -1174,10 +1201,16 @@ static int vmw_gb_surface_bind(struct > vmw_resource *res, > cmd2->header.id = SVGA_3D_CMD_UPDATE_GB_SURFACE; > cmd2->header.size = sizeof(cmd2->body); > cmd2->body.sid = res->id; > - res->backup_dirty = false; > } > vmw_fifo_commit(dev_priv, submit_size); > > + if (res->backup->dirty && res->backup_dirty) { > + /* We've just made a full upload. Cear dirty regions. > */ > + vmw_bo_dirty_clear_res(res); > + } > + > + res->backup_dirty = false; > + > return 0; > } > > @@ -1642,7 +1675,8 @@ vmw_gb_surface_define_internal(struct > drm_device *dev, > } > } > } else if (req->base.drm_surface_flags & > - drm_vmw_surface_flag_create_buffer) > + (drm_vmw_surface_flag_create_buffer | > + drm_vmw_surface_flag_coherent)) > ret = vmw_user_bo_alloc(dev_priv, tfile, > res->backup_size, > req->base.drm_surface_flags & > @@ -1656,6 +1690,26 @@ vmw_gb_surface_define_internal(struct > drm_device *dev, > goto out_unlock; > } > > + if (req->base.drm_surface_flags & > drm_vmw_surface_flag_coherent) { > + struct vmw_buffer_object *backup = res->backup; > + > + ttm_bo_reserve(&backup->base, false, false, NULL); > + if (!res->func->dirty_alloc) > + ret = -EINVAL; > + if (!ret) > + ret = vmw_bo_dirty_add(backup); > + if (!ret) { > + res->coherent = true; > + ret = res->func->dirty_alloc(res); > + } > + ttm_bo_unreserve(&backup->base); > + if (ret) { > + vmw_resource_unreference(&res); > + goto out_unlock; > + } > + > + } > + > tmp = vmw_resource_reference(res); > ret = ttm_prime_object_init(tfile, res->backup_size, &user_srf- > >prime, > req->base.drm_surface_flags & > @@ -1764,3 +1818,333 @@ vmw_gb_surface_reference_internal(struct > drm_device *dev, > > return ret; > } > + > +/** > + * vmw_subres_dirty_add - Add a dirty region to a subresource > + * @dirty: The surfaces's dirty tracker. > + * @loc_start: The location corresponding to the start of the > region. > + * @loc_end: The location corresponding to the end of the region. > + * > + * As we are assuming that @loc_start and @loc_end represent a > sequential > + * range of backing store memory, if the region spans multiple lines > then > + * regardless of the x coordinate, the full lines are dirtied. > + * Correspondingly if the region spans multiple z slices, then full > rather > + * than partial z slices are dirtied. > + */ > +static void vmw_subres_dirty_add(struct vmw_surface_dirty *dirty, > + const struct svga3dsurface_loc > *loc_start, > + const struct svga3dsurface_loc > *loc_end) > +{ > + const struct svga3dsurface_cache *cache = &dirty->cache; > + SVGA3dBox *box = &dirty->boxes[loc_start->sub_resource]; > + u32 mip = loc_start->sub_resource % cache->num_mip_levels; > + const struct drm_vmw_size *size = &cache->mip[mip].size; > + u32 box_c2 = box->z + box->d; > + > + if (WARN_ON(loc_start->sub_resource >= dirty->num_subres)) > + return; > + > + if (box->d == 0 || box->z > loc_start->z) > + box->z = loc_start->z; > + if (box_c2 < loc_end->z) > + box->d = loc_end->z - box->z; > + > + if (loc_start->z + 1 == loc_end->z) { > + box_c2 = box->y + box->h; > + if (box->h == 0 || box->y > loc_start->y) > + box->y = loc_start->y; > + if (box_c2 < loc_end->y) > + box->h = loc_end->y - box->y; > + > + if (loc_start->y + 1 == loc_end->y) { > + box_c2 = box->x + box->w; > + if (box->w == 0 || box->x > loc_start->x) > + box->x = loc_start->x; > + if (box_c2 < loc_end->x) > + box->w = loc_end->x - box->x; > + } else { > + box->x = 0; > + box->w = size->width; > + } > + } else { > + box->y = 0; > + box->h = size->height; > + box->x = 0; > + box->w = size->width; > + } > +} > + > +/** > + * vmw_subres_dirty_full - Mark a full subresource as dirty > + * @dirty: The surface's dirty tracker. > + * @subres: The subresource > + */ > +static void vmw_subres_dirty_full(struct vmw_surface_dirty *dirty, > u32 subres) > +{ > + const struct svga3dsurface_cache *cache = &dirty->cache; > + u32 mip = subres % cache->num_mip_levels; > + const struct drm_vmw_size *size = &cache->mip[mip].size; > + SVGA3dBox *box = &dirty->boxes[subres]; > + > + box->x = 0; > + box->y = 0; > + box->z = 0; > + box->w = size->width; > + box->h = size->height; > + box->d = size->depth; > +} > + > +/* > + * vmw_surface_tex_dirty_add_range - The dirty_add_range callback > for texture > + * surfaces. > + */ > +static void vmw_surface_tex_dirty_range_add(struct vmw_resource > *res, > + size_t start, size_t end) > +{ > + struct vmw_surface_dirty *dirty = > + (struct vmw_surface_dirty *) res->dirty; > + size_t backup_end = res->backup_offset + res->backup_size; > + struct svga3dsurface_loc loc1, loc2; > + const struct svga3dsurface_cache *cache; > + > + start = max(start, res->backup_offset) - res->backup_offset; > + end = min(end, backup_end) - res->backup_offset; > + cache = &dirty->cache; > + svga3dsurface_get_loc(cache, &loc1, start); > + svga3dsurface_get_loc(cache, &loc2, end - 1); > + svga3dsurface_inc_loc(cache, &loc2); > + > + if (loc1.sub_resource + 1 == loc2.sub_resource) { > + /* Dirty range covers a single sub-resource */ > + vmw_subres_dirty_add(dirty, &loc1, &loc2); > + } else { > + /* Dirty range covers multiple sub-resources */ > + struct svga3dsurface_loc loc_min, loc_max; > + u32 sub_res = loc1.sub_resource; > + > + svga3dsurface_max_loc(cache, loc1.sub_resource, > &loc_max); > + vmw_subres_dirty_add(dirty, &loc1, &loc_max); > + svga3dsurface_min_loc(cache, loc2.sub_resource - 1, > &loc_min); > + vmw_subres_dirty_add(dirty, &loc_min, &loc2); > + for (sub_res = loc1.sub_resource + 1; > + sub_res < loc2.sub_resource - 1; ++sub_res) > + vmw_subres_dirty_full(dirty, sub_res); > + } > +} > + > +/* > + * vmw_surface_tex_dirty_add_range - The dirty_add_range callback > for buffer > + * surfaces. > + */ > +static void vmw_surface_buf_dirty_range_add(struct vmw_resource > *res, > + size_t start, size_t end) > +{ > + struct vmw_surface_dirty *dirty = > + (struct vmw_surface_dirty *) res->dirty; > + const struct svga3dsurface_cache *cache = &dirty->cache; > + size_t backup_end = res->backup_offset + cache- > >mip_chain_bytes; > + SVGA3dBox *box = &dirty->boxes[0]; > + u32 box_c2; > + > + box->h = box->d = 1; > + start = max(start, res->backup_offset) - res->backup_offset; > + end = min(end, backup_end) - res->backup_offset; > + box_c2 = box->x + box->w; > + if (box->w == 0 || box->x > start) > + box->x = start; > + if (box_c2 < end) > + box->w = end - box->x; > +} > + > +/* > + * vmw_surface_tex_dirty_add_range - The dirty_add_range callback > for surfaces > + */ > +static void vmw_surface_dirty_range_add(struct vmw_resource *res, > size_t start, > + size_t end) > +{ > + struct vmw_surface *srf = vmw_res_to_srf(res); > + > + if (WARN_ON(end <= res->backup_offset || > + start >= res->backup_offset + res->backup_size)) > + return; > + > + if (srf->format == SVGA3D_BUFFER) > + vmw_surface_buf_dirty_range_add(res, start, end); > + else > + vmw_surface_tex_dirty_range_add(res, start, end); > +} > + > +/* > + * vmw_surface_dirty_sync - The surface's dirty_sync callback. > + */ > +static int vmw_surface_dirty_sync(struct vmw_resource *res) > +{ > + struct vmw_private *dev_priv = res->dev_priv; > + bool has_dx = 0; > + u32 i, num_dirty; > + struct vmw_surface_dirty *dirty = > + (struct vmw_surface_dirty *) res->dirty; > + size_t alloc_size; > + const struct svga3dsurface_cache *cache = &dirty->cache; > + struct { > + SVGA3dCmdHeader header; > + SVGA3dCmdDXUpdateSubResource body; > + } *cmd1; > + struct { > + SVGA3dCmdHeader header; > + SVGA3dCmdUpdateGBImage body; > + } *cmd2; > + void *cmd; > + > + num_dirty = 0; > + for (i = 0; i < dirty->num_subres; ++i) { > + const SVGA3dBox *box = &dirty->boxes[i]; > + > + if (box->d) > + num_dirty++; > + } > + > + if (!num_dirty) > + goto out; > + > + alloc_size = num_dirty * ((has_dx) ? sizeof(*cmd1) : > sizeof(*cmd2)); > + cmd = VMW_FIFO_RESERVE(dev_priv, alloc_size); > + if (!cmd) > + return -ENOMEM; > + > + cmd1 = cmd; > + cmd2 = cmd; > + > + for (i = 0; i < dirty->num_subres; ++i) { > + const SVGA3dBox *box = &dirty->boxes[i]; > + > + if (!box->d) > + continue; > + > + /* > + * DX_UPDATE_SUBRESOURCE is aware of array surfaces. > + * UPDATE_GB_IMAGE is not. > + */ > + if (has_dx) { > + cmd1->header.id = > SVGA_3D_CMD_DX_UPDATE_SUBRESOURCE; > + cmd1->header.size = sizeof(cmd1->body); > + cmd1->body.sid = res->id; > + cmd1->body.subResource = i; > + cmd1->body.box = *box; > + cmd1++; > + } else { > + cmd2->header.id = SVGA_3D_CMD_UPDATE_GB_IMAGE; > + cmd2->header.size = sizeof(cmd2->body); > + cmd2->body.image.sid = res->id; > + cmd2->body.image.face = i / cache- > >num_mip_levels; > + cmd2->body.image.mipmap = i - > + (cache->num_mip_levels * cmd2- > >body.image.face); > + cmd2->body.box = *box; > + cmd2++; > + } > + > + } > + vmw_fifo_commit(dev_priv, alloc_size); > + out: > + memset(&dirty->boxes[0], 0, sizeof(dirty->boxes[0]) * > + dirty->num_subres); > + > + return 0; > +} > + > +/* > + * vmw_surface_dirty_alloc - The surface's dirty_alloc callback. > + */ > +static int vmw_surface_dirty_alloc(struct vmw_resource *res) > +{ > + struct vmw_surface *srf = vmw_res_to_srf(res); > + struct vmw_surface_dirty *dirty; > + u32 num_layers = 1; > + u32 num_mip; > + u32 num_subres; > + u32 num_samples; > + size_t dirty_size, acc_size; > + static struct ttm_operation_ctx ctx = { > + .interruptible = false, > + .no_wait_gpu = false > + }; > + int ret; > + > + if (srf->array_size) > + num_layers = srf->array_size; > + else if (srf->flags & SVGA3D_SURFACE_CUBEMAP) > + num_layers *= SVGA3D_MAX_SURFACE_FACES; > + > + num_mip = srf->mip_levels[0]; > + if (!num_mip) > + num_mip = 1; > + > + num_subres = num_layers * num_mip; > + dirty_size = sizeof(*dirty) + num_subres * sizeof(dirty- > >boxes[0]); > + acc_size = ttm_round_pot(dirty_size); > + ret = ttm_mem_global_alloc(vmw_mem_glob(res->dev_priv), > + acc_size, &ctx); > + if (ret) { > + VMW_DEBUG_USER("Out of graphics memory for surface " > + "dirty tracker.\n"); > + return ret; > + } > + > + dirty = kvzalloc(dirty_size, GFP_KERNEL); > + if (!dirty) { > + ret = -ENOMEM; > + goto out_no_dirty; > + } > + > + num_samples = max_t(u32, 1, srf->multisample_count); > + svga3dsurface_setup_cache(&srf->base_size, srf->format, > num_mip, > + num_layers, num_samples, &dirty- > >cache); > + dirty->num_subres = num_subres; > + dirty->size = acc_size; > + res->dirty = (struct vmw_resource_dirty *) dirty; > + > + return 0; > + > +out_no_dirty: > + ttm_mem_global_free(vmw_mem_glob(res->dev_priv), acc_size); > + return ret; > +} > + > +/* > + * vmw_surface_dirty_free - The surface's dirty_free callback > + */ > +static void vmw_surface_dirty_free(struct vmw_resource *res) > +{ > + struct vmw_surface_dirty *dirty = > + (struct vmw_surface_dirty *) res->dirty; > + size_t acc_size = dirty->size; > + > + kvfree(dirty); > + ttm_mem_global_free(vmw_mem_glob(res->dev_priv), acc_size); > + res->dirty = NULL; > +} > + > +/* > + * vmw_surface_clean - The surface's clean callback > + */ > +static int vmw_surface_clean(struct vmw_resource *res) > +{ > + struct vmw_private *dev_priv = res->dev_priv; > + size_t alloc_size; > + struct { > + SVGA3dCmdHeader header; > + SVGA3dCmdReadbackGBSurface body; > + } *cmd; > + > + alloc_size = sizeof(*cmd); > + cmd = VMW_FIFO_RESERVE(dev_priv, alloc_size); > + if (!cmd) > + return -ENOMEM; > + > + cmd->header.id = SVGA_3D_CMD_READBACK_GB_SURFACE; > + cmd->header.size = sizeof(cmd->body); > + cmd->body.sid = res->id; > + vmw_fifo_commit(dev_priv, alloc_size); > + > + return 0; > +} > diff --git a/include/uapi/drm/vmwgfx_drm.h > b/include/uapi/drm/vmwgfx_drm.h > index 399f58317cff..02cab33f2f25 100644 > --- a/include/uapi/drm/vmwgfx_drm.h > +++ b/include/uapi/drm/vmwgfx_drm.h > @@ -891,11 +891,13 @@ struct drm_vmw_shader_arg { > * surface. > * @drm_vmw_surface_flag_create_buffer: Create a backup buffer if > none is > * given. > + * @drm_vmw_surface_flag_coherent: Back surface with coherent > memory. > */ > enum drm_vmw_surface_flags { > drm_vmw_surface_flag_shareable = (1 << 0), > drm_vmw_surface_flag_scanout = (1 << 1), > - drm_vmw_surface_flag_create_buffer = (1 << 2) > + drm_vmw_surface_flag_create_buffer = (1 << 2), > + drm_vmw_surface_flag_coherent = (1 << 3), > }; > > /** > -- > 2.20.1 >
diff --git a/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h b/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h index f2bfd3d80598..d901206c04e3 100644 --- a/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h +++ b/drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h @@ -1280,7 +1280,6 @@ svga3dsurface_get_pixel_offset(SVGA3dSurfaceFormat format, return offset; } - static inline u32 svga3dsurface_get_image_offset(SVGA3dSurfaceFormat format, surf_size_struct baseLevelSize, @@ -1375,4 +1374,212 @@ svga3dsurface_is_screen_target_format(SVGA3dSurfaceFormat format) return svga3dsurface_is_dx_screen_target_format(format); } +/** + * struct svga3dsurface_mip - Mimpmap level information + * @bytes: Bytes required in the backing store of this mipmap level. + * @img_stride: Byte stride per image. + * @row_stride: Byte stride per block row. + * @size: The size of the mipmap. + */ +struct svga3dsurface_mip { + size_t bytes; + size_t img_stride; + size_t row_stride; + struct drm_vmw_size size; + +}; + +/** + * struct svga3dsurface_cache - Cached surface information + * @desc: Pointer to the surface descriptor + * @mip: Array of mipmap level information. Valid size is @num_mip_levels. + * @mip_chain_bytes: Bytes required in the backing store for the whole chain + * of mip levels. + * @num_mip_levels: Valid size of the @mip array. Number of mipmap levels in + * a chain. + * @num_layers: Number of slices in an array texture or number of faces in + * a cubemap texture. + */ +struct svga3dsurface_cache { + const struct svga3d_surface_desc *desc; + struct svga3dsurface_mip mip[DRM_VMW_MAX_MIP_LEVELS]; + size_t mip_chain_bytes; + u32 num_mip_levels; + u32 num_layers; +}; + +/** + * struct svga3dsurface_loc - Surface location + * @sub_resource: Surface subresource. Defined as layer * num_mip_levels + + * mip_level. + * @x: X coordinate. + * @y: Y coordinate. + * @z: Z coordinate. + */ +struct svga3dsurface_loc { + u32 sub_resource; + u32 x, y, z; +}; + +/** + * svga3dsurface_subres - Compute the subresource from layer and mipmap. + * @cache: Surface layout data. + * @mip_level: The mipmap level. + * @layer: The surface layer (face or array slice). + * + * Return: The subresource. + */ +static inline u32 svga3dsurface_subres(const struct svga3dsurface_cache *cache, + u32 mip_level, u32 layer) +{ + return cache->num_mip_levels * layer + mip_level; +} + +/** + * svga3dsurface_setup_cache - Build a surface cache entry + * @size: The surface base level dimensions. + * @format: The surface format. + * @num_mip_levels: Number of mipmap levels. + * @num_layers: Number of layers. + * @cache: Pointer to a struct svga3dsurface_cach object to be filled in. + */ +static inline void svga3dsurface_setup_cache(const struct drm_vmw_size *size, + SVGA3dSurfaceFormat format, + u32 num_mip_levels, + u32 num_layers, + u32 num_samples, + struct svga3dsurface_cache *cache) +{ + const struct svga3d_surface_desc *desc; + u32 i; + + memset(cache, 0, sizeof(*cache)); + cache->desc = desc = svga3dsurface_get_desc(format); + cache->num_mip_levels = num_mip_levels; + cache->num_layers = num_layers; + for (i = 0; i < cache->num_mip_levels; i++) { + struct svga3dsurface_mip *mip = &cache->mip[i]; + + mip->size = svga3dsurface_get_mip_size(*size, i); + mip->bytes = svga3dsurface_get_image_buffer_size + (desc, &mip->size, 0) * num_samples; + mip->row_stride = + __KERNEL_DIV_ROUND_UP(mip->size.width, + desc->block_size.width) * + desc->bytes_per_block * num_samples; + mip->img_stride = + __KERNEL_DIV_ROUND_UP(mip->size.height, + desc->block_size.height) * + mip->row_stride; + cache->mip_chain_bytes += mip->bytes; + } +} + +/** + * svga3dsurface_get_loc - Get a surface location from an offset into the + * backing store + * @cache: Surface layout data. + * @loc: Pointer to a struct svga3dsurface_loc to be filled in. + * @offset: Offset into the surface backing store. + */ +static inline void +svga3dsurface_get_loc(const struct svga3dsurface_cache *cache, + struct svga3dsurface_loc *loc, + size_t offset) +{ + u32 layer = offset / cache->mip_chain_bytes; + const struct svga3dsurface_mip *mip = &cache->mip[0]; + const struct svga3d_surface_desc *desc = cache->desc; + int i; + + offset -= layer * cache->mip_chain_bytes; + for (i = 0; i < cache->num_mip_levels; ++i, ++mip) { + if (mip->bytes > offset) + break; + offset -= mip->bytes; + } + + loc->sub_resource = svga3dsurface_subres(cache, i, layer); + loc->z = offset / mip->img_stride; + offset -= loc->z * mip->img_stride; + loc->z *= desc->block_size.depth; + loc->y = offset / mip->row_stride; + offset -= loc->y * mip->row_stride; + loc->y *= desc->block_size.height; + loc->x = offset / desc->bytes_per_block; + loc->x *= desc->block_size.width; +} + +/** + * svga3dsurface_inc_loc - Clamp increment a surface location with one block + * size + * in each dimension. + * @loc: Pointer to a struct svga3dsurface_loc to be incremented. + * + * When computing the size of a range as size = end - start, the range does not + * include the end element. However a location representing the last byte + * of a touched region in the backing store *is* included in the range. + * This function motifies such a location to match the end definition + * given as start + size which is the one used in a SVGA3dBox. + */ +static inline void +svga3dsurface_inc_loc(const struct svga3dsurface_cache *cache, + struct svga3dsurface_loc *loc) +{ + const struct svga3d_surface_desc *desc = cache->desc; + u32 mip = loc->sub_resource % cache->num_mip_levels; + const struct drm_vmw_size *size = &cache->mip[mip].size; + + loc->sub_resource++; + loc->x += desc->block_size.width; + if (loc->x > size->width) + loc->x = size->width; + loc->y += desc->block_size.height; + if (loc->y > size->height) + loc->y = size->height; + loc->z += desc->block_size.depth; + if (loc->z > size->depth) + loc->z = size->depth; +} + +/** + * svga3dsurface_min_loc - The start location in a subresorce + * @cache: Surface layout data. + * @sub_resource: The subresource. + * @loc: Pointer to a struct svga3dsurface_loc to be filled in. + */ +static inline void +svga3dsurface_min_loc(const struct svga3dsurface_cache *cache, + u32 sub_resource, + struct svga3dsurface_loc *loc) +{ + loc->sub_resource = sub_resource; + loc->x = loc->y = loc->z = 0; +} + +/** + * svga3dsurface_min_loc - The end location in a subresorce + * @cache: Surface layout data. + * @sub_resource: The subresource. + * @loc: Pointer to a struct svga3dsurface_loc to be filled in. + * + * Following the end definition given in svga3dsurface_inc_loc(), + * Compute the end location of a surface subresource. + */ +static inline void +svga3dsurface_max_loc(const struct svga3dsurface_cache *cache, + u32 sub_resource, + struct svga3dsurface_loc *loc) +{ + const struct drm_vmw_size *size; + u32 mip; + + loc->sub_resource = sub_resource + 1; + mip = sub_resource % cache->num_mip_levels; + size = &cache->mip[mip].size; + loc->x = size->width; + loc->y = size->height; + loc->z = size->depth; +} + #endif /* _SVGA3D_SURFACEDEFS_H_ */ diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h index 00794415335e..630a01d75a41 100644 --- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h @@ -44,9 +44,9 @@ #include <linux/sync_file.h> #define VMWGFX_DRIVER_NAME "vmwgfx" -#define VMWGFX_DRIVER_DATE "20180704" +#define VMWGFX_DRIVER_DATE "20190328" #define VMWGFX_DRIVER_MAJOR 2 -#define VMWGFX_DRIVER_MINOR 15 +#define VMWGFX_DRIVER_MINOR 16 #define VMWGFX_DRIVER_PATCHLEVEL 0 #define VMWGFX_FILE_PAGE_OFFSET 0x00100000 #define VMWGFX_FIFO_STATIC_SIZE (1024*1024) diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c b/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c index c40d44f4d9af..f56141529da5 100644 --- a/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_surface.c @@ -68,6 +68,20 @@ struct vmw_surface_offset { uint32_t bo_offset; }; +/** + * vmw_surface_dirty - Surface dirty-tracker + * @cache: Cached layout information of the surface. + * @size: Accounting size for the struct vmw_surface_dirty. + * @num_subres: Number of subresources. + * @boxes: Array of SVGA3dBoxes indicating dirty regions. One per subresource. + */ +struct vmw_surface_dirty { + struct svga3dsurface_cache cache; + size_t size; + u32 num_subres; + SVGA3dBox boxes[0]; +}; + static void vmw_user_surface_free(struct vmw_resource *res); static struct vmw_resource * vmw_user_surface_base_to_res(struct ttm_base_object *base); @@ -96,6 +110,13 @@ vmw_gb_surface_reference_internal(struct drm_device *dev, struct drm_vmw_gb_surface_ref_ext_rep *rep, struct drm_file *file_priv); +static void vmw_surface_dirty_free(struct vmw_resource *res); +static int vmw_surface_dirty_alloc(struct vmw_resource *res); +static int vmw_surface_dirty_sync(struct vmw_resource *res); +static void vmw_surface_dirty_range_add(struct vmw_resource *res, size_t start, + size_t end); +static int vmw_surface_clean(struct vmw_resource *res); + static const struct vmw_user_resource_conv user_surface_conv = { .object_type = VMW_RES_SURFACE, .base_obj_to_res = vmw_user_surface_base_to_res, @@ -133,7 +154,12 @@ static const struct vmw_res_func vmw_gb_surface_func = { .create = vmw_gb_surface_create, .destroy = vmw_gb_surface_destroy, .bind = vmw_gb_surface_bind, - .unbind = vmw_gb_surface_unbind + .unbind = vmw_gb_surface_unbind, + .dirty_alloc = vmw_surface_dirty_alloc, + .dirty_free = vmw_surface_dirty_free, + .dirty_sync = vmw_surface_dirty_sync, + .dirty_range_add = vmw_surface_dirty_range_add, + .clean = vmw_surface_clean, }; /** @@ -641,6 +667,7 @@ static void vmw_user_surface_free(struct vmw_resource *res) struct vmw_private *dev_priv = srf->res.dev_priv; uint32_t size = user_srf->size; + WARN_ON_ONCE(res->dirty); if (user_srf->master) drm_master_put(&user_srf->master); kfree(srf->offsets); @@ -1174,10 +1201,16 @@ static int vmw_gb_surface_bind(struct vmw_resource *res, cmd2->header.id = SVGA_3D_CMD_UPDATE_GB_SURFACE; cmd2->header.size = sizeof(cmd2->body); cmd2->body.sid = res->id; - res->backup_dirty = false; } vmw_fifo_commit(dev_priv, submit_size); + if (res->backup->dirty && res->backup_dirty) { + /* We've just made a full upload. Cear dirty regions. */ + vmw_bo_dirty_clear_res(res); + } + + res->backup_dirty = false; + return 0; } @@ -1642,7 +1675,8 @@ vmw_gb_surface_define_internal(struct drm_device *dev, } } } else if (req->base.drm_surface_flags & - drm_vmw_surface_flag_create_buffer) + (drm_vmw_surface_flag_create_buffer | + drm_vmw_surface_flag_coherent)) ret = vmw_user_bo_alloc(dev_priv, tfile, res->backup_size, req->base.drm_surface_flags & @@ -1656,6 +1690,26 @@ vmw_gb_surface_define_internal(struct drm_device *dev, goto out_unlock; } + if (req->base.drm_surface_flags & drm_vmw_surface_flag_coherent) { + struct vmw_buffer_object *backup = res->backup; + + ttm_bo_reserve(&backup->base, false, false, NULL); + if (!res->func->dirty_alloc) + ret = -EINVAL; + if (!ret) + ret = vmw_bo_dirty_add(backup); + if (!ret) { + res->coherent = true; + ret = res->func->dirty_alloc(res); + } + ttm_bo_unreserve(&backup->base); + if (ret) { + vmw_resource_unreference(&res); + goto out_unlock; + } + + } + tmp = vmw_resource_reference(res); ret = ttm_prime_object_init(tfile, res->backup_size, &user_srf->prime, req->base.drm_surface_flags & @@ -1764,3 +1818,333 @@ vmw_gb_surface_reference_internal(struct drm_device *dev, return ret; } + +/** + * vmw_subres_dirty_add - Add a dirty region to a subresource + * @dirty: The surfaces's dirty tracker. + * @loc_start: The location corresponding to the start of the region. + * @loc_end: The location corresponding to the end of the region. + * + * As we are assuming that @loc_start and @loc_end represent a sequential + * range of backing store memory, if the region spans multiple lines then + * regardless of the x coordinate, the full lines are dirtied. + * Correspondingly if the region spans multiple z slices, then full rather + * than partial z slices are dirtied. + */ +static void vmw_subres_dirty_add(struct vmw_surface_dirty *dirty, + const struct svga3dsurface_loc *loc_start, + const struct svga3dsurface_loc *loc_end) +{ + const struct svga3dsurface_cache *cache = &dirty->cache; + SVGA3dBox *box = &dirty->boxes[loc_start->sub_resource]; + u32 mip = loc_start->sub_resource % cache->num_mip_levels; + const struct drm_vmw_size *size = &cache->mip[mip].size; + u32 box_c2 = box->z + box->d; + + if (WARN_ON(loc_start->sub_resource >= dirty->num_subres)) + return; + + if (box->d == 0 || box->z > loc_start->z) + box->z = loc_start->z; + if (box_c2 < loc_end->z) + box->d = loc_end->z - box->z; + + if (loc_start->z + 1 == loc_end->z) { + box_c2 = box->y + box->h; + if (box->h == 0 || box->y > loc_start->y) + box->y = loc_start->y; + if (box_c2 < loc_end->y) + box->h = loc_end->y - box->y; + + if (loc_start->y + 1 == loc_end->y) { + box_c2 = box->x + box->w; + if (box->w == 0 || box->x > loc_start->x) + box->x = loc_start->x; + if (box_c2 < loc_end->x) + box->w = loc_end->x - box->x; + } else { + box->x = 0; + box->w = size->width; + } + } else { + box->y = 0; + box->h = size->height; + box->x = 0; + box->w = size->width; + } +} + +/** + * vmw_subres_dirty_full - Mark a full subresource as dirty + * @dirty: The surface's dirty tracker. + * @subres: The subresource + */ +static void vmw_subres_dirty_full(struct vmw_surface_dirty *dirty, u32 subres) +{ + const struct svga3dsurface_cache *cache = &dirty->cache; + u32 mip = subres % cache->num_mip_levels; + const struct drm_vmw_size *size = &cache->mip[mip].size; + SVGA3dBox *box = &dirty->boxes[subres]; + + box->x = 0; + box->y = 0; + box->z = 0; + box->w = size->width; + box->h = size->height; + box->d = size->depth; +} + +/* + * vmw_surface_tex_dirty_add_range - The dirty_add_range callback for texture + * surfaces. + */ +static void vmw_surface_tex_dirty_range_add(struct vmw_resource *res, + size_t start, size_t end) +{ + struct vmw_surface_dirty *dirty = + (struct vmw_surface_dirty *) res->dirty; + size_t backup_end = res->backup_offset + res->backup_size; + struct svga3dsurface_loc loc1, loc2; + const struct svga3dsurface_cache *cache; + + start = max(start, res->backup_offset) - res->backup_offset; + end = min(end, backup_end) - res->backup_offset; + cache = &dirty->cache; + svga3dsurface_get_loc(cache, &loc1, start); + svga3dsurface_get_loc(cache, &loc2, end - 1); + svga3dsurface_inc_loc(cache, &loc2); + + if (loc1.sub_resource + 1 == loc2.sub_resource) { + /* Dirty range covers a single sub-resource */ + vmw_subres_dirty_add(dirty, &loc1, &loc2); + } else { + /* Dirty range covers multiple sub-resources */ + struct svga3dsurface_loc loc_min, loc_max; + u32 sub_res = loc1.sub_resource; + + svga3dsurface_max_loc(cache, loc1.sub_resource, &loc_max); + vmw_subres_dirty_add(dirty, &loc1, &loc_max); + svga3dsurface_min_loc(cache, loc2.sub_resource - 1, &loc_min); + vmw_subres_dirty_add(dirty, &loc_min, &loc2); + for (sub_res = loc1.sub_resource + 1; + sub_res < loc2.sub_resource - 1; ++sub_res) + vmw_subres_dirty_full(dirty, sub_res); + } +} + +/* + * vmw_surface_tex_dirty_add_range - The dirty_add_range callback for buffer + * surfaces. + */ +static void vmw_surface_buf_dirty_range_add(struct vmw_resource *res, + size_t start, size_t end) +{ + struct vmw_surface_dirty *dirty = + (struct vmw_surface_dirty *) res->dirty; + const struct svga3dsurface_cache *cache = &dirty->cache; + size_t backup_end = res->backup_offset + cache->mip_chain_bytes; + SVGA3dBox *box = &dirty->boxes[0]; + u32 box_c2; + + box->h = box->d = 1; + start = max(start, res->backup_offset) - res->backup_offset; + end = min(end, backup_end) - res->backup_offset; + box_c2 = box->x + box->w; + if (box->w == 0 || box->x > start) + box->x = start; + if (box_c2 < end) + box->w = end - box->x; +} + +/* + * vmw_surface_tex_dirty_add_range - The dirty_add_range callback for surfaces + */ +static void vmw_surface_dirty_range_add(struct vmw_resource *res, size_t start, + size_t end) +{ + struct vmw_surface *srf = vmw_res_to_srf(res); + + if (WARN_ON(end <= res->backup_offset || + start >= res->backup_offset + res->backup_size)) + return; + + if (srf->format == SVGA3D_BUFFER) + vmw_surface_buf_dirty_range_add(res, start, end); + else + vmw_surface_tex_dirty_range_add(res, start, end); +} + +/* + * vmw_surface_dirty_sync - The surface's dirty_sync callback. + */ +static int vmw_surface_dirty_sync(struct vmw_resource *res) +{ + struct vmw_private *dev_priv = res->dev_priv; + bool has_dx = 0; + u32 i, num_dirty; + struct vmw_surface_dirty *dirty = + (struct vmw_surface_dirty *) res->dirty; + size_t alloc_size; + const struct svga3dsurface_cache *cache = &dirty->cache; + struct { + SVGA3dCmdHeader header; + SVGA3dCmdDXUpdateSubResource body; + } *cmd1; + struct { + SVGA3dCmdHeader header; + SVGA3dCmdUpdateGBImage body; + } *cmd2; + void *cmd; + + num_dirty = 0; + for (i = 0; i < dirty->num_subres; ++i) { + const SVGA3dBox *box = &dirty->boxes[i]; + + if (box->d) + num_dirty++; + } + + if (!num_dirty) + goto out; + + alloc_size = num_dirty * ((has_dx) ? sizeof(*cmd1) : sizeof(*cmd2)); + cmd = VMW_FIFO_RESERVE(dev_priv, alloc_size); + if (!cmd) + return -ENOMEM; + + cmd1 = cmd; + cmd2 = cmd; + + for (i = 0; i < dirty->num_subres; ++i) { + const SVGA3dBox *box = &dirty->boxes[i]; + + if (!box->d) + continue; + + /* + * DX_UPDATE_SUBRESOURCE is aware of array surfaces. + * UPDATE_GB_IMAGE is not. + */ + if (has_dx) { + cmd1->header.id = SVGA_3D_CMD_DX_UPDATE_SUBRESOURCE; + cmd1->header.size = sizeof(cmd1->body); + cmd1->body.sid = res->id; + cmd1->body.subResource = i; + cmd1->body.box = *box; + cmd1++; + } else { + cmd2->header.id = SVGA_3D_CMD_UPDATE_GB_IMAGE; + cmd2->header.size = sizeof(cmd2->body); + cmd2->body.image.sid = res->id; + cmd2->body.image.face = i / cache->num_mip_levels; + cmd2->body.image.mipmap = i - + (cache->num_mip_levels * cmd2->body.image.face); + cmd2->body.box = *box; + cmd2++; + } + + } + vmw_fifo_commit(dev_priv, alloc_size); + out: + memset(&dirty->boxes[0], 0, sizeof(dirty->boxes[0]) * + dirty->num_subres); + + return 0; +} + +/* + * vmw_surface_dirty_alloc - The surface's dirty_alloc callback. + */ +static int vmw_surface_dirty_alloc(struct vmw_resource *res) +{ + struct vmw_surface *srf = vmw_res_to_srf(res); + struct vmw_surface_dirty *dirty; + u32 num_layers = 1; + u32 num_mip; + u32 num_subres; + u32 num_samples; + size_t dirty_size, acc_size; + static struct ttm_operation_ctx ctx = { + .interruptible = false, + .no_wait_gpu = false + }; + int ret; + + if (srf->array_size) + num_layers = srf->array_size; + else if (srf->flags & SVGA3D_SURFACE_CUBEMAP) + num_layers *= SVGA3D_MAX_SURFACE_FACES; + + num_mip = srf->mip_levels[0]; + if (!num_mip) + num_mip = 1; + + num_subres = num_layers * num_mip; + dirty_size = sizeof(*dirty) + num_subres * sizeof(dirty->boxes[0]); + acc_size = ttm_round_pot(dirty_size); + ret = ttm_mem_global_alloc(vmw_mem_glob(res->dev_priv), + acc_size, &ctx); + if (ret) { + VMW_DEBUG_USER("Out of graphics memory for surface " + "dirty tracker.\n"); + return ret; + } + + dirty = kvzalloc(dirty_size, GFP_KERNEL); + if (!dirty) { + ret = -ENOMEM; + goto out_no_dirty; + } + + num_samples = max_t(u32, 1, srf->multisample_count); + svga3dsurface_setup_cache(&srf->base_size, srf->format, num_mip, + num_layers, num_samples, &dirty->cache); + dirty->num_subres = num_subres; + dirty->size = acc_size; + res->dirty = (struct vmw_resource_dirty *) dirty; + + return 0; + +out_no_dirty: + ttm_mem_global_free(vmw_mem_glob(res->dev_priv), acc_size); + return ret; +} + +/* + * vmw_surface_dirty_free - The surface's dirty_free callback + */ +static void vmw_surface_dirty_free(struct vmw_resource *res) +{ + struct vmw_surface_dirty *dirty = + (struct vmw_surface_dirty *) res->dirty; + size_t acc_size = dirty->size; + + kvfree(dirty); + ttm_mem_global_free(vmw_mem_glob(res->dev_priv), acc_size); + res->dirty = NULL; +} + +/* + * vmw_surface_clean - The surface's clean callback + */ +static int vmw_surface_clean(struct vmw_resource *res) +{ + struct vmw_private *dev_priv = res->dev_priv; + size_t alloc_size; + struct { + SVGA3dCmdHeader header; + SVGA3dCmdReadbackGBSurface body; + } *cmd; + + alloc_size = sizeof(*cmd); + cmd = VMW_FIFO_RESERVE(dev_priv, alloc_size); + if (!cmd) + return -ENOMEM; + + cmd->header.id = SVGA_3D_CMD_READBACK_GB_SURFACE; + cmd->header.size = sizeof(cmd->body); + cmd->body.sid = res->id; + vmw_fifo_commit(dev_priv, alloc_size); + + return 0; +} diff --git a/include/uapi/drm/vmwgfx_drm.h b/include/uapi/drm/vmwgfx_drm.h index 399f58317cff..02cab33f2f25 100644 --- a/include/uapi/drm/vmwgfx_drm.h +++ b/include/uapi/drm/vmwgfx_drm.h @@ -891,11 +891,13 @@ struct drm_vmw_shader_arg { * surface. * @drm_vmw_surface_flag_create_buffer: Create a backup buffer if none is * given. + * @drm_vmw_surface_flag_coherent: Back surface with coherent memory. */ enum drm_vmw_surface_flags { drm_vmw_surface_flag_shareable = (1 << 0), drm_vmw_surface_flag_scanout = (1 << 1), - drm_vmw_surface_flag_create_buffer = (1 << 2) + drm_vmw_surface_flag_create_buffer = (1 << 2), + drm_vmw_surface_flag_coherent = (1 << 3), }; /**
Add the callbacks necessary to implement emulated coherent memory for surfaces. Add a flag to the gb_surface_create ioctl to indicate that surface memory should be coherent. Also bump the drm minor version to signal the availability of coherent surfaces. Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com> --- .../device_include/svga3d_surfacedefs.h | 209 +++++++++- drivers/gpu/drm/vmwgfx/vmwgfx_drv.h | 4 +- drivers/gpu/drm/vmwgfx/vmwgfx_surface.c | 390 +++++++++++++++++- include/uapi/drm/vmwgfx_drm.h | 4 +- 4 files changed, 600 insertions(+), 7 deletions(-)