@@ -24,16 +24,14 @@
#include "etnaviv_mmu.h"
/* called with dev->struct_mutex held */
-static struct page **get_pages(struct drm_gem_object *obj)
+static struct page **get_pages(struct etnaviv_gem_object *etnaviv_obj)
{
- struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
-
if (!etnaviv_obj->pages) {
- struct drm_device *dev = obj->dev;
+ struct drm_device *dev = etnaviv_obj->base.dev;
struct page **p;
- int npages = obj->size >> PAGE_SHIFT;
+ int npages = etnaviv_obj->base.size >> PAGE_SHIFT;
- p = drm_gem_get_pages(obj);
+ p = drm_gem_get_pages(&etnaviv_obj->base);
if (IS_ERR(p)) {
dev_err(dev->dev, "could not get pages: %ld\n",
@@ -44,7 +42,7 @@ static struct page **get_pages(struct drm_gem_object *obj)
etnaviv_obj->sgt = drm_prime_pages_to_sg(p, npages);
if (IS_ERR(etnaviv_obj->sgt)) {
dev_err(dev->dev, "failed to allocate sgt\n");
- drm_gem_put_pages(obj, p, false, false);
+ drm_gem_put_pages(&etnaviv_obj->base, p, false, false);
return ERR_CAST(etnaviv_obj->sgt);
}
@@ -61,22 +59,23 @@ static struct page **get_pages(struct drm_gem_object *obj)
return etnaviv_obj->pages;
}
-static void put_pages(struct drm_gem_object *obj)
+static void put_pages(struct etnaviv_gem_object *etnaviv_obj)
{
- struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
-
if (etnaviv_obj->pages) {
+ struct drm_device *dev = etnaviv_obj->base.dev;
+
/* For non-cached buffers, ensure the new pages are clean
* because display controller, GPU, etc. are not coherent:
*/
if (etnaviv_obj->flags & (ETNA_BO_WC|ETNA_BO_UNCACHED))
- dma_unmap_sg(obj->dev->dev, etnaviv_obj->sgt->sgl,
+ dma_unmap_sg(dev->dev, etnaviv_obj->sgt->sgl,
etnaviv_obj->sgt->nents,
DMA_BIDIRECTIONAL);
sg_free_table(etnaviv_obj->sgt);
kfree(etnaviv_obj->sgt);
- drm_gem_put_pages(obj, etnaviv_obj->pages, true, false);
+ drm_gem_put_pages(&etnaviv_obj->base, etnaviv_obj->pages,
+ true, false);
etnaviv_obj->pages = NULL;
}
@@ -88,7 +87,7 @@ struct page **etnaviv_gem_get_pages(struct drm_gem_object *obj)
struct page **p;
mutex_lock(&dev->struct_mutex);
- p = get_pages(obj);
+ p = get_pages(to_etnaviv_bo(obj));
mutex_unlock(&dev->struct_mutex);
return p;
@@ -189,7 +188,7 @@ int etnaviv_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
goto out;
/* make sure we have pages attached now */
- pages = get_pages(obj);
+ pages = get_pages(to_etnaviv_bo(obj));
if (IS_ERR(pages)) {
ret = PTR_ERR(pages);
goto out_unlock;
@@ -273,7 +272,7 @@ int etnaviv_gem_get_iova_locked(struct etnaviv_gpu *gpu,
if (!etnaviv_obj->iova && !(etnaviv_obj->flags & ETNA_BO_CMDSTREAM)) {
struct etnaviv_drm_private *priv = obj->dev->dev_private;
struct etnaviv_iommu *mmu = priv->mmu;
- struct page **pages = get_pages(obj);
+ struct page **pages = get_pages(etnaviv_obj);
uint32_t offset;
struct drm_mm_node *node = NULL;
@@ -375,7 +374,7 @@ void *etnaviv_gem_vaddr_locked(struct drm_gem_object *obj)
WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
if (!etnaviv_obj->vaddr) {
- struct page **pages = get_pages(obj);
+ struct page **pages = get_pages(etnaviv_obj);
if (IS_ERR(pages))
return ERR_CAST(pages);
@@ -536,7 +535,7 @@ static void etnaviv_gem_shmem_release(struct etnaviv_gem_object *etnaviv_obj)
{
if (etnaviv_obj->vaddr)
vunmap(etnaviv_obj->vaddr);
- put_pages(&etnaviv_obj->base);
+ put_pages(etnaviv_obj);
}
static const struct etnaviv_gem_ops etnaviv_gem_shmem_ops = {