diff mbox

gpu: drm: ttm: Adding new return type vm_fault_t

Message ID 20180523185538.GA975@jordon-HP-15-Notebook-PC (mailing list archive)
State New, archived
Headers show

Commit Message

Souptick Joarder May 23, 2018, 6:55 p.m. UTC
Use new return type vm_fault_t for fault handler. For
now, this is just documenting that the function returns
a VM_FAULT value rather than an errno. Once all instances
are converted, vm_fault_t will become a distinct type.

Ref-> commit 1c8f422059ae ("mm: change return type to vm_fault_t")

Previously vm_insert_{mixed,pfn} returns err which driver
mapped into VM_FAULT_* type. The new function
vmf_insert_{mixed,pfn} will replace this inefficiency by
returning VM_FAULT_* type.

Signed-off-by: Souptick Joarder <jrdr.linux@gmail.com>
---
 drivers/gpu/drm/ttm/ttm_bo_vm.c | 45 ++++++++++++++++++++---------------------
 1 file changed, 22 insertions(+), 23 deletions(-)

Comments

Souptick Joarder May 31, 2018, 5:07 a.m. UTC | #1
On Thu, May 24, 2018 at 12:25 AM, Souptick Joarder <jrdr.linux@gmail.com> wrote:
> Use new return type vm_fault_t for fault handler. For
> now, this is just documenting that the function returns
> a VM_FAULT value rather than an errno. Once all instances
> are converted, vm_fault_t will become a distinct type.
>
> Ref-> commit 1c8f422059ae ("mm: change return type to vm_fault_t")
>
> Previously vm_insert_{mixed,pfn} returns err which driver
> mapped into VM_FAULT_* type. The new function
> vmf_insert_{mixed,pfn} will replace this inefficiency by
> returning VM_FAULT_* type.
>
> Signed-off-by: Souptick Joarder <jrdr.linux@gmail.com>
> ---
>  drivers/gpu/drm/ttm/ttm_bo_vm.c | 45 ++++++++++++++++++++---------------------
>  1 file changed, 22 insertions(+), 23 deletions(-)
>
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> index 8eba95b..2d13f03 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> @@ -43,10 +43,11 @@
>
>  #define TTM_BO_VM_NUM_PREFAULT 16
>
> -static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
> +static vm_fault_t ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>                                 struct vm_fault *vmf)
>  {
> -       int ret = 0;
> +       int err = 0;
> +       vm_fault_t ret = 0;
>
>         if (likely(!bo->moving))
>                 goto out_unlock;
> @@ -77,9 +78,9 @@ static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>         /*
>          * Ordinary wait.
>          */
> -       ret = dma_fence_wait(bo->moving, true);
> -       if (unlikely(ret != 0)) {
> -               ret = (ret != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
> +       err = dma_fence_wait(bo->moving, true);
> +       if (unlikely(err != 0)) {
> +               ret = (err != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
>                         VM_FAULT_NOPAGE;
>                 goto out_unlock;
>         }
> @@ -104,7 +105,7 @@ static unsigned long ttm_bo_io_mem_pfn(struct ttm_buffer_object *bo,
>                 + page_offset;
>  }
>
> -static int ttm_bo_vm_fault(struct vm_fault *vmf)
> +static vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf)
>  {
>         struct vm_area_struct *vma = vmf->vma;
>         struct ttm_buffer_object *bo = (struct ttm_buffer_object *)
> @@ -115,8 +116,9 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>         unsigned long pfn;
>         struct ttm_tt *ttm = NULL;
>         struct page *page;
> -       int ret;
> +       int err;
>         int i;
> +       vm_fault_t ret = VM_FAULT_NOPAGE;
>         unsigned long address = vmf->address;
>         struct ttm_mem_type_manager *man =
>                 &bdev->man[bo->mem.mem_type];
> @@ -128,9 +130,9 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>          * for reserve, and if it fails, retry the fault after waiting
>          * for the buffer to become unreserved.
>          */
> -       ret = ttm_bo_reserve(bo, true, true, NULL);
> -       if (unlikely(ret != 0)) {
> -               if (ret != -EBUSY)
> +       err = ttm_bo_reserve(bo, true, true, NULL);
> +       if (unlikely(err != 0)) {
> +               if (err != -EBUSY)
>                         return VM_FAULT_NOPAGE;
>
>                 if (vmf->flags & FAULT_FLAG_ALLOW_RETRY) {
> @@ -162,8 +164,8 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>         }
>
>         if (bdev->driver->fault_reserve_notify) {
> -               ret = bdev->driver->fault_reserve_notify(bo);
> -               switch (ret) {
> +               err = bdev->driver->fault_reserve_notify(bo);
> +               switch (err) {
>                 case 0:
>                         break;
>                 case -EBUSY:
> @@ -191,13 +193,13 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>                 goto out_unlock;
>         }
>
> -       ret = ttm_mem_io_lock(man, true);
> -       if (unlikely(ret != 0)) {
> +       err = ttm_mem_io_lock(man, true);
> +       if (unlikely(err != 0)) {
>                 ret = VM_FAULT_NOPAGE;
>                 goto out_unlock;
>         }
> -       ret = ttm_mem_io_reserve_vm(bo);
> -       if (unlikely(ret != 0)) {
> +       err = ttm_mem_io_reserve_vm(bo);
> +       if (unlikely(err != 0)) {
>                 ret = VM_FAULT_SIGBUS;
>                 goto out_io_unlock;
>         }
> @@ -265,23 +267,20 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>                 }
>
>                 if (vma->vm_flags & VM_MIXEDMAP)
> -                       ret = vm_insert_mixed(&cvma, address,
> +                       ret = vmf_insert_mixed(&cvma, address,
>                                         __pfn_to_pfn_t(pfn, PFN_DEV));
>                 else
> -                       ret = vm_insert_pfn(&cvma, address, pfn);
> +                       ret = vmf_insert_pfn(&cvma, address, pfn);
>
>                 /*
>                  * Somebody beat us to this PTE or prefaulting to
>                  * an already populated PTE, or prefaulting error.
>                  */
>
> -               if (unlikely((ret == -EBUSY) || (ret != 0 && i > 0)))
> +               if (unlikely((ret == VM_FAULT_NOPAGE && i > 0)))
>                         break;
> -               else if (unlikely(ret != 0)) {
> -                       ret =
> -                           (ret == -ENOMEM) ? VM_FAULT_OOM : VM_FAULT_SIGBUS;
> +               else if (unlikely(ret & VM_FAULT_ERROR))
>                         goto out_io_unlock;
> -               }
>
>                 address += PAGE_SIZE;
>                 if (unlikely(++page_offset >= page_last))
> --
> 1.9.1
>

Any comment for this patch ?
Christian König June 1, 2018, 11:48 a.m. UTC | #2
Am 31.05.2018 um 07:07 schrieb Souptick Joarder:
> On Thu, May 24, 2018 at 12:25 AM, Souptick Joarder <jrdr.linux@gmail.com> wrote:
>> Use new return type vm_fault_t for fault handler. For
>> now, this is just documenting that the function returns
>> a VM_FAULT value rather than an errno. Once all instances
>> are converted, vm_fault_t will become a distinct type.
>>
>> Ref-> commit 1c8f422059ae ("mm: change return type to vm_fault_t")
>>
>> Previously vm_insert_{mixed,pfn} returns err which driver
>> mapped into VM_FAULT_* type. The new function
>> vmf_insert_{mixed,pfn} will replace this inefficiency by
>> returning VM_FAULT_* type.
>>
>> Signed-off-by: Souptick Joarder <jrdr.linux@gmail.com>
>> ---
>>   drivers/gpu/drm/ttm/ttm_bo_vm.c | 45 ++++++++++++++++++++---------------------
>>   1 file changed, 22 insertions(+), 23 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
>> index 8eba95b..2d13f03 100644
>> --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
>> +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
>> @@ -43,10 +43,11 @@
>>
>>   #define TTM_BO_VM_NUM_PREFAULT 16
>>
>> -static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>> +static vm_fault_t ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>>                                  struct vm_fault *vmf)
>>   {
>> -       int ret = 0;
>> +       int err = 0;
>> +       vm_fault_t ret = 0;

Please keep reverse xmas tree order for variable declarations.

Except for that it looks good to me,
Christian.

>>
>>          if (likely(!bo->moving))
>>                  goto out_unlock;
>> @@ -77,9 +78,9 @@ static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>>          /*
>>           * Ordinary wait.
>>           */
>> -       ret = dma_fence_wait(bo->moving, true);
>> -       if (unlikely(ret != 0)) {
>> -               ret = (ret != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
>> +       err = dma_fence_wait(bo->moving, true);
>> +       if (unlikely(err != 0)) {
>> +               ret = (err != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
>>                          VM_FAULT_NOPAGE;
>>                  goto out_unlock;
>>          }
>> @@ -104,7 +105,7 @@ static unsigned long ttm_bo_io_mem_pfn(struct ttm_buffer_object *bo,
>>                  + page_offset;
>>   }
>>
>> -static int ttm_bo_vm_fault(struct vm_fault *vmf)
>> +static vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf)
>>   {
>>          struct vm_area_struct *vma = vmf->vma;
>>          struct ttm_buffer_object *bo = (struct ttm_buffer_object *)
>> @@ -115,8 +116,9 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>>          unsigned long pfn;
>>          struct ttm_tt *ttm = NULL;
>>          struct page *page;
>> -       int ret;
>> +       int err;
>>          int i;
>> +       vm_fault_t ret = VM_FAULT_NOPAGE;
>>          unsigned long address = vmf->address;
>>          struct ttm_mem_type_manager *man =
>>                  &bdev->man[bo->mem.mem_type];
>> @@ -128,9 +130,9 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>>           * for reserve, and if it fails, retry the fault after waiting
>>           * for the buffer to become unreserved.
>>           */
>> -       ret = ttm_bo_reserve(bo, true, true, NULL);
>> -       if (unlikely(ret != 0)) {
>> -               if (ret != -EBUSY)
>> +       err = ttm_bo_reserve(bo, true, true, NULL);
>> +       if (unlikely(err != 0)) {
>> +               if (err != -EBUSY)
>>                          return VM_FAULT_NOPAGE;
>>
>>                  if (vmf->flags & FAULT_FLAG_ALLOW_RETRY) {
>> @@ -162,8 +164,8 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>>          }
>>
>>          if (bdev->driver->fault_reserve_notify) {
>> -               ret = bdev->driver->fault_reserve_notify(bo);
>> -               switch (ret) {
>> +               err = bdev->driver->fault_reserve_notify(bo);
>> +               switch (err) {
>>                  case 0:
>>                          break;
>>                  case -EBUSY:
>> @@ -191,13 +193,13 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>>                  goto out_unlock;
>>          }
>>
>> -       ret = ttm_mem_io_lock(man, true);
>> -       if (unlikely(ret != 0)) {
>> +       err = ttm_mem_io_lock(man, true);
>> +       if (unlikely(err != 0)) {
>>                  ret = VM_FAULT_NOPAGE;
>>                  goto out_unlock;
>>          }
>> -       ret = ttm_mem_io_reserve_vm(bo);
>> -       if (unlikely(ret != 0)) {
>> +       err = ttm_mem_io_reserve_vm(bo);
>> +       if (unlikely(err != 0)) {
>>                  ret = VM_FAULT_SIGBUS;
>>                  goto out_io_unlock;
>>          }
>> @@ -265,23 +267,20 @@ static int ttm_bo_vm_fault(struct vm_fault *vmf)
>>                  }
>>
>>                  if (vma->vm_flags & VM_MIXEDMAP)
>> -                       ret = vm_insert_mixed(&cvma, address,
>> +                       ret = vmf_insert_mixed(&cvma, address,
>>                                          __pfn_to_pfn_t(pfn, PFN_DEV));
>>                  else
>> -                       ret = vm_insert_pfn(&cvma, address, pfn);
>> +                       ret = vmf_insert_pfn(&cvma, address, pfn);
>>
>>                  /*
>>                   * Somebody beat us to this PTE or prefaulting to
>>                   * an already populated PTE, or prefaulting error.
>>                   */
>>
>> -               if (unlikely((ret == -EBUSY) || (ret != 0 && i > 0)))
>> +               if (unlikely((ret == VM_FAULT_NOPAGE && i > 0)))
>>                          break;
>> -               else if (unlikely(ret != 0)) {
>> -                       ret =
>> -                           (ret == -ENOMEM) ? VM_FAULT_OOM : VM_FAULT_SIGBUS;
>> +               else if (unlikely(ret & VM_FAULT_ERROR))
>>                          goto out_io_unlock;
>> -               }
>>
>>                  address += PAGE_SIZE;
>>                  if (unlikely(++page_offset >= page_last))
>> --
>> 1.9.1
>>
> Any comment for this patch ?
Souptick Joarder June 1, 2018, 7:01 p.m. UTC | #3
On Fri, Jun 1, 2018 at 5:18 PM, Christian König
<christian.koenig@amd.com> wrote:
> Am 31.05.2018 um 07:07 schrieb Souptick Joarder:
>>
>> On Thu, May 24, 2018 at 12:25 AM, Souptick Joarder <jrdr.linux@gmail.com>
>> wrote:
>>>
>>> Use new return type vm_fault_t for fault handler. For
>>> now, this is just documenting that the function returns
>>> a VM_FAULT value rather than an errno. Once all instances
>>> are converted, vm_fault_t will become a distinct type.
>>>
>>> Ref-> commit 1c8f422059ae ("mm: change return type to vm_fault_t")
>>>
>>> Previously vm_insert_{mixed,pfn} returns err which driver
>>> mapped into VM_FAULT_* type. The new function
>>> vmf_insert_{mixed,pfn} will replace this inefficiency by
>>> returning VM_FAULT_* type.
>>>
>>> Signed-off-by: Souptick Joarder <jrdr.linux@gmail.com>
>>> ---
>>>   drivers/gpu/drm/ttm/ttm_bo_vm.c | 45
>>> ++++++++++++++++++++---------------------
>>>   1 file changed, 22 insertions(+), 23 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c
>>> b/drivers/gpu/drm/ttm/ttm_bo_vm.c
>>> index 8eba95b..2d13f03 100644
>>> --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
>>> +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
>>> @@ -43,10 +43,11 @@
>>>
>>>   #define TTM_BO_VM_NUM_PREFAULT 16
>>>
>>> -static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>>> +static vm_fault_t ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
>>>                                  struct vm_fault *vmf)
>>>   {
>>> -       int ret = 0;
>>> +       int err = 0;
>>> +       vm_fault_t ret = 0;
>
>
> Please keep reverse xmas tree order for variable declarations.
>
> Except for that it looks good to me,
> Christian.
>

Sure, I will send v2. We would like to get this patch in
queue for 4.18.
diff mbox

Patch

diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
index 8eba95b..2d13f03 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
@@ -43,10 +43,11 @@ 
 
 #define TTM_BO_VM_NUM_PREFAULT 16
 
-static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
+static vm_fault_t ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
 				struct vm_fault *vmf)
 {
-	int ret = 0;
+	int err = 0;
+	vm_fault_t ret = 0;
 
 	if (likely(!bo->moving))
 		goto out_unlock;
@@ -77,9 +78,9 @@  static int ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo,
 	/*
 	 * Ordinary wait.
 	 */
-	ret = dma_fence_wait(bo->moving, true);
-	if (unlikely(ret != 0)) {
-		ret = (ret != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
+	err = dma_fence_wait(bo->moving, true);
+	if (unlikely(err != 0)) {
+		ret = (err != -ERESTARTSYS) ? VM_FAULT_SIGBUS :
 			VM_FAULT_NOPAGE;
 		goto out_unlock;
 	}
@@ -104,7 +105,7 @@  static unsigned long ttm_bo_io_mem_pfn(struct ttm_buffer_object *bo,
 		+ page_offset;
 }
 
-static int ttm_bo_vm_fault(struct vm_fault *vmf)
+static vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf)
 {
 	struct vm_area_struct *vma = vmf->vma;
 	struct ttm_buffer_object *bo = (struct ttm_buffer_object *)
@@ -115,8 +116,9 @@  static int ttm_bo_vm_fault(struct vm_fault *vmf)
 	unsigned long pfn;
 	struct ttm_tt *ttm = NULL;
 	struct page *page;
-	int ret;
+	int err;
 	int i;
+	vm_fault_t ret = VM_FAULT_NOPAGE;
 	unsigned long address = vmf->address;
 	struct ttm_mem_type_manager *man =
 		&bdev->man[bo->mem.mem_type];
@@ -128,9 +130,9 @@  static int ttm_bo_vm_fault(struct vm_fault *vmf)
 	 * for reserve, and if it fails, retry the fault after waiting
 	 * for the buffer to become unreserved.
 	 */
-	ret = ttm_bo_reserve(bo, true, true, NULL);
-	if (unlikely(ret != 0)) {
-		if (ret != -EBUSY)
+	err = ttm_bo_reserve(bo, true, true, NULL);
+	if (unlikely(err != 0)) {
+		if (err != -EBUSY)
 			return VM_FAULT_NOPAGE;
 
 		if (vmf->flags & FAULT_FLAG_ALLOW_RETRY) {
@@ -162,8 +164,8 @@  static int ttm_bo_vm_fault(struct vm_fault *vmf)
 	}
 
 	if (bdev->driver->fault_reserve_notify) {
-		ret = bdev->driver->fault_reserve_notify(bo);
-		switch (ret) {
+		err = bdev->driver->fault_reserve_notify(bo);
+		switch (err) {
 		case 0:
 			break;
 		case -EBUSY:
@@ -191,13 +193,13 @@  static int ttm_bo_vm_fault(struct vm_fault *vmf)
 		goto out_unlock;
 	}
 
-	ret = ttm_mem_io_lock(man, true);
-	if (unlikely(ret != 0)) {
+	err = ttm_mem_io_lock(man, true);
+	if (unlikely(err != 0)) {
 		ret = VM_FAULT_NOPAGE;
 		goto out_unlock;
 	}
-	ret = ttm_mem_io_reserve_vm(bo);
-	if (unlikely(ret != 0)) {
+	err = ttm_mem_io_reserve_vm(bo);
+	if (unlikely(err != 0)) {
 		ret = VM_FAULT_SIGBUS;
 		goto out_io_unlock;
 	}
@@ -265,23 +267,20 @@  static int ttm_bo_vm_fault(struct vm_fault *vmf)
 		}
 
 		if (vma->vm_flags & VM_MIXEDMAP)
-			ret = vm_insert_mixed(&cvma, address,
+			ret = vmf_insert_mixed(&cvma, address,
 					__pfn_to_pfn_t(pfn, PFN_DEV));
 		else
-			ret = vm_insert_pfn(&cvma, address, pfn);
+			ret = vmf_insert_pfn(&cvma, address, pfn);
 
 		/*
 		 * Somebody beat us to this PTE or prefaulting to
 		 * an already populated PTE, or prefaulting error.
 		 */
 
-		if (unlikely((ret == -EBUSY) || (ret != 0 && i > 0)))
+		if (unlikely((ret == VM_FAULT_NOPAGE && i > 0)))
 			break;
-		else if (unlikely(ret != 0)) {
-			ret =
-			    (ret == -ENOMEM) ? VM_FAULT_OOM : VM_FAULT_SIGBUS;
+		else if (unlikely(ret & VM_FAULT_ERROR))
 			goto out_io_unlock;
-		}
 
 		address += PAGE_SIZE;
 		if (unlikely(++page_offset >= page_last))