diff mbox

[2/4] drm/vmwgfx: Defer fence irq processing to a tasklet

Message ID 1446198166-3068-2-git-send-email-thellstrom@vmware.com (mailing list archive)
State New, archived
Headers show

Commit Message

Thomas Hellstrom Oct. 30, 2015, 9:42 a.m. UTC
Reduce the time in hardware irq context and hardware irq latency.

Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com>
Reviewed-by: Sinclair Yeh <syeh@vmware.com>
---
 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c | 108 ++++++++++++++++++++--------------
 drivers/gpu/drm/vmwgfx/vmwgfx_fence.h |   2 +
 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c   |   6 +-
 3 files changed, 68 insertions(+), 48 deletions(-)

Comments

Daniel Vetter Oct. 30, 2015, 10:23 a.m. UTC | #1
On Fri, Oct 30, 2015 at 02:42:44AM -0700, Thomas Hellstrom wrote:
> Reduce the time in hardware irq context and hardware irq latency.
> 
> Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com>
> Reviewed-by: Sinclair Yeh <syeh@vmware.com>
> ---
>  drivers/gpu/drm/vmwgfx/vmwgfx_fence.c | 108 ++++++++++++++++++++--------------
>  drivers/gpu/drm/vmwgfx/vmwgfx_fence.h |   2 +
>  drivers/gpu/drm/vmwgfx/vmwgfx_irq.c   |   6 +-
>  3 files changed, 68 insertions(+), 48 deletions(-)
> 
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> index 8e689b4..f40c36e 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> @@ -47,6 +47,7 @@ struct vmw_fence_manager {
>  	bool seqno_valid; /* Protected by @lock, and may not be set to true
>  			     without the @goal_irq_mutex held. */
>  	unsigned ctx;
> +	struct tasklet_struct tasklet;

Bottom halves are super-deprecated except for giant existing users like
networking. I think the recommended way to do this is to either use
threaded interrupts or work-queues. The reason for that seems to be that
locking is funky around them, which is a major pain for RT. And RT is
going mainline now for real.
-Daniel

>  };
>  
>  struct vmw_user_fence {
> @@ -81,6 +82,8 @@ struct vmw_event_fence_action {
>  	uint32_t *tv_usec;
>  };
>  
> +static void vmw_fence_tasklet(unsigned long data);
> +
>  static struct vmw_fence_manager *
>  fman_from_fence(struct vmw_fence_obj *fence)
>  {
> @@ -115,12 +118,11 @@ static void vmw_fence_obj_destroy(struct fence *f)
>  		container_of(f, struct vmw_fence_obj, base);
>  
>  	struct vmw_fence_manager *fman = fman_from_fence(fence);
> -	unsigned long irq_flags;
>  
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  	list_del_init(&fence->head);
>  	--fman->num_fence_objects;
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  	fence->destroy(fence);
>  }
>  
> @@ -177,7 +179,6 @@ static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
>  	struct vmw_private *dev_priv = fman->dev_priv;
>  	struct vmwgfx_wait_cb cb;
>  	long ret = timeout;
> -	unsigned long irq_flags;
>  
>  	if (likely(vmw_fence_obj_signaled(fence)))
>  		return timeout;
> @@ -185,7 +186,7 @@ static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
>  	vmw_fifo_ping_host(dev_priv, SVGA_SYNC_GENERIC);
>  	vmw_seqno_waiter_add(dev_priv);
>  
> -	spin_lock_irqsave(f->lock, irq_flags);
> +	spin_lock_bh(f->lock);
>  
>  	if (intr && signal_pending(current)) {
>  		ret = -ERESTARTSYS;
> @@ -205,11 +206,11 @@ static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
>  			__set_current_state(TASK_INTERRUPTIBLE);
>  		else
>  			__set_current_state(TASK_UNINTERRUPTIBLE);
> -		spin_unlock_irqrestore(f->lock, irq_flags);
> +		spin_unlock_bh(f->lock);
>  
>  		ret = schedule_timeout(ret);
>  
> -		spin_lock_irqsave(f->lock, irq_flags);
> +		spin_lock_bh(f->lock);
>  		if (ret > 0 && intr && signal_pending(current))
>  			ret = -ERESTARTSYS;
>  	}
> @@ -219,7 +220,7 @@ static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
>  	__set_current_state(TASK_RUNNING);
>  
>  out:
> -	spin_unlock_irqrestore(f->lock, irq_flags);
> +	spin_unlock_bh(f->lock);
>  
>  	vmw_seqno_waiter_remove(dev_priv);
>  
> @@ -300,21 +301,22 @@ struct vmw_fence_manager *vmw_fence_manager_init(struct vmw_private *dev_priv)
>  		ttm_round_pot(sizeof(struct vmw_event_fence_action));
>  	mutex_init(&fman->goal_irq_mutex);
>  	fman->ctx = fence_context_alloc(1);
> +	tasklet_init(&fman->tasklet, vmw_fence_tasklet,
> +		     (unsigned long) fman);
>  
>  	return fman;
>  }
>  
>  void vmw_fence_manager_takedown(struct vmw_fence_manager *fman)
>  {
> -	unsigned long irq_flags;
>  	bool lists_empty;
>  
>  	(void) cancel_work_sync(&fman->work);
>  
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  	lists_empty = list_empty(&fman->fence_list) &&
>  		list_empty(&fman->cleanup_list);
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  
>  	BUG_ON(!lists_empty);
>  	kfree(fman);
> @@ -324,7 +326,6 @@ static int vmw_fence_obj_init(struct vmw_fence_manager *fman,
>  			      struct vmw_fence_obj *fence, u32 seqno,
>  			      void (*destroy) (struct vmw_fence_obj *fence))
>  {
> -	unsigned long irq_flags;
>  	int ret = 0;
>  
>  	fence_init(&fence->base, &vmw_fence_ops, &fman->lock,
> @@ -332,7 +333,7 @@ static int vmw_fence_obj_init(struct vmw_fence_manager *fman,
>  	INIT_LIST_HEAD(&fence->seq_passed_actions);
>  	fence->destroy = destroy;
>  
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  	if (unlikely(fman->fifo_down)) {
>  		ret = -EBUSY;
>  		goto out_unlock;
> @@ -341,7 +342,7 @@ static int vmw_fence_obj_init(struct vmw_fence_manager *fman,
>  	++fman->num_fence_objects;
>  
>  out_unlock:
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  	return ret;
>  
>  }
> @@ -490,11 +491,9 @@ rerun:
>  
>  void vmw_fences_update(struct vmw_fence_manager *fman)
>  {
> -	unsigned long irq_flags;
> -
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  	__vmw_fences_update(fman);
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  }
>  
>  bool vmw_fence_obj_signaled(struct vmw_fence_obj *fence)
> @@ -694,11 +693,9 @@ void vmw_fence_fifo_down(struct vmw_fence_manager *fman)
>  
>  void vmw_fence_fifo_up(struct vmw_fence_manager *fman)
>  {
> -	unsigned long irq_flags;
> -
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  	fman->fifo_down = false;
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  }
>  
>  
> @@ -825,10 +822,9 @@ void vmw_event_fence_fpriv_gone(struct vmw_fence_manager *fman,
>  {
>  	struct vmw_event_fence_action *eaction;
>  	struct drm_pending_event *event;
> -	unsigned long irq_flags;
>  
>  	while (1) {
> -		spin_lock_irqsave(&fman->lock, irq_flags);
> +		spin_lock_bh(&fman->lock);
>  		if (list_empty(event_list))
>  			goto out_unlock;
>  		eaction = list_first_entry(event_list,
> @@ -837,11 +833,11 @@ void vmw_event_fence_fpriv_gone(struct vmw_fence_manager *fman,
>  		list_del_init(&eaction->fpriv_head);
>  		event = eaction->event;
>  		eaction->event = NULL;
> -		spin_unlock_irqrestore(&fman->lock, irq_flags);
> +		spin_unlock_bh(&fman->lock);
>  		event->destroy(event);
>  	}
>  out_unlock:
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  }
>  
>  
> @@ -854,7 +850,7 @@ out_unlock:
>   * This function is called when the seqno of the fence where @action is
>   * attached has passed. It queues the event on the submitter's event list.
>   * This function is always called from atomic context, and may be called
> - * from irq context.
> + * from tasklet context.
>   */
>  static void vmw_event_fence_action_seq_passed(struct vmw_fence_action *action)
>  {
> @@ -863,13 +859,12 @@ static void vmw_event_fence_action_seq_passed(struct vmw_fence_action *action)
>  	struct drm_device *dev = eaction->dev;
>  	struct drm_pending_event *event = eaction->event;
>  	struct drm_file *file_priv;
> -	unsigned long irq_flags;
>  
>  	if (unlikely(event == NULL))
>  		return;
>  
>  	file_priv = event->file_priv;
> -	spin_lock_irqsave(&dev->event_lock, irq_flags);
> +	spin_lock_bh(&dev->event_lock);
>  
>  	if (likely(eaction->tv_sec != NULL)) {
>  		struct timeval tv;
> @@ -883,7 +878,7 @@ static void vmw_event_fence_action_seq_passed(struct vmw_fence_action *action)
>  	list_add_tail(&eaction->event->link, &file_priv->event_list);
>  	eaction->event = NULL;
>  	wake_up_all(&file_priv->event_wait);
> -	spin_unlock_irqrestore(&dev->event_lock, irq_flags);
> +	spin_unlock_bh(&dev->event_lock);
>  }
>  
>  /**
> @@ -900,11 +895,10 @@ static void vmw_event_fence_action_cleanup(struct vmw_fence_action *action)
>  	struct vmw_event_fence_action *eaction =
>  		container_of(action, struct vmw_event_fence_action, action);
>  	struct vmw_fence_manager *fman = fman_from_fence(eaction->fence);
> -	unsigned long irq_flags;
>  
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  	list_del(&eaction->fpriv_head);
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  
>  	vmw_fence_obj_unreference(&eaction->fence);
>  	kfree(eaction);
> @@ -924,11 +918,10 @@ static void vmw_fence_obj_add_action(struct vmw_fence_obj *fence,
>  			      struct vmw_fence_action *action)
>  {
>  	struct vmw_fence_manager *fman = fman_from_fence(fence);
> -	unsigned long irq_flags;
>  	bool run_update = false;
>  
>  	mutex_lock(&fman->goal_irq_mutex);
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  
>  	fman->pending_actions[action->type]++;
>  	if (fence_is_signaled_locked(&fence->base)) {
> @@ -947,7 +940,7 @@ static void vmw_fence_obj_add_action(struct vmw_fence_obj *fence,
>  		run_update = vmw_fence_goal_check_locked(fence);
>  	}
>  
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  
>  	if (run_update) {
>  		if (!fman->goal_irq_on) {
> @@ -985,7 +978,6 @@ int vmw_event_fence_action_queue(struct drm_file *file_priv,
>  	struct vmw_event_fence_action *eaction;
>  	struct vmw_fence_manager *fman = fman_from_fence(fence);
>  	struct vmw_fpriv *vmw_fp = vmw_fpriv(file_priv);
> -	unsigned long irq_flags;
>  
>  	eaction = kzalloc(sizeof(*eaction), GFP_KERNEL);
>  	if (unlikely(eaction == NULL))
> @@ -1002,9 +994,9 @@ int vmw_event_fence_action_queue(struct drm_file *file_priv,
>  	eaction->tv_sec = tv_sec;
>  	eaction->tv_usec = tv_usec;
>  
> -	spin_lock_irqsave(&fman->lock, irq_flags);
> +	spin_lock_bh(&fman->lock);
>  	list_add_tail(&eaction->fpriv_head, &vmw_fp->fence_events);
> -	spin_unlock_irqrestore(&fman->lock, irq_flags);
> +	spin_unlock_bh(&fman->lock);
>  
>  	vmw_fence_obj_add_action(fence, &eaction->action);
>  
> @@ -1025,16 +1017,15 @@ static int vmw_event_fence_action_create(struct drm_file *file_priv,
>  	struct vmw_event_fence_pending *event;
>  	struct vmw_fence_manager *fman = fman_from_fence(fence);
>  	struct drm_device *dev = fman->dev_priv->dev;
> -	unsigned long irq_flags;
>  	int ret;
>  
> -	spin_lock_irqsave(&dev->event_lock, irq_flags);
> +	spin_lock_bh(&dev->event_lock);
>  
>  	ret = (file_priv->event_space < sizeof(event->event)) ? -EBUSY : 0;
>  	if (likely(ret == 0))
>  		file_priv->event_space -= sizeof(event->event);
>  
> -	spin_unlock_irqrestore(&dev->event_lock, irq_flags);
> +	spin_unlock_bh(&dev->event_lock);
>  
>  	if (unlikely(ret != 0)) {
>  		DRM_ERROR("Failed to allocate event space for this file.\n");
> @@ -1078,9 +1069,9 @@ static int vmw_event_fence_action_create(struct drm_file *file_priv,
>  out_no_queue:
>  	event->base.destroy(&event->base);
>  out_no_event:
> -	spin_lock_irqsave(&dev->event_lock, irq_flags);
> +	spin_lock_bh(&dev->event_lock);
>  	file_priv->event_space += sizeof(*event);
> -	spin_unlock_irqrestore(&dev->event_lock, irq_flags);
> +	spin_unlock_bh(&dev->event_lock);
>  out_no_space:
>  	return ret;
>  }
> @@ -1172,3 +1163,32 @@ out_no_ref_obj:
>  	vmw_fence_obj_unreference(&fence);
>  	return ret;
>  }
> +
> +/**
> + * vmw_fence_tasklet - Fence manager tasklet entry point
> + *
> + * @data: The tasklet closure - A pointer to the fence manager cast to an
> + * unsigned long.
> + */
> +static void vmw_fence_tasklet(unsigned long data)
> +{
> +	struct vmw_fence_manager *fman = (struct vmw_fence_manager *) data;
> +
> +	spin_lock(&fman->lock);
> +	__vmw_fences_update(fman);
> +	spin_unlock(&fman->lock);
> +	wake_up_all(&fman->dev_priv->fence_queue);
> +}
> +
> +/**
> + * vmw_fence_tasklet_schedule - Schedule a fence manager tasklet run
> + *
> + * @fman: Pointer to a fence manager
> + */
> +void vmw_fence_tasklet_schedule(struct vmw_fence_manager *fman)
> +{
> +	if (!fman)
> +		return;
> +
> +	tasklet_schedule(&fman->tasklet);
> +}
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
> index 8be6c29..e55b2c9 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
> @@ -124,4 +124,6 @@ extern int vmw_event_fence_action_queue(struct drm_file *filee_priv,
>  					uint32_t *tv_sec,
>  					uint32_t *tv_usec,
>  					bool interruptible);
> +extern void vmw_fence_tasklet_schedule(struct vmw_fence_manager *fman);
> +
>  #endif /* _VMWGFX_FENCE_H_ */
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c b/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
> index ac3eccd..b0a6e65 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
> @@ -48,10 +48,8 @@ irqreturn_t vmw_irq_handler(int irq, void *arg)
>  		return IRQ_NONE;
>  
>  	if (masked_status & (SVGA_IRQFLAG_ANY_FENCE |
> -			     SVGA_IRQFLAG_FENCE_GOAL)) {
> -		vmw_fences_update(dev_priv->fman);
> -		wake_up_all(&dev_priv->fence_queue);
> -	}
> +			     SVGA_IRQFLAG_FENCE_GOAL))
> +		vmw_fence_tasklet_schedule(dev_priv->fman);
>  
>  	if (masked_status & SVGA_IRQFLAG_FIFO_PROGRESS)
>  		wake_up_all(&dev_priv->fifo_queue);
> -- 
> 2.4.3
> 
> _______________________________________________
> dri-devel mailing list
> dri-devel@lists.freedesktop.org
> http://lists.freedesktop.org/mailman/listinfo/dri-devel
Thomas Hellstrom Oct. 30, 2015, 11:08 a.m. UTC | #2
On 10/30/2015 11:23 AM, Daniel Vetter wrote:
> On Fri, Oct 30, 2015 at 02:42:44AM -0700, Thomas Hellstrom wrote:
>> Reduce the time in hardware irq context and hardware irq latency.
>>
>> Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com>
>> Reviewed-by: Sinclair Yeh <syeh@vmware.com>
>> ---
>>  drivers/gpu/drm/vmwgfx/vmwgfx_fence.c | 108 ++++++++++++++++++++--------------
>>  drivers/gpu/drm/vmwgfx/vmwgfx_fence.h |   2 +
>>  drivers/gpu/drm/vmwgfx/vmwgfx_irq.c   |   6 +-
>>  3 files changed, 68 insertions(+), 48 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
>> index 8e689b4..f40c36e 100644
>> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
>> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
>> @@ -47,6 +47,7 @@ struct vmw_fence_manager {
>>  	bool seqno_valid; /* Protected by @lock, and may not be set to true
>>  			     without the @goal_irq_mutex held. */
>>  	unsigned ctx;
>> +	struct tasklet_struct tasklet;
> Bottom halves are super-deprecated except for giant existing users like
> networking. I think the recommended way to do this is to either use
> threaded interrupts or work-queues. The reason for that seems to be that
> locking is funky around them, which is a major pain for RT. And RT is
> going mainline now for real.
> -Daniel
>
>

Thanks for the heads up. Unfortunately work-queues introduce too much
latency for this use-case.
Given that we (vmwgfx) already is an existing user (albeit not giant
:)), and RT in a VM guest is somewhat unlikely
I wonder how much of an issue this is.

I'll read up on threaded interrupts.

/Thomas
diff mbox

Patch

diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
index 8e689b4..f40c36e 100644
--- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
+++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
@@ -47,6 +47,7 @@  struct vmw_fence_manager {
 	bool seqno_valid; /* Protected by @lock, and may not be set to true
 			     without the @goal_irq_mutex held. */
 	unsigned ctx;
+	struct tasklet_struct tasklet;
 };
 
 struct vmw_user_fence {
@@ -81,6 +82,8 @@  struct vmw_event_fence_action {
 	uint32_t *tv_usec;
 };
 
+static void vmw_fence_tasklet(unsigned long data);
+
 static struct vmw_fence_manager *
 fman_from_fence(struct vmw_fence_obj *fence)
 {
@@ -115,12 +118,11 @@  static void vmw_fence_obj_destroy(struct fence *f)
 		container_of(f, struct vmw_fence_obj, base);
 
 	struct vmw_fence_manager *fman = fman_from_fence(fence);
-	unsigned long irq_flags;
 
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 	list_del_init(&fence->head);
 	--fman->num_fence_objects;
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 	fence->destroy(fence);
 }
 
@@ -177,7 +179,6 @@  static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
 	struct vmw_private *dev_priv = fman->dev_priv;
 	struct vmwgfx_wait_cb cb;
 	long ret = timeout;
-	unsigned long irq_flags;
 
 	if (likely(vmw_fence_obj_signaled(fence)))
 		return timeout;
@@ -185,7 +186,7 @@  static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
 	vmw_fifo_ping_host(dev_priv, SVGA_SYNC_GENERIC);
 	vmw_seqno_waiter_add(dev_priv);
 
-	spin_lock_irqsave(f->lock, irq_flags);
+	spin_lock_bh(f->lock);
 
 	if (intr && signal_pending(current)) {
 		ret = -ERESTARTSYS;
@@ -205,11 +206,11 @@  static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
 			__set_current_state(TASK_INTERRUPTIBLE);
 		else
 			__set_current_state(TASK_UNINTERRUPTIBLE);
-		spin_unlock_irqrestore(f->lock, irq_flags);
+		spin_unlock_bh(f->lock);
 
 		ret = schedule_timeout(ret);
 
-		spin_lock_irqsave(f->lock, irq_flags);
+		spin_lock_bh(f->lock);
 		if (ret > 0 && intr && signal_pending(current))
 			ret = -ERESTARTSYS;
 	}
@@ -219,7 +220,7 @@  static long vmw_fence_wait(struct fence *f, bool intr, signed long timeout)
 	__set_current_state(TASK_RUNNING);
 
 out:
-	spin_unlock_irqrestore(f->lock, irq_flags);
+	spin_unlock_bh(f->lock);
 
 	vmw_seqno_waiter_remove(dev_priv);
 
@@ -300,21 +301,22 @@  struct vmw_fence_manager *vmw_fence_manager_init(struct vmw_private *dev_priv)
 		ttm_round_pot(sizeof(struct vmw_event_fence_action));
 	mutex_init(&fman->goal_irq_mutex);
 	fman->ctx = fence_context_alloc(1);
+	tasklet_init(&fman->tasklet, vmw_fence_tasklet,
+		     (unsigned long) fman);
 
 	return fman;
 }
 
 void vmw_fence_manager_takedown(struct vmw_fence_manager *fman)
 {
-	unsigned long irq_flags;
 	bool lists_empty;
 
 	(void) cancel_work_sync(&fman->work);
 
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 	lists_empty = list_empty(&fman->fence_list) &&
 		list_empty(&fman->cleanup_list);
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 
 	BUG_ON(!lists_empty);
 	kfree(fman);
@@ -324,7 +326,6 @@  static int vmw_fence_obj_init(struct vmw_fence_manager *fman,
 			      struct vmw_fence_obj *fence, u32 seqno,
 			      void (*destroy) (struct vmw_fence_obj *fence))
 {
-	unsigned long irq_flags;
 	int ret = 0;
 
 	fence_init(&fence->base, &vmw_fence_ops, &fman->lock,
@@ -332,7 +333,7 @@  static int vmw_fence_obj_init(struct vmw_fence_manager *fman,
 	INIT_LIST_HEAD(&fence->seq_passed_actions);
 	fence->destroy = destroy;
 
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 	if (unlikely(fman->fifo_down)) {
 		ret = -EBUSY;
 		goto out_unlock;
@@ -341,7 +342,7 @@  static int vmw_fence_obj_init(struct vmw_fence_manager *fman,
 	++fman->num_fence_objects;
 
 out_unlock:
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 	return ret;
 
 }
@@ -490,11 +491,9 @@  rerun:
 
 void vmw_fences_update(struct vmw_fence_manager *fman)
 {
-	unsigned long irq_flags;
-
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 	__vmw_fences_update(fman);
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 }
 
 bool vmw_fence_obj_signaled(struct vmw_fence_obj *fence)
@@ -694,11 +693,9 @@  void vmw_fence_fifo_down(struct vmw_fence_manager *fman)
 
 void vmw_fence_fifo_up(struct vmw_fence_manager *fman)
 {
-	unsigned long irq_flags;
-
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 	fman->fifo_down = false;
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 }
 
 
@@ -825,10 +822,9 @@  void vmw_event_fence_fpriv_gone(struct vmw_fence_manager *fman,
 {
 	struct vmw_event_fence_action *eaction;
 	struct drm_pending_event *event;
-	unsigned long irq_flags;
 
 	while (1) {
-		spin_lock_irqsave(&fman->lock, irq_flags);
+		spin_lock_bh(&fman->lock);
 		if (list_empty(event_list))
 			goto out_unlock;
 		eaction = list_first_entry(event_list,
@@ -837,11 +833,11 @@  void vmw_event_fence_fpriv_gone(struct vmw_fence_manager *fman,
 		list_del_init(&eaction->fpriv_head);
 		event = eaction->event;
 		eaction->event = NULL;
-		spin_unlock_irqrestore(&fman->lock, irq_flags);
+		spin_unlock_bh(&fman->lock);
 		event->destroy(event);
 	}
 out_unlock:
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 }
 
 
@@ -854,7 +850,7 @@  out_unlock:
  * This function is called when the seqno of the fence where @action is
  * attached has passed. It queues the event on the submitter's event list.
  * This function is always called from atomic context, and may be called
- * from irq context.
+ * from tasklet context.
  */
 static void vmw_event_fence_action_seq_passed(struct vmw_fence_action *action)
 {
@@ -863,13 +859,12 @@  static void vmw_event_fence_action_seq_passed(struct vmw_fence_action *action)
 	struct drm_device *dev = eaction->dev;
 	struct drm_pending_event *event = eaction->event;
 	struct drm_file *file_priv;
-	unsigned long irq_flags;
 
 	if (unlikely(event == NULL))
 		return;
 
 	file_priv = event->file_priv;
-	spin_lock_irqsave(&dev->event_lock, irq_flags);
+	spin_lock_bh(&dev->event_lock);
 
 	if (likely(eaction->tv_sec != NULL)) {
 		struct timeval tv;
@@ -883,7 +878,7 @@  static void vmw_event_fence_action_seq_passed(struct vmw_fence_action *action)
 	list_add_tail(&eaction->event->link, &file_priv->event_list);
 	eaction->event = NULL;
 	wake_up_all(&file_priv->event_wait);
-	spin_unlock_irqrestore(&dev->event_lock, irq_flags);
+	spin_unlock_bh(&dev->event_lock);
 }
 
 /**
@@ -900,11 +895,10 @@  static void vmw_event_fence_action_cleanup(struct vmw_fence_action *action)
 	struct vmw_event_fence_action *eaction =
 		container_of(action, struct vmw_event_fence_action, action);
 	struct vmw_fence_manager *fman = fman_from_fence(eaction->fence);
-	unsigned long irq_flags;
 
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 	list_del(&eaction->fpriv_head);
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 
 	vmw_fence_obj_unreference(&eaction->fence);
 	kfree(eaction);
@@ -924,11 +918,10 @@  static void vmw_fence_obj_add_action(struct vmw_fence_obj *fence,
 			      struct vmw_fence_action *action)
 {
 	struct vmw_fence_manager *fman = fman_from_fence(fence);
-	unsigned long irq_flags;
 	bool run_update = false;
 
 	mutex_lock(&fman->goal_irq_mutex);
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 
 	fman->pending_actions[action->type]++;
 	if (fence_is_signaled_locked(&fence->base)) {
@@ -947,7 +940,7 @@  static void vmw_fence_obj_add_action(struct vmw_fence_obj *fence,
 		run_update = vmw_fence_goal_check_locked(fence);
 	}
 
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 
 	if (run_update) {
 		if (!fman->goal_irq_on) {
@@ -985,7 +978,6 @@  int vmw_event_fence_action_queue(struct drm_file *file_priv,
 	struct vmw_event_fence_action *eaction;
 	struct vmw_fence_manager *fman = fman_from_fence(fence);
 	struct vmw_fpriv *vmw_fp = vmw_fpriv(file_priv);
-	unsigned long irq_flags;
 
 	eaction = kzalloc(sizeof(*eaction), GFP_KERNEL);
 	if (unlikely(eaction == NULL))
@@ -1002,9 +994,9 @@  int vmw_event_fence_action_queue(struct drm_file *file_priv,
 	eaction->tv_sec = tv_sec;
 	eaction->tv_usec = tv_usec;
 
-	spin_lock_irqsave(&fman->lock, irq_flags);
+	spin_lock_bh(&fman->lock);
 	list_add_tail(&eaction->fpriv_head, &vmw_fp->fence_events);
-	spin_unlock_irqrestore(&fman->lock, irq_flags);
+	spin_unlock_bh(&fman->lock);
 
 	vmw_fence_obj_add_action(fence, &eaction->action);
 
@@ -1025,16 +1017,15 @@  static int vmw_event_fence_action_create(struct drm_file *file_priv,
 	struct vmw_event_fence_pending *event;
 	struct vmw_fence_manager *fman = fman_from_fence(fence);
 	struct drm_device *dev = fman->dev_priv->dev;
-	unsigned long irq_flags;
 	int ret;
 
-	spin_lock_irqsave(&dev->event_lock, irq_flags);
+	spin_lock_bh(&dev->event_lock);
 
 	ret = (file_priv->event_space < sizeof(event->event)) ? -EBUSY : 0;
 	if (likely(ret == 0))
 		file_priv->event_space -= sizeof(event->event);
 
-	spin_unlock_irqrestore(&dev->event_lock, irq_flags);
+	spin_unlock_bh(&dev->event_lock);
 
 	if (unlikely(ret != 0)) {
 		DRM_ERROR("Failed to allocate event space for this file.\n");
@@ -1078,9 +1069,9 @@  static int vmw_event_fence_action_create(struct drm_file *file_priv,
 out_no_queue:
 	event->base.destroy(&event->base);
 out_no_event:
-	spin_lock_irqsave(&dev->event_lock, irq_flags);
+	spin_lock_bh(&dev->event_lock);
 	file_priv->event_space += sizeof(*event);
-	spin_unlock_irqrestore(&dev->event_lock, irq_flags);
+	spin_unlock_bh(&dev->event_lock);
 out_no_space:
 	return ret;
 }
@@ -1172,3 +1163,32 @@  out_no_ref_obj:
 	vmw_fence_obj_unreference(&fence);
 	return ret;
 }
+
+/**
+ * vmw_fence_tasklet - Fence manager tasklet entry point
+ *
+ * @data: The tasklet closure - A pointer to the fence manager cast to an
+ * unsigned long.
+ */
+static void vmw_fence_tasklet(unsigned long data)
+{
+	struct vmw_fence_manager *fman = (struct vmw_fence_manager *) data;
+
+	spin_lock(&fman->lock);
+	__vmw_fences_update(fman);
+	spin_unlock(&fman->lock);
+	wake_up_all(&fman->dev_priv->fence_queue);
+}
+
+/**
+ * vmw_fence_tasklet_schedule - Schedule a fence manager tasklet run
+ *
+ * @fman: Pointer to a fence manager
+ */
+void vmw_fence_tasklet_schedule(struct vmw_fence_manager *fman)
+{
+	if (!fman)
+		return;
+
+	tasklet_schedule(&fman->tasklet);
+}
diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
index 8be6c29..e55b2c9 100644
--- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
+++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.h
@@ -124,4 +124,6 @@  extern int vmw_event_fence_action_queue(struct drm_file *filee_priv,
 					uint32_t *tv_sec,
 					uint32_t *tv_usec,
 					bool interruptible);
+extern void vmw_fence_tasklet_schedule(struct vmw_fence_manager *fman);
+
 #endif /* _VMWGFX_FENCE_H_ */
diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c b/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
index ac3eccd..b0a6e65 100644
--- a/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
+++ b/drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
@@ -48,10 +48,8 @@  irqreturn_t vmw_irq_handler(int irq, void *arg)
 		return IRQ_NONE;
 
 	if (masked_status & (SVGA_IRQFLAG_ANY_FENCE |
-			     SVGA_IRQFLAG_FENCE_GOAL)) {
-		vmw_fences_update(dev_priv->fman);
-		wake_up_all(&dev_priv->fence_queue);
-	}
+			     SVGA_IRQFLAG_FENCE_GOAL))
+		vmw_fence_tasklet_schedule(dev_priv->fman);
 
 	if (masked_status & SVGA_IRQFLAG_FIFO_PROGRESS)
 		wake_up_all(&dev_priv->fifo_queue);