diff mbox series

scsi: target: tcmu: userspace must not complete queued commands

Message ID 20200512154250.12310-1-bstroesser@ts.fujitsu.com (mailing list archive)
State Superseded
Headers show
Series scsi: target: tcmu: userspace must not complete queued commands | expand

Commit Message

Bodo Stroesser May 12, 2020, 3:42 p.m. UTC
When tcmu queues a new command - no matter whether in command
ring or in qfull_queue - a cmd_id from IDR udev->commands is
assigned to the command.

If userspaces sends a wrong command completion containing the
cmd_id of a command on the qfull_queue, tcmu_handle_completions()
finds the command in the IDR and calls tcmu_handle_completion()
for it. This might do some nasty things, because commands in
qfull_queue do not have a valid dbi list.

To fix this bug, we no longer add queued commands to the idr.
Instead the cmd_id is assign when a command is written to
the command ring.

Due to this change I had to adapt the source code at several
places where up to now an idr_for_each had been done.

Signed-off-by: Bodo Stroesser <bstroesser@ts.fujitsu.com>
---
 drivers/target/target_core_user.c | 154 ++++++++++++++++++--------------------
 1 file changed, 71 insertions(+), 83 deletions(-)

Comments

Mike Christie May 17, 2020, 8:38 p.m. UTC | #1
On 5/12/20 10:42 AM, Bodo Stroesser wrote:
> When tcmu queues a new command - no matter whether in command
> ring or in qfull_queue - a cmd_id from IDR udev->commands is
> assigned to the command.
> 
> If userspaces sends a wrong command completion containing the
> cmd_id of a command on the qfull_queue, tcmu_handle_completions()
> finds the command in the IDR and calls tcmu_handle_completion()
> for it. This might do some nasty things, because commands in
> qfull_queue do not have a valid dbi list.
> 
> To fix this bug, we no longer add queued commands to the idr.
> Instead the cmd_id is assign when a command is written to
> the command ring.
> 
> Due to this change I had to adapt the source code at several
> places where up to now an idr_for_each had been done.
> 
> Signed-off-by: Bodo Stroesser <bstroesser@ts.fujitsu.com>
> ---
>  drivers/target/target_core_user.c | 154 ++++++++++++++++++--------------------
>  1 file changed, 71 insertions(+), 83 deletions(-)
> 
> diff --git a/drivers/target/target_core_user.c b/drivers/target/target_core_user.c
> index f769bb1e3735..de75583b74c6 100644
> --- a/drivers/target/target_core_user.c
> +++ b/drivers/target/target_core_user.c
> @@ -882,41 +882,24 @@ static inline size_t tcmu_cmd_get_cmd_size(struct tcmu_cmd *tcmu_cmd,
>  	return command_size;
>  }
>  
> -static int tcmu_setup_cmd_timer(struct tcmu_cmd *tcmu_cmd, unsigned int tmo,
> -				struct timer_list *timer)
> +static void tcmu_setup_cmd_timer(struct tcmu_cmd *tcmu_cmd, unsigned int tmo,
> +				 struct timer_list *timer)
>  {
> -	struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
> -	int cmd_id;
> -
> -	if (tcmu_cmd->cmd_id)
> -		goto setup_timer;
> -
> -	cmd_id = idr_alloc(&udev->commands, tcmu_cmd, 1, USHRT_MAX, GFP_NOWAIT);
> -	if (cmd_id < 0) {
> -		pr_err("tcmu: Could not allocate cmd id.\n");
> -		return cmd_id;
> -	}
> -	tcmu_cmd->cmd_id = cmd_id;
> -
> -	pr_debug("allocated cmd %u for dev %s tmo %lu\n", tcmu_cmd->cmd_id,
> -		 udev->name, tmo / MSEC_PER_SEC);
> -
> -setup_timer:
>  	if (!tmo)
> -		return 0;
> +		return;
>  
>  	tcmu_cmd->deadline = round_jiffies_up(jiffies + msecs_to_jiffies(tmo));
>  	if (!timer_pending(timer))
>  		mod_timer(timer, tcmu_cmd->deadline);
>  
> -	return 0;
> +	pr_debug("Timeout set up for cmd %p, dev = %s, tmo = %lu\n", tcmu_cmd,
> +		 tcmu_cmd->tcmu_dev->name, tmo / MSEC_PER_SEC);
>  }
>  
>  static int add_to_qfull_queue(struct tcmu_cmd *tcmu_cmd)
>  {
>  	struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
>  	unsigned int tmo;
> -	int ret;
>  
>  	/*
>  	 * For backwards compat if qfull_time_out is not set use
> @@ -931,13 +914,11 @@ static int add_to_qfull_queue(struct tcmu_cmd *tcmu_cmd)
>  	else
>  		tmo = TCMU_TIME_OUT;
>  
> -	ret = tcmu_setup_cmd_timer(tcmu_cmd, tmo, &udev->qfull_timer);
> -	if (ret)
> -		return ret;
> +	tcmu_setup_cmd_timer(tcmu_cmd, tmo, &udev->qfull_timer);
>  
>  	list_add_tail(&tcmu_cmd->queue_entry, &udev->qfull_queue);
> -	pr_debug("adding cmd %u on dev %s to ring space wait queue\n",
> -		 tcmu_cmd->cmd_id, udev->name);
> +	pr_debug("adding cmd %p on dev %s to ring space wait queue\n",
> +		 tcmu_cmd, udev->name);
>  	return 0;
>  }
>  
> @@ -959,7 +940,7 @@ static int queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, sense_reason_t *scsi_err)
>  	struct tcmu_mailbox *mb;
>  	struct tcmu_cmd_entry *entry;
>  	struct iovec *iov;
> -	int iov_cnt, ret;
> +	int iov_cnt, cmd_id;
>  	uint32_t cmd_head;
>  	uint64_t cdb_off;
>  	bool copy_to_data_area;
> @@ -1060,14 +1041,21 @@ static int queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, sense_reason_t *scsi_err)
>  	}
>  	entry->req.iov_bidi_cnt = iov_cnt;
>  
> -	ret = tcmu_setup_cmd_timer(tcmu_cmd, udev->cmd_time_out,
> -				   &udev->cmd_timer);
> -	if (ret) {
> -		tcmu_cmd_free_data(tcmu_cmd, tcmu_cmd->dbi_cnt);
> +	cmd_id = idr_alloc(&udev->commands, tcmu_cmd, 1, USHRT_MAX, GFP_NOWAIT);
> +	if (cmd_id < 0) {
> +		pr_err("tcmu: Could not allocate cmd id.\n");
>  
> +		tcmu_cmd_free_data(tcmu_cmd, tcmu_cmd->dbi_cnt);
>  		*scsi_err = TCM_OUT_OF_RESOURCES;
>  		return -1;
>  	}
> +	tcmu_cmd->cmd_id = cmd_id;
> +
> +	pr_debug("allocated cmd id %u for cmd %p dev %s\n", tcmu_cmd->cmd_id,
> +		 tcmu_cmd, udev->name);
> +
> +	tcmu_setup_cmd_timer(tcmu_cmd, udev->cmd_time_out, &udev->cmd_timer);
> +
>  	entry->hdr.cmd_id = tcmu_cmd->cmd_id;
>  
>  	/*
> @@ -1279,50 +1267,39 @@ static unsigned int tcmu_handle_completions(struct tcmu_dev *udev)
>  	return handled;
>  }
>  
> -static int tcmu_check_expired_cmd(int id, void *p, void *data)
> +static void tcmu_check_expired_ring_cmd(struct tcmu_cmd *cmd)
>  {
> -	struct tcmu_cmd *cmd = p;
> -	struct tcmu_dev *udev = cmd->tcmu_dev;
> -	u8 scsi_status;
>  	struct se_cmd *se_cmd;
> -	bool is_running;
> -
> -	if (test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags))
> -		return 0;
>  
>  	if (!time_after(jiffies, cmd->deadline))
> -		return 0;
> +		return;
>  
> -	is_running = test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags);
> +	set_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags);
> +	list_del_init(&cmd->queue_entry);
>  	se_cmd = cmd->se_cmd;
> +	cmd->se_cmd = NULL;
>  
> -	if (is_running) {
> -		/*
> -		 * If cmd_time_out is disabled but qfull is set deadline
> -		 * will only reflect the qfull timeout. Ignore it.
> -		 */
> -		if (!udev->cmd_time_out)
> -			return 0;
> +	pr_debug("Timing out inflight cmd %u on dev %s.\n",
> +		 cmd->cmd_id, cmd->tcmu_dev->name);
>  
> -		set_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags);
> -		/*
> -		 * target_complete_cmd will translate this to LUN COMM FAILURE
> -		 */
> -		scsi_status = SAM_STAT_CHECK_CONDITION;
> -		list_del_init(&cmd->queue_entry);
> -		cmd->se_cmd = NULL;
> -	} else {
> -		list_del_init(&cmd->queue_entry);
> -		idr_remove(&udev->commands, id);
> -		tcmu_free_cmd(cmd);
> -		scsi_status = SAM_STAT_TASK_SET_FULL;
> -	}
> +	target_complete_cmd(se_cmd, SAM_STAT_CHECK_CONDITION);
> +}
>  
> -	pr_debug("Timing out cmd %u on dev %s that is %s.\n",
> -		 id, udev->name, is_running ? "inflight" : "queued");
> +static void tcmu_check_expired_queue_cmd(struct tcmu_cmd *cmd)
> +{
> +	struct se_cmd *se_cmd;
>  
> -	target_complete_cmd(se_cmd, scsi_status);
> -	return 0;
> +	if (!time_after(jiffies, cmd->deadline))
> +		return;
> +
> +	list_del_init(&cmd->queue_entry);
> +	se_cmd = cmd->se_cmd;
> +	tcmu_free_cmd(cmd);
> +
> +	pr_debug("Timing out queued cmd %p on dev %s.\n",
> +		  cmd, cmd->tcmu_dev->name);
> +
> +	target_complete_cmd(se_cmd, SAM_STAT_TASK_SET_FULL);
>  }
>  
>  static void tcmu_device_timedout(struct tcmu_dev *udev)
> @@ -1407,16 +1384,15 @@ static struct se_device *tcmu_alloc_device(struct se_hba *hba, const char *name)
>  	return &udev->se_dev;
>  }
>  
> -static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
> +static void run_qfull_queue(struct tcmu_dev *udev, bool fail)
>  {
>  	struct tcmu_cmd *tcmu_cmd, *tmp_cmd;
>  	LIST_HEAD(cmds);
> -	bool drained = true;
>  	sense_reason_t scsi_ret;
>  	int ret;
>  
>  	if (list_empty(&udev->qfull_queue))
> -		return true;
> +		return;
>  
>  	pr_debug("running %s's cmdr queue forcefail %d\n", udev->name, fail);
>  
> @@ -1425,11 +1401,10 @@ static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
>  	list_for_each_entry_safe(tcmu_cmd, tmp_cmd, &cmds, queue_entry) {
>  		list_del_init(&tcmu_cmd->queue_entry);
>  
> -	        pr_debug("removing cmd %u on dev %s from queue\n",
> -		         tcmu_cmd->cmd_id, udev->name);
> +	        pr_debug("removing cmd %p on dev %s from queue\n",
> +		         tcmu_cmd, udev->name);
>  
>  		if (fail) {
> -			idr_remove(&udev->commands, tcmu_cmd->cmd_id);
>  			/*
>  			 * We were not able to even start the command, so
>  			 * fail with busy to allow a retry in case runner
> @@ -1444,10 +1419,8 @@ static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
>  
>  		ret = queue_cmd_ring(tcmu_cmd, &scsi_ret);
>  		if (ret < 0) {
> -		        pr_debug("cmd %u on dev %s failed with %u\n",
> -			         tcmu_cmd->cmd_id, udev->name, scsi_ret);
> -
> -			idr_remove(&udev->commands, tcmu_cmd->cmd_id);
> +		        pr_debug("cmd %p on dev %s failed with %u\n",
> +			         tcmu_cmd, udev->name, scsi_ret);
>  			/*
>  			 * Ignore scsi_ret for now. target_complete_cmd
>  			 * drops it.
> @@ -1462,13 +1435,11 @@ static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
>  			 * the queue
>  			 */
>  			list_splice_tail(&cmds, &udev->qfull_queue);
> -			drained = false;
>  			break;
>  		}
>  	}
>  
>  	tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer);
> -	return drained;
>  }
>  
>  static int tcmu_irqcontrol(struct uio_info *info, s32 irq_on)
> @@ -1652,6 +1623,8 @@ static void tcmu_dev_kref_release(struct kref *kref)
>  		if (tcmu_check_and_free_pending_cmd(cmd) != 0)
>  			all_expired = false;
>  	}
> +	if (!list_empty(&udev->qfull_queue))
> +		all_expired = false;

Just drop. It's never going to happen and doesn't tell us a lot. We
don't know what list it was on and why LIO core's refcounts got messed up.

The WARN_ON should just be removed in some other patchset one day. If we
wanted to detect if non timed out commands were running at this time
then it should go in lio core, because it would be a problem that
affects all drivers.

I think we actually want the opposite and should WARN or BUG if commands
have timed out and we are freeing the device. In this case userspace
could still be accessing the buffers we are going to free. It probably
won't happen now that we run this from the release function and so the
device is closed though.



>  	idr_destroy(&udev->commands);
>  	WARN_ON(!all_expired);
>  
> @@ -2037,9 +2010,6 @@ static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
>  	mutex_lock(&udev->cmdr_lock);
>  
>  	idr_for_each_entry(&udev->commands, cmd, i) {
> -		if (!test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags))
> -			continue;
> -
>  		pr_debug("removing cmd %u on dev %s from ring (is expired %d)\n",
>  			  cmd->cmd_id, udev->name,
>  			  test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags));
> @@ -2077,6 +2047,8 @@ static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
>  
>  	del_timer(&udev->cmd_timer);
>  
> +	run_qfull_queue(udev, false);
> +
>  	mutex_unlock(&udev->cmdr_lock);
>  }
>  
> @@ -2698,6 +2670,7 @@ static void find_free_blocks(void)
>  static void check_timedout_devices(void)
>  {
>  	struct tcmu_dev *udev, *tmp_dev;
> +	struct tcmu_cmd *ucmd, *tmp_cmd;

ucmd probably would have been best when the code was first written. Now,
you should use tcmu_cmd or cmd to match the existing naming style.


>  	LIST_HEAD(devs);
>  
>  	spin_lock_bh(&timed_out_udevs_lock);
> @@ -2708,9 +2681,24 @@ static void check_timedout_devices(void)
>  		spin_unlock_bh(&timed_out_udevs_lock);
>  
>  		mutex_lock(&udev->cmdr_lock);
> -		idr_for_each(&udev->commands, tcmu_check_expired_cmd, NULL);
>  
> -		tcmu_set_next_deadline(&udev->inflight_queue, &udev->cmd_timer);
> +		/*
> +		 * If cmd_time_out is disabled but qfull is set deadline
> +		 * will only reflect the qfull timeout. Ignore it.
> +		 */
> +		if (udev->cmd_time_out) {
> +			list_for_each_entry_safe(ucmd, tmp_cmd,
> +			                         &udev->inflight_queue,
> +			                         queue_entry) {
> +				tcmu_check_expired_ring_cmd(ucmd);
> +			}
> +			tcmu_set_next_deadline(&udev->inflight_queue,
> +			                       &udev->cmd_timer);
> +		}
> +		list_for_each_entry_safe(ucmd, tmp_cmd, &udev->qfull_queue,
> +			                 queue_entry) {
> +			tcmu_check_expired_queue_cmd(ucmd);
> +		}
>  		tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer);
>  
>  		mutex_unlock(&udev->cmdr_lock);
>
Bodo Stroesser May 18, 2020, 2:49 p.m. UTC | #2
On 05/17/20 22:38, Mike Christie wrote:
> On 5/12/20 10:42 AM, Bodo Stroesser wrote:
>> When tcmu queues a new command - no matter whether in command
>> ring or in qfull_queue - a cmd_id from IDR udev->commands is
>> assigned to the command.
>>
>> If userspaces sends a wrong command completion containing the
>> cmd_id of a command on the qfull_queue, tcmu_handle_completions()
>> finds the command in the IDR and calls tcmu_handle_completion()
>> for it. This might do some nasty things, because commands in
>> qfull_queue do not have a valid dbi list.
>>
>> To fix this bug, we no longer add queued commands to the idr.
>> Instead the cmd_id is assign when a command is written to
>> the command ring.
>>
>> Due to this change I had to adapt the source code at several
>> places where up to now an idr_for_each had been done.
>>
>> Signed-off-by: Bodo Stroesser <bstroesser@ts.fujitsu.com>
>> ---
>>   drivers/target/target_core_user.c | 154 ++++++++++++++++++--------------------
>>   1 file changed, 71 insertions(+), 83 deletions(-)
>>
>> diff --git a/drivers/target/target_core_user.c b/drivers/target/target_core_user.c
>> index f769bb1e3735..de75583b74c6 100644
>> --- a/drivers/target/target_core_user.c
>> +++ b/drivers/target/target_core_user.c
>> @@ -882,41 +882,24 @@ static inline size_t tcmu_cmd_get_cmd_size(struct tcmu_cmd *tcmu_cmd,
>>   	return command_size;
>>   }
>>   
>> -static int tcmu_setup_cmd_timer(struct tcmu_cmd *tcmu_cmd, unsigned int tmo,
>> -				struct timer_list *timer)
>> +static void tcmu_setup_cmd_timer(struct tcmu_cmd *tcmu_cmd, unsigned int tmo,
>> +				 struct timer_list *timer)
>>   {
>> -	struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
>> -	int cmd_id;
>> -
>> -	if (tcmu_cmd->cmd_id)
>> -		goto setup_timer;
>> -
>> -	cmd_id = idr_alloc(&udev->commands, tcmu_cmd, 1, USHRT_MAX, GFP_NOWAIT);
>> -	if (cmd_id < 0) {
>> -		pr_err("tcmu: Could not allocate cmd id.\n");
>> -		return cmd_id;
>> -	}
>> -	tcmu_cmd->cmd_id = cmd_id;
>> -
>> -	pr_debug("allocated cmd %u for dev %s tmo %lu\n", tcmu_cmd->cmd_id,
>> -		 udev->name, tmo / MSEC_PER_SEC);
>> -
>> -setup_timer:
>>   	if (!tmo)
>> -		return 0;
>> +		return;
>>   
>>   	tcmu_cmd->deadline = round_jiffies_up(jiffies + msecs_to_jiffies(tmo));
>>   	if (!timer_pending(timer))
>>   		mod_timer(timer, tcmu_cmd->deadline);
>>   
>> -	return 0;
>> +	pr_debug("Timeout set up for cmd %p, dev = %s, tmo = %lu\n", tcmu_cmd,
>> +		 tcmu_cmd->tcmu_dev->name, tmo / MSEC_PER_SEC);
>>   }
>>   
>>   static int add_to_qfull_queue(struct tcmu_cmd *tcmu_cmd)
>>   {
>>   	struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
>>   	unsigned int tmo;
>> -	int ret;
>>   
>>   	/*
>>   	 * For backwards compat if qfull_time_out is not set use
>> @@ -931,13 +914,11 @@ static int add_to_qfull_queue(struct tcmu_cmd *tcmu_cmd)
>>   	else
>>   		tmo = TCMU_TIME_OUT;
>>   
>> -	ret = tcmu_setup_cmd_timer(tcmu_cmd, tmo, &udev->qfull_timer);
>> -	if (ret)
>> -		return ret;
>> +	tcmu_setup_cmd_timer(tcmu_cmd, tmo, &udev->qfull_timer);
>>   
>>   	list_add_tail(&tcmu_cmd->queue_entry, &udev->qfull_queue);
>> -	pr_debug("adding cmd %u on dev %s to ring space wait queue\n",
>> -		 tcmu_cmd->cmd_id, udev->name);
>> +	pr_debug("adding cmd %p on dev %s to ring space wait queue\n",
>> +		 tcmu_cmd, udev->name);
>>   	return 0;
>>   }
>>   
>> @@ -959,7 +940,7 @@ static int queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, sense_reason_t *scsi_err)
>>   	struct tcmu_mailbox *mb;
>>   	struct tcmu_cmd_entry *entry;
>>   	struct iovec *iov;
>> -	int iov_cnt, ret;
>> +	int iov_cnt, cmd_id;
>>   	uint32_t cmd_head;
>>   	uint64_t cdb_off;
>>   	bool copy_to_data_area;
>> @@ -1060,14 +1041,21 @@ static int queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, sense_reason_t *scsi_err)
>>   	}
>>   	entry->req.iov_bidi_cnt = iov_cnt;
>>   
>> -	ret = tcmu_setup_cmd_timer(tcmu_cmd, udev->cmd_time_out,
>> -				   &udev->cmd_timer);
>> -	if (ret) {
>> -		tcmu_cmd_free_data(tcmu_cmd, tcmu_cmd->dbi_cnt);
>> +	cmd_id = idr_alloc(&udev->commands, tcmu_cmd, 1, USHRT_MAX, GFP_NOWAIT);
>> +	if (cmd_id < 0) {
>> +		pr_err("tcmu: Could not allocate cmd id.\n");
>>   
>> +		tcmu_cmd_free_data(tcmu_cmd, tcmu_cmd->dbi_cnt);
>>   		*scsi_err = TCM_OUT_OF_RESOURCES;
>>   		return -1;
>>   	}
>> +	tcmu_cmd->cmd_id = cmd_id;
>> +
>> +	pr_debug("allocated cmd id %u for cmd %p dev %s\n", tcmu_cmd->cmd_id,
>> +		 tcmu_cmd, udev->name);
>> +
>> +	tcmu_setup_cmd_timer(tcmu_cmd, udev->cmd_time_out, &udev->cmd_timer);
>> +
>>   	entry->hdr.cmd_id = tcmu_cmd->cmd_id;
>>   
>>   	/*
>> @@ -1279,50 +1267,39 @@ static unsigned int tcmu_handle_completions(struct tcmu_dev *udev)
>>   	return handled;
>>   }
>>   
>> -static int tcmu_check_expired_cmd(int id, void *p, void *data)
>> +static void tcmu_check_expired_ring_cmd(struct tcmu_cmd *cmd)
>>   {
>> -	struct tcmu_cmd *cmd = p;
>> -	struct tcmu_dev *udev = cmd->tcmu_dev;
>> -	u8 scsi_status;
>>   	struct se_cmd *se_cmd;
>> -	bool is_running;
>> -
>> -	if (test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags))
>> -		return 0;
>>   
>>   	if (!time_after(jiffies, cmd->deadline))
>> -		return 0;
>> +		return;
>>   
>> -	is_running = test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags);
>> +	set_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags);
>> +	list_del_init(&cmd->queue_entry);
>>   	se_cmd = cmd->se_cmd;
>> +	cmd->se_cmd = NULL;
>>   
>> -	if (is_running) {
>> -		/*
>> -		 * If cmd_time_out is disabled but qfull is set deadline
>> -		 * will only reflect the qfull timeout. Ignore it.
>> -		 */
>> -		if (!udev->cmd_time_out)
>> -			return 0;
>> +	pr_debug("Timing out inflight cmd %u on dev %s.\n",
>> +		 cmd->cmd_id, cmd->tcmu_dev->name);
>>   
>> -		set_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags);
>> -		/*
>> -		 * target_complete_cmd will translate this to LUN COMM FAILURE
>> -		 */
>> -		scsi_status = SAM_STAT_CHECK_CONDITION;
>> -		list_del_init(&cmd->queue_entry);
>> -		cmd->se_cmd = NULL;
>> -	} else {
>> -		list_del_init(&cmd->queue_entry);
>> -		idr_remove(&udev->commands, id);
>> -		tcmu_free_cmd(cmd);
>> -		scsi_status = SAM_STAT_TASK_SET_FULL;
>> -	}
>> +	target_complete_cmd(se_cmd, SAM_STAT_CHECK_CONDITION);
>> +}
>>   
>> -	pr_debug("Timing out cmd %u on dev %s that is %s.\n",
>> -		 id, udev->name, is_running ? "inflight" : "queued");
>> +static void tcmu_check_expired_queue_cmd(struct tcmu_cmd *cmd)
>> +{
>> +	struct se_cmd *se_cmd;
>>   
>> -	target_complete_cmd(se_cmd, scsi_status);
>> -	return 0;
>> +	if (!time_after(jiffies, cmd->deadline))
>> +		return;
>> +
>> +	list_del_init(&cmd->queue_entry);
>> +	se_cmd = cmd->se_cmd;
>> +	tcmu_free_cmd(cmd);
>> +
>> +	pr_debug("Timing out queued cmd %p on dev %s.\n",
>> +		  cmd, cmd->tcmu_dev->name);
>> +
>> +	target_complete_cmd(se_cmd, SAM_STAT_TASK_SET_FULL);
>>   }
>>   
>>   static void tcmu_device_timedout(struct tcmu_dev *udev)
>> @@ -1407,16 +1384,15 @@ static struct se_device *tcmu_alloc_device(struct se_hba *hba, const char *name)
>>   	return &udev->se_dev;
>>   }
>>   
>> -static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
>> +static void run_qfull_queue(struct tcmu_dev *udev, bool fail)
>>   {
>>   	struct tcmu_cmd *tcmu_cmd, *tmp_cmd;
>>   	LIST_HEAD(cmds);
>> -	bool drained = true;
>>   	sense_reason_t scsi_ret;
>>   	int ret;
>>   
>>   	if (list_empty(&udev->qfull_queue))
>> -		return true;
>> +		return;
>>   
>>   	pr_debug("running %s's cmdr queue forcefail %d\n", udev->name, fail);
>>   
>> @@ -1425,11 +1401,10 @@ static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
>>   	list_for_each_entry_safe(tcmu_cmd, tmp_cmd, &cmds, queue_entry) {
>>   		list_del_init(&tcmu_cmd->queue_entry);
>>   
>> -	        pr_debug("removing cmd %u on dev %s from queue\n",
>> -		         tcmu_cmd->cmd_id, udev->name);
>> +	        pr_debug("removing cmd %p on dev %s from queue\n",
>> +		         tcmu_cmd, udev->name);
>>   
>>   		if (fail) {
>> -			idr_remove(&udev->commands, tcmu_cmd->cmd_id);
>>   			/*
>>   			 * We were not able to even start the command, so
>>   			 * fail with busy to allow a retry in case runner
>> @@ -1444,10 +1419,8 @@ static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
>>   
>>   		ret = queue_cmd_ring(tcmu_cmd, &scsi_ret);
>>   		if (ret < 0) {
>> -		        pr_debug("cmd %u on dev %s failed with %u\n",
>> -			         tcmu_cmd->cmd_id, udev->name, scsi_ret);
>> -
>> -			idr_remove(&udev->commands, tcmu_cmd->cmd_id);
>> +		        pr_debug("cmd %p on dev %s failed with %u\n",
>> +			         tcmu_cmd, udev->name, scsi_ret);
>>   			/*
>>   			 * Ignore scsi_ret for now. target_complete_cmd
>>   			 * drops it.
>> @@ -1462,13 +1435,11 @@ static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
>>   			 * the queue
>>   			 */
>>   			list_splice_tail(&cmds, &udev->qfull_queue);
>> -			drained = false;
>>   			break;
>>   		}
>>   	}
>>   
>>   	tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer);
>> -	return drained;
>>   }
>>   
>>   static int tcmu_irqcontrol(struct uio_info *info, s32 irq_on)
>> @@ -1652,6 +1623,8 @@ static void tcmu_dev_kref_release(struct kref *kref)
>>   		if (tcmu_check_and_free_pending_cmd(cmd) != 0)
>>   			all_expired = false;
>>   	}
>> +	if (!list_empty(&udev->qfull_queue))
>> +		all_expired = false;
> 
> Just drop.

This patch should just fix a bug, but not change behavior of
tcmu. Therefore I think the two additional lines are necessary
to keep a consistent behavior.

> It's never going to happen and doesn't tell us a lot. We
> don't know what list it was on and why LIO core's refcounts got messed up
I agree.

> 
> The WARN_ON should just be removed in some other patchset one day. If we
> wanted to detect if non timed out commands were running at this time
> then it should go in lio core, because it would be a problem that
> affects all drivers.

I agree.

In this case we can remove or change the additional two lines again
together with the other stuff.

> 
> I think we actually want the opposite and should WARN or BUG if commands
> have timed out and we are freeing the device.

Better not BUG, but WARN but be ok, I think.

> In this case userspace
> could still be accessing the buffers we are going to free. It probably

No. Userspace can access the uio device only, if it has mapped the
uio space. But in that case userspace holds a ref_cnt of the tcmu_dev
and therefore tcmu_dev_kref_release cannot happen at that time.

> won't happen now that we run this from the release function and so the
> device is closed though.
> 
> 
> 
>>   	idr_destroy(&udev->commands);
>>   	WARN_ON(!all_expired);
>>   
>> @@ -2037,9 +2010,6 @@ static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
>>   	mutex_lock(&udev->cmdr_lock);
>>   
>>   	idr_for_each_entry(&udev->commands, cmd, i) {
>> -		if (!test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags))
>> -			continue;
>> -
>>   		pr_debug("removing cmd %u on dev %s from ring (is expired %d)\n",
>>   			  cmd->cmd_id, udev->name,
>>   			  test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags));
>> @@ -2077,6 +2047,8 @@ static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
>>   
>>   	del_timer(&udev->cmd_timer);
>>   
>> +	run_qfull_queue(udev, false);
>> +
>>   	mutex_unlock(&udev->cmdr_lock);
>>   }
>>   
>> @@ -2698,6 +2670,7 @@ static void find_free_blocks(void)
>>   static void check_timedout_devices(void)
>>   {
>>   	struct tcmu_dev *udev, *tmp_dev;
>> +	struct tcmu_cmd *ucmd, *tmp_cmd;
> 
> ucmd probably would have been best when the code was first written. Now,
> you should use tcmu_cmd or cmd to match the existing naming style.

Ok. I'll change it.

> 
> 
>>   	LIST_HEAD(devs);
>>   
>>   	spin_lock_bh(&timed_out_udevs_lock);
>> @@ -2708,9 +2681,24 @@ static void check_timedout_devices(void)
>>   		spin_unlock_bh(&timed_out_udevs_lock);
>>   
>>   		mutex_lock(&udev->cmdr_lock);
>> -		idr_for_each(&udev->commands, tcmu_check_expired_cmd, NULL);
>>   
>> -		tcmu_set_next_deadline(&udev->inflight_queue, &udev->cmd_timer);
>> +		/*
>> +		 * If cmd_time_out is disabled but qfull is set deadline
>> +		 * will only reflect the qfull timeout. Ignore it.
>> +		 */
>> +		if (udev->cmd_time_out) {
>> +			list_for_each_entry_safe(ucmd, tmp_cmd,
>> +			                         &udev->inflight_queue,
>> +			                         queue_entry) {
>> +				tcmu_check_expired_ring_cmd(ucmd);
>> +			}
>> +			tcmu_set_next_deadline(&udev->inflight_queue,
>> +			                       &udev->cmd_timer);
>> +		}
>> +		list_for_each_entry_safe(ucmd, tmp_cmd, &udev->qfull_queue,
>> +			                 queue_entry) {
>> +			tcmu_check_expired_queue_cmd(ucmd);
>> +		}
>>   		tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer);
>>   
>>   		mutex_unlock(&udev->cmdr_lock);
>>
>
diff mbox series

Patch

diff --git a/drivers/target/target_core_user.c b/drivers/target/target_core_user.c
index f769bb1e3735..de75583b74c6 100644
--- a/drivers/target/target_core_user.c
+++ b/drivers/target/target_core_user.c
@@ -882,41 +882,24 @@  static inline size_t tcmu_cmd_get_cmd_size(struct tcmu_cmd *tcmu_cmd,
 	return command_size;
 }
 
-static int tcmu_setup_cmd_timer(struct tcmu_cmd *tcmu_cmd, unsigned int tmo,
-				struct timer_list *timer)
+static void tcmu_setup_cmd_timer(struct tcmu_cmd *tcmu_cmd, unsigned int tmo,
+				 struct timer_list *timer)
 {
-	struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
-	int cmd_id;
-
-	if (tcmu_cmd->cmd_id)
-		goto setup_timer;
-
-	cmd_id = idr_alloc(&udev->commands, tcmu_cmd, 1, USHRT_MAX, GFP_NOWAIT);
-	if (cmd_id < 0) {
-		pr_err("tcmu: Could not allocate cmd id.\n");
-		return cmd_id;
-	}
-	tcmu_cmd->cmd_id = cmd_id;
-
-	pr_debug("allocated cmd %u for dev %s tmo %lu\n", tcmu_cmd->cmd_id,
-		 udev->name, tmo / MSEC_PER_SEC);
-
-setup_timer:
 	if (!tmo)
-		return 0;
+		return;
 
 	tcmu_cmd->deadline = round_jiffies_up(jiffies + msecs_to_jiffies(tmo));
 	if (!timer_pending(timer))
 		mod_timer(timer, tcmu_cmd->deadline);
 
-	return 0;
+	pr_debug("Timeout set up for cmd %p, dev = %s, tmo = %lu\n", tcmu_cmd,
+		 tcmu_cmd->tcmu_dev->name, tmo / MSEC_PER_SEC);
 }
 
 static int add_to_qfull_queue(struct tcmu_cmd *tcmu_cmd)
 {
 	struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
 	unsigned int tmo;
-	int ret;
 
 	/*
 	 * For backwards compat if qfull_time_out is not set use
@@ -931,13 +914,11 @@  static int add_to_qfull_queue(struct tcmu_cmd *tcmu_cmd)
 	else
 		tmo = TCMU_TIME_OUT;
 
-	ret = tcmu_setup_cmd_timer(tcmu_cmd, tmo, &udev->qfull_timer);
-	if (ret)
-		return ret;
+	tcmu_setup_cmd_timer(tcmu_cmd, tmo, &udev->qfull_timer);
 
 	list_add_tail(&tcmu_cmd->queue_entry, &udev->qfull_queue);
-	pr_debug("adding cmd %u on dev %s to ring space wait queue\n",
-		 tcmu_cmd->cmd_id, udev->name);
+	pr_debug("adding cmd %p on dev %s to ring space wait queue\n",
+		 tcmu_cmd, udev->name);
 	return 0;
 }
 
@@ -959,7 +940,7 @@  static int queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, sense_reason_t *scsi_err)
 	struct tcmu_mailbox *mb;
 	struct tcmu_cmd_entry *entry;
 	struct iovec *iov;
-	int iov_cnt, ret;
+	int iov_cnt, cmd_id;
 	uint32_t cmd_head;
 	uint64_t cdb_off;
 	bool copy_to_data_area;
@@ -1060,14 +1041,21 @@  static int queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, sense_reason_t *scsi_err)
 	}
 	entry->req.iov_bidi_cnt = iov_cnt;
 
-	ret = tcmu_setup_cmd_timer(tcmu_cmd, udev->cmd_time_out,
-				   &udev->cmd_timer);
-	if (ret) {
-		tcmu_cmd_free_data(tcmu_cmd, tcmu_cmd->dbi_cnt);
+	cmd_id = idr_alloc(&udev->commands, tcmu_cmd, 1, USHRT_MAX, GFP_NOWAIT);
+	if (cmd_id < 0) {
+		pr_err("tcmu: Could not allocate cmd id.\n");
 
+		tcmu_cmd_free_data(tcmu_cmd, tcmu_cmd->dbi_cnt);
 		*scsi_err = TCM_OUT_OF_RESOURCES;
 		return -1;
 	}
+	tcmu_cmd->cmd_id = cmd_id;
+
+	pr_debug("allocated cmd id %u for cmd %p dev %s\n", tcmu_cmd->cmd_id,
+		 tcmu_cmd, udev->name);
+
+	tcmu_setup_cmd_timer(tcmu_cmd, udev->cmd_time_out, &udev->cmd_timer);
+
 	entry->hdr.cmd_id = tcmu_cmd->cmd_id;
 
 	/*
@@ -1279,50 +1267,39 @@  static unsigned int tcmu_handle_completions(struct tcmu_dev *udev)
 	return handled;
 }
 
-static int tcmu_check_expired_cmd(int id, void *p, void *data)
+static void tcmu_check_expired_ring_cmd(struct tcmu_cmd *cmd)
 {
-	struct tcmu_cmd *cmd = p;
-	struct tcmu_dev *udev = cmd->tcmu_dev;
-	u8 scsi_status;
 	struct se_cmd *se_cmd;
-	bool is_running;
-
-	if (test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags))
-		return 0;
 
 	if (!time_after(jiffies, cmd->deadline))
-		return 0;
+		return;
 
-	is_running = test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags);
+	set_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags);
+	list_del_init(&cmd->queue_entry);
 	se_cmd = cmd->se_cmd;
+	cmd->se_cmd = NULL;
 
-	if (is_running) {
-		/*
-		 * If cmd_time_out is disabled but qfull is set deadline
-		 * will only reflect the qfull timeout. Ignore it.
-		 */
-		if (!udev->cmd_time_out)
-			return 0;
+	pr_debug("Timing out inflight cmd %u on dev %s.\n",
+		 cmd->cmd_id, cmd->tcmu_dev->name);
 
-		set_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags);
-		/*
-		 * target_complete_cmd will translate this to LUN COMM FAILURE
-		 */
-		scsi_status = SAM_STAT_CHECK_CONDITION;
-		list_del_init(&cmd->queue_entry);
-		cmd->se_cmd = NULL;
-	} else {
-		list_del_init(&cmd->queue_entry);
-		idr_remove(&udev->commands, id);
-		tcmu_free_cmd(cmd);
-		scsi_status = SAM_STAT_TASK_SET_FULL;
-	}
+	target_complete_cmd(se_cmd, SAM_STAT_CHECK_CONDITION);
+}
 
-	pr_debug("Timing out cmd %u on dev %s that is %s.\n",
-		 id, udev->name, is_running ? "inflight" : "queued");
+static void tcmu_check_expired_queue_cmd(struct tcmu_cmd *cmd)
+{
+	struct se_cmd *se_cmd;
 
-	target_complete_cmd(se_cmd, scsi_status);
-	return 0;
+	if (!time_after(jiffies, cmd->deadline))
+		return;
+
+	list_del_init(&cmd->queue_entry);
+	se_cmd = cmd->se_cmd;
+	tcmu_free_cmd(cmd);
+
+	pr_debug("Timing out queued cmd %p on dev %s.\n",
+		  cmd, cmd->tcmu_dev->name);
+
+	target_complete_cmd(se_cmd, SAM_STAT_TASK_SET_FULL);
 }
 
 static void tcmu_device_timedout(struct tcmu_dev *udev)
@@ -1407,16 +1384,15 @@  static struct se_device *tcmu_alloc_device(struct se_hba *hba, const char *name)
 	return &udev->se_dev;
 }
 
-static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
+static void run_qfull_queue(struct tcmu_dev *udev, bool fail)
 {
 	struct tcmu_cmd *tcmu_cmd, *tmp_cmd;
 	LIST_HEAD(cmds);
-	bool drained = true;
 	sense_reason_t scsi_ret;
 	int ret;
 
 	if (list_empty(&udev->qfull_queue))
-		return true;
+		return;
 
 	pr_debug("running %s's cmdr queue forcefail %d\n", udev->name, fail);
 
@@ -1425,11 +1401,10 @@  static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
 	list_for_each_entry_safe(tcmu_cmd, tmp_cmd, &cmds, queue_entry) {
 		list_del_init(&tcmu_cmd->queue_entry);
 
-	        pr_debug("removing cmd %u on dev %s from queue\n",
-		         tcmu_cmd->cmd_id, udev->name);
+	        pr_debug("removing cmd %p on dev %s from queue\n",
+		         tcmu_cmd, udev->name);
 
 		if (fail) {
-			idr_remove(&udev->commands, tcmu_cmd->cmd_id);
 			/*
 			 * We were not able to even start the command, so
 			 * fail with busy to allow a retry in case runner
@@ -1444,10 +1419,8 @@  static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
 
 		ret = queue_cmd_ring(tcmu_cmd, &scsi_ret);
 		if (ret < 0) {
-		        pr_debug("cmd %u on dev %s failed with %u\n",
-			         tcmu_cmd->cmd_id, udev->name, scsi_ret);
-
-			idr_remove(&udev->commands, tcmu_cmd->cmd_id);
+		        pr_debug("cmd %p on dev %s failed with %u\n",
+			         tcmu_cmd, udev->name, scsi_ret);
 			/*
 			 * Ignore scsi_ret for now. target_complete_cmd
 			 * drops it.
@@ -1462,13 +1435,11 @@  static bool run_qfull_queue(struct tcmu_dev *udev, bool fail)
 			 * the queue
 			 */
 			list_splice_tail(&cmds, &udev->qfull_queue);
-			drained = false;
 			break;
 		}
 	}
 
 	tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer);
-	return drained;
 }
 
 static int tcmu_irqcontrol(struct uio_info *info, s32 irq_on)
@@ -1652,6 +1623,8 @@  static void tcmu_dev_kref_release(struct kref *kref)
 		if (tcmu_check_and_free_pending_cmd(cmd) != 0)
 			all_expired = false;
 	}
+	if (!list_empty(&udev->qfull_queue))
+		all_expired = false;
 	idr_destroy(&udev->commands);
 	WARN_ON(!all_expired);
 
@@ -2037,9 +2010,6 @@  static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
 	mutex_lock(&udev->cmdr_lock);
 
 	idr_for_each_entry(&udev->commands, cmd, i) {
-		if (!test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags))
-			continue;
-
 		pr_debug("removing cmd %u on dev %s from ring (is expired %d)\n",
 			  cmd->cmd_id, udev->name,
 			  test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags));
@@ -2077,6 +2047,8 @@  static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
 
 	del_timer(&udev->cmd_timer);
 
+	run_qfull_queue(udev, false);
+
 	mutex_unlock(&udev->cmdr_lock);
 }
 
@@ -2698,6 +2670,7 @@  static void find_free_blocks(void)
 static void check_timedout_devices(void)
 {
 	struct tcmu_dev *udev, *tmp_dev;
+	struct tcmu_cmd *ucmd, *tmp_cmd;
 	LIST_HEAD(devs);
 
 	spin_lock_bh(&timed_out_udevs_lock);
@@ -2708,9 +2681,24 @@  static void check_timedout_devices(void)
 		spin_unlock_bh(&timed_out_udevs_lock);
 
 		mutex_lock(&udev->cmdr_lock);
-		idr_for_each(&udev->commands, tcmu_check_expired_cmd, NULL);
 
-		tcmu_set_next_deadline(&udev->inflight_queue, &udev->cmd_timer);
+		/*
+		 * If cmd_time_out is disabled but qfull is set deadline
+		 * will only reflect the qfull timeout. Ignore it.
+		 */
+		if (udev->cmd_time_out) {
+			list_for_each_entry_safe(ucmd, tmp_cmd,
+			                         &udev->inflight_queue,
+			                         queue_entry) {
+				tcmu_check_expired_ring_cmd(ucmd);
+			}
+			tcmu_set_next_deadline(&udev->inflight_queue,
+			                       &udev->cmd_timer);
+		}
+		list_for_each_entry_safe(ucmd, tmp_cmd, &udev->qfull_queue,
+			                 queue_entry) {
+			tcmu_check_expired_queue_cmd(ucmd);
+		}
 		tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer);
 
 		mutex_unlock(&udev->cmdr_lock);