diff mbox series

[1/3] target: core: Send mtl in blocks

Message ID 20220630122241.1658-2-a.kovaleva@yadro.com (mailing list archive)
State Changes Requested
Headers show
Series Make target send correct io limits | expand

Commit Message

Anastasia Kovaleva June 30, 2022, 12:22 p.m. UTC
A MAXIMUM TRANSFER LENGTH value indicates the maximum transfer length in
logical blocks that the device server accepts for a single command. Fix
function sending the length in sectors instead of blocks.

This patch also removes the special casing for fileio in
block_size_store since this logic in now unified in
spc_emulate_evpd_b0() for all backends.

Signed-off-by: Anastasia Kovaleva <a.kovaleva@yadro.com>
Reviewed-by: Konstantin Shelekhin <k.shelekhin@yadro.com>
Reviewed-by: Dmitriy Bogdanov <d.bogdanov@yadro.com>
---
 drivers/target/target_core_configfs.c | 2 --
 drivers/target/target_core_file.c     | 1 -
 drivers/target/target_core_spc.c      | 7 ++++++-
 include/target/target_core_base.h     | 1 -
 4 files changed, 6 insertions(+), 5 deletions(-)

Comments

Mike Christie July 7, 2022, 9:34 p.m. UTC | #1
On 6/30/22 7:22 AM, Anastasia Kovaleva wrote:
> --- a/drivers/target/target_core_spc.c
> +++ b/drivers/target/target_core_spc.c
> @@ -513,6 +513,7 @@ spc_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
>  	struct se_device *dev = cmd->se_dev;
>  	u32 mtl = 0;
>  	int have_tp = 0, opt, min;
> +	u32 io_max_blocks;
>  
>  	/*
>  	 * Following spc3r22 section 6.5.3 Block Limits VPD page, when
> @@ -547,11 +548,15 @@ spc_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
>  	 * XXX: Currently assumes single PAGE_SIZE per scatterlist for fabrics
>  	 * enforcing maximum HW scatter-gather-list entry limit
>  	 */
> +	io_max_blocks = mult_frac(
> +			dev->dev_attrib.hw_max_sectors,
> +			dev->dev_attrib.hw_block_size,
> +			dev->dev_attrib.block_size);
Just some trivial comments.

This chunk might be better after the max_data_sg_nents chunk below because
the comment above is for the max_data_sg_nents chunk.

Also, fix up the tabbing/coding style like:

	io_max_blocks = mult_frac(dev->dev_attrib.hw_max_sectors,
				  dev->dev_attrib.hw_block_size,
				  dev->dev_attrib.block_size);



>  	if (cmd->se_tfo->max_data_sg_nents) {
>  		mtl = (cmd->se_tfo->max_data_sg_nents * PAGE_SIZE) /
>  		       dev->dev_attrib.block_size;
>  	}
> -	put_unaligned_be32(min_not_zero(mtl, dev->dev_attrib.hw_max_sectors), &buf[8]);
> +	put_unaligned_be32(min_not_zero(mtl, io_max_blocks), &buf[8]);
>
diff mbox series

Patch

diff --git a/drivers/target/target_core_configfs.c b/drivers/target/target_core_configfs.c
index bbcbbfa72b07..51943369887a 100644
--- a/drivers/target/target_core_configfs.c
+++ b/drivers/target/target_core_configfs.c
@@ -1089,8 +1089,6 @@  static ssize_t block_size_store(struct config_item *item,
 	}
 
 	da->block_size = val;
-	if (da->max_bytes_per_io)
-		da->hw_max_sectors = da->max_bytes_per_io / val;
 
 	pr_debug("dev[%p]: SE Device block_size changed to %u\n",
 			da->da_dev, val);
diff --git a/drivers/target/target_core_file.c b/drivers/target/target_core_file.c
index e68f1cc8ef98..dcf4c8e47e8b 100644
--- a/drivers/target/target_core_file.c
+++ b/drivers/target/target_core_file.c
@@ -189,7 +189,6 @@  static int fd_configure_device(struct se_device *dev)
 	}
 
 	dev->dev_attrib.hw_block_size = fd_dev->fd_block_size;
-	dev->dev_attrib.max_bytes_per_io = FD_MAX_BYTES;
 	dev->dev_attrib.hw_max_sectors = FD_MAX_BYTES / fd_dev->fd_block_size;
 	dev->dev_attrib.hw_queue_depth = FD_MAX_DEVICE_QUEUE_DEPTH;
 
diff --git a/drivers/target/target_core_spc.c b/drivers/target/target_core_spc.c
index c14441c89bed..c3d90e811617 100644
--- a/drivers/target/target_core_spc.c
+++ b/drivers/target/target_core_spc.c
@@ -513,6 +513,7 @@  spc_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
 	struct se_device *dev = cmd->se_dev;
 	u32 mtl = 0;
 	int have_tp = 0, opt, min;
+	u32 io_max_blocks;
 
 	/*
 	 * Following spc3r22 section 6.5.3 Block Limits VPD page, when
@@ -547,11 +548,15 @@  spc_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
 	 * XXX: Currently assumes single PAGE_SIZE per scatterlist for fabrics
 	 * enforcing maximum HW scatter-gather-list entry limit
 	 */
+	io_max_blocks = mult_frac(
+			dev->dev_attrib.hw_max_sectors,
+			dev->dev_attrib.hw_block_size,
+			dev->dev_attrib.block_size);
 	if (cmd->se_tfo->max_data_sg_nents) {
 		mtl = (cmd->se_tfo->max_data_sg_nents * PAGE_SIZE) /
 		       dev->dev_attrib.block_size;
 	}
-	put_unaligned_be32(min_not_zero(mtl, dev->dev_attrib.hw_max_sectors), &buf[8]);
+	put_unaligned_be32(min_not_zero(mtl, io_max_blocks), &buf[8]);
 
 	/*
 	 * Set OPTIMAL TRANSFER LENGTH
diff --git a/include/target/target_core_base.h b/include/target/target_core_base.h
index c2b36f7d917d..748d1afbadaa 100644
--- a/include/target/target_core_base.h
+++ b/include/target/target_core_base.h
@@ -709,7 +709,6 @@  struct se_dev_attrib {
 	u32		unmap_granularity;
 	u32		unmap_granularity_alignment;
 	u32		max_write_same_len;
-	u32		max_bytes_per_io;
 	struct se_device *da_dev;
 	struct config_group da_group;
 };