Message ID | 20160607081810.6640-3-vigneshr@ti.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On 06/07/16 11:18, Vignesh R wrote: > Use mem-to-mem DMA to read from flash when reading in mmap mode. This > gives improved read performance and reduces CPU load. > > With this patch the raw-read throughput is ~16MB/s on DRA74 EVM. And CPU > load is <20%. UBIFS read ~13 MB/s. > > Signed-off-by: Vignesh R <vigneshr@ti.com> > --- > > v3: Cleanup code based on review comments for v2. > v2: Handle kmap'd buffers of JFFS2 FS. > > drivers/spi/spi-ti-qspi.c | 189 ++++++++++++++++++++++++++++++++++++++++++---- > 1 file changed, 176 insertions(+), 13 deletions(-) > > diff --git a/drivers/spi/spi-ti-qspi.c b/drivers/spi/spi-ti-qspi.c > index 29ea8d2f9824..1f6c59c29157 100644 > --- a/drivers/spi/spi-ti-qspi.c > +++ b/drivers/spi/spi-ti-qspi.c > @@ -33,6 +33,7 @@ > #include <linux/pinctrl/consumer.h> > #include <linux/mfd/syscon.h> > #include <linux/regmap.h> > +#include <linux/highmem.h> > > #include <linux/spi/spi.h> > > @@ -41,6 +42,8 @@ struct ti_qspi_regs { > }; > > struct ti_qspi { > + struct completion transfer_complete; > + > /* list synchronization */ > struct mutex list_lock; > > @@ -54,6 +57,9 @@ struct ti_qspi { > > struct ti_qspi_regs ctx_reg; > > + dma_addr_t mmap_phys_base; > + struct dma_chan *rx_chan; > + > u32 spi_max_frequency; > u32 cmd; > u32 dc; > @@ -379,6 +385,72 @@ static int qspi_transfer_msg(struct ti_qspi *qspi, struct spi_transfer *t, > return 0; > } > > +static void ti_qspi_dma_callback(void *param) > +{ > + struct ti_qspi *qspi = param; > + > + complete(&qspi->transfer_complete); > +} > + > +static int ti_qspi_dma_xfer(struct ti_qspi *qspi, dma_addr_t dma_dst, > + dma_addr_t dma_src, size_t len) > +{ > + struct dma_chan *chan = qspi->rx_chan; > + struct dma_device *dma_dev = chan->device; > + dma_cookie_t cookie; > + enum dma_ctrl_flags flags = DMA_CTRL_ACK | DMA_PREP_INTERRUPT; > + struct dma_async_tx_descriptor *tx; > + int ret; > + > + tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src, > + len, flags); > + if (!tx) { > + dev_err(qspi->dev, "device_prep_dma_memcpy error\n"); > + return -EIO; > + } > + > + tx->callback = ti_qspi_dma_callback; > + tx->callback_param = qspi; > + cookie = tx->tx_submit(tx); > + > + ret = dma_submit_error(cookie); > + if (ret) { > + dev_err(qspi->dev, "dma_submit_error %d\n", cookie); > + return -EIO; > + } > + > + dma_async_issue_pending(chan); > + ret = wait_for_completion_timeout(&qspi->transfer_complete, > + msecs_to_jiffies(len)); > + if (ret <= 0) { > + dmaengine_terminate_sync(chan); > + dev_err(qspi->dev, "DMA wait_for_completion_timeout\n"); > + return -ETIMEDOUT; > + } > + > + return 0; > +} > + > +static int ti_qspi_dma_xfer_sg(struct ti_qspi *qspi, struct sg_table rx_sg, > + loff_t from) > +{ > + struct scatterlist *sg; > + dma_addr_t dma_src = qspi->mmap_phys_base + from; > + dma_addr_t dma_dst; > + int i, len, ret; > + > + for_each_sg(rx_sg.sgl, sg, rx_sg.nents, i) { > + dma_dst = sg_dma_address(sg); > + len = sg_dma_len(sg); > + ret = ti_qspi_dma_xfer(qspi, dma_dst, dma_src, len); > + if (ret) > + return ret; > + dma_src += len; > + } > + > + return 0; > +} > + > static void ti_qspi_enable_memory_map(struct spi_device *spi) > { > struct ti_qspi *qspi = spi_master_get_devdata(spi->master); > @@ -426,7 +498,40 @@ static void ti_qspi_setup_mmap_read(struct spi_device *spi, > QSPI_SPI_SETUP_REG(spi->chip_select)); > } > > -static int ti_qspi_spi_flash_read(struct spi_device *spi, > +#ifdef CONFIG_HIGHMEM > +static int ti_qspi_map_buf(struct ti_qspi *qspi, void *buf, > + unsigned int len, struct sg_table *sgt) > +{ > + unsigned int max_seg_size = > + dma_get_max_seg_size(qspi->rx_chan->device->dev); > + unsigned int desc_len = min_t(int, max_seg_size, PAGE_SIZE); > + int sgs = DIV_ROUND_UP(len + offset_in_page(buf), desc_len); > + struct page *vm_page; > + size_t min; > + int i, ret; > + > + ret = sg_alloc_table(sgt, sgs, GFP_KERNEL); > + if (ret) > + return ret; > + > + for (i = 0; i < sgs; i++) { > + min = min_t(size_t, len, desc_len - > + offset_in_page(buf)); > + vm_page = kmap_to_page(buf); > + if (!vm_page) { > + sg_free_table(sgt); > + return -ENOMEM; > + } > + sg_set_page(&sgt->sgl[i], vm_page, min, > + offset_in_page(buf)); > + buf += min; > + len -= min; > + } > + return 0; > +} > +#endif > + > +static int ti_qspi_spi_flash_read(struct spi_device *spi, > struct spi_flash_read_message *msg) > { > struct ti_qspi *qspi = spi_master_get_devdata(spi->master); > @@ -437,9 +542,46 @@ static int ti_qspi_spi_flash_read(struct spi_device *spi, > if (!qspi->mmap_enabled) > ti_qspi_enable_memory_map(spi); > ti_qspi_setup_mmap_read(spi, msg); > - memcpy_fromio(msg->buf, qspi->mmap_base + msg->from, msg->len); > + > + if (qspi->rx_chan) { > + struct device *dev = qspi->rx_chan->device->dev; > + void *buf = msg->buf; > + struct sg_table sgt; > + > + if (msg->cur_msg_mapped) { > + ret = ti_qspi_dma_xfer_sg(qspi, msg->rx_sg, msg->from); > + if (ret) > + goto err_unlock; > +#ifdef CONFIG_HIGHMEM > + } else if ((unsigned long)buf >= PKMAP_BASE && > + (unsigned long)buf < (PKMAP_BASE + > + (LAST_PKMAP * PAGE_SIZE))) { > + /* Generate sg_table for kmap buffers */ > + ret = ti_qspi_map_buf(qspi, buf, msg->len, &sgt); > + if (ret) > + goto err_unlock; > + ret = dma_map_sg(dev, sgt.sgl, sgt.nents, > + DMA_FROM_DEVICE); > + if (!ret) { > + ret = -ENOMEM; > + goto err_unlock; > + } > + ret = ti_qspi_dma_xfer_sg(qspi, sgt, msg->from); > + dma_unmap_sg(dev, sgt.sgl, sgt.orig_nents, > + DMA_FROM_DEVICE); > + sg_free_table(&sgt); > +#endif > + } else { > + dev_err(qspi->dev, "Invalid address for DMA\n"); > + ret = -EIO; > + goto err_unlock; > + } > + } else { > + memcpy_fromio(msg->buf, qspi->mmap_base + msg->from, msg->len); > + } > msg->retlen = msg->len; > > +err_unlock: > mutex_unlock(&qspi->list_lock); > > return ret; > @@ -536,6 +678,7 @@ static int ti_qspi_probe(struct platform_device *pdev) > struct device_node *np = pdev->dev.of_node; > u32 max_freq; > int ret = 0, num_cs, irq; > + dma_cap_mask_t mask; > > master = spi_alloc_master(&pdev->dev, sizeof(*qspi)); > if (!master) > @@ -550,6 +693,7 @@ static int ti_qspi_probe(struct platform_device *pdev) > master->dev.of_node = pdev->dev.of_node; > master->bits_per_word_mask = SPI_BPW_MASK(32) | SPI_BPW_MASK(16) | > SPI_BPW_MASK(8); > + master->spi_flash_read = ti_qspi_spi_flash_read; > > if (!of_property_read_u32(np, "num-cs", &num_cs)) > master->num_chipselect = num_cs; > @@ -592,17 +736,6 @@ static int ti_qspi_probe(struct platform_device *pdev) > goto free_master; > } > > - if (res_mmap) { > - qspi->mmap_base = devm_ioremap_resource(&pdev->dev, > - res_mmap); > - master->spi_flash_read = ti_qspi_spi_flash_read; > - if (IS_ERR(qspi->mmap_base)) { > - dev_err(&pdev->dev, > - "falling back to PIO mode\n"); > - master->spi_flash_read = NULL; > - } > - } > - qspi->mmap_enabled = false; > > if (of_property_read_bool(np, "syscon-chipselects")) { > qspi->ctrl_base = > @@ -637,6 +770,33 @@ static int ti_qspi_probe(struct platform_device *pdev) > if (ret) > goto free_master; > > + dma_cap_zero(mask); > + dma_cap_set(DMA_MEMCPY, mask); > + > + qspi->rx_chan = dma_request_channel(mask, NULL, NULL); dma_request_channel is deprecated, please use the: dma_request_chan_by_mask() > + if (!qspi->rx_chan) { > + dev_err(qspi->dev, > + "No Rx DMA available, trying mmap mode\n"); > + ret = 0; > + goto no_dma; > + } > + master->dma_rx = qspi->rx_chan; > + init_completion(&qspi->transfer_complete); > + if (res_mmap) > + qspi->mmap_phys_base = (dma_addr_t)res_mmap->start; > + > +no_dma: > + if (!qspi->rx_chan && res_mmap) { > + qspi->mmap_base = devm_ioremap_resource(&pdev->dev, res_mmap); > + if (IS_ERR(qspi->mmap_base)) { > + dev_info(&pdev->dev, > + "mmap failed with error %ld using PIO mode\n", > + PTR_ERR(qspi->mmap_base)); > + qspi->mmap_base = NULL; > + master->spi_flash_read = NULL; > + } > + } > + qspi->mmap_enabled = false; > return 0; > > free_master: > @@ -656,6 +816,9 @@ static int ti_qspi_remove(struct platform_device *pdev) > pm_runtime_put_sync(&pdev->dev); > pm_runtime_disable(&pdev->dev); > > + if (qspi->rx_chan) > + dma_release_channel(qspi->rx_chan); > + > return 0; > } > >
Hi, [auto build test WARNING on spi/for-next] [also build test WARNING on v4.7-rc2 next-20160606] [if your patch is applied to the wrong git tree, please drop us a note to help improve the system] url: https://github.com/0day-ci/linux/commits/Vignesh-R/spi-Add-DMA-support-for-ti-qspi/20160607-162134 base: https://git.kernel.org/pub/scm/linux/kernel/git/broonie/spi for-next config: mn10300-allmodconfig (attached as .config) compiler: am33_2.0-linux-gcc (GCC) 4.9.0 reproduce: wget https://git.kernel.org/cgit/linux/kernel/git/wfg/lkp-tests.git/plain/sbin/make.cross -O ~/bin/make.cross chmod +x ~/bin/make.cross # save the attached .config to linux build tree make.cross ARCH=mn10300 All warnings (new ones prefixed by >>): drivers/spi/spi-ti-qspi.c: In function 'ti_qspi_spi_flash_read': >> drivers/spi/spi-ti-qspi.c:549:19: warning: unused variable 'sgt' [-Wunused-variable] struct sg_table sgt; ^ >> drivers/spi/spi-ti-qspi.c:548:9: warning: unused variable 'buf' [-Wunused-variable] void *buf = msg->buf; ^ >> drivers/spi/spi-ti-qspi.c:547:18: warning: unused variable 'dev' [-Wunused-variable] struct device *dev = qspi->rx_chan->device->dev; ^ vim +/sgt +549 drivers/spi/spi-ti-qspi.c 541 542 if (!qspi->mmap_enabled) 543 ti_qspi_enable_memory_map(spi); 544 ti_qspi_setup_mmap_read(spi, msg); 545 546 if (qspi->rx_chan) { > 547 struct device *dev = qspi->rx_chan->device->dev; > 548 void *buf = msg->buf; > 549 struct sg_table sgt; 550 551 if (msg->cur_msg_mapped) { 552 ret = ti_qspi_dma_xfer_sg(qspi, msg->rx_sg, msg->from); --- 0-DAY kernel test infrastructure Open Source Technology Center https://lists.01.org/pipermail/kbuild-all Intel Corporation
On Tuesday 07 June 2016 02:47 PM, Peter Ujfalusi wrote: [...] >> @@ -637,6 +770,33 @@ static int ti_qspi_probe(struct platform_device *pdev) >> if (ret) >> goto free_master; >> >> + dma_cap_zero(mask); >> + dma_cap_set(DMA_MEMCPY, mask); >> + >> + qspi->rx_chan = dma_request_channel(mask, NULL, NULL); > > dma_request_channel is deprecated, please use the: > dma_request_chan_by_mask() > Updated to use dma_request_chan_by_mask() in v4, Thanks.
diff --git a/drivers/spi/spi-ti-qspi.c b/drivers/spi/spi-ti-qspi.c index 29ea8d2f9824..1f6c59c29157 100644 --- a/drivers/spi/spi-ti-qspi.c +++ b/drivers/spi/spi-ti-qspi.c @@ -33,6 +33,7 @@ #include <linux/pinctrl/consumer.h> #include <linux/mfd/syscon.h> #include <linux/regmap.h> +#include <linux/highmem.h> #include <linux/spi/spi.h> @@ -41,6 +42,8 @@ struct ti_qspi_regs { }; struct ti_qspi { + struct completion transfer_complete; + /* list synchronization */ struct mutex list_lock; @@ -54,6 +57,9 @@ struct ti_qspi { struct ti_qspi_regs ctx_reg; + dma_addr_t mmap_phys_base; + struct dma_chan *rx_chan; + u32 spi_max_frequency; u32 cmd; u32 dc; @@ -379,6 +385,72 @@ static int qspi_transfer_msg(struct ti_qspi *qspi, struct spi_transfer *t, return 0; } +static void ti_qspi_dma_callback(void *param) +{ + struct ti_qspi *qspi = param; + + complete(&qspi->transfer_complete); +} + +static int ti_qspi_dma_xfer(struct ti_qspi *qspi, dma_addr_t dma_dst, + dma_addr_t dma_src, size_t len) +{ + struct dma_chan *chan = qspi->rx_chan; + struct dma_device *dma_dev = chan->device; + dma_cookie_t cookie; + enum dma_ctrl_flags flags = DMA_CTRL_ACK | DMA_PREP_INTERRUPT; + struct dma_async_tx_descriptor *tx; + int ret; + + tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src, + len, flags); + if (!tx) { + dev_err(qspi->dev, "device_prep_dma_memcpy error\n"); + return -EIO; + } + + tx->callback = ti_qspi_dma_callback; + tx->callback_param = qspi; + cookie = tx->tx_submit(tx); + + ret = dma_submit_error(cookie); + if (ret) { + dev_err(qspi->dev, "dma_submit_error %d\n", cookie); + return -EIO; + } + + dma_async_issue_pending(chan); + ret = wait_for_completion_timeout(&qspi->transfer_complete, + msecs_to_jiffies(len)); + if (ret <= 0) { + dmaengine_terminate_sync(chan); + dev_err(qspi->dev, "DMA wait_for_completion_timeout\n"); + return -ETIMEDOUT; + } + + return 0; +} + +static int ti_qspi_dma_xfer_sg(struct ti_qspi *qspi, struct sg_table rx_sg, + loff_t from) +{ + struct scatterlist *sg; + dma_addr_t dma_src = qspi->mmap_phys_base + from; + dma_addr_t dma_dst; + int i, len, ret; + + for_each_sg(rx_sg.sgl, sg, rx_sg.nents, i) { + dma_dst = sg_dma_address(sg); + len = sg_dma_len(sg); + ret = ti_qspi_dma_xfer(qspi, dma_dst, dma_src, len); + if (ret) + return ret; + dma_src += len; + } + + return 0; +} + static void ti_qspi_enable_memory_map(struct spi_device *spi) { struct ti_qspi *qspi = spi_master_get_devdata(spi->master); @@ -426,7 +498,40 @@ static void ti_qspi_setup_mmap_read(struct spi_device *spi, QSPI_SPI_SETUP_REG(spi->chip_select)); } -static int ti_qspi_spi_flash_read(struct spi_device *spi, +#ifdef CONFIG_HIGHMEM +static int ti_qspi_map_buf(struct ti_qspi *qspi, void *buf, + unsigned int len, struct sg_table *sgt) +{ + unsigned int max_seg_size = + dma_get_max_seg_size(qspi->rx_chan->device->dev); + unsigned int desc_len = min_t(int, max_seg_size, PAGE_SIZE); + int sgs = DIV_ROUND_UP(len + offset_in_page(buf), desc_len); + struct page *vm_page; + size_t min; + int i, ret; + + ret = sg_alloc_table(sgt, sgs, GFP_KERNEL); + if (ret) + return ret; + + for (i = 0; i < sgs; i++) { + min = min_t(size_t, len, desc_len - + offset_in_page(buf)); + vm_page = kmap_to_page(buf); + if (!vm_page) { + sg_free_table(sgt); + return -ENOMEM; + } + sg_set_page(&sgt->sgl[i], vm_page, min, + offset_in_page(buf)); + buf += min; + len -= min; + } + return 0; +} +#endif + +static int ti_qspi_spi_flash_read(struct spi_device *spi, struct spi_flash_read_message *msg) { struct ti_qspi *qspi = spi_master_get_devdata(spi->master); @@ -437,9 +542,46 @@ static int ti_qspi_spi_flash_read(struct spi_device *spi, if (!qspi->mmap_enabled) ti_qspi_enable_memory_map(spi); ti_qspi_setup_mmap_read(spi, msg); - memcpy_fromio(msg->buf, qspi->mmap_base + msg->from, msg->len); + + if (qspi->rx_chan) { + struct device *dev = qspi->rx_chan->device->dev; + void *buf = msg->buf; + struct sg_table sgt; + + if (msg->cur_msg_mapped) { + ret = ti_qspi_dma_xfer_sg(qspi, msg->rx_sg, msg->from); + if (ret) + goto err_unlock; +#ifdef CONFIG_HIGHMEM + } else if ((unsigned long)buf >= PKMAP_BASE && + (unsigned long)buf < (PKMAP_BASE + + (LAST_PKMAP * PAGE_SIZE))) { + /* Generate sg_table for kmap buffers */ + ret = ti_qspi_map_buf(qspi, buf, msg->len, &sgt); + if (ret) + goto err_unlock; + ret = dma_map_sg(dev, sgt.sgl, sgt.nents, + DMA_FROM_DEVICE); + if (!ret) { + ret = -ENOMEM; + goto err_unlock; + } + ret = ti_qspi_dma_xfer_sg(qspi, sgt, msg->from); + dma_unmap_sg(dev, sgt.sgl, sgt.orig_nents, + DMA_FROM_DEVICE); + sg_free_table(&sgt); +#endif + } else { + dev_err(qspi->dev, "Invalid address for DMA\n"); + ret = -EIO; + goto err_unlock; + } + } else { + memcpy_fromio(msg->buf, qspi->mmap_base + msg->from, msg->len); + } msg->retlen = msg->len; +err_unlock: mutex_unlock(&qspi->list_lock); return ret; @@ -536,6 +678,7 @@ static int ti_qspi_probe(struct platform_device *pdev) struct device_node *np = pdev->dev.of_node; u32 max_freq; int ret = 0, num_cs, irq; + dma_cap_mask_t mask; master = spi_alloc_master(&pdev->dev, sizeof(*qspi)); if (!master) @@ -550,6 +693,7 @@ static int ti_qspi_probe(struct platform_device *pdev) master->dev.of_node = pdev->dev.of_node; master->bits_per_word_mask = SPI_BPW_MASK(32) | SPI_BPW_MASK(16) | SPI_BPW_MASK(8); + master->spi_flash_read = ti_qspi_spi_flash_read; if (!of_property_read_u32(np, "num-cs", &num_cs)) master->num_chipselect = num_cs; @@ -592,17 +736,6 @@ static int ti_qspi_probe(struct platform_device *pdev) goto free_master; } - if (res_mmap) { - qspi->mmap_base = devm_ioremap_resource(&pdev->dev, - res_mmap); - master->spi_flash_read = ti_qspi_spi_flash_read; - if (IS_ERR(qspi->mmap_base)) { - dev_err(&pdev->dev, - "falling back to PIO mode\n"); - master->spi_flash_read = NULL; - } - } - qspi->mmap_enabled = false; if (of_property_read_bool(np, "syscon-chipselects")) { qspi->ctrl_base = @@ -637,6 +770,33 @@ static int ti_qspi_probe(struct platform_device *pdev) if (ret) goto free_master; + dma_cap_zero(mask); + dma_cap_set(DMA_MEMCPY, mask); + + qspi->rx_chan = dma_request_channel(mask, NULL, NULL); + if (!qspi->rx_chan) { + dev_err(qspi->dev, + "No Rx DMA available, trying mmap mode\n"); + ret = 0; + goto no_dma; + } + master->dma_rx = qspi->rx_chan; + init_completion(&qspi->transfer_complete); + if (res_mmap) + qspi->mmap_phys_base = (dma_addr_t)res_mmap->start; + +no_dma: + if (!qspi->rx_chan && res_mmap) { + qspi->mmap_base = devm_ioremap_resource(&pdev->dev, res_mmap); + if (IS_ERR(qspi->mmap_base)) { + dev_info(&pdev->dev, + "mmap failed with error %ld using PIO mode\n", + PTR_ERR(qspi->mmap_base)); + qspi->mmap_base = NULL; + master->spi_flash_read = NULL; + } + } + qspi->mmap_enabled = false; return 0; free_master: @@ -656,6 +816,9 @@ static int ti_qspi_remove(struct platform_device *pdev) pm_runtime_put_sync(&pdev->dev); pm_runtime_disable(&pdev->dev); + if (qspi->rx_chan) + dma_release_channel(qspi->rx_chan); + return 0; }
Use mem-to-mem DMA to read from flash when reading in mmap mode. This gives improved read performance and reduces CPU load. With this patch the raw-read throughput is ~16MB/s on DRA74 EVM. And CPU load is <20%. UBIFS read ~13 MB/s. Signed-off-by: Vignesh R <vigneshr@ti.com> --- v3: Cleanup code based on review comments for v2. v2: Handle kmap'd buffers of JFFS2 FS. drivers/spi/spi-ti-qspi.c | 189 ++++++++++++++++++++++++++++++++++++++++++---- 1 file changed, 176 insertions(+), 13 deletions(-)