Message ID | 1509715220-31885-6-git-send-email-adrian.hunter@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Fri, Nov 3, 2017 at 2:20 PM, Adrian Hunter <adrian.hunter@intel.com> wrote: > From: Venkat Gopalakrishnan <venkatg@codeaurora.org> > > This patch adds CMDQ support for command-queue compatible > hosts. > > Command queue is added in eMMC-5.1 specification. This > enables the controller to process upto 32 requests at > a time. > > Adrian Hunter contributed renaming to cqhci, recovery, suspend > and resume, cqhci_off, cqhci_wait_for_idle, and external timeout > handling. > > Signed-off-by: Asutosh Das <asutoshd@codeaurora.org> > Signed-off-by: Sujit Reddy Thumma <sthumma@codeaurora.org> > Signed-off-by: Konstantin Dorfman <kdorfman@codeaurora.org> > Signed-off-by: Venkat Gopalakrishnan <venkatg@codeaurora.org> > Signed-off-by: Subhash Jadavani <subhashj@codeaurora.org> > Signed-off-by: Ritesh Harjani <riteshh@codeaurora.org> > Signed-off-by: Adrian Hunter <adrian.hunter@intel.com> (...) > +config MMC_CQHCI > + tristate "Command Queue Host Controller Interface support" > + depends on HAS_DMA If the CQE implementation depends on MQ, should this not also select MMC_MQ_DEFAULT? And if the CQE implementation depens on MQ, isn't it dangerous to even have a boot option to disable it? Again I push the point that things are easier to keep in line if we just support one block request path (MQ), sorry for hammering. > +int cqhci_suspend(struct mmc_host *mmc) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + > + if (cq_host->enabled) > + __cqhci_disable(cq_host); > + > + return 0; > +} > +EXPORT_SYMBOL(cqhci_suspend); > + > +int cqhci_resume(struct mmc_host *mmc) > +{ > + /* Re-enable is done upon first request */ > + return 0; > +} > +EXPORT_SYMBOL(cqhci_resume); Why would the CQE case require special suspend/resume functionality? This seems two much like on the side CQE-silo engineering, just use the device .[runtime]_suspend/resume callbacks like everyone else, make it possible for the host to figure out if it is in CQE mode or not (I guess it should know already since cqhci .enable() has been called?) and handle it from there. > +struct cqhci_host_ops { > + void (*dumpregs)(struct mmc_host *mmc); > + void (*write_l)(struct cqhci_host *host, u32 val, int reg); > + u32 (*read_l)(struct cqhci_host *host, int reg); > + void (*enable)(struct mmc_host *mmc); > + void (*disable)(struct mmc_host *mmc, bool recovery); > +}; > + > +static inline void cqhci_writel(struct cqhci_host *host, u32 val, int reg) > +{ > + if (unlikely(host->ops->write_l)) > + host->ops->write_l(host, val, reg); > + else > + writel_relaxed(val, host->mmio + reg); > +} Why would CQE hosts need special accessors and the rest of the host not need it? This seems to be a bit hackish "just for CQE" approach, abstract callbacks like ->dumpregs(), ->write_l() and ->read_l() seems to be something that should be in the generic struct mmc_host_ops and not tied to CQE, if needed at all. ->enable and ->disable() for just CQE seem reasonable. But that leaves just two new ops. So why not just put .cqe_enable() and .cqe_disable() ops into mmc_host_ops as optional and be done with it? > +irqreturn_t cqhci_irq(struct mmc_host *mmc, u32 intmask, int cmd_error, > + int data_error); > +int cqhci_init(struct cqhci_host *cq_host, struct mmc_host *mmc, bool dma64); > +struct cqhci_host *cqhci_pltfm_init(struct platform_device *pdev); > +int cqhci_suspend(struct mmc_host *mmc); > +int cqhci_resume(struct mmc_host *mmc); This seems overall too much bolted on the side. I think the above approach to put any CQE-specific callbacks directly into the struct mmc_host_ops is way more viable. If special CQE init is needed, why a special cqhci_init() call? And cqhci_pltfm_init()? It's confusing. Can't you just call this by default from the core if the host is CQE capable? Ass a .cqhci_init() callback into mmc_host_ops if need be. cqhci_irq() seems necessary though, I see something like this is probably necessary. Yours, Linus Walleij
On 08/11/17 11:22, Linus Walleij wrote: > On Fri, Nov 3, 2017 at 2:20 PM, Adrian Hunter <adrian.hunter@intel.com> wrote: > >> From: Venkat Gopalakrishnan <venkatg@codeaurora.org> >> >> This patch adds CMDQ support for command-queue compatible >> hosts. >> >> Command queue is added in eMMC-5.1 specification. This >> enables the controller to process upto 32 requests at >> a time. >> >> Adrian Hunter contributed renaming to cqhci, recovery, suspend >> and resume, cqhci_off, cqhci_wait_for_idle, and external timeout >> handling. >> >> Signed-off-by: Asutosh Das <asutoshd@codeaurora.org> >> Signed-off-by: Sujit Reddy Thumma <sthumma@codeaurora.org> >> Signed-off-by: Konstantin Dorfman <kdorfman@codeaurora.org> >> Signed-off-by: Venkat Gopalakrishnan <venkatg@codeaurora.org> >> Signed-off-by: Subhash Jadavani <subhashj@codeaurora.org> >> Signed-off-by: Ritesh Harjani <riteshh@codeaurora.org> >> Signed-off-by: Adrian Hunter <adrian.hunter@intel.com> > > (...) >> +config MMC_CQHCI >> + tristate "Command Queue Host Controller Interface support" >> + depends on HAS_DMA > > If the CQE implementation depends on MQ, should this not > also select MMC_MQ_DEFAULT? No, there is no dependency on MMC_MQ_DEFAULT. > > And if the CQE implementation depens on MQ, isn't it dangerous > to even have a boot option to disable it? No, there is no dependency on MMC_MQ_DEFAULT. > > Again I push the point that things are easier to keep in line > if we just support one block request path (MQ), sorry for > hammering. > >> +int cqhci_suspend(struct mmc_host *mmc) >> +{ >> + struct cqhci_host *cq_host = mmc->cqe_private; >> + >> + if (cq_host->enabled) >> + __cqhci_disable(cq_host); >> + >> + return 0; >> +} >> +EXPORT_SYMBOL(cqhci_suspend); >> + >> +int cqhci_resume(struct mmc_host *mmc) >> +{ >> + /* Re-enable is done upon first request */ >> + return 0; >> +} >> +EXPORT_SYMBOL(cqhci_resume); > > Why would the CQE case require special suspend/resume > functionality? Seems like a very strange question. Obviously CQHCI has to be configured after suspend. Also please don't confuse CQE and CQHCI. CQHCI is an implementation of a CQE. We currently do not expect to have another implementation, but it is not impossible. > > This seems two much like on the side CQE-silo engineering, > just use the device .[runtime]_suspend/resume callbacks like > everyone else, make it possible for the host to figure out > if it is in CQE mode or not (I guess it should know already > since cqhci .enable() has been called?) and handle it > from there. That is how it works! The host controller has to decide how to handle suspend / resume. cqhci_suspend() / cqhci_resume() are helper functions that the host controller can use, but doesn't have to. > >> +struct cqhci_host_ops { >> + void (*dumpregs)(struct mmc_host *mmc); >> + void (*write_l)(struct cqhci_host *host, u32 val, int reg); >> + u32 (*read_l)(struct cqhci_host *host, int reg); >> + void (*enable)(struct mmc_host *mmc); >> + void (*disable)(struct mmc_host *mmc, bool recovery); >> +}; >> + >> +static inline void cqhci_writel(struct cqhci_host *host, u32 val, int reg) >> +{ >> + if (unlikely(host->ops->write_l)) >> + host->ops->write_l(host, val, reg); >> + else >> + writel_relaxed(val, host->mmio + reg); >> +} > > Why would CQE hosts need special accessors and the rest > of the host not need it? Special accessors can be used to fix up registers that don't work exactly the way the standard specified. > > This seems to be a bit hackish "just for CQE" approach, > abstract callbacks like ->dumpregs(), ->write_l() and > ->read_l() seems to be something that should be in the > generic struct mmc_host_ops and not tied to CQE, if > needed at all. They are nothing to do with CQE or mmc ops. These are for hosts that have a CQHCI compatible CQE. > > ->enable and ->disable() for just CQE seem reasonable. > But that leaves just two new ops. > > So why not just put .cqe_enable() and .cqe_disable() > ops into mmc_host_ops as optional and be done with it? Ok so you are not understanding this at all. As a CQE implementation, CQHCI interfaces with the upper layers through the CQE ops etc. But CQHCI also has to work with any host controller driver, so it needs an interface for that, which is what cqhci_host_ops is for. All the ops serve useful purposes. > >> +irqreturn_t cqhci_irq(struct mmc_host *mmc, u32 intmask, int cmd_error, >> + int data_error); >> +int cqhci_init(struct cqhci_host *cq_host, struct mmc_host *mmc, bool dma64); >> +struct cqhci_host *cqhci_pltfm_init(struct platform_device *pdev); >> +int cqhci_suspend(struct mmc_host *mmc); >> +int cqhci_resume(struct mmc_host *mmc); > > This seems overall too much bolted on the side. The whole point is to prove a library that can work with any host controller driver. That means it must provide functions and callbacks. > > I think the above approach to put any CQE-specific callbacks > directly into the struct mmc_host_ops is way more viable. Nothing to do with CQE. This is CQHCI. Please try to get the difference. > > If special CQE init is needed, why a special cqhci_init() > call? And cqhci_pltfm_init()? It's confusing. Can't > you just call this by default from the core if the host is > CQE capable? Ass a .cqhci_init() callback into mmc_host_ops > if need be. Yeah, so CQHCI is just one of theoretically any number of CQE implementations. This has nothing to do with the core. It is entirely up to the host driver. cqhci_pltfm_init() allows the mmio space to be defined by platform resources, whereas cqhci_init() does all the rest of the initialization. > > cqhci_irq() seems necessary though, I see something like this is > probably necessary.
On Wed, Nov 8, 2017 at 3:14 PM, Adrian Hunter <adrian.hunter@intel.com> wrote: > On 08/11/17 11:22, Linus Walleij wrote: >> On Fri, Nov 3, 2017 at 2:20 PM, Adrian Hunter <adrian.hunter@intel.com> wrote: >> (...) >>> +EXPORT_SYMBOL(cqhci_resume); >> >> Why would the CQE case require special suspend/resume >> functionality? > > Seems like a very strange question. Please realize that patch review is partly about education. Educating me or anyone about your patch set involves being humbe and not seeing your peers as lesser. Making your reviewer feel stupid by saying the ask "strange" or outright stupid questions is not helping your cause. > Obviously CQHCI has to be configured > after suspend. Yeah. I think I misunderstood is such that: > Also please don't confuse CQE and CQHCI. CQHCI is an implementation of a > CQE. We currently do not expect to have another implementation, but it is > not impossible. OK now you educated me, see it's not that hard without using belitteling language. >> This seems two much like on the side CQE-silo engineering, >> just use the device .[runtime]_suspend/resume callbacks like >> everyone else, make it possible for the host to figure out >> if it is in CQE mode or not (I guess it should know already >> since cqhci .enable() has been called?) and handle it >> from there. > > That is how it works! The host controller has to decide how to handle > suspend / resume. OK. > cqhci_suspend() / cqhci_resume() are helper functions that the host > controller can use, but doesn't have to. OK. >> Why would CQE hosts need special accessors and the rest >> of the host not need it? > > Special accessors can be used to fix up registers that don't work exactly > the way the standard specified. Yeah this is fine as it is for CQHCI, I didn't get that part :) >> ->enable and ->disable() for just CQE seem reasonable. >> But that leaves just two new ops. >> >> So why not just put .cqe_enable() and .cqe_disable() >> ops into mmc_host_ops as optional and be done with it? > > Ok so you are not understanding this at all. No I did not get it. But I do now (I think). > As a CQE implementation, CQHCI interfaces with the upper layers through the > CQE ops etc. > > But CQHCI also has to work with any host controller driver, so it needs an > interface for that, which is what cqhci_host_ops is for. All the ops serve > useful purposes. (...) > The whole point is to prove a library that can work with any host controller > driver. That means it must provide functions and callbacks. OK >> I think the above approach to put any CQE-specific callbacks >> directly into the struct mmc_host_ops is way more viable. > > Nothing to do with CQE. This is CQHCI. Please try to get the difference. I am trying, please try to think about your language. >> If special CQE init is needed, why a special cqhci_init() >> call? And cqhci_pltfm_init()? It's confusing. Can't >> you just call this by default from the core if the host is >> CQE capable? Ass a .cqhci_init() callback into mmc_host_ops >> if need be. > > Yeah, so CQHCI is just one of theoretically any number of CQE > implementations. This has nothing to do with the core. It is entirely up > to the host driver. cqhci_pltfm_init() allows the mmio space to be defined > by platform resources, whereas cqhci_init() does all the rest of the > initialization. It's fair. Yours, Linus Walleij
On 09/11/17 14:26, Linus Walleij wrote: > On Wed, Nov 8, 2017 at 3:14 PM, Adrian Hunter <adrian.hunter@intel.com> wrote: >> On 08/11/17 11:22, Linus Walleij wrote: >>> On Fri, Nov 3, 2017 at 2:20 PM, Adrian Hunter <adrian.hunter@intel.com> wrote: > >>> (...) > >>>> +EXPORT_SYMBOL(cqhci_resume); >>> >>> Why would the CQE case require special suspend/resume >>> functionality? >> >> Seems like a very strange question. > > Please realize that patch review is partly about education. > > Educating me or anyone about your patch set involves > being humbe and not seeing your peers as lesser. > > Making your reviewer feel stupid by saying the ask > "strange" or outright stupid questions is not helping > your cause. Please try to forgive me for being rude, it is just frustration. > >> Obviously CQHCI has to be configured >> after suspend. > > Yeah. I think I misunderstood is such that: > >> Also please don't confuse CQE and CQHCI. CQHCI is an implementation of a >> CQE. We currently do not expect to have another implementation, but it is >> not impossible. > > OK now you educated me, see it's not that hard without > using belitteling language. > >>> This seems two much like on the side CQE-silo engineering, >>> just use the device .[runtime]_suspend/resume callbacks like >>> everyone else, make it possible for the host to figure out >>> if it is in CQE mode or not (I guess it should know already >>> since cqhci .enable() has been called?) and handle it >>> from there. >> >> That is how it works! The host controller has to decide how to handle >> suspend / resume. > > OK. > >> cqhci_suspend() / cqhci_resume() are helper functions that the host >> controller can use, but doesn't have to. > > OK. > >>> Why would CQE hosts need special accessors and the rest >>> of the host not need it? >> >> Special accessors can be used to fix up registers that don't work exactly >> the way the standard specified. > > Yeah this is fine as it is for CQHCI, I didn't get that > part :) > >>> ->enable and ->disable() for just CQE seem reasonable. >>> But that leaves just two new ops. >>> >>> So why not just put .cqe_enable() and .cqe_disable() >>> ops into mmc_host_ops as optional and be done with it? >> >> Ok so you are not understanding this at all. > > No I did not get it. But I do now (I think). > >> As a CQE implementation, CQHCI interfaces with the upper layers through the >> CQE ops etc. >> >> But CQHCI also has to work with any host controller driver, so it needs an >> interface for that, which is what cqhci_host_ops is for. All the ops serve >> useful purposes. > (...) >> The whole point is to prove a library that can work with any host controller >> driver. That means it must provide functions and callbacks. > > OK > >>> I think the above approach to put any CQE-specific callbacks >>> directly into the struct mmc_host_ops is way more viable. >> >> Nothing to do with CQE. This is CQHCI. Please try to get the difference. > > I am trying, please try to think about your language. I strongly disapprove of being rude but sadly it seems to get results.
On 3 November 2017 at 14:20, Adrian Hunter <adrian.hunter@intel.com> wrote: > From: Venkat Gopalakrishnan <venkatg@codeaurora.org> > > This patch adds CMDQ support for command-queue compatible > hosts. > > Command queue is added in eMMC-5.1 specification. This > enables the controller to process upto 32 requests at > a time. > > Adrian Hunter contributed renaming to cqhci, recovery, suspend > and resume, cqhci_off, cqhci_wait_for_idle, and external timeout > handling. > > Signed-off-by: Asutosh Das <asutoshd@codeaurora.org> > Signed-off-by: Sujit Reddy Thumma <sthumma@codeaurora.org> > Signed-off-by: Konstantin Dorfman <kdorfman@codeaurora.org> > Signed-off-by: Venkat Gopalakrishnan <venkatg@codeaurora.org> > Signed-off-by: Subhash Jadavani <subhashj@codeaurora.org> > Signed-off-by: Ritesh Harjani <riteshh@codeaurora.org> > Signed-off-by: Adrian Hunter <adrian.hunter@intel.com> Overall this looks good to me. However, I didn't see MAINTAINERS being updated. Is anybody above volunteering to maintain cqhci.*? Kind regards Uffe > --- > drivers/mmc/host/Kconfig | 13 + > drivers/mmc/host/Makefile | 1 + > drivers/mmc/host/cqhci.c | 1150 +++++++++++++++++++++++++++++++++++++++++++++ > drivers/mmc/host/cqhci.h | 240 ++++++++++ > 4 files changed, 1404 insertions(+) > create mode 100644 drivers/mmc/host/cqhci.c > create mode 100644 drivers/mmc/host/cqhci.h > > diff --git a/drivers/mmc/host/Kconfig b/drivers/mmc/host/Kconfig > index 567028c9219a..3092b7085cb5 100644 > --- a/drivers/mmc/host/Kconfig > +++ b/drivers/mmc/host/Kconfig > @@ -857,6 +857,19 @@ config MMC_SUNXI > This selects support for the SD/MMC Host Controller on > Allwinner sunxi SoCs. > > +config MMC_CQHCI > + tristate "Command Queue Host Controller Interface support" > + depends on HAS_DMA > + help > + This selects the Command Queue Host Controller Interface (CQHCI) > + support present in host controllers of Qualcomm Technologies, Inc > + amongst others. > + This controller supports eMMC devices with command queue support. > + > + If you have a controller with this interface, say Y or M here. > + > + If unsure, say N. > + > config MMC_TOSHIBA_PCI > tristate "Toshiba Type A SD/MMC Card Interface Driver" > depends on PCI > diff --git a/drivers/mmc/host/Makefile b/drivers/mmc/host/Makefile > index ab61a3e39c0b..de140e3ef402 100644 > --- a/drivers/mmc/host/Makefile > +++ b/drivers/mmc/host/Makefile > @@ -91,6 +91,7 @@ obj-$(CONFIG_MMC_SDHCI_ST) += sdhci-st.o > obj-$(CONFIG_MMC_SDHCI_MICROCHIP_PIC32) += sdhci-pic32.o > obj-$(CONFIG_MMC_SDHCI_BRCMSTB) += sdhci-brcmstb.o > obj-$(CONFIG_MMC_SDHCI_OMAP) += sdhci-omap.o > +obj-$(CONFIG_MMC_CQHCI) += cqhci.o > > ifeq ($(CONFIG_CB710_DEBUG),y) > CFLAGS-cb710-mmc += -DDEBUG > diff --git a/drivers/mmc/host/cqhci.c b/drivers/mmc/host/cqhci.c > new file mode 100644 > index 000000000000..159270e947cf > --- /dev/null > +++ b/drivers/mmc/host/cqhci.c > @@ -0,0 +1,1150 @@ > +/* Copyright (c) 2015, The Linux Foundation. All rights reserved. > + * > + * This program is free software; you can redistribute it and/or modify > + * it under the terms of the GNU General Public License version 2 and > + * only version 2 as published by the Free Software Foundation. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + */ > + > +#include <linux/delay.h> > +#include <linux/highmem.h> > +#include <linux/io.h> > +#include <linux/module.h> > +#include <linux/dma-mapping.h> > +#include <linux/slab.h> > +#include <linux/scatterlist.h> > +#include <linux/platform_device.h> > +#include <linux/ktime.h> > + > +#include <linux/mmc/mmc.h> > +#include <linux/mmc/host.h> > +#include <linux/mmc/card.h> > + > +#include "cqhci.h" > + > +#define DCMD_SLOT 31 > +#define NUM_SLOTS 32 > + > +struct cqhci_slot { > + struct mmc_request *mrq; > + unsigned int flags; > +#define CQHCI_EXTERNAL_TIMEOUT BIT(0) > +#define CQHCI_COMPLETED BIT(1) > +#define CQHCI_HOST_CRC BIT(2) > +#define CQHCI_HOST_TIMEOUT BIT(3) > +#define CQHCI_HOST_OTHER BIT(4) > +}; > + > +static inline u8 *get_desc(struct cqhci_host *cq_host, u8 tag) > +{ > + return cq_host->desc_base + (tag * cq_host->slot_sz); > +} > + > +static inline u8 *get_link_desc(struct cqhci_host *cq_host, u8 tag) > +{ > + u8 *desc = get_desc(cq_host, tag); > + > + return desc + cq_host->task_desc_len; > +} > + > +static inline dma_addr_t get_trans_desc_dma(struct cqhci_host *cq_host, u8 tag) > +{ > + return cq_host->trans_desc_dma_base + > + (cq_host->mmc->max_segs * tag * > + cq_host->trans_desc_len); > +} > + > +static inline u8 *get_trans_desc(struct cqhci_host *cq_host, u8 tag) > +{ > + return cq_host->trans_desc_base + > + (cq_host->trans_desc_len * cq_host->mmc->max_segs * tag); > +} > + > +static void setup_trans_desc(struct cqhci_host *cq_host, u8 tag) > +{ > + u8 *link_temp; > + dma_addr_t trans_temp; > + > + link_temp = get_link_desc(cq_host, tag); > + trans_temp = get_trans_desc_dma(cq_host, tag); > + > + memset(link_temp, 0, cq_host->link_desc_len); > + if (cq_host->link_desc_len > 8) > + *(link_temp + 8) = 0; > + > + if (tag == DCMD_SLOT && (cq_host->mmc->caps2 & MMC_CAP2_CQE_DCMD)) { > + *link_temp = CQHCI_VALID(0) | CQHCI_ACT(0) | CQHCI_END(1); > + return; > + } > + > + *link_temp = CQHCI_VALID(1) | CQHCI_ACT(0x6) | CQHCI_END(0); > + > + if (cq_host->dma64) { > + __le64 *data_addr = (__le64 __force *)(link_temp + 4); > + > + data_addr[0] = cpu_to_le64(trans_temp); > + } else { > + __le32 *data_addr = (__le32 __force *)(link_temp + 4); > + > + data_addr[0] = cpu_to_le32(trans_temp); > + } > +} > + > +static void cqhci_set_irqs(struct cqhci_host *cq_host, u32 set) > +{ > + cqhci_writel(cq_host, set, CQHCI_ISTE); > + cqhci_writel(cq_host, set, CQHCI_ISGE); > +} > + > +#define DRV_NAME "cqhci" > + > +#define CQHCI_DUMP(f, x...) \ > + pr_err("%s: " DRV_NAME ": " f, mmc_hostname(mmc), ## x) > + > +static void cqhci_dumpregs(struct cqhci_host *cq_host) > +{ > + struct mmc_host *mmc = cq_host->mmc; > + > + CQHCI_DUMP("============ CQHCI REGISTER DUMP ===========\n"); > + > + CQHCI_DUMP("Caps: 0x%08x | Version: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_CAP), > + cqhci_readl(cq_host, CQHCI_VER)); > + CQHCI_DUMP("Config: 0x%08x | Control: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_CFG), > + cqhci_readl(cq_host, CQHCI_CTL)); > + CQHCI_DUMP("Int stat: 0x%08x | Int enab: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_IS), > + cqhci_readl(cq_host, CQHCI_ISTE)); > + CQHCI_DUMP("Int sig: 0x%08x | Int Coal: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_ISGE), > + cqhci_readl(cq_host, CQHCI_IC)); > + CQHCI_DUMP("TDL base: 0x%08x | TDL up32: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_TDLBA), > + cqhci_readl(cq_host, CQHCI_TDLBAU)); > + CQHCI_DUMP("Doorbell: 0x%08x | TCN: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_TDBR), > + cqhci_readl(cq_host, CQHCI_TCN)); > + CQHCI_DUMP("Dev queue: 0x%08x | Dev Pend: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_DQS), > + cqhci_readl(cq_host, CQHCI_DPT)); > + CQHCI_DUMP("Task clr: 0x%08x | SSC1: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_TCLR), > + cqhci_readl(cq_host, CQHCI_SSC1)); > + CQHCI_DUMP("SSC2: 0x%08x | DCMD rsp: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_SSC2), > + cqhci_readl(cq_host, CQHCI_CRDCT)); > + CQHCI_DUMP("RED mask: 0x%08x | TERRI: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_RMEM), > + cqhci_readl(cq_host, CQHCI_TERRI)); > + CQHCI_DUMP("Resp idx: 0x%08x | Resp arg: 0x%08x\n", > + cqhci_readl(cq_host, CQHCI_CRI), > + cqhci_readl(cq_host, CQHCI_CRA)); > + > + if (cq_host->ops->dumpregs) > + cq_host->ops->dumpregs(mmc); > + else > + CQHCI_DUMP(": ===========================================\n"); > +} > + > +/** > + * The allocated descriptor table for task, link & transfer descritors > + * looks like: > + * |----------| > + * |task desc | |->|----------| > + * |----------| | |trans desc| > + * |link desc-|->| |----------| > + * |----------| . > + * . . > + * no. of slots max-segs > + * . |----------| > + * |----------| > + * The idea here is to create the [task+trans] table and mark & point the > + * link desc to the transfer desc table on a per slot basis. > + */ > +static int cqhci_host_alloc_tdl(struct cqhci_host *cq_host) > +{ > + int i = 0; > + > + /* task descriptor can be 64/128 bit irrespective of arch */ > + if (cq_host->caps & CQHCI_TASK_DESC_SZ_128) { > + cqhci_writel(cq_host, cqhci_readl(cq_host, CQHCI_CFG) | > + CQHCI_TASK_DESC_SZ, CQHCI_CFG); > + cq_host->task_desc_len = 16; > + } else { > + cq_host->task_desc_len = 8; > + } > + > + /* > + * 96 bits length of transfer desc instead of 128 bits which means > + * ADMA would expect next valid descriptor at the 96th bit > + * or 128th bit > + */ > + if (cq_host->dma64) { > + if (cq_host->quirks & CQHCI_QUIRK_SHORT_TXFR_DESC_SZ) > + cq_host->trans_desc_len = 12; > + else > + cq_host->trans_desc_len = 16; > + cq_host->link_desc_len = 16; > + } else { > + cq_host->trans_desc_len = 8; > + cq_host->link_desc_len = 8; > + } > + > + /* total size of a slot: 1 task & 1 transfer (link) */ > + cq_host->slot_sz = cq_host->task_desc_len + cq_host->link_desc_len; > + > + cq_host->desc_size = cq_host->slot_sz * cq_host->num_slots; > + > + cq_host->data_size = cq_host->trans_desc_len * cq_host->mmc->max_segs * > + (cq_host->num_slots - 1); > + > + pr_debug("%s: cqhci: desc_size: %zu data_sz: %zu slot-sz: %d\n", > + mmc_hostname(cq_host->mmc), cq_host->desc_size, cq_host->data_size, > + cq_host->slot_sz); > + > + /* > + * allocate a dma-mapped chunk of memory for the descriptors > + * allocate a dma-mapped chunk of memory for link descriptors > + * setup each link-desc memory offset per slot-number to > + * the descriptor table. > + */ > + cq_host->desc_base = dmam_alloc_coherent(mmc_dev(cq_host->mmc), > + cq_host->desc_size, > + &cq_host->desc_dma_base, > + GFP_KERNEL); > + cq_host->trans_desc_base = dmam_alloc_coherent(mmc_dev(cq_host->mmc), > + cq_host->data_size, > + &cq_host->trans_desc_dma_base, > + GFP_KERNEL); > + if (!cq_host->desc_base || !cq_host->trans_desc_base) > + return -ENOMEM; > + > + pr_debug("%s: cqhci: desc-base: 0x%p trans-base: 0x%p\n desc_dma 0x%llx trans_dma: 0x%llx\n", > + mmc_hostname(cq_host->mmc), cq_host->desc_base, cq_host->trans_desc_base, > + (unsigned long long)cq_host->desc_dma_base, > + (unsigned long long)cq_host->trans_desc_dma_base); > + > + for (; i < (cq_host->num_slots); i++) > + setup_trans_desc(cq_host, i); > + > + return 0; > +} > + > +static void __cqhci_enable(struct cqhci_host *cq_host) > +{ > + struct mmc_host *mmc = cq_host->mmc; > + u32 cqcfg; > + > + cqcfg = cqhci_readl(cq_host, CQHCI_CFG); > + > + /* Configuration must not be changed while enabled */ > + if (cqcfg & CQHCI_ENABLE) { > + cqcfg &= ~CQHCI_ENABLE; > + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); > + } > + > + cqcfg &= ~(CQHCI_DCMD | CQHCI_TASK_DESC_SZ); > + > + if (mmc->caps2 & MMC_CAP2_CQE_DCMD) > + cqcfg |= CQHCI_DCMD; > + > + if (cq_host->caps & CQHCI_TASK_DESC_SZ_128) > + cqcfg |= CQHCI_TASK_DESC_SZ; > + > + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); > + > + cqhci_writel(cq_host, lower_32_bits(cq_host->desc_dma_base), > + CQHCI_TDLBA); > + cqhci_writel(cq_host, upper_32_bits(cq_host->desc_dma_base), > + CQHCI_TDLBAU); > + > + cqhci_writel(cq_host, cq_host->rca, CQHCI_SSC2); > + > + cqhci_set_irqs(cq_host, 0); > + > + cqcfg |= CQHCI_ENABLE; > + > + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); > + > + mmc->cqe_on = true; > + > + if (cq_host->ops->enable) > + cq_host->ops->enable(mmc); > + > + /* Ensure all writes are done before interrupts are enabled */ > + wmb(); > + > + cqhci_set_irqs(cq_host, CQHCI_IS_MASK); > + > + cq_host->activated = true; > +} > + > +static void __cqhci_disable(struct cqhci_host *cq_host) > +{ > + u32 cqcfg; > + > + cqcfg = cqhci_readl(cq_host, CQHCI_CFG); > + cqcfg &= ~CQHCI_ENABLE; > + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); > + > + cq_host->mmc->cqe_on = false; > + > + cq_host->activated = false; > +} > + > +int cqhci_suspend(struct mmc_host *mmc) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + > + if (cq_host->enabled) > + __cqhci_disable(cq_host); > + > + return 0; > +} > +EXPORT_SYMBOL(cqhci_suspend); > + > +int cqhci_resume(struct mmc_host *mmc) > +{ > + /* Re-enable is done upon first request */ > + return 0; > +} > +EXPORT_SYMBOL(cqhci_resume); > + > +static int cqhci_enable(struct mmc_host *mmc, struct mmc_card *card) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + int err; > + > + if (cq_host->enabled) > + return 0; > + > + cq_host->rca = card->rca; > + > + err = cqhci_host_alloc_tdl(cq_host); > + if (err) > + return err; > + > + __cqhci_enable(cq_host); > + > + cq_host->enabled = true; > + > +#ifdef DEBUG > + cqhci_dumpregs(cq_host); > +#endif > + return 0; > +} > + > +/* CQHCI is idle and should halt immediately, so set a small timeout */ > +#define CQHCI_OFF_TIMEOUT 100 > + > +static void cqhci_off(struct mmc_host *mmc) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + ktime_t timeout; > + bool timed_out; > + u32 reg; > + > + if (!cq_host->enabled || !mmc->cqe_on || cq_host->recovery_halt) > + return; > + > + if (cq_host->ops->disable) > + cq_host->ops->disable(mmc, false); > + > + cqhci_writel(cq_host, CQHCI_HALT, CQHCI_CTL); > + > + timeout = ktime_add_us(ktime_get(), CQHCI_OFF_TIMEOUT); > + while (1) { > + timed_out = ktime_compare(ktime_get(), timeout) > 0; > + reg = cqhci_readl(cq_host, CQHCI_CTL); > + if ((reg & CQHCI_HALT) || timed_out) > + break; > + } > + > + if (timed_out) > + pr_err("%s: cqhci: CQE stuck on\n", mmc_hostname(mmc)); > + else > + pr_debug("%s: cqhci: CQE off\n", mmc_hostname(mmc)); > + > + mmc->cqe_on = false; > +} > + > +static void cqhci_disable(struct mmc_host *mmc) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + > + if (!cq_host->enabled) > + return; > + > + cqhci_off(mmc); > + > + __cqhci_disable(cq_host); > + > + dmam_free_coherent(mmc_dev(mmc), cq_host->data_size, > + cq_host->trans_desc_base, > + cq_host->trans_desc_dma_base); > + > + dmam_free_coherent(mmc_dev(mmc), cq_host->desc_size, > + cq_host->desc_base, > + cq_host->desc_dma_base); > + > + cq_host->trans_desc_base = NULL; > + cq_host->desc_base = NULL; > + > + cq_host->enabled = false; > +} > + > +static void cqhci_prep_task_desc(struct mmc_request *mrq, > + u64 *data, bool intr) > +{ > + u32 req_flags = mrq->data->flags; > + > + *data = CQHCI_VALID(1) | > + CQHCI_END(1) | > + CQHCI_INT(intr) | > + CQHCI_ACT(0x5) | > + CQHCI_FORCED_PROG(!!(req_flags & MMC_DATA_FORCED_PRG)) | > + CQHCI_DATA_TAG(!!(req_flags & MMC_DATA_DAT_TAG)) | > + CQHCI_DATA_DIR(!!(req_flags & MMC_DATA_READ)) | > + CQHCI_PRIORITY(!!(req_flags & MMC_DATA_PRIO)) | > + CQHCI_QBAR(!!(req_flags & MMC_DATA_QBR)) | > + CQHCI_REL_WRITE(!!(req_flags & MMC_DATA_REL_WR)) | > + CQHCI_BLK_COUNT(mrq->data->blocks) | > + CQHCI_BLK_ADDR((u64)mrq->data->blk_addr); > + > + pr_debug("%s: cqhci: tag %d task descriptor 0x016%llx\n", > + mmc_hostname(mrq->host), mrq->tag, (unsigned long long)*data); > +} > + > +static int cqhci_dma_map(struct mmc_host *host, struct mmc_request *mrq) > +{ > + int sg_count; > + struct mmc_data *data = mrq->data; > + > + if (!data) > + return -EINVAL; > + > + sg_count = dma_map_sg(mmc_dev(host), data->sg, > + data->sg_len, > + (data->flags & MMC_DATA_WRITE) ? > + DMA_TO_DEVICE : DMA_FROM_DEVICE); > + if (!sg_count) { > + pr_err("%s: sg-len: %d\n", __func__, data->sg_len); > + return -ENOMEM; > + } > + > + return sg_count; > +} > + > +static void cqhci_set_tran_desc(u8 *desc, dma_addr_t addr, int len, bool end, > + bool dma64) > +{ > + __le32 *attr = (__le32 __force *)desc; > + > + *attr = (CQHCI_VALID(1) | > + CQHCI_END(end ? 1 : 0) | > + CQHCI_INT(0) | > + CQHCI_ACT(0x4) | > + CQHCI_DAT_LENGTH(len)); > + > + if (dma64) { > + __le64 *dataddr = (__le64 __force *)(desc + 4); > + > + dataddr[0] = cpu_to_le64(addr); > + } else { > + __le32 *dataddr = (__le32 __force *)(desc + 4); > + > + dataddr[0] = cpu_to_le32(addr); > + } > +} > + > +static int cqhci_prep_tran_desc(struct mmc_request *mrq, > + struct cqhci_host *cq_host, int tag) > +{ > + struct mmc_data *data = mrq->data; > + int i, sg_count, len; > + bool end = false; > + bool dma64 = cq_host->dma64; > + dma_addr_t addr; > + u8 *desc; > + struct scatterlist *sg; > + > + sg_count = cqhci_dma_map(mrq->host, mrq); > + if (sg_count < 0) { > + pr_err("%s: %s: unable to map sg lists, %d\n", > + mmc_hostname(mrq->host), __func__, sg_count); > + return sg_count; > + } > + > + desc = get_trans_desc(cq_host, tag); > + > + for_each_sg(data->sg, sg, sg_count, i) { > + addr = sg_dma_address(sg); > + len = sg_dma_len(sg); > + > + if ((i+1) == sg_count) > + end = true; > + cqhci_set_tran_desc(desc, addr, len, end, dma64); > + desc += cq_host->trans_desc_len; > + } > + > + return 0; > +} > + > +static void cqhci_prep_dcmd_desc(struct mmc_host *mmc, > + struct mmc_request *mrq) > +{ > + u64 *task_desc = NULL; > + u64 data = 0; > + u8 resp_type; > + u8 *desc; > + __le64 *dataddr; > + struct cqhci_host *cq_host = mmc->cqe_private; > + u8 timing; > + > + if (!(mrq->cmd->flags & MMC_RSP_PRESENT)) { > + resp_type = 0x0; > + timing = 0x1; > + } else { > + if (mrq->cmd->flags & MMC_RSP_R1B) { > + resp_type = 0x3; > + timing = 0x0; > + } else { > + resp_type = 0x2; > + timing = 0x1; > + } > + } > + > + task_desc = (__le64 __force *)get_desc(cq_host, cq_host->dcmd_slot); > + memset(task_desc, 0, cq_host->task_desc_len); > + data |= (CQHCI_VALID(1) | > + CQHCI_END(1) | > + CQHCI_INT(1) | > + CQHCI_QBAR(1) | > + CQHCI_ACT(0x5) | > + CQHCI_CMD_INDEX(mrq->cmd->opcode) | > + CQHCI_CMD_TIMING(timing) | CQHCI_RESP_TYPE(resp_type)); > + *task_desc |= data; > + desc = (u8 *)task_desc; > + pr_debug("%s: cqhci: dcmd: cmd: %d timing: %d resp: %d\n", > + mmc_hostname(mmc), mrq->cmd->opcode, timing, resp_type); > + dataddr = (__le64 __force *)(desc + 4); > + dataddr[0] = cpu_to_le64((u64)mrq->cmd->arg); > + > +} > + > +static void cqhci_post_req(struct mmc_host *host, struct mmc_request *mrq) > +{ > + struct mmc_data *data = mrq->data; > + > + if (data) { > + dma_unmap_sg(mmc_dev(host), data->sg, data->sg_len, > + (data->flags & MMC_DATA_READ) ? > + DMA_FROM_DEVICE : DMA_TO_DEVICE); > + } > +} > + > +static inline int cqhci_tag(struct mmc_request *mrq) > +{ > + return mrq->cmd ? DCMD_SLOT : mrq->tag; > +} > + > +static int cqhci_request(struct mmc_host *mmc, struct mmc_request *mrq) > +{ > + int err = 0; > + u64 data = 0; > + u64 *task_desc = NULL; > + int tag = cqhci_tag(mrq); > + struct cqhci_host *cq_host = mmc->cqe_private; > + unsigned long flags; > + > + if (!cq_host->enabled) { > + pr_err("%s: cqhci: not enabled\n", mmc_hostname(mmc)); > + return -EINVAL; > + } > + > + /* First request after resume has to re-enable */ > + if (!cq_host->activated) > + __cqhci_enable(cq_host); > + > + if (!mmc->cqe_on) { > + cqhci_writel(cq_host, 0, CQHCI_CTL); > + mmc->cqe_on = true; > + pr_debug("%s: cqhci: CQE on\n", mmc_hostname(mmc)); > + if (cqhci_readl(cq_host, CQHCI_CTL) && CQHCI_HALT) { > + pr_err("%s: cqhci: CQE failed to exit halt state\n", > + mmc_hostname(mmc)); > + } > + if (cq_host->ops->enable) > + cq_host->ops->enable(mmc); > + } > + > + if (mrq->data) { > + task_desc = (__le64 __force *)get_desc(cq_host, tag); > + cqhci_prep_task_desc(mrq, &data, 1); > + *task_desc = cpu_to_le64(data); > + err = cqhci_prep_tran_desc(mrq, cq_host, tag); > + if (err) { > + pr_err("%s: cqhci: failed to setup tx desc: %d\n", > + mmc_hostname(mmc), err); > + return err; > + } > + } else { > + cqhci_prep_dcmd_desc(mmc, mrq); > + } > + > + spin_lock_irqsave(&cq_host->lock, flags); > + > + if (cq_host->recovery_halt) { > + err = -EBUSY; > + goto out_unlock; > + } > + > + cq_host->slot[tag].mrq = mrq; > + cq_host->slot[tag].flags = 0; > + > + cq_host->qcnt += 1; > + > + cqhci_writel(cq_host, 1 << tag, CQHCI_TDBR); > + if (!(cqhci_readl(cq_host, CQHCI_TDBR) & (1 << tag))) > + pr_debug("%s: cqhci: doorbell not set for tag %d\n", > + mmc_hostname(mmc), tag); > +out_unlock: > + spin_unlock_irqrestore(&cq_host->lock, flags); > + > + if (err) > + cqhci_post_req(mmc, mrq); > + > + return err; > +} > + > +static void cqhci_recovery_needed(struct mmc_host *mmc, struct mmc_request *mrq, > + bool notify) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + > + if (!cq_host->recovery_halt) { > + cq_host->recovery_halt = true; > + pr_debug("%s: cqhci: recovery needed\n", mmc_hostname(mmc)); > + wake_up(&cq_host->wait_queue); > + if (notify && mrq->recovery_notifier) > + mrq->recovery_notifier(mrq); > + } > +} > + > +static unsigned int cqhci_error_flags(int error1, int error2) > +{ > + int error = error1 ? error1 : error2; > + > + switch (error) { > + case -EILSEQ: > + return CQHCI_HOST_CRC; > + case -ETIMEDOUT: > + return CQHCI_HOST_TIMEOUT; > + default: > + return CQHCI_HOST_OTHER; > + } > +} > + > +static void cqhci_error_irq(struct mmc_host *mmc, u32 status, int cmd_error, > + int data_error) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + struct cqhci_slot *slot; > + u32 terri; > + int tag; > + > + spin_lock(&cq_host->lock); > + > + terri = cqhci_readl(cq_host, CQHCI_TERRI); > + > + pr_debug("%s: cqhci: error IRQ status: 0x%08x cmd error %d data error %d TERRI: 0x%08x\n", > + mmc_hostname(mmc), status, cmd_error, data_error, terri); > + > + /* Forget about errors when recovery has already been triggered */ > + if (cq_host->recovery_halt) > + goto out_unlock; > + > + if (!cq_host->qcnt) { > + WARN_ONCE(1, "%s: cqhci: error when idle. IRQ status: 0x%08x cmd error %d data error %d TERRI: 0x%08x\n", > + mmc_hostname(mmc), status, cmd_error, data_error, > + terri); > + goto out_unlock; > + } > + > + if (CQHCI_TERRI_C_VALID(terri)) { > + tag = CQHCI_TERRI_C_TASK(terri); > + slot = &cq_host->slot[tag]; > + if (slot->mrq) { > + slot->flags = cqhci_error_flags(cmd_error, data_error); > + cqhci_recovery_needed(mmc, slot->mrq, true); > + } > + } > + > + if (CQHCI_TERRI_D_VALID(terri)) { > + tag = CQHCI_TERRI_D_TASK(terri); > + slot = &cq_host->slot[tag]; > + if (slot->mrq) { > + slot->flags = cqhci_error_flags(data_error, cmd_error); > + cqhci_recovery_needed(mmc, slot->mrq, true); > + } > + } > + > + if (!cq_host->recovery_halt) { > + /* > + * The only way to guarantee forward progress is to mark at > + * least one task in error, so if none is indicated, pick one. > + */ > + for (tag = 0; tag < NUM_SLOTS; tag++) { > + slot = &cq_host->slot[tag]; > + if (!slot->mrq) > + continue; > + slot->flags = cqhci_error_flags(data_error, cmd_error); > + cqhci_recovery_needed(mmc, slot->mrq, true); > + break; > + } > + } > + > +out_unlock: > + spin_unlock(&cq_host->lock); > +} > + > +static void cqhci_finish_mrq(struct mmc_host *mmc, unsigned int tag) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + struct cqhci_slot *slot = &cq_host->slot[tag]; > + struct mmc_request *mrq = slot->mrq; > + struct mmc_data *data; > + > + if (!mrq) { > + WARN_ONCE(1, "%s: cqhci: spurious TCN for tag %d\n", > + mmc_hostname(mmc), tag); > + return; > + } > + > + /* No completions allowed during recovery */ > + if (cq_host->recovery_halt) { > + slot->flags |= CQHCI_COMPLETED; > + return; > + } > + > + slot->mrq = NULL; > + > + cq_host->qcnt -= 1; > + > + data = mrq->data; > + if (data) { > + if (data->error) > + data->bytes_xfered = 0; > + else > + data->bytes_xfered = data->blksz * data->blocks; > + } > + > + mmc_cqe_request_done(mmc, mrq); > +} > + > +irqreturn_t cqhci_irq(struct mmc_host *mmc, u32 intmask, int cmd_error, > + int data_error) > +{ > + u32 status; > + unsigned long tag = 0, comp_status; > + struct cqhci_host *cq_host = mmc->cqe_private; > + > + status = cqhci_readl(cq_host, CQHCI_IS); > + cqhci_writel(cq_host, status, CQHCI_IS); > + > + pr_debug("%s: cqhci: IRQ status: 0x%08x\n", mmc_hostname(mmc), status); > + > + if ((status & CQHCI_IS_RED) || cmd_error || data_error) > + cqhci_error_irq(mmc, status, cmd_error, data_error); > + > + if (status & CQHCI_IS_TCC) { > + /* read TCN and complete the request */ > + comp_status = cqhci_readl(cq_host, CQHCI_TCN); > + cqhci_writel(cq_host, comp_status, CQHCI_TCN); > + pr_debug("%s: cqhci: TCN: 0x%08lx\n", > + mmc_hostname(mmc), comp_status); > + > + spin_lock(&cq_host->lock); > + > + for_each_set_bit(tag, &comp_status, cq_host->num_slots) { > + /* complete the corresponding mrq */ > + pr_debug("%s: cqhci: completing tag %lu\n", > + mmc_hostname(mmc), tag); > + cqhci_finish_mrq(mmc, tag); > + } > + > + if (cq_host->waiting_for_idle && !cq_host->qcnt) { > + cq_host->waiting_for_idle = false; > + wake_up(&cq_host->wait_queue); > + } > + > + spin_unlock(&cq_host->lock); > + } > + > + if (status & CQHCI_IS_TCL) > + wake_up(&cq_host->wait_queue); > + > + if (status & CQHCI_IS_HAC) > + wake_up(&cq_host->wait_queue); > + > + return IRQ_HANDLED; > +} > +EXPORT_SYMBOL(cqhci_irq); > + > +static bool cqhci_is_idle(struct cqhci_host *cq_host, int *ret) > +{ > + unsigned long flags; > + bool is_idle; > + > + spin_lock_irqsave(&cq_host->lock, flags); > + is_idle = !cq_host->qcnt || cq_host->recovery_halt; > + *ret = cq_host->recovery_halt ? -EBUSY : 0; > + cq_host->waiting_for_idle = !is_idle; > + spin_unlock_irqrestore(&cq_host->lock, flags); > + > + return is_idle; > +} > + > +static int cqhci_wait_for_idle(struct mmc_host *mmc) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + int ret; > + > + wait_event(cq_host->wait_queue, cqhci_is_idle(cq_host, &ret)); > + > + return ret; > +} > + > +static bool cqhci_timeout(struct mmc_host *mmc, struct mmc_request *mrq, > + bool *recovery_needed) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + int tag = cqhci_tag(mrq); > + struct cqhci_slot *slot = &cq_host->slot[tag]; > + unsigned long flags; > + bool timed_out; > + > + spin_lock_irqsave(&cq_host->lock, flags); > + timed_out = slot->mrq == mrq; > + if (timed_out) { > + slot->flags |= CQHCI_EXTERNAL_TIMEOUT; > + cqhci_recovery_needed(mmc, mrq, false); > + *recovery_needed = cq_host->recovery_halt; > + } > + spin_unlock_irqrestore(&cq_host->lock, flags); > + > + if (timed_out) { > + pr_err("%s: cqhci: timeout for tag %d\n", > + mmc_hostname(mmc), tag); > + cqhci_dumpregs(cq_host); > + } > + > + return timed_out; > +} > + > +static bool cqhci_tasks_cleared(struct cqhci_host *cq_host) > +{ > + return !(cqhci_readl(cq_host, CQHCI_CTL) & CQHCI_CLEAR_ALL_TASKS); > +} > + > +static bool cqhci_clear_all_tasks(struct mmc_host *mmc, unsigned int timeout) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + bool ret; > + u32 ctl; > + > + cqhci_set_irqs(cq_host, CQHCI_IS_TCL); > + > + ctl = cqhci_readl(cq_host, CQHCI_CTL); > + ctl |= CQHCI_CLEAR_ALL_TASKS; > + cqhci_writel(cq_host, ctl, CQHCI_CTL); > + > + wait_event_timeout(cq_host->wait_queue, cqhci_tasks_cleared(cq_host), > + msecs_to_jiffies(timeout) + 1); > + > + cqhci_set_irqs(cq_host, 0); > + > + ret = cqhci_tasks_cleared(cq_host); > + > + if (!ret) > + pr_debug("%s: cqhci: Failed to clear tasks\n", > + mmc_hostname(mmc)); > + > + return ret; > +} > + > +static bool cqhci_halted(struct cqhci_host *cq_host) > +{ > + return cqhci_readl(cq_host, CQHCI_CTL) & CQHCI_HALT; > +} > + > +static bool cqhci_halt(struct mmc_host *mmc, unsigned int timeout) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + bool ret; > + u32 ctl; > + > + if (cqhci_halted(cq_host)) > + return true; > + > + cqhci_set_irqs(cq_host, CQHCI_IS_HAC); > + > + ctl = cqhci_readl(cq_host, CQHCI_CTL); > + ctl |= CQHCI_HALT; > + cqhci_writel(cq_host, ctl, CQHCI_CTL); > + > + wait_event_timeout(cq_host->wait_queue, cqhci_halted(cq_host), > + msecs_to_jiffies(timeout) + 1); > + > + cqhci_set_irqs(cq_host, 0); > + > + ret = cqhci_halted(cq_host); > + > + if (!ret) > + pr_debug("%s: cqhci: Failed to halt\n", mmc_hostname(mmc)); > + > + return ret; > +} > + > +/* > + * After halting we expect to be able to use the command line. We interpret the > + * failure to halt to mean the data lines might still be in use (and the upper > + * layers will need to send a STOP command), so we set the timeout based on a > + * generous command timeout. > + */ > +#define CQHCI_START_HALT_TIMEOUT 5 > + > +static void cqhci_recovery_start(struct mmc_host *mmc) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + > + pr_debug("%s: cqhci: %s\n", mmc_hostname(mmc), __func__); > + > + WARN_ON(!cq_host->recovery_halt); > + > + cqhci_halt(mmc, CQHCI_START_HALT_TIMEOUT); > + > + if (cq_host->ops->disable) > + cq_host->ops->disable(mmc, true); > + > + mmc->cqe_on = false; > +} > + > +static int cqhci_error_from_flags(unsigned int flags) > +{ > + if (!flags) > + return 0; > + > + /* CRC errors might indicate re-tuning so prefer to report that */ > + if (flags & CQHCI_HOST_CRC) > + return -EILSEQ; > + > + if (flags & (CQHCI_EXTERNAL_TIMEOUT | CQHCI_HOST_TIMEOUT)) > + return -ETIMEDOUT; > + > + return -EIO; > +} > + > +static void cqhci_recover_mrq(struct cqhci_host *cq_host, unsigned int tag) > +{ > + struct cqhci_slot *slot = &cq_host->slot[tag]; > + struct mmc_request *mrq = slot->mrq; > + struct mmc_data *data; > + > + if (!mrq) > + return; > + > + slot->mrq = NULL; > + > + cq_host->qcnt -= 1; > + > + data = mrq->data; > + if (data) { > + data->bytes_xfered = 0; > + data->error = cqhci_error_from_flags(slot->flags); > + } else { > + mrq->cmd->error = cqhci_error_from_flags(slot->flags); > + } > + > + mmc_cqe_request_done(cq_host->mmc, mrq); > +} > + > +static void cqhci_recover_mrqs(struct cqhci_host *cq_host) > +{ > + int i; > + > + for (i = 0; i < cq_host->num_slots; i++) > + cqhci_recover_mrq(cq_host, i); > +} > + > +/* > + * By now the command and data lines should be unused so there is no reason for > + * CQHCI to take a long time to halt, but if it doesn't halt there could be > + * problems clearing tasks, so be generous. > + */ > +#define CQHCI_FINISH_HALT_TIMEOUT 20 > + > +/* CQHCI could be expected to clear it's internal state pretty quickly */ > +#define CQHCI_CLEAR_TIMEOUT 20 > + > +static void cqhci_recovery_finish(struct mmc_host *mmc) > +{ > + struct cqhci_host *cq_host = mmc->cqe_private; > + unsigned long flags; > + u32 cqcfg; > + bool ok; > + > + pr_debug("%s: cqhci: %s\n", mmc_hostname(mmc), __func__); > + > + WARN_ON(!cq_host->recovery_halt); > + > + ok = cqhci_halt(mmc, CQHCI_FINISH_HALT_TIMEOUT); > + > + if (!cqhci_clear_all_tasks(mmc, CQHCI_CLEAR_TIMEOUT)) > + ok = false; > + > + /* > + * The specification contradicts itself, by saying that tasks cannot be > + * cleared if CQHCI does not halt, but if CQHCI does not halt, it should > + * be disabled/re-enabled, but not to disable before clearing tasks. > + * Have a go anyway. > + */ > + if (!ok) { > + pr_debug("%s: cqhci: disable / re-enable\n", mmc_hostname(mmc)); > + cqcfg = cqhci_readl(cq_host, CQHCI_CFG); > + cqcfg &= ~CQHCI_ENABLE; > + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); > + cqcfg |= CQHCI_ENABLE; > + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); > + /* Be sure that there are no tasks */ > + ok = cqhci_halt(mmc, CQHCI_FINISH_HALT_TIMEOUT); > + if (!cqhci_clear_all_tasks(mmc, CQHCI_CLEAR_TIMEOUT)) > + ok = false; > + WARN_ON(!ok); > + } > + > + cqhci_recover_mrqs(cq_host); > + > + WARN_ON(cq_host->qcnt); > + > + spin_lock_irqsave(&cq_host->lock, flags); > + cq_host->qcnt = 0; > + cq_host->recovery_halt = false; > + mmc->cqe_on = false; > + spin_unlock_irqrestore(&cq_host->lock, flags); > + > + /* Ensure all writes are done before interrupts are re-enabled */ > + wmb(); > + > + cqhci_writel(cq_host, CQHCI_IS_HAC | CQHCI_IS_TCL, CQHCI_IS); > + > + cqhci_set_irqs(cq_host, CQHCI_IS_MASK); > + > + pr_debug("%s: cqhci: recovery done\n", mmc_hostname(mmc)); > +} > + > +static const struct mmc_cqe_ops cqhci_cqe_ops = { > + .cqe_enable = cqhci_enable, > + .cqe_disable = cqhci_disable, > + .cqe_request = cqhci_request, > + .cqe_post_req = cqhci_post_req, > + .cqe_off = cqhci_off, > + .cqe_wait_for_idle = cqhci_wait_for_idle, > + .cqe_timeout = cqhci_timeout, > + .cqe_recovery_start = cqhci_recovery_start, > + .cqe_recovery_finish = cqhci_recovery_finish, > +}; > + > +struct cqhci_host *cqhci_pltfm_init(struct platform_device *pdev) > +{ > + struct cqhci_host *cq_host; > + struct resource *cqhci_memres = NULL; > + > + /* check and setup CMDQ interface */ > + cqhci_memres = platform_get_resource_byname(pdev, IORESOURCE_MEM, > + "cqhci_mem"); > + if (!cqhci_memres) { > + dev_dbg(&pdev->dev, "CMDQ not supported\n"); > + return ERR_PTR(-EINVAL); > + } > + > + cq_host = devm_kzalloc(&pdev->dev, sizeof(*cq_host), GFP_KERNEL); > + if (!cq_host) > + return ERR_PTR(-ENOMEM); > + cq_host->mmio = devm_ioremap(&pdev->dev, > + cqhci_memres->start, > + resource_size(cqhci_memres)); > + if (!cq_host->mmio) { > + dev_err(&pdev->dev, "failed to remap cqhci regs\n"); > + return ERR_PTR(-EBUSY); > + } > + dev_dbg(&pdev->dev, "CMDQ ioremap: done\n"); > + > + return cq_host; > +} > +EXPORT_SYMBOL(cqhci_pltfm_init); > + > +static unsigned int cqhci_ver_major(struct cqhci_host *cq_host) > +{ > + return CQHCI_VER_MAJOR(cqhci_readl(cq_host, CQHCI_VER)); > +} > + > +static unsigned int cqhci_ver_minor(struct cqhci_host *cq_host) > +{ > + u32 ver = cqhci_readl(cq_host, CQHCI_VER); > + > + return CQHCI_VER_MINOR1(ver) * 10 + CQHCI_VER_MINOR2(ver); > +} > + > +int cqhci_init(struct cqhci_host *cq_host, struct mmc_host *mmc, > + bool dma64) > +{ > + int err; > + > + cq_host->dma64 = dma64; > + cq_host->mmc = mmc; > + cq_host->mmc->cqe_private = cq_host; > + > + cq_host->num_slots = NUM_SLOTS; > + cq_host->dcmd_slot = DCMD_SLOT; > + > + mmc->cqe_ops = &cqhci_cqe_ops; > + > + mmc->cqe_qdepth = NUM_SLOTS; > + if (mmc->caps2 & MMC_CAP2_CQE_DCMD) > + mmc->cqe_qdepth -= 1; > + > + cq_host->slot = devm_kcalloc(mmc_dev(mmc), cq_host->num_slots, > + sizeof(*cq_host->slot), GFP_KERNEL); > + if (!cq_host->slot) { > + err = -ENOMEM; > + goto out_err; > + } > + > + spin_lock_init(&cq_host->lock); > + > + init_completion(&cq_host->halt_comp); > + init_waitqueue_head(&cq_host->wait_queue); > + > + pr_info("%s: CQHCI version %u.%02u\n", > + mmc_hostname(mmc), cqhci_ver_major(cq_host), > + cqhci_ver_minor(cq_host)); > + > + return 0; > + > +out_err: > + pr_err("%s: CQHCI version %u.%02u failed to initialize, error %d\n", > + mmc_hostname(mmc), cqhci_ver_major(cq_host), > + cqhci_ver_minor(cq_host), err); > + return err; > +} > +EXPORT_SYMBOL(cqhci_init); > + > +MODULE_AUTHOR("Venkat Gopalakrishnan <venkatg@codeaurora.org>"); > +MODULE_DESCRIPTION("Command Queue Host Controller Interface driver"); > +MODULE_LICENSE("GPL v2"); > diff --git a/drivers/mmc/host/cqhci.h b/drivers/mmc/host/cqhci.h > new file mode 100644 > index 000000000000..2d39d361b322 > --- /dev/null > +++ b/drivers/mmc/host/cqhci.h > @@ -0,0 +1,240 @@ > +/* Copyright (c) 2015, The Linux Foundation. All rights reserved. > + * > + * This program is free software; you can redistribute it and/or modify > + * it under the terms of the GNU General Public License version 2 and > + * only version 2 as published by the Free Software Foundation. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + */ > +#ifndef LINUX_MMC_CQHCI_H > +#define LINUX_MMC_CQHCI_H > + > +#include <linux/compiler.h> > +#include <linux/bitops.h> > +#include <linux/spinlock_types.h> > +#include <linux/types.h> > +#include <linux/completion.h> > +#include <linux/wait.h> > +#include <linux/irqreturn.h> > +#include <asm/io.h> > + > +/* registers */ > +/* version */ > +#define CQHCI_VER 0x00 > +#define CQHCI_VER_MAJOR(x) (((x) & GENMASK(11, 8)) >> 8) > +#define CQHCI_VER_MINOR1(x) (((x) & GENMASK(7, 4)) >> 4) > +#define CQHCI_VER_MINOR2(x) ((x) & GENMASK(3, 0)) > + > +/* capabilities */ > +#define CQHCI_CAP 0x04 > +/* configuration */ > +#define CQHCI_CFG 0x08 > +#define CQHCI_DCMD 0x00001000 > +#define CQHCI_TASK_DESC_SZ 0x00000100 > +#define CQHCI_ENABLE 0x00000001 > + > +/* control */ > +#define CQHCI_CTL 0x0C > +#define CQHCI_CLEAR_ALL_TASKS 0x00000100 > +#define CQHCI_HALT 0x00000001 > + > +/* interrupt status */ > +#define CQHCI_IS 0x10 > +#define CQHCI_IS_HAC BIT(0) > +#define CQHCI_IS_TCC BIT(1) > +#define CQHCI_IS_RED BIT(2) > +#define CQHCI_IS_TCL BIT(3) > + > +#define CQHCI_IS_MASK (CQHCI_IS_TCC | CQHCI_IS_RED) > + > +/* interrupt status enable */ > +#define CQHCI_ISTE 0x14 > + > +/* interrupt signal enable */ > +#define CQHCI_ISGE 0x18 > + > +/* interrupt coalescing */ > +#define CQHCI_IC 0x1C > +#define CQHCI_IC_ENABLE BIT(31) > +#define CQHCI_IC_RESET BIT(16) > +#define CQHCI_IC_ICCTHWEN BIT(15) > +#define CQHCI_IC_ICCTH(x) ((x & 0x1F) << 8) > +#define CQHCI_IC_ICTOVALWEN BIT(7) > +#define CQHCI_IC_ICTOVAL(x) (x & 0x7F) > + > +/* task list base address */ > +#define CQHCI_TDLBA 0x20 > + > +/* task list base address upper */ > +#define CQHCI_TDLBAU 0x24 > + > +/* door-bell */ > +#define CQHCI_TDBR 0x28 > + > +/* task completion notification */ > +#define CQHCI_TCN 0x2C > + > +/* device queue status */ > +#define CQHCI_DQS 0x30 > + > +/* device pending tasks */ > +#define CQHCI_DPT 0x34 > + > +/* task clear */ > +#define CQHCI_TCLR 0x38 > + > +/* send status config 1 */ > +#define CQHCI_SSC1 0x40 > + > +/* send status config 2 */ > +#define CQHCI_SSC2 0x44 > + > +/* response for dcmd */ > +#define CQHCI_CRDCT 0x48 > + > +/* response mode error mask */ > +#define CQHCI_RMEM 0x50 > + > +/* task error info */ > +#define CQHCI_TERRI 0x54 > + > +#define CQHCI_TERRI_C_INDEX(x) ((x) & GENMASK(5, 0)) > +#define CQHCI_TERRI_C_TASK(x) (((x) & GENMASK(12, 8)) >> 8) > +#define CQHCI_TERRI_C_VALID(x) ((x) & BIT(15)) > +#define CQHCI_TERRI_D_INDEX(x) (((x) & GENMASK(21, 16)) >> 16) > +#define CQHCI_TERRI_D_TASK(x) (((x) & GENMASK(28, 24)) >> 24) > +#define CQHCI_TERRI_D_VALID(x) ((x) & BIT(31)) > + > +/* command response index */ > +#define CQHCI_CRI 0x58 > + > +/* command response argument */ > +#define CQHCI_CRA 0x5C > + > +#define CQHCI_INT_ALL 0xF > +#define CQHCI_IC_DEFAULT_ICCTH 31 > +#define CQHCI_IC_DEFAULT_ICTOVAL 1 > + > +/* attribute fields */ > +#define CQHCI_VALID(x) ((x & 1) << 0) > +#define CQHCI_END(x) ((x & 1) << 1) > +#define CQHCI_INT(x) ((x & 1) << 2) > +#define CQHCI_ACT(x) ((x & 0x7) << 3) > + > +/* data command task descriptor fields */ > +#define CQHCI_FORCED_PROG(x) ((x & 1) << 6) > +#define CQHCI_CONTEXT(x) ((x & 0xF) << 7) > +#define CQHCI_DATA_TAG(x) ((x & 1) << 11) > +#define CQHCI_DATA_DIR(x) ((x & 1) << 12) > +#define CQHCI_PRIORITY(x) ((x & 1) << 13) > +#define CQHCI_QBAR(x) ((x & 1) << 14) > +#define CQHCI_REL_WRITE(x) ((x & 1) << 15) > +#define CQHCI_BLK_COUNT(x) ((x & 0xFFFF) << 16) > +#define CQHCI_BLK_ADDR(x) ((x & 0xFFFFFFFF) << 32) > + > +/* direct command task descriptor fields */ > +#define CQHCI_CMD_INDEX(x) ((x & 0x3F) << 16) > +#define CQHCI_CMD_TIMING(x) ((x & 1) << 22) > +#define CQHCI_RESP_TYPE(x) ((x & 0x3) << 23) > + > +/* transfer descriptor fields */ > +#define CQHCI_DAT_LENGTH(x) ((x & 0xFFFF) << 16) > +#define CQHCI_DAT_ADDR_LO(x) ((x & 0xFFFFFFFF) << 32) > +#define CQHCI_DAT_ADDR_HI(x) ((x & 0xFFFFFFFF) << 0) > + > +struct cqhci_host_ops; > +struct mmc_host; > +struct cqhci_slot; > + > +struct cqhci_host { > + const struct cqhci_host_ops *ops; > + void __iomem *mmio; > + struct mmc_host *mmc; > + > + spinlock_t lock; > + > + /* relative card address of device */ > + unsigned int rca; > + > + /* 64 bit DMA */ > + bool dma64; > + int num_slots; > + int qcnt; > + > + u32 dcmd_slot; > + u32 caps; > +#define CQHCI_TASK_DESC_SZ_128 0x1 > + > + u32 quirks; > +#define CQHCI_QUIRK_SHORT_TXFR_DESC_SZ 0x1 > + > + bool enabled; > + bool halted; > + bool init_done; > + bool activated; > + bool waiting_for_idle; > + bool recovery_halt; > + > + size_t desc_size; > + size_t data_size; > + > + u8 *desc_base; > + > + /* total descriptor size */ > + u8 slot_sz; > + > + /* 64/128 bit depends on CQHCI_CFG */ > + u8 task_desc_len; > + > + /* 64 bit on 32-bit arch, 128 bit on 64-bit */ > + u8 link_desc_len; > + > + u8 *trans_desc_base; > + /* same length as transfer descriptor */ > + u8 trans_desc_len; > + > + dma_addr_t desc_dma_base; > + dma_addr_t trans_desc_dma_base; > + > + struct completion halt_comp; > + wait_queue_head_t wait_queue; > + struct cqhci_slot *slot; > +}; > + > +struct cqhci_host_ops { > + void (*dumpregs)(struct mmc_host *mmc); > + void (*write_l)(struct cqhci_host *host, u32 val, int reg); > + u32 (*read_l)(struct cqhci_host *host, int reg); > + void (*enable)(struct mmc_host *mmc); > + void (*disable)(struct mmc_host *mmc, bool recovery); > +}; > + > +static inline void cqhci_writel(struct cqhci_host *host, u32 val, int reg) > +{ > + if (unlikely(host->ops->write_l)) > + host->ops->write_l(host, val, reg); > + else > + writel_relaxed(val, host->mmio + reg); > +} > + > +static inline u32 cqhci_readl(struct cqhci_host *host, int reg) > +{ > + if (unlikely(host->ops->read_l)) > + return host->ops->read_l(host, reg); > + else > + return readl_relaxed(host->mmio + reg); > +} > + > +struct platform_device; > + > +irqreturn_t cqhci_irq(struct mmc_host *mmc, u32 intmask, int cmd_error, > + int data_error); > +int cqhci_init(struct cqhci_host *cq_host, struct mmc_host *mmc, bool dma64); > +struct cqhci_host *cqhci_pltfm_init(struct platform_device *pdev); > +int cqhci_suspend(struct mmc_host *mmc); > +int cqhci_resume(struct mmc_host *mmc); > + > +#endif > -- > 1.9.1 >
On 09/11/17 15:41, Ulf Hansson wrote: > On 3 November 2017 at 14:20, Adrian Hunter <adrian.hunter@intel.com> wrote: >> From: Venkat Gopalakrishnan <venkatg@codeaurora.org> >> >> This patch adds CMDQ support for command-queue compatible >> hosts. >> >> Command queue is added in eMMC-5.1 specification. This >> enables the controller to process upto 32 requests at >> a time. >> >> Adrian Hunter contributed renaming to cqhci, recovery, suspend >> and resume, cqhci_off, cqhci_wait_for_idle, and external timeout >> handling. >> >> Signed-off-by: Asutosh Das <asutoshd@codeaurora.org> >> Signed-off-by: Sujit Reddy Thumma <sthumma@codeaurora.org> >> Signed-off-by: Konstantin Dorfman <kdorfman@codeaurora.org> >> Signed-off-by: Venkat Gopalakrishnan <venkatg@codeaurora.org> >> Signed-off-by: Subhash Jadavani <subhashj@codeaurora.org> >> Signed-off-by: Ritesh Harjani <riteshh@codeaurora.org> >> Signed-off-by: Adrian Hunter <adrian.hunter@intel.com> > > Overall this looks good to me. > > However, I didn't see MAINTAINERS being updated. Is anybody above > volunteering to maintain cqhci.*? I have hardware that I can use for testing so I can maintain it if no one else wants to, however if anyone else is keen please feel free to offer too.
On Thu, Nov 9, 2017 at 1:55 PM, Adrian Hunter <adrian.hunter@intel.com> wrote: > On 09/11/17 14:26, Linus Walleij wrote: >>>> I think the above approach to put any CQE-specific callbacks >>>> directly into the struct mmc_host_ops is way more viable. >>> >>> Nothing to do with CQE. This is CQHCI. Please try to get the difference. >> >> I am trying, please try to think about your language. > > I strongly disapprove of being rude but sadly it seems to get results. Conflict is a natural part of human existence. Using it deliberately to get one's way is manipulation. Admitting to being manipulative is losing one's face in public. Yours, Linus Walleij
diff --git a/drivers/mmc/host/Kconfig b/drivers/mmc/host/Kconfig index 567028c9219a..3092b7085cb5 100644 --- a/drivers/mmc/host/Kconfig +++ b/drivers/mmc/host/Kconfig @@ -857,6 +857,19 @@ config MMC_SUNXI This selects support for the SD/MMC Host Controller on Allwinner sunxi SoCs. +config MMC_CQHCI + tristate "Command Queue Host Controller Interface support" + depends on HAS_DMA + help + This selects the Command Queue Host Controller Interface (CQHCI) + support present in host controllers of Qualcomm Technologies, Inc + amongst others. + This controller supports eMMC devices with command queue support. + + If you have a controller with this interface, say Y or M here. + + If unsure, say N. + config MMC_TOSHIBA_PCI tristate "Toshiba Type A SD/MMC Card Interface Driver" depends on PCI diff --git a/drivers/mmc/host/Makefile b/drivers/mmc/host/Makefile index ab61a3e39c0b..de140e3ef402 100644 --- a/drivers/mmc/host/Makefile +++ b/drivers/mmc/host/Makefile @@ -91,6 +91,7 @@ obj-$(CONFIG_MMC_SDHCI_ST) += sdhci-st.o obj-$(CONFIG_MMC_SDHCI_MICROCHIP_PIC32) += sdhci-pic32.o obj-$(CONFIG_MMC_SDHCI_BRCMSTB) += sdhci-brcmstb.o obj-$(CONFIG_MMC_SDHCI_OMAP) += sdhci-omap.o +obj-$(CONFIG_MMC_CQHCI) += cqhci.o ifeq ($(CONFIG_CB710_DEBUG),y) CFLAGS-cb710-mmc += -DDEBUG diff --git a/drivers/mmc/host/cqhci.c b/drivers/mmc/host/cqhci.c new file mode 100644 index 000000000000..159270e947cf --- /dev/null +++ b/drivers/mmc/host/cqhci.c @@ -0,0 +1,1150 @@ +/* Copyright (c) 2015, The Linux Foundation. All rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License version 2 and + * only version 2 as published by the Free Software Foundation. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + */ + +#include <linux/delay.h> +#include <linux/highmem.h> +#include <linux/io.h> +#include <linux/module.h> +#include <linux/dma-mapping.h> +#include <linux/slab.h> +#include <linux/scatterlist.h> +#include <linux/platform_device.h> +#include <linux/ktime.h> + +#include <linux/mmc/mmc.h> +#include <linux/mmc/host.h> +#include <linux/mmc/card.h> + +#include "cqhci.h" + +#define DCMD_SLOT 31 +#define NUM_SLOTS 32 + +struct cqhci_slot { + struct mmc_request *mrq; + unsigned int flags; +#define CQHCI_EXTERNAL_TIMEOUT BIT(0) +#define CQHCI_COMPLETED BIT(1) +#define CQHCI_HOST_CRC BIT(2) +#define CQHCI_HOST_TIMEOUT BIT(3) +#define CQHCI_HOST_OTHER BIT(4) +}; + +static inline u8 *get_desc(struct cqhci_host *cq_host, u8 tag) +{ + return cq_host->desc_base + (tag * cq_host->slot_sz); +} + +static inline u8 *get_link_desc(struct cqhci_host *cq_host, u8 tag) +{ + u8 *desc = get_desc(cq_host, tag); + + return desc + cq_host->task_desc_len; +} + +static inline dma_addr_t get_trans_desc_dma(struct cqhci_host *cq_host, u8 tag) +{ + return cq_host->trans_desc_dma_base + + (cq_host->mmc->max_segs * tag * + cq_host->trans_desc_len); +} + +static inline u8 *get_trans_desc(struct cqhci_host *cq_host, u8 tag) +{ + return cq_host->trans_desc_base + + (cq_host->trans_desc_len * cq_host->mmc->max_segs * tag); +} + +static void setup_trans_desc(struct cqhci_host *cq_host, u8 tag) +{ + u8 *link_temp; + dma_addr_t trans_temp; + + link_temp = get_link_desc(cq_host, tag); + trans_temp = get_trans_desc_dma(cq_host, tag); + + memset(link_temp, 0, cq_host->link_desc_len); + if (cq_host->link_desc_len > 8) + *(link_temp + 8) = 0; + + if (tag == DCMD_SLOT && (cq_host->mmc->caps2 & MMC_CAP2_CQE_DCMD)) { + *link_temp = CQHCI_VALID(0) | CQHCI_ACT(0) | CQHCI_END(1); + return; + } + + *link_temp = CQHCI_VALID(1) | CQHCI_ACT(0x6) | CQHCI_END(0); + + if (cq_host->dma64) { + __le64 *data_addr = (__le64 __force *)(link_temp + 4); + + data_addr[0] = cpu_to_le64(trans_temp); + } else { + __le32 *data_addr = (__le32 __force *)(link_temp + 4); + + data_addr[0] = cpu_to_le32(trans_temp); + } +} + +static void cqhci_set_irqs(struct cqhci_host *cq_host, u32 set) +{ + cqhci_writel(cq_host, set, CQHCI_ISTE); + cqhci_writel(cq_host, set, CQHCI_ISGE); +} + +#define DRV_NAME "cqhci" + +#define CQHCI_DUMP(f, x...) \ + pr_err("%s: " DRV_NAME ": " f, mmc_hostname(mmc), ## x) + +static void cqhci_dumpregs(struct cqhci_host *cq_host) +{ + struct mmc_host *mmc = cq_host->mmc; + + CQHCI_DUMP("============ CQHCI REGISTER DUMP ===========\n"); + + CQHCI_DUMP("Caps: 0x%08x | Version: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_CAP), + cqhci_readl(cq_host, CQHCI_VER)); + CQHCI_DUMP("Config: 0x%08x | Control: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_CFG), + cqhci_readl(cq_host, CQHCI_CTL)); + CQHCI_DUMP("Int stat: 0x%08x | Int enab: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_IS), + cqhci_readl(cq_host, CQHCI_ISTE)); + CQHCI_DUMP("Int sig: 0x%08x | Int Coal: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_ISGE), + cqhci_readl(cq_host, CQHCI_IC)); + CQHCI_DUMP("TDL base: 0x%08x | TDL up32: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_TDLBA), + cqhci_readl(cq_host, CQHCI_TDLBAU)); + CQHCI_DUMP("Doorbell: 0x%08x | TCN: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_TDBR), + cqhci_readl(cq_host, CQHCI_TCN)); + CQHCI_DUMP("Dev queue: 0x%08x | Dev Pend: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_DQS), + cqhci_readl(cq_host, CQHCI_DPT)); + CQHCI_DUMP("Task clr: 0x%08x | SSC1: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_TCLR), + cqhci_readl(cq_host, CQHCI_SSC1)); + CQHCI_DUMP("SSC2: 0x%08x | DCMD rsp: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_SSC2), + cqhci_readl(cq_host, CQHCI_CRDCT)); + CQHCI_DUMP("RED mask: 0x%08x | TERRI: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_RMEM), + cqhci_readl(cq_host, CQHCI_TERRI)); + CQHCI_DUMP("Resp idx: 0x%08x | Resp arg: 0x%08x\n", + cqhci_readl(cq_host, CQHCI_CRI), + cqhci_readl(cq_host, CQHCI_CRA)); + + if (cq_host->ops->dumpregs) + cq_host->ops->dumpregs(mmc); + else + CQHCI_DUMP(": ===========================================\n"); +} + +/** + * The allocated descriptor table for task, link & transfer descritors + * looks like: + * |----------| + * |task desc | |->|----------| + * |----------| | |trans desc| + * |link desc-|->| |----------| + * |----------| . + * . . + * no. of slots max-segs + * . |----------| + * |----------| + * The idea here is to create the [task+trans] table and mark & point the + * link desc to the transfer desc table on a per slot basis. + */ +static int cqhci_host_alloc_tdl(struct cqhci_host *cq_host) +{ + int i = 0; + + /* task descriptor can be 64/128 bit irrespective of arch */ + if (cq_host->caps & CQHCI_TASK_DESC_SZ_128) { + cqhci_writel(cq_host, cqhci_readl(cq_host, CQHCI_CFG) | + CQHCI_TASK_DESC_SZ, CQHCI_CFG); + cq_host->task_desc_len = 16; + } else { + cq_host->task_desc_len = 8; + } + + /* + * 96 bits length of transfer desc instead of 128 bits which means + * ADMA would expect next valid descriptor at the 96th bit + * or 128th bit + */ + if (cq_host->dma64) { + if (cq_host->quirks & CQHCI_QUIRK_SHORT_TXFR_DESC_SZ) + cq_host->trans_desc_len = 12; + else + cq_host->trans_desc_len = 16; + cq_host->link_desc_len = 16; + } else { + cq_host->trans_desc_len = 8; + cq_host->link_desc_len = 8; + } + + /* total size of a slot: 1 task & 1 transfer (link) */ + cq_host->slot_sz = cq_host->task_desc_len + cq_host->link_desc_len; + + cq_host->desc_size = cq_host->slot_sz * cq_host->num_slots; + + cq_host->data_size = cq_host->trans_desc_len * cq_host->mmc->max_segs * + (cq_host->num_slots - 1); + + pr_debug("%s: cqhci: desc_size: %zu data_sz: %zu slot-sz: %d\n", + mmc_hostname(cq_host->mmc), cq_host->desc_size, cq_host->data_size, + cq_host->slot_sz); + + /* + * allocate a dma-mapped chunk of memory for the descriptors + * allocate a dma-mapped chunk of memory for link descriptors + * setup each link-desc memory offset per slot-number to + * the descriptor table. + */ + cq_host->desc_base = dmam_alloc_coherent(mmc_dev(cq_host->mmc), + cq_host->desc_size, + &cq_host->desc_dma_base, + GFP_KERNEL); + cq_host->trans_desc_base = dmam_alloc_coherent(mmc_dev(cq_host->mmc), + cq_host->data_size, + &cq_host->trans_desc_dma_base, + GFP_KERNEL); + if (!cq_host->desc_base || !cq_host->trans_desc_base) + return -ENOMEM; + + pr_debug("%s: cqhci: desc-base: 0x%p trans-base: 0x%p\n desc_dma 0x%llx trans_dma: 0x%llx\n", + mmc_hostname(cq_host->mmc), cq_host->desc_base, cq_host->trans_desc_base, + (unsigned long long)cq_host->desc_dma_base, + (unsigned long long)cq_host->trans_desc_dma_base); + + for (; i < (cq_host->num_slots); i++) + setup_trans_desc(cq_host, i); + + return 0; +} + +static void __cqhci_enable(struct cqhci_host *cq_host) +{ + struct mmc_host *mmc = cq_host->mmc; + u32 cqcfg; + + cqcfg = cqhci_readl(cq_host, CQHCI_CFG); + + /* Configuration must not be changed while enabled */ + if (cqcfg & CQHCI_ENABLE) { + cqcfg &= ~CQHCI_ENABLE; + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); + } + + cqcfg &= ~(CQHCI_DCMD | CQHCI_TASK_DESC_SZ); + + if (mmc->caps2 & MMC_CAP2_CQE_DCMD) + cqcfg |= CQHCI_DCMD; + + if (cq_host->caps & CQHCI_TASK_DESC_SZ_128) + cqcfg |= CQHCI_TASK_DESC_SZ; + + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); + + cqhci_writel(cq_host, lower_32_bits(cq_host->desc_dma_base), + CQHCI_TDLBA); + cqhci_writel(cq_host, upper_32_bits(cq_host->desc_dma_base), + CQHCI_TDLBAU); + + cqhci_writel(cq_host, cq_host->rca, CQHCI_SSC2); + + cqhci_set_irqs(cq_host, 0); + + cqcfg |= CQHCI_ENABLE; + + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); + + mmc->cqe_on = true; + + if (cq_host->ops->enable) + cq_host->ops->enable(mmc); + + /* Ensure all writes are done before interrupts are enabled */ + wmb(); + + cqhci_set_irqs(cq_host, CQHCI_IS_MASK); + + cq_host->activated = true; +} + +static void __cqhci_disable(struct cqhci_host *cq_host) +{ + u32 cqcfg; + + cqcfg = cqhci_readl(cq_host, CQHCI_CFG); + cqcfg &= ~CQHCI_ENABLE; + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); + + cq_host->mmc->cqe_on = false; + + cq_host->activated = false; +} + +int cqhci_suspend(struct mmc_host *mmc) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + + if (cq_host->enabled) + __cqhci_disable(cq_host); + + return 0; +} +EXPORT_SYMBOL(cqhci_suspend); + +int cqhci_resume(struct mmc_host *mmc) +{ + /* Re-enable is done upon first request */ + return 0; +} +EXPORT_SYMBOL(cqhci_resume); + +static int cqhci_enable(struct mmc_host *mmc, struct mmc_card *card) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + int err; + + if (cq_host->enabled) + return 0; + + cq_host->rca = card->rca; + + err = cqhci_host_alloc_tdl(cq_host); + if (err) + return err; + + __cqhci_enable(cq_host); + + cq_host->enabled = true; + +#ifdef DEBUG + cqhci_dumpregs(cq_host); +#endif + return 0; +} + +/* CQHCI is idle and should halt immediately, so set a small timeout */ +#define CQHCI_OFF_TIMEOUT 100 + +static void cqhci_off(struct mmc_host *mmc) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + ktime_t timeout; + bool timed_out; + u32 reg; + + if (!cq_host->enabled || !mmc->cqe_on || cq_host->recovery_halt) + return; + + if (cq_host->ops->disable) + cq_host->ops->disable(mmc, false); + + cqhci_writel(cq_host, CQHCI_HALT, CQHCI_CTL); + + timeout = ktime_add_us(ktime_get(), CQHCI_OFF_TIMEOUT); + while (1) { + timed_out = ktime_compare(ktime_get(), timeout) > 0; + reg = cqhci_readl(cq_host, CQHCI_CTL); + if ((reg & CQHCI_HALT) || timed_out) + break; + } + + if (timed_out) + pr_err("%s: cqhci: CQE stuck on\n", mmc_hostname(mmc)); + else + pr_debug("%s: cqhci: CQE off\n", mmc_hostname(mmc)); + + mmc->cqe_on = false; +} + +static void cqhci_disable(struct mmc_host *mmc) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + + if (!cq_host->enabled) + return; + + cqhci_off(mmc); + + __cqhci_disable(cq_host); + + dmam_free_coherent(mmc_dev(mmc), cq_host->data_size, + cq_host->trans_desc_base, + cq_host->trans_desc_dma_base); + + dmam_free_coherent(mmc_dev(mmc), cq_host->desc_size, + cq_host->desc_base, + cq_host->desc_dma_base); + + cq_host->trans_desc_base = NULL; + cq_host->desc_base = NULL; + + cq_host->enabled = false; +} + +static void cqhci_prep_task_desc(struct mmc_request *mrq, + u64 *data, bool intr) +{ + u32 req_flags = mrq->data->flags; + + *data = CQHCI_VALID(1) | + CQHCI_END(1) | + CQHCI_INT(intr) | + CQHCI_ACT(0x5) | + CQHCI_FORCED_PROG(!!(req_flags & MMC_DATA_FORCED_PRG)) | + CQHCI_DATA_TAG(!!(req_flags & MMC_DATA_DAT_TAG)) | + CQHCI_DATA_DIR(!!(req_flags & MMC_DATA_READ)) | + CQHCI_PRIORITY(!!(req_flags & MMC_DATA_PRIO)) | + CQHCI_QBAR(!!(req_flags & MMC_DATA_QBR)) | + CQHCI_REL_WRITE(!!(req_flags & MMC_DATA_REL_WR)) | + CQHCI_BLK_COUNT(mrq->data->blocks) | + CQHCI_BLK_ADDR((u64)mrq->data->blk_addr); + + pr_debug("%s: cqhci: tag %d task descriptor 0x016%llx\n", + mmc_hostname(mrq->host), mrq->tag, (unsigned long long)*data); +} + +static int cqhci_dma_map(struct mmc_host *host, struct mmc_request *mrq) +{ + int sg_count; + struct mmc_data *data = mrq->data; + + if (!data) + return -EINVAL; + + sg_count = dma_map_sg(mmc_dev(host), data->sg, + data->sg_len, + (data->flags & MMC_DATA_WRITE) ? + DMA_TO_DEVICE : DMA_FROM_DEVICE); + if (!sg_count) { + pr_err("%s: sg-len: %d\n", __func__, data->sg_len); + return -ENOMEM; + } + + return sg_count; +} + +static void cqhci_set_tran_desc(u8 *desc, dma_addr_t addr, int len, bool end, + bool dma64) +{ + __le32 *attr = (__le32 __force *)desc; + + *attr = (CQHCI_VALID(1) | + CQHCI_END(end ? 1 : 0) | + CQHCI_INT(0) | + CQHCI_ACT(0x4) | + CQHCI_DAT_LENGTH(len)); + + if (dma64) { + __le64 *dataddr = (__le64 __force *)(desc + 4); + + dataddr[0] = cpu_to_le64(addr); + } else { + __le32 *dataddr = (__le32 __force *)(desc + 4); + + dataddr[0] = cpu_to_le32(addr); + } +} + +static int cqhci_prep_tran_desc(struct mmc_request *mrq, + struct cqhci_host *cq_host, int tag) +{ + struct mmc_data *data = mrq->data; + int i, sg_count, len; + bool end = false; + bool dma64 = cq_host->dma64; + dma_addr_t addr; + u8 *desc; + struct scatterlist *sg; + + sg_count = cqhci_dma_map(mrq->host, mrq); + if (sg_count < 0) { + pr_err("%s: %s: unable to map sg lists, %d\n", + mmc_hostname(mrq->host), __func__, sg_count); + return sg_count; + } + + desc = get_trans_desc(cq_host, tag); + + for_each_sg(data->sg, sg, sg_count, i) { + addr = sg_dma_address(sg); + len = sg_dma_len(sg); + + if ((i+1) == sg_count) + end = true; + cqhci_set_tran_desc(desc, addr, len, end, dma64); + desc += cq_host->trans_desc_len; + } + + return 0; +} + +static void cqhci_prep_dcmd_desc(struct mmc_host *mmc, + struct mmc_request *mrq) +{ + u64 *task_desc = NULL; + u64 data = 0; + u8 resp_type; + u8 *desc; + __le64 *dataddr; + struct cqhci_host *cq_host = mmc->cqe_private; + u8 timing; + + if (!(mrq->cmd->flags & MMC_RSP_PRESENT)) { + resp_type = 0x0; + timing = 0x1; + } else { + if (mrq->cmd->flags & MMC_RSP_R1B) { + resp_type = 0x3; + timing = 0x0; + } else { + resp_type = 0x2; + timing = 0x1; + } + } + + task_desc = (__le64 __force *)get_desc(cq_host, cq_host->dcmd_slot); + memset(task_desc, 0, cq_host->task_desc_len); + data |= (CQHCI_VALID(1) | + CQHCI_END(1) | + CQHCI_INT(1) | + CQHCI_QBAR(1) | + CQHCI_ACT(0x5) | + CQHCI_CMD_INDEX(mrq->cmd->opcode) | + CQHCI_CMD_TIMING(timing) | CQHCI_RESP_TYPE(resp_type)); + *task_desc |= data; + desc = (u8 *)task_desc; + pr_debug("%s: cqhci: dcmd: cmd: %d timing: %d resp: %d\n", + mmc_hostname(mmc), mrq->cmd->opcode, timing, resp_type); + dataddr = (__le64 __force *)(desc + 4); + dataddr[0] = cpu_to_le64((u64)mrq->cmd->arg); + +} + +static void cqhci_post_req(struct mmc_host *host, struct mmc_request *mrq) +{ + struct mmc_data *data = mrq->data; + + if (data) { + dma_unmap_sg(mmc_dev(host), data->sg, data->sg_len, + (data->flags & MMC_DATA_READ) ? + DMA_FROM_DEVICE : DMA_TO_DEVICE); + } +} + +static inline int cqhci_tag(struct mmc_request *mrq) +{ + return mrq->cmd ? DCMD_SLOT : mrq->tag; +} + +static int cqhci_request(struct mmc_host *mmc, struct mmc_request *mrq) +{ + int err = 0; + u64 data = 0; + u64 *task_desc = NULL; + int tag = cqhci_tag(mrq); + struct cqhci_host *cq_host = mmc->cqe_private; + unsigned long flags; + + if (!cq_host->enabled) { + pr_err("%s: cqhci: not enabled\n", mmc_hostname(mmc)); + return -EINVAL; + } + + /* First request after resume has to re-enable */ + if (!cq_host->activated) + __cqhci_enable(cq_host); + + if (!mmc->cqe_on) { + cqhci_writel(cq_host, 0, CQHCI_CTL); + mmc->cqe_on = true; + pr_debug("%s: cqhci: CQE on\n", mmc_hostname(mmc)); + if (cqhci_readl(cq_host, CQHCI_CTL) && CQHCI_HALT) { + pr_err("%s: cqhci: CQE failed to exit halt state\n", + mmc_hostname(mmc)); + } + if (cq_host->ops->enable) + cq_host->ops->enable(mmc); + } + + if (mrq->data) { + task_desc = (__le64 __force *)get_desc(cq_host, tag); + cqhci_prep_task_desc(mrq, &data, 1); + *task_desc = cpu_to_le64(data); + err = cqhci_prep_tran_desc(mrq, cq_host, tag); + if (err) { + pr_err("%s: cqhci: failed to setup tx desc: %d\n", + mmc_hostname(mmc), err); + return err; + } + } else { + cqhci_prep_dcmd_desc(mmc, mrq); + } + + spin_lock_irqsave(&cq_host->lock, flags); + + if (cq_host->recovery_halt) { + err = -EBUSY; + goto out_unlock; + } + + cq_host->slot[tag].mrq = mrq; + cq_host->slot[tag].flags = 0; + + cq_host->qcnt += 1; + + cqhci_writel(cq_host, 1 << tag, CQHCI_TDBR); + if (!(cqhci_readl(cq_host, CQHCI_TDBR) & (1 << tag))) + pr_debug("%s: cqhci: doorbell not set for tag %d\n", + mmc_hostname(mmc), tag); +out_unlock: + spin_unlock_irqrestore(&cq_host->lock, flags); + + if (err) + cqhci_post_req(mmc, mrq); + + return err; +} + +static void cqhci_recovery_needed(struct mmc_host *mmc, struct mmc_request *mrq, + bool notify) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + + if (!cq_host->recovery_halt) { + cq_host->recovery_halt = true; + pr_debug("%s: cqhci: recovery needed\n", mmc_hostname(mmc)); + wake_up(&cq_host->wait_queue); + if (notify && mrq->recovery_notifier) + mrq->recovery_notifier(mrq); + } +} + +static unsigned int cqhci_error_flags(int error1, int error2) +{ + int error = error1 ? error1 : error2; + + switch (error) { + case -EILSEQ: + return CQHCI_HOST_CRC; + case -ETIMEDOUT: + return CQHCI_HOST_TIMEOUT; + default: + return CQHCI_HOST_OTHER; + } +} + +static void cqhci_error_irq(struct mmc_host *mmc, u32 status, int cmd_error, + int data_error) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + struct cqhci_slot *slot; + u32 terri; + int tag; + + spin_lock(&cq_host->lock); + + terri = cqhci_readl(cq_host, CQHCI_TERRI); + + pr_debug("%s: cqhci: error IRQ status: 0x%08x cmd error %d data error %d TERRI: 0x%08x\n", + mmc_hostname(mmc), status, cmd_error, data_error, terri); + + /* Forget about errors when recovery has already been triggered */ + if (cq_host->recovery_halt) + goto out_unlock; + + if (!cq_host->qcnt) { + WARN_ONCE(1, "%s: cqhci: error when idle. IRQ status: 0x%08x cmd error %d data error %d TERRI: 0x%08x\n", + mmc_hostname(mmc), status, cmd_error, data_error, + terri); + goto out_unlock; + } + + if (CQHCI_TERRI_C_VALID(terri)) { + tag = CQHCI_TERRI_C_TASK(terri); + slot = &cq_host->slot[tag]; + if (slot->mrq) { + slot->flags = cqhci_error_flags(cmd_error, data_error); + cqhci_recovery_needed(mmc, slot->mrq, true); + } + } + + if (CQHCI_TERRI_D_VALID(terri)) { + tag = CQHCI_TERRI_D_TASK(terri); + slot = &cq_host->slot[tag]; + if (slot->mrq) { + slot->flags = cqhci_error_flags(data_error, cmd_error); + cqhci_recovery_needed(mmc, slot->mrq, true); + } + } + + if (!cq_host->recovery_halt) { + /* + * The only way to guarantee forward progress is to mark at + * least one task in error, so if none is indicated, pick one. + */ + for (tag = 0; tag < NUM_SLOTS; tag++) { + slot = &cq_host->slot[tag]; + if (!slot->mrq) + continue; + slot->flags = cqhci_error_flags(data_error, cmd_error); + cqhci_recovery_needed(mmc, slot->mrq, true); + break; + } + } + +out_unlock: + spin_unlock(&cq_host->lock); +} + +static void cqhci_finish_mrq(struct mmc_host *mmc, unsigned int tag) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + struct cqhci_slot *slot = &cq_host->slot[tag]; + struct mmc_request *mrq = slot->mrq; + struct mmc_data *data; + + if (!mrq) { + WARN_ONCE(1, "%s: cqhci: spurious TCN for tag %d\n", + mmc_hostname(mmc), tag); + return; + } + + /* No completions allowed during recovery */ + if (cq_host->recovery_halt) { + slot->flags |= CQHCI_COMPLETED; + return; + } + + slot->mrq = NULL; + + cq_host->qcnt -= 1; + + data = mrq->data; + if (data) { + if (data->error) + data->bytes_xfered = 0; + else + data->bytes_xfered = data->blksz * data->blocks; + } + + mmc_cqe_request_done(mmc, mrq); +} + +irqreturn_t cqhci_irq(struct mmc_host *mmc, u32 intmask, int cmd_error, + int data_error) +{ + u32 status; + unsigned long tag = 0, comp_status; + struct cqhci_host *cq_host = mmc->cqe_private; + + status = cqhci_readl(cq_host, CQHCI_IS); + cqhci_writel(cq_host, status, CQHCI_IS); + + pr_debug("%s: cqhci: IRQ status: 0x%08x\n", mmc_hostname(mmc), status); + + if ((status & CQHCI_IS_RED) || cmd_error || data_error) + cqhci_error_irq(mmc, status, cmd_error, data_error); + + if (status & CQHCI_IS_TCC) { + /* read TCN and complete the request */ + comp_status = cqhci_readl(cq_host, CQHCI_TCN); + cqhci_writel(cq_host, comp_status, CQHCI_TCN); + pr_debug("%s: cqhci: TCN: 0x%08lx\n", + mmc_hostname(mmc), comp_status); + + spin_lock(&cq_host->lock); + + for_each_set_bit(tag, &comp_status, cq_host->num_slots) { + /* complete the corresponding mrq */ + pr_debug("%s: cqhci: completing tag %lu\n", + mmc_hostname(mmc), tag); + cqhci_finish_mrq(mmc, tag); + } + + if (cq_host->waiting_for_idle && !cq_host->qcnt) { + cq_host->waiting_for_idle = false; + wake_up(&cq_host->wait_queue); + } + + spin_unlock(&cq_host->lock); + } + + if (status & CQHCI_IS_TCL) + wake_up(&cq_host->wait_queue); + + if (status & CQHCI_IS_HAC) + wake_up(&cq_host->wait_queue); + + return IRQ_HANDLED; +} +EXPORT_SYMBOL(cqhci_irq); + +static bool cqhci_is_idle(struct cqhci_host *cq_host, int *ret) +{ + unsigned long flags; + bool is_idle; + + spin_lock_irqsave(&cq_host->lock, flags); + is_idle = !cq_host->qcnt || cq_host->recovery_halt; + *ret = cq_host->recovery_halt ? -EBUSY : 0; + cq_host->waiting_for_idle = !is_idle; + spin_unlock_irqrestore(&cq_host->lock, flags); + + return is_idle; +} + +static int cqhci_wait_for_idle(struct mmc_host *mmc) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + int ret; + + wait_event(cq_host->wait_queue, cqhci_is_idle(cq_host, &ret)); + + return ret; +} + +static bool cqhci_timeout(struct mmc_host *mmc, struct mmc_request *mrq, + bool *recovery_needed) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + int tag = cqhci_tag(mrq); + struct cqhci_slot *slot = &cq_host->slot[tag]; + unsigned long flags; + bool timed_out; + + spin_lock_irqsave(&cq_host->lock, flags); + timed_out = slot->mrq == mrq; + if (timed_out) { + slot->flags |= CQHCI_EXTERNAL_TIMEOUT; + cqhci_recovery_needed(mmc, mrq, false); + *recovery_needed = cq_host->recovery_halt; + } + spin_unlock_irqrestore(&cq_host->lock, flags); + + if (timed_out) { + pr_err("%s: cqhci: timeout for tag %d\n", + mmc_hostname(mmc), tag); + cqhci_dumpregs(cq_host); + } + + return timed_out; +} + +static bool cqhci_tasks_cleared(struct cqhci_host *cq_host) +{ + return !(cqhci_readl(cq_host, CQHCI_CTL) & CQHCI_CLEAR_ALL_TASKS); +} + +static bool cqhci_clear_all_tasks(struct mmc_host *mmc, unsigned int timeout) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + bool ret; + u32 ctl; + + cqhci_set_irqs(cq_host, CQHCI_IS_TCL); + + ctl = cqhci_readl(cq_host, CQHCI_CTL); + ctl |= CQHCI_CLEAR_ALL_TASKS; + cqhci_writel(cq_host, ctl, CQHCI_CTL); + + wait_event_timeout(cq_host->wait_queue, cqhci_tasks_cleared(cq_host), + msecs_to_jiffies(timeout) + 1); + + cqhci_set_irqs(cq_host, 0); + + ret = cqhci_tasks_cleared(cq_host); + + if (!ret) + pr_debug("%s: cqhci: Failed to clear tasks\n", + mmc_hostname(mmc)); + + return ret; +} + +static bool cqhci_halted(struct cqhci_host *cq_host) +{ + return cqhci_readl(cq_host, CQHCI_CTL) & CQHCI_HALT; +} + +static bool cqhci_halt(struct mmc_host *mmc, unsigned int timeout) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + bool ret; + u32 ctl; + + if (cqhci_halted(cq_host)) + return true; + + cqhci_set_irqs(cq_host, CQHCI_IS_HAC); + + ctl = cqhci_readl(cq_host, CQHCI_CTL); + ctl |= CQHCI_HALT; + cqhci_writel(cq_host, ctl, CQHCI_CTL); + + wait_event_timeout(cq_host->wait_queue, cqhci_halted(cq_host), + msecs_to_jiffies(timeout) + 1); + + cqhci_set_irqs(cq_host, 0); + + ret = cqhci_halted(cq_host); + + if (!ret) + pr_debug("%s: cqhci: Failed to halt\n", mmc_hostname(mmc)); + + return ret; +} + +/* + * After halting we expect to be able to use the command line. We interpret the + * failure to halt to mean the data lines might still be in use (and the upper + * layers will need to send a STOP command), so we set the timeout based on a + * generous command timeout. + */ +#define CQHCI_START_HALT_TIMEOUT 5 + +static void cqhci_recovery_start(struct mmc_host *mmc) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + + pr_debug("%s: cqhci: %s\n", mmc_hostname(mmc), __func__); + + WARN_ON(!cq_host->recovery_halt); + + cqhci_halt(mmc, CQHCI_START_HALT_TIMEOUT); + + if (cq_host->ops->disable) + cq_host->ops->disable(mmc, true); + + mmc->cqe_on = false; +} + +static int cqhci_error_from_flags(unsigned int flags) +{ + if (!flags) + return 0; + + /* CRC errors might indicate re-tuning so prefer to report that */ + if (flags & CQHCI_HOST_CRC) + return -EILSEQ; + + if (flags & (CQHCI_EXTERNAL_TIMEOUT | CQHCI_HOST_TIMEOUT)) + return -ETIMEDOUT; + + return -EIO; +} + +static void cqhci_recover_mrq(struct cqhci_host *cq_host, unsigned int tag) +{ + struct cqhci_slot *slot = &cq_host->slot[tag]; + struct mmc_request *mrq = slot->mrq; + struct mmc_data *data; + + if (!mrq) + return; + + slot->mrq = NULL; + + cq_host->qcnt -= 1; + + data = mrq->data; + if (data) { + data->bytes_xfered = 0; + data->error = cqhci_error_from_flags(slot->flags); + } else { + mrq->cmd->error = cqhci_error_from_flags(slot->flags); + } + + mmc_cqe_request_done(cq_host->mmc, mrq); +} + +static void cqhci_recover_mrqs(struct cqhci_host *cq_host) +{ + int i; + + for (i = 0; i < cq_host->num_slots; i++) + cqhci_recover_mrq(cq_host, i); +} + +/* + * By now the command and data lines should be unused so there is no reason for + * CQHCI to take a long time to halt, but if it doesn't halt there could be + * problems clearing tasks, so be generous. + */ +#define CQHCI_FINISH_HALT_TIMEOUT 20 + +/* CQHCI could be expected to clear it's internal state pretty quickly */ +#define CQHCI_CLEAR_TIMEOUT 20 + +static void cqhci_recovery_finish(struct mmc_host *mmc) +{ + struct cqhci_host *cq_host = mmc->cqe_private; + unsigned long flags; + u32 cqcfg; + bool ok; + + pr_debug("%s: cqhci: %s\n", mmc_hostname(mmc), __func__); + + WARN_ON(!cq_host->recovery_halt); + + ok = cqhci_halt(mmc, CQHCI_FINISH_HALT_TIMEOUT); + + if (!cqhci_clear_all_tasks(mmc, CQHCI_CLEAR_TIMEOUT)) + ok = false; + + /* + * The specification contradicts itself, by saying that tasks cannot be + * cleared if CQHCI does not halt, but if CQHCI does not halt, it should + * be disabled/re-enabled, but not to disable before clearing tasks. + * Have a go anyway. + */ + if (!ok) { + pr_debug("%s: cqhci: disable / re-enable\n", mmc_hostname(mmc)); + cqcfg = cqhci_readl(cq_host, CQHCI_CFG); + cqcfg &= ~CQHCI_ENABLE; + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); + cqcfg |= CQHCI_ENABLE; + cqhci_writel(cq_host, cqcfg, CQHCI_CFG); + /* Be sure that there are no tasks */ + ok = cqhci_halt(mmc, CQHCI_FINISH_HALT_TIMEOUT); + if (!cqhci_clear_all_tasks(mmc, CQHCI_CLEAR_TIMEOUT)) + ok = false; + WARN_ON(!ok); + } + + cqhci_recover_mrqs(cq_host); + + WARN_ON(cq_host->qcnt); + + spin_lock_irqsave(&cq_host->lock, flags); + cq_host->qcnt = 0; + cq_host->recovery_halt = false; + mmc->cqe_on = false; + spin_unlock_irqrestore(&cq_host->lock, flags); + + /* Ensure all writes are done before interrupts are re-enabled */ + wmb(); + + cqhci_writel(cq_host, CQHCI_IS_HAC | CQHCI_IS_TCL, CQHCI_IS); + + cqhci_set_irqs(cq_host, CQHCI_IS_MASK); + + pr_debug("%s: cqhci: recovery done\n", mmc_hostname(mmc)); +} + +static const struct mmc_cqe_ops cqhci_cqe_ops = { + .cqe_enable = cqhci_enable, + .cqe_disable = cqhci_disable, + .cqe_request = cqhci_request, + .cqe_post_req = cqhci_post_req, + .cqe_off = cqhci_off, + .cqe_wait_for_idle = cqhci_wait_for_idle, + .cqe_timeout = cqhci_timeout, + .cqe_recovery_start = cqhci_recovery_start, + .cqe_recovery_finish = cqhci_recovery_finish, +}; + +struct cqhci_host *cqhci_pltfm_init(struct platform_device *pdev) +{ + struct cqhci_host *cq_host; + struct resource *cqhci_memres = NULL; + + /* check and setup CMDQ interface */ + cqhci_memres = platform_get_resource_byname(pdev, IORESOURCE_MEM, + "cqhci_mem"); + if (!cqhci_memres) { + dev_dbg(&pdev->dev, "CMDQ not supported\n"); + return ERR_PTR(-EINVAL); + } + + cq_host = devm_kzalloc(&pdev->dev, sizeof(*cq_host), GFP_KERNEL); + if (!cq_host) + return ERR_PTR(-ENOMEM); + cq_host->mmio = devm_ioremap(&pdev->dev, + cqhci_memres->start, + resource_size(cqhci_memres)); + if (!cq_host->mmio) { + dev_err(&pdev->dev, "failed to remap cqhci regs\n"); + return ERR_PTR(-EBUSY); + } + dev_dbg(&pdev->dev, "CMDQ ioremap: done\n"); + + return cq_host; +} +EXPORT_SYMBOL(cqhci_pltfm_init); + +static unsigned int cqhci_ver_major(struct cqhci_host *cq_host) +{ + return CQHCI_VER_MAJOR(cqhci_readl(cq_host, CQHCI_VER)); +} + +static unsigned int cqhci_ver_minor(struct cqhci_host *cq_host) +{ + u32 ver = cqhci_readl(cq_host, CQHCI_VER); + + return CQHCI_VER_MINOR1(ver) * 10 + CQHCI_VER_MINOR2(ver); +} + +int cqhci_init(struct cqhci_host *cq_host, struct mmc_host *mmc, + bool dma64) +{ + int err; + + cq_host->dma64 = dma64; + cq_host->mmc = mmc; + cq_host->mmc->cqe_private = cq_host; + + cq_host->num_slots = NUM_SLOTS; + cq_host->dcmd_slot = DCMD_SLOT; + + mmc->cqe_ops = &cqhci_cqe_ops; + + mmc->cqe_qdepth = NUM_SLOTS; + if (mmc->caps2 & MMC_CAP2_CQE_DCMD) + mmc->cqe_qdepth -= 1; + + cq_host->slot = devm_kcalloc(mmc_dev(mmc), cq_host->num_slots, + sizeof(*cq_host->slot), GFP_KERNEL); + if (!cq_host->slot) { + err = -ENOMEM; + goto out_err; + } + + spin_lock_init(&cq_host->lock); + + init_completion(&cq_host->halt_comp); + init_waitqueue_head(&cq_host->wait_queue); + + pr_info("%s: CQHCI version %u.%02u\n", + mmc_hostname(mmc), cqhci_ver_major(cq_host), + cqhci_ver_minor(cq_host)); + + return 0; + +out_err: + pr_err("%s: CQHCI version %u.%02u failed to initialize, error %d\n", + mmc_hostname(mmc), cqhci_ver_major(cq_host), + cqhci_ver_minor(cq_host), err); + return err; +} +EXPORT_SYMBOL(cqhci_init); + +MODULE_AUTHOR("Venkat Gopalakrishnan <venkatg@codeaurora.org>"); +MODULE_DESCRIPTION("Command Queue Host Controller Interface driver"); +MODULE_LICENSE("GPL v2"); diff --git a/drivers/mmc/host/cqhci.h b/drivers/mmc/host/cqhci.h new file mode 100644 index 000000000000..2d39d361b322 --- /dev/null +++ b/drivers/mmc/host/cqhci.h @@ -0,0 +1,240 @@ +/* Copyright (c) 2015, The Linux Foundation. All rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License version 2 and + * only version 2 as published by the Free Software Foundation. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + */ +#ifndef LINUX_MMC_CQHCI_H +#define LINUX_MMC_CQHCI_H + +#include <linux/compiler.h> +#include <linux/bitops.h> +#include <linux/spinlock_types.h> +#include <linux/types.h> +#include <linux/completion.h> +#include <linux/wait.h> +#include <linux/irqreturn.h> +#include <asm/io.h> + +/* registers */ +/* version */ +#define CQHCI_VER 0x00 +#define CQHCI_VER_MAJOR(x) (((x) & GENMASK(11, 8)) >> 8) +#define CQHCI_VER_MINOR1(x) (((x) & GENMASK(7, 4)) >> 4) +#define CQHCI_VER_MINOR2(x) ((x) & GENMASK(3, 0)) + +/* capabilities */ +#define CQHCI_CAP 0x04 +/* configuration */ +#define CQHCI_CFG 0x08 +#define CQHCI_DCMD 0x00001000 +#define CQHCI_TASK_DESC_SZ 0x00000100 +#define CQHCI_ENABLE 0x00000001 + +/* control */ +#define CQHCI_CTL 0x0C +#define CQHCI_CLEAR_ALL_TASKS 0x00000100 +#define CQHCI_HALT 0x00000001 + +/* interrupt status */ +#define CQHCI_IS 0x10 +#define CQHCI_IS_HAC BIT(0) +#define CQHCI_IS_TCC BIT(1) +#define CQHCI_IS_RED BIT(2) +#define CQHCI_IS_TCL BIT(3) + +#define CQHCI_IS_MASK (CQHCI_IS_TCC | CQHCI_IS_RED) + +/* interrupt status enable */ +#define CQHCI_ISTE 0x14 + +/* interrupt signal enable */ +#define CQHCI_ISGE 0x18 + +/* interrupt coalescing */ +#define CQHCI_IC 0x1C +#define CQHCI_IC_ENABLE BIT(31) +#define CQHCI_IC_RESET BIT(16) +#define CQHCI_IC_ICCTHWEN BIT(15) +#define CQHCI_IC_ICCTH(x) ((x & 0x1F) << 8) +#define CQHCI_IC_ICTOVALWEN BIT(7) +#define CQHCI_IC_ICTOVAL(x) (x & 0x7F) + +/* task list base address */ +#define CQHCI_TDLBA 0x20 + +/* task list base address upper */ +#define CQHCI_TDLBAU 0x24 + +/* door-bell */ +#define CQHCI_TDBR 0x28 + +/* task completion notification */ +#define CQHCI_TCN 0x2C + +/* device queue status */ +#define CQHCI_DQS 0x30 + +/* device pending tasks */ +#define CQHCI_DPT 0x34 + +/* task clear */ +#define CQHCI_TCLR 0x38 + +/* send status config 1 */ +#define CQHCI_SSC1 0x40 + +/* send status config 2 */ +#define CQHCI_SSC2 0x44 + +/* response for dcmd */ +#define CQHCI_CRDCT 0x48 + +/* response mode error mask */ +#define CQHCI_RMEM 0x50 + +/* task error info */ +#define CQHCI_TERRI 0x54 + +#define CQHCI_TERRI_C_INDEX(x) ((x) & GENMASK(5, 0)) +#define CQHCI_TERRI_C_TASK(x) (((x) & GENMASK(12, 8)) >> 8) +#define CQHCI_TERRI_C_VALID(x) ((x) & BIT(15)) +#define CQHCI_TERRI_D_INDEX(x) (((x) & GENMASK(21, 16)) >> 16) +#define CQHCI_TERRI_D_TASK(x) (((x) & GENMASK(28, 24)) >> 24) +#define CQHCI_TERRI_D_VALID(x) ((x) & BIT(31)) + +/* command response index */ +#define CQHCI_CRI 0x58 + +/* command response argument */ +#define CQHCI_CRA 0x5C + +#define CQHCI_INT_ALL 0xF +#define CQHCI_IC_DEFAULT_ICCTH 31 +#define CQHCI_IC_DEFAULT_ICTOVAL 1 + +/* attribute fields */ +#define CQHCI_VALID(x) ((x & 1) << 0) +#define CQHCI_END(x) ((x & 1) << 1) +#define CQHCI_INT(x) ((x & 1) << 2) +#define CQHCI_ACT(x) ((x & 0x7) << 3) + +/* data command task descriptor fields */ +#define CQHCI_FORCED_PROG(x) ((x & 1) << 6) +#define CQHCI_CONTEXT(x) ((x & 0xF) << 7) +#define CQHCI_DATA_TAG(x) ((x & 1) << 11) +#define CQHCI_DATA_DIR(x) ((x & 1) << 12) +#define CQHCI_PRIORITY(x) ((x & 1) << 13) +#define CQHCI_QBAR(x) ((x & 1) << 14) +#define CQHCI_REL_WRITE(x) ((x & 1) << 15) +#define CQHCI_BLK_COUNT(x) ((x & 0xFFFF) << 16) +#define CQHCI_BLK_ADDR(x) ((x & 0xFFFFFFFF) << 32) + +/* direct command task descriptor fields */ +#define CQHCI_CMD_INDEX(x) ((x & 0x3F) << 16) +#define CQHCI_CMD_TIMING(x) ((x & 1) << 22) +#define CQHCI_RESP_TYPE(x) ((x & 0x3) << 23) + +/* transfer descriptor fields */ +#define CQHCI_DAT_LENGTH(x) ((x & 0xFFFF) << 16) +#define CQHCI_DAT_ADDR_LO(x) ((x & 0xFFFFFFFF) << 32) +#define CQHCI_DAT_ADDR_HI(x) ((x & 0xFFFFFFFF) << 0) + +struct cqhci_host_ops; +struct mmc_host; +struct cqhci_slot; + +struct cqhci_host { + const struct cqhci_host_ops *ops; + void __iomem *mmio; + struct mmc_host *mmc; + + spinlock_t lock; + + /* relative card address of device */ + unsigned int rca; + + /* 64 bit DMA */ + bool dma64; + int num_slots; + int qcnt; + + u32 dcmd_slot; + u32 caps; +#define CQHCI_TASK_DESC_SZ_128 0x1 + + u32 quirks; +#define CQHCI_QUIRK_SHORT_TXFR_DESC_SZ 0x1 + + bool enabled; + bool halted; + bool init_done; + bool activated; + bool waiting_for_idle; + bool recovery_halt; + + size_t desc_size; + size_t data_size; + + u8 *desc_base; + + /* total descriptor size */ + u8 slot_sz; + + /* 64/128 bit depends on CQHCI_CFG */ + u8 task_desc_len; + + /* 64 bit on 32-bit arch, 128 bit on 64-bit */ + u8 link_desc_len; + + u8 *trans_desc_base; + /* same length as transfer descriptor */ + u8 trans_desc_len; + + dma_addr_t desc_dma_base; + dma_addr_t trans_desc_dma_base; + + struct completion halt_comp; + wait_queue_head_t wait_queue; + struct cqhci_slot *slot; +}; + +struct cqhci_host_ops { + void (*dumpregs)(struct mmc_host *mmc); + void (*write_l)(struct cqhci_host *host, u32 val, int reg); + u32 (*read_l)(struct cqhci_host *host, int reg); + void (*enable)(struct mmc_host *mmc); + void (*disable)(struct mmc_host *mmc, bool recovery); +}; + +static inline void cqhci_writel(struct cqhci_host *host, u32 val, int reg) +{ + if (unlikely(host->ops->write_l)) + host->ops->write_l(host, val, reg); + else + writel_relaxed(val, host->mmio + reg); +} + +static inline u32 cqhci_readl(struct cqhci_host *host, int reg) +{ + if (unlikely(host->ops->read_l)) + return host->ops->read_l(host, reg); + else + return readl_relaxed(host->mmio + reg); +} + +struct platform_device; + +irqreturn_t cqhci_irq(struct mmc_host *mmc, u32 intmask, int cmd_error, + int data_error); +int cqhci_init(struct cqhci_host *cq_host, struct mmc_host *mmc, bool dma64); +struct cqhci_host *cqhci_pltfm_init(struct platform_device *pdev); +int cqhci_suspend(struct mmc_host *mmc); +int cqhci_resume(struct mmc_host *mmc); + +#endif