Message ID | ca98e05f8a8dabdc6cc5b6e8c3ce50cf4e900fac.1425497218.git.dhdang@apm.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On 04/03/15 19:39, Duc Dang wrote: > X-Gene v1 SOC supports total 2688 MSI/MSIX vectors coalesced into > 16 HW IRQ lines. > > Signed-off-by: Duc Dang <dhdang@apm.com> > Signed-off-by: Tanmay Inamdar <tinamdar@apm.com> I just had a quick look at this, and this seems to be going in the exact opposite direction compared to what we now have on arm64, where we move away from using struct msi_controller for most thing, and implement PCI MSI/MSIX in a generic way, using MSI domains. I suggest you have a look at how GICv2m and GICv3 ITS implement the MSI support. You can also have a look at what I did for the Tegra MSI controller in this patch: https://git.kernel.org/cgit/linux/kernel/git/maz/arm-platforms.git/commit/?h=irq/kill-msi-controller&id=83b3602fcee7972b9d549ed729b56ec28de16081 Eventually, the plan is to kill msi_controller entirely, so introducing new drivers that rely on it is not something I'm eager to see. Thanks, M. > --- > drivers/pci/host/Kconfig | 4 + > drivers/pci/host/Makefile | 1 + > drivers/pci/host/pci-xgene-msi.c | 393 +++++++++++++++++++++++++++++++++++++++ > drivers/pci/host/pci-xgene.c | 25 +++ > 4 files changed, 423 insertions(+) > create mode 100644 drivers/pci/host/pci-xgene-msi.c > > diff --git a/drivers/pci/host/Kconfig b/drivers/pci/host/Kconfig > index 7b892a9..6c0f98c 100644 > --- a/drivers/pci/host/Kconfig > +++ b/drivers/pci/host/Kconfig > @@ -84,11 +84,15 @@ config PCIE_XILINX > Say 'Y' here if you want kernel to support the Xilinx AXI PCIe > Host Bridge driver. > > +config PCI_XGENE_MSI > + bool > + > config PCI_XGENE > bool "X-Gene PCIe controller" > depends on ARCH_XGENE > depends on OF > select PCIEPORTBUS > + select PCI_XGENE_MSI if PCI_MSI > help > Say Y here if you want internal PCI support on APM X-Gene SoC. > There are 5 internal PCIe ports available. Each port is GEN3 capable > diff --git a/drivers/pci/host/Makefile b/drivers/pci/host/Makefile > index e61d91c..f39bde3 100644 > --- a/drivers/pci/host/Makefile > +++ b/drivers/pci/host/Makefile > @@ -11,5 +11,6 @@ obj-$(CONFIG_PCIE_SPEAR13XX) += pcie-spear13xx.o > obj-$(CONFIG_PCI_KEYSTONE) += pci-keystone-dw.o pci-keystone.o > obj-$(CONFIG_PCIE_XILINX) += pcie-xilinx.o > obj-$(CONFIG_PCI_XGENE) += pci-xgene.o > +obj-$(CONFIG_PCI_XGENE_MSI) += pci-xgene-msi.o > obj-$(CONFIG_PCI_LAYERSCAPE) += pci-layerscape.o > obj-$(CONFIG_PCI_VERSATILE) += pci-versatile.o > diff --git a/drivers/pci/host/pci-xgene-msi.c b/drivers/pci/host/pci-xgene-msi.c > new file mode 100644 > index 0000000..e1cab39 > --- /dev/null > +++ b/drivers/pci/host/pci-xgene-msi.c > @@ -0,0 +1,393 @@ > +/* > + * APM X-Gene MSI Driver > + * > + * Copyright (c) 2014, Applied Micro Circuits Corporation > + * Author: Tanmay Inamdar <tinamdar@apm.com> > + * Duc Dang <dhdang@apm.com> > + * > + * This program is free software; you can redistribute it and/or modify it > + * under the terms of the GNU General Public License as published by the > + * Free Software Foundation; either version 2 of the License, or (at your > + * option) any later version. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + */ > +#include <linux/interrupt.h> > +#include <linux/module.h> > +#include <linux/msi.h> > +#include <linux/of_irq.h> > +#include <linux/pci.h> > +#include <linux/platform_device.h> > +#include <linux/of_pci.h> > + > +#define MSI_INDEX0 0x000000 > +#define MSI_INT0 0x800000 > + > +struct xgene_msi_settings { > + u32 index_per_group; > + u32 irqs_per_index; > + u32 nr_msi_vec; > + u32 nr_hw_irqs; > +}; > + > +struct xgene_msi { > + struct irq_domain *irqhost; > + struct msi_controller msi_chip; > + struct xgene_msi_settings *settings; > + u32 msi_addr_lo; > + u32 msi_addr_hi; > + void __iomem *msi_regs; > + unsigned long *bitmap; > + struct mutex bitmap_lock; > + int *msi_virqs; > +}; > + > +static inline struct xgene_msi *to_xgene_msi(struct msi_controller *msi_chip) > +{ > + return container_of(msi_chip, struct xgene_msi, msi_chip); > +} > + > +struct xgene_msi_settings storm_msi_settings = { > + .index_per_group = 8, > + .irqs_per_index = 21, > + .nr_msi_vec = 2688, > + .nr_hw_irqs = 16, > +}; > + > +typedef int (*xgene_msi_initcall_t)(struct xgene_msi *); > + > +static inline irq_hw_number_t virq_to_hw(unsigned int virq) > +{ > + struct irq_data *irq_data = irq_get_irq_data(virq); > + > + return WARN_ON(!irq_data) ? 0 : irq_data->hwirq; > +} > + > +static int xgene_msi_init_storm_settings(struct xgene_msi *xgene_msi) > +{ > + xgene_msi->settings = &storm_msi_settings; > + return 0; > +} > + > +static struct irq_chip xgene_msi_chip = { > + .name = "X-Gene-1 MSI", > + .irq_enable = pci_msi_unmask_irq, > + .irq_disable = pci_msi_mask_irq, > + .irq_mask = pci_msi_mask_irq, > + .irq_unmask = pci_msi_unmask_irq, > +}; > + > +static int xgene_msi_host_map(struct irq_domain *h, unsigned int virq, > + irq_hw_number_t hw) > +{ > + irq_set_chip_and_handler(virq, &xgene_msi_chip, handle_simple_irq); > + irq_set_chip_data(virq, h->host_data); > + set_irq_flags(irq, IRQF_VALID); > + > + return 0; > +} > + > +static const struct irq_domain_ops xgene_msi_host_ops = { > + .map = xgene_msi_host_map, > +}; > + > +static int xgene_msi_alloc(struct xgene_msi *xgene_msi) > +{ > + u32 msi_irq_count = xgene_msi->settings->nr_msi_vec; > + int msi; > + > + mutex_lock(&xgene_msi->bitmap_lock); > + > + msi = find_first_zero_bit(xgene_msi->bitmap, msi_irq_count); > + if (msi < msi_irq_count) > + set_bit(msi, xgene_msi->bitmap); > + else > + msi = -ENOSPC; > + > + mutex_unlock(&xgene_msi->bitmap_lock); > + > + return msi; > +} > + > +static void xgene_msi_free(struct xgene_msi *xgene_msi, unsigned long irq) > +{ > + mutex_lock(&xgene_msi->bitmap_lock); > + > + if (!test_bit(irq, xgene_msi->bitmap)) > + pr_err("trying to free unused MSI#%lu\n", irq); > + else > + clear_bit(irq, xgene_msi->bitmap); > + > + mutex_unlock(&xgene_msi->bitmap_lock); > +} > + > +static int xgene_msi_init_allocator(struct xgene_msi *xgene_msi) > +{ > + u32 msi_irq_count = xgene_msi->settings->nr_msi_vec; > + u32 hw_irq_count = xgene_msi->settings->nr_hw_irqs; > + int size = BITS_TO_LONGS(msi_irq_count) * sizeof(long); > + > + xgene_msi->bitmap = kzalloc(size, GFP_KERNEL); > + if (!xgene_msi->bitmap) > + return -ENOMEM; > + mutex_init(&xgene_msi->bitmap_lock); > + > + xgene_msi->msi_virqs = kcalloc(hw_irq_count, sizeof(int), GFP_KERNEL); > + if (!xgene_msi->msi_virqs) > + return -ENOMEM; > + return 0; > +} > + > +static void xgene_msi_teardown_irq(struct msi_controller *chip, > + unsigned int irq) > +{ > + struct xgene_msi *xgene_msi = to_xgene_msi(chip); > + > + irq_set_msi_desc(irq, NULL); > + xgene_msi_free(xgene_msi, virq_to_hw(irq)); > +} > + > +static void xgene_compose_msi_msg(struct pci_dev *dev, int hwirq, > + struct msi_msg *msg, > + struct xgene_msi *xgene_msi) > +{ > + u32 nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; > + u32 irqs_per_index = xgene_msi->settings->irqs_per_index; > + u32 reg_set = hwirq / (nr_hw_irqs * irqs_per_index); > + u32 group = hwirq % nr_hw_irqs; > + > + msg->address_hi = xgene_msi->msi_addr_hi; > + msg->address_lo = xgene_msi->msi_addr_lo + > + (((8 * group) + reg_set) << 16); > + msg->data = (hwirq / nr_hw_irqs) % irqs_per_index; > +} > + > +static int xgene_msi_setup_irq(struct msi_controller *chip, > + struct pci_dev *pdev, struct msi_desc *desc) > +{ > + struct xgene_msi *xgene_msi = to_xgene_msi(chip); > + struct msi_msg msg; > + unsigned long virq, gic_irq; > + int hwirq; > + > + hwirq = xgene_msi_alloc(xgene_msi); > + if (hwirq < 0) { > + dev_err(&pdev->dev, "failed to allocate MSI\n"); > + return -ENOSPC; > + } > + > + virq = irq_create_mapping(xgene_msi->irqhost, hwirq); > + if (virq == 0) { > + dev_err(&pdev->dev, "failed to map hwirq %i\n", hwirq); > + return -ENOSPC; > + } > + > + gic_irq = xgene_msi->msi_virqs[hwirq % > + xgene_msi->settings->nr_hw_irqs]; > + pr_debug("Mapp HWIRQ %d on GIC IRQ %lu TO VIRQ %lu\n", > + hwirq, gic_irq, virq); > + irq_set_msi_desc(virq, desc); > + xgene_compose_msi_msg(pdev, hwirq, &msg, xgene_msi); > + irq_set_handler_data(virq, (void *)gic_irq); > + write_msi_msg(virq, &msg); > + > + return 0; > +} > + > +static irqreturn_t xgene_msi_isr(int irq, void *data) > +{ > + struct xgene_msi *xgene_msi = (struct xgene_msi *) data; > + unsigned int virq; > + int msir_index, msir_reg, msir_val, hw_irq; > + u32 intr_index, grp_select, msi_grp, processed = 0; > + u32 nr_hw_irqs, irqs_per_index, index_per_group; > + > + msi_grp = irq - xgene_msi->msi_virqs[0]; > + if (msi_grp >= xgene_msi->settings->nr_hw_irqs) { > + pr_err("invalid msi received\n"); > + return IRQ_NONE; > + } > + > + nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; > + irqs_per_index = xgene_msi->settings->irqs_per_index; > + index_per_group = xgene_msi->settings->index_per_group; > + > + grp_select = readl(xgene_msi->msi_regs + MSI_INT0 + (msi_grp << 16)); > + while (grp_select) { > + msir_index = ffs(grp_select) - 1; > + msir_reg = (msi_grp << 19) + (msir_index << 16); > + msir_val = readl(xgene_msi->msi_regs + MSI_INDEX0 + msir_reg); > + while (msir_val) { > + intr_index = ffs(msir_val) - 1; > + hw_irq = (((msir_index * irqs_per_index) + intr_index) * > + nr_hw_irqs) + msi_grp; > + virq = irq_find_mapping(xgene_msi->irqhost, hw_irq); > + if (virq != 0) > + generic_handle_irq(virq); > + msir_val &= ~(1 << intr_index); > + processed++; > + } > + grp_select &= ~(1 << msir_index); > + } > + > + return processed > 0 ? IRQ_HANDLED : IRQ_NONE; > +} > + > +static int xgene_msi_remove(struct platform_device *pdev) > +{ > + int virq, i; > + struct xgene_msi *msi = platform_get_drvdata(pdev); > + u32 nr_hw_irqs = msi->settings->nr_hw_irqs; > + > + for (i = 0; i < nr_hw_irqs; i++) { > + virq = msi->msi_virqs[i]; > + if (virq != 0) > + free_irq(virq, msi); > + } > + > + kfree(msi->bitmap); > + msi->bitmap = NULL; > + > + return 0; > +} > + > +static int xgene_msi_setup_hwirq(struct xgene_msi *msi, > + struct platform_device *pdev, > + int irq_index) > +{ > + int virt_msir; > + cpumask_var_t mask; > + int err; > + > + virt_msir = platform_get_irq(pdev, irq_index); > + if (virt_msir < 0) { > + dev_err(&pdev->dev, "Cannot translate IRQ index %d\n", > + irq_index); > + return -EINVAL; > + } > + > + err = request_irq(virt_msir, xgene_msi_isr, 0, "xgene-msi", msi); > + if (err) { > + dev_err(&pdev->dev, "request irq failed\n"); > + return err; > + } > + > + if (alloc_cpumask_var(&mask, GFP_KERNEL)) { > + cpumask_setall(mask); > + irq_set_affinity(virt_msir, mask); > + free_cpumask_var(mask); > + } > + > + msi->msi_virqs[irq_index] = virt_msir; > + > + return 0; > +} > + > +static const struct of_device_id xgene_msi_match_table[] = { > + {.compatible = "apm,xgene1-msi", > + .data = xgene_msi_init_storm_settings}, > + {}, > +}; > + > +static int xgene_msi_probe(struct platform_device *pdev) > +{ > + struct resource *res; > + int rc, irq_index; > + struct device_node *np; > + const struct of_device_id *matched_np; > + struct xgene_msi *xgene_msi; > + xgene_msi_initcall_t init_fn; > + u32 nr_hw_irqs, nr_msi_vecs; > + > + np = of_find_matching_node_and_match(NULL, > + xgene_msi_match_table, &matched_np); > + if (!np) > + return -ENODEV; > + > + xgene_msi = kzalloc(sizeof(struct xgene_msi), GFP_KERNEL); > + if (!xgene_msi) { > + dev_err(&pdev->dev, "failed to allocate X-Gene MSI data\n"); > + return -ENOMEM; > + } > + > + init_fn = (xgene_msi_initcall_t) matched_np->data; > + rc = init_fn(xgene_msi); > + if (rc) > + return rc; > + > + platform_set_drvdata(pdev, xgene_msi); > + > + nr_msi_vecs = xgene_msi->settings->nr_msi_vec; > + xgene_msi->irqhost = irq_domain_add_linear(pdev->dev.of_node, > + nr_msi_vecs, &xgene_msi_host_ops, xgene_msi); > + if (!xgene_msi->irqhost) { > + dev_err(&pdev->dev, "No memory for MSI irqhost\n"); > + rc = -ENOMEM; > + goto error; > + } > + > + res = platform_get_resource(pdev, IORESOURCE_MEM, 0); > + xgene_msi->msi_regs = devm_ioremap_resource(&pdev->dev, res); > + if (IS_ERR(xgene_msi->msi_regs)) { > + dev_err(&pdev->dev, "no reg space\n"); > + rc = -EINVAL; > + goto error; > + } > + > + xgene_msi->msi_addr_hi = upper_32_bits(res->start); > + xgene_msi->msi_addr_lo = lower_32_bits(res->start); > + > + rc = xgene_msi_init_allocator(xgene_msi); > + if (rc) { > + dev_err(&pdev->dev, "Error allocating MSI bitmap\n"); > + goto error; > + } > + > + nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; > + for (irq_index = 0; irq_index < nr_hw_irqs; irq_index++) { > + rc = xgene_msi_setup_hwirq(xgene_msi, pdev, irq_index); > + if (rc) > + goto error; > + } > + > + xgene_msi->msi_chip.dev = &pdev->dev; > + xgene_msi->msi_chip.of_node = np; > + xgene_msi->msi_chip.setup_irq = xgene_msi_setup_irq; > + xgene_msi->msi_chip.teardown_irq = xgene_msi_teardown_irq; > + > + rc = of_pci_msi_chip_add(&xgene_msi->msi_chip); > + if (rc) { > + dev_err(&pdev->dev, "failed to add MSI controller chip\n"); > + goto error; > + } > + > + dev_info(&pdev->dev, "APM X-Gene PCIe MSI driver loaded\n"); > + > + return 0; > +error: > + xgene_msi_remove(pdev); > + return rc; > +} > + > +static struct platform_driver xgene_msi_driver = { > + .driver = { > + .name = "xgene-msi", > + .owner = THIS_MODULE, > + .of_match_table = xgene_msi_match_table, > + }, > + .probe = xgene_msi_probe, > + .remove = xgene_msi_remove, > +}; > + > +static int __init xgene_pcie_msi_init(void) > +{ > + return platform_driver_register(&xgene_msi_driver); > +} > +subsys_initcall(xgene_pcie_msi_init); > + > +MODULE_AUTHOR("Duc Dang <dhdang@apm.com>"); > +MODULE_DESCRIPTION("APM X-Gene PCIe MSI driver"); > +MODULE_LICENSE("GPL v2"); > diff --git a/drivers/pci/host/pci-xgene.c b/drivers/pci/host/pci-xgene.c > index ee082c0..63d58e6 100644 > --- a/drivers/pci/host/pci-xgene.c > +++ b/drivers/pci/host/pci-xgene.c > @@ -468,6 +468,23 @@ static int xgene_pcie_setup(struct xgene_pcie_port *port, > return 0; > } > > +static int xgene_pcie_msi_enable(struct pci_bus *bus) > +{ > + struct device_node *msi_node; > + > + msi_node = of_parse_phandle(bus->dev.of_node, > + "msi-parent", 0); > + if (!msi_node) > + return -ENODEV; > + > + bus->msi = of_pci_find_msi_chip_by_node(msi_node); > + if (bus->msi) > + bus->msi->dev = &bus->dev; > + else > + return -ENODEV; > + return 0; > +} > + > static int xgene_pcie_probe_bridge(struct platform_device *pdev) > { > struct device_node *dn = pdev->dev.of_node; > @@ -504,6 +521,14 @@ static int xgene_pcie_probe_bridge(struct platform_device *pdev) > if (!bus) > return -ENOMEM; > > + if (IS_ENABLED(CONFIG_PCI_MSI)) { > + ret = xgene_pcie_msi_enable(bus); > + if (ret) { > + dev_err(port->dev, "failed to enable MSI\n"); > + return ret; > + } > + } > + > pci_scan_child_bus(bus); > pci_assign_unassigned_bus_resources(bus); > pci_bus_add_devices(bus); >
On Wed, Mar 18, 2015 at 11:05 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: > On 04/03/15 19:39, Duc Dang wrote: >> X-Gene v1 SOC supports total 2688 MSI/MSIX vectors coalesced into >> 16 HW IRQ lines. >> >> Signed-off-by: Duc Dang <dhdang@apm.com> >> Signed-off-by: Tanmay Inamdar <tinamdar@apm.com> > > I just had a quick look at this, and this seems to be going in the exact > opposite direction compared to what we now have on arm64, where we move > away from using struct msi_controller for most thing, and implement PCI > MSI/MSIX in a generic way, using MSI domains. > > I suggest you have a look at how GICv2m and GICv3 ITS implement the MSI > support. You can also have a look at what I did for the Tegra MSI > controller in this patch: > > https://git.kernel.org/cgit/linux/kernel/git/maz/arm-platforms.git/commit/?h=irq/kill-msi-controller&id=83b3602fcee7972b9d549ed729b56ec28de16081 > > Eventually, the plan is to kill msi_controller entirely, so introducing > new drivers that rely on it is not something I'm eager to see. Thanks, Marc. X-Gene 1 MSI is handled by separate MSI controller block, so its driver implementation is different from GICv2m and GICv3. I will refer to what you did for Tegra MSI, but I don't see your latest changes in 4.0-rc4. Is the change you made for Tegra MSI going to mainline soon? Regards, Duc Dang. > > Thanks, > > M. > >> --- >> drivers/pci/host/Kconfig | 4 + >> drivers/pci/host/Makefile | 1 + >> drivers/pci/host/pci-xgene-msi.c | 393 +++++++++++++++++++++++++++++++++++++++ >> drivers/pci/host/pci-xgene.c | 25 +++ >> 4 files changed, 423 insertions(+) >> create mode 100644 drivers/pci/host/pci-xgene-msi.c >> >> diff --git a/drivers/pci/host/Kconfig b/drivers/pci/host/Kconfig >> index 7b892a9..6c0f98c 100644 >> --- a/drivers/pci/host/Kconfig >> +++ b/drivers/pci/host/Kconfig >> @@ -84,11 +84,15 @@ config PCIE_XILINX >> Say 'Y' here if you want kernel to support the Xilinx AXI PCIe >> Host Bridge driver. >> >> +config PCI_XGENE_MSI >> + bool >> + >> config PCI_XGENE >> bool "X-Gene PCIe controller" >> depends on ARCH_XGENE >> depends on OF >> select PCIEPORTBUS >> + select PCI_XGENE_MSI if PCI_MSI >> help >> Say Y here if you want internal PCI support on APM X-Gene SoC. >> There are 5 internal PCIe ports available. Each port is GEN3 capable >> diff --git a/drivers/pci/host/Makefile b/drivers/pci/host/Makefile >> index e61d91c..f39bde3 100644 >> --- a/drivers/pci/host/Makefile >> +++ b/drivers/pci/host/Makefile >> @@ -11,5 +11,6 @@ obj-$(CONFIG_PCIE_SPEAR13XX) += pcie-spear13xx.o >> obj-$(CONFIG_PCI_KEYSTONE) += pci-keystone-dw.o pci-keystone.o >> obj-$(CONFIG_PCIE_XILINX) += pcie-xilinx.o >> obj-$(CONFIG_PCI_XGENE) += pci-xgene.o >> +obj-$(CONFIG_PCI_XGENE_MSI) += pci-xgene-msi.o >> obj-$(CONFIG_PCI_LAYERSCAPE) += pci-layerscape.o >> obj-$(CONFIG_PCI_VERSATILE) += pci-versatile.o >> diff --git a/drivers/pci/host/pci-xgene-msi.c b/drivers/pci/host/pci-xgene-msi.c >> new file mode 100644 >> index 0000000..e1cab39 >> --- /dev/null >> +++ b/drivers/pci/host/pci-xgene-msi.c >> @@ -0,0 +1,393 @@ >> +/* >> + * APM X-Gene MSI Driver >> + * >> + * Copyright (c) 2014, Applied Micro Circuits Corporation >> + * Author: Tanmay Inamdar <tinamdar@apm.com> >> + * Duc Dang <dhdang@apm.com> >> + * >> + * This program is free software; you can redistribute it and/or modify it >> + * under the terms of the GNU General Public License as published by the >> + * Free Software Foundation; either version 2 of the License, or (at your >> + * option) any later version. >> + * >> + * This program is distributed in the hope that it will be useful, >> + * but WITHOUT ANY WARRANTY; without even the implied warranty of >> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the >> + * GNU General Public License for more details. >> + */ >> +#include <linux/interrupt.h> >> +#include <linux/module.h> >> +#include <linux/msi.h> >> +#include <linux/of_irq.h> >> +#include <linux/pci.h> >> +#include <linux/platform_device.h> >> +#include <linux/of_pci.h> >> + >> +#define MSI_INDEX0 0x000000 >> +#define MSI_INT0 0x800000 >> + >> +struct xgene_msi_settings { >> + u32 index_per_group; >> + u32 irqs_per_index; >> + u32 nr_msi_vec; >> + u32 nr_hw_irqs; >> +}; >> + >> +struct xgene_msi { >> + struct irq_domain *irqhost; >> + struct msi_controller msi_chip; >> + struct xgene_msi_settings *settings; >> + u32 msi_addr_lo; >> + u32 msi_addr_hi; >> + void __iomem *msi_regs; >> + unsigned long *bitmap; >> + struct mutex bitmap_lock; >> + int *msi_virqs; >> +}; >> + >> +static inline struct xgene_msi *to_xgene_msi(struct msi_controller *msi_chip) >> +{ >> + return container_of(msi_chip, struct xgene_msi, msi_chip); >> +} >> + >> +struct xgene_msi_settings storm_msi_settings = { >> + .index_per_group = 8, >> + .irqs_per_index = 21, >> + .nr_msi_vec = 2688, >> + .nr_hw_irqs = 16, >> +}; >> + >> +typedef int (*xgene_msi_initcall_t)(struct xgene_msi *); >> + >> +static inline irq_hw_number_t virq_to_hw(unsigned int virq) >> +{ >> + struct irq_data *irq_data = irq_get_irq_data(virq); >> + >> + return WARN_ON(!irq_data) ? 0 : irq_data->hwirq; >> +} >> + >> +static int xgene_msi_init_storm_settings(struct xgene_msi *xgene_msi) >> +{ >> + xgene_msi->settings = &storm_msi_settings; >> + return 0; >> +} >> + >> +static struct irq_chip xgene_msi_chip = { >> + .name = "X-Gene-1 MSI", >> + .irq_enable = pci_msi_unmask_irq, >> + .irq_disable = pci_msi_mask_irq, >> + .irq_mask = pci_msi_mask_irq, >> + .irq_unmask = pci_msi_unmask_irq, >> +}; >> + >> +static int xgene_msi_host_map(struct irq_domain *h, unsigned int virq, >> + irq_hw_number_t hw) >> +{ >> + irq_set_chip_and_handler(virq, &xgene_msi_chip, handle_simple_irq); >> + irq_set_chip_data(virq, h->host_data); >> + set_irq_flags(irq, IRQF_VALID); >> + >> + return 0; >> +} >> + >> +static const struct irq_domain_ops xgene_msi_host_ops = { >> + .map = xgene_msi_host_map, >> +}; >> + >> +static int xgene_msi_alloc(struct xgene_msi *xgene_msi) >> +{ >> + u32 msi_irq_count = xgene_msi->settings->nr_msi_vec; >> + int msi; >> + >> + mutex_lock(&xgene_msi->bitmap_lock); >> + >> + msi = find_first_zero_bit(xgene_msi->bitmap, msi_irq_count); >> + if (msi < msi_irq_count) >> + set_bit(msi, xgene_msi->bitmap); >> + else >> + msi = -ENOSPC; >> + >> + mutex_unlock(&xgene_msi->bitmap_lock); >> + >> + return msi; >> +} >> + >> +static void xgene_msi_free(struct xgene_msi *xgene_msi, unsigned long irq) >> +{ >> + mutex_lock(&xgene_msi->bitmap_lock); >> + >> + if (!test_bit(irq, xgene_msi->bitmap)) >> + pr_err("trying to free unused MSI#%lu\n", irq); >> + else >> + clear_bit(irq, xgene_msi->bitmap); >> + >> + mutex_unlock(&xgene_msi->bitmap_lock); >> +} >> + >> +static int xgene_msi_init_allocator(struct xgene_msi *xgene_msi) >> +{ >> + u32 msi_irq_count = xgene_msi->settings->nr_msi_vec; >> + u32 hw_irq_count = xgene_msi->settings->nr_hw_irqs; >> + int size = BITS_TO_LONGS(msi_irq_count) * sizeof(long); >> + >> + xgene_msi->bitmap = kzalloc(size, GFP_KERNEL); >> + if (!xgene_msi->bitmap) >> + return -ENOMEM; >> + mutex_init(&xgene_msi->bitmap_lock); >> + >> + xgene_msi->msi_virqs = kcalloc(hw_irq_count, sizeof(int), GFP_KERNEL); >> + if (!xgene_msi->msi_virqs) >> + return -ENOMEM; >> + return 0; >> +} >> + >> +static void xgene_msi_teardown_irq(struct msi_controller *chip, >> + unsigned int irq) >> +{ >> + struct xgene_msi *xgene_msi = to_xgene_msi(chip); >> + >> + irq_set_msi_desc(irq, NULL); >> + xgene_msi_free(xgene_msi, virq_to_hw(irq)); >> +} >> + >> +static void xgene_compose_msi_msg(struct pci_dev *dev, int hwirq, >> + struct msi_msg *msg, >> + struct xgene_msi *xgene_msi) >> +{ >> + u32 nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; >> + u32 irqs_per_index = xgene_msi->settings->irqs_per_index; >> + u32 reg_set = hwirq / (nr_hw_irqs * irqs_per_index); >> + u32 group = hwirq % nr_hw_irqs; >> + >> + msg->address_hi = xgene_msi->msi_addr_hi; >> + msg->address_lo = xgene_msi->msi_addr_lo + >> + (((8 * group) + reg_set) << 16); >> + msg->data = (hwirq / nr_hw_irqs) % irqs_per_index; >> +} >> + >> +static int xgene_msi_setup_irq(struct msi_controller *chip, >> + struct pci_dev *pdev, struct msi_desc *desc) >> +{ >> + struct xgene_msi *xgene_msi = to_xgene_msi(chip); >> + struct msi_msg msg; >> + unsigned long virq, gic_irq; >> + int hwirq; >> + >> + hwirq = xgene_msi_alloc(xgene_msi); >> + if (hwirq < 0) { >> + dev_err(&pdev->dev, "failed to allocate MSI\n"); >> + return -ENOSPC; >> + } >> + >> + virq = irq_create_mapping(xgene_msi->irqhost, hwirq); >> + if (virq == 0) { >> + dev_err(&pdev->dev, "failed to map hwirq %i\n", hwirq); >> + return -ENOSPC; >> + } >> + >> + gic_irq = xgene_msi->msi_virqs[hwirq % >> + xgene_msi->settings->nr_hw_irqs]; >> + pr_debug("Mapp HWIRQ %d on GIC IRQ %lu TO VIRQ %lu\n", >> + hwirq, gic_irq, virq); >> + irq_set_msi_desc(virq, desc); >> + xgene_compose_msi_msg(pdev, hwirq, &msg, xgene_msi); >> + irq_set_handler_data(virq, (void *)gic_irq); >> + write_msi_msg(virq, &msg); >> + >> + return 0; >> +} >> + >> +static irqreturn_t xgene_msi_isr(int irq, void *data) >> +{ >> + struct xgene_msi *xgene_msi = (struct xgene_msi *) data; >> + unsigned int virq; >> + int msir_index, msir_reg, msir_val, hw_irq; >> + u32 intr_index, grp_select, msi_grp, processed = 0; >> + u32 nr_hw_irqs, irqs_per_index, index_per_group; >> + >> + msi_grp = irq - xgene_msi->msi_virqs[0]; >> + if (msi_grp >= xgene_msi->settings->nr_hw_irqs) { >> + pr_err("invalid msi received\n"); >> + return IRQ_NONE; >> + } >> + >> + nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; >> + irqs_per_index = xgene_msi->settings->irqs_per_index; >> + index_per_group = xgene_msi->settings->index_per_group; >> + >> + grp_select = readl(xgene_msi->msi_regs + MSI_INT0 + (msi_grp << 16)); >> + while (grp_select) { >> + msir_index = ffs(grp_select) - 1; >> + msir_reg = (msi_grp << 19) + (msir_index << 16); >> + msir_val = readl(xgene_msi->msi_regs + MSI_INDEX0 + msir_reg); >> + while (msir_val) { >> + intr_index = ffs(msir_val) - 1; >> + hw_irq = (((msir_index * irqs_per_index) + intr_index) * >> + nr_hw_irqs) + msi_grp; >> + virq = irq_find_mapping(xgene_msi->irqhost, hw_irq); >> + if (virq != 0) >> + generic_handle_irq(virq); >> + msir_val &= ~(1 << intr_index); >> + processed++; >> + } >> + grp_select &= ~(1 << msir_index); >> + } >> + >> + return processed > 0 ? IRQ_HANDLED : IRQ_NONE; >> +} >> + >> +static int xgene_msi_remove(struct platform_device *pdev) >> +{ >> + int virq, i; >> + struct xgene_msi *msi = platform_get_drvdata(pdev); >> + u32 nr_hw_irqs = msi->settings->nr_hw_irqs; >> + >> + for (i = 0; i < nr_hw_irqs; i++) { >> + virq = msi->msi_virqs[i]; >> + if (virq != 0) >> + free_irq(virq, msi); >> + } >> + >> + kfree(msi->bitmap); >> + msi->bitmap = NULL; >> + >> + return 0; >> +} >> + >> +static int xgene_msi_setup_hwirq(struct xgene_msi *msi, >> + struct platform_device *pdev, >> + int irq_index) >> +{ >> + int virt_msir; >> + cpumask_var_t mask; >> + int err; >> + >> + virt_msir = platform_get_irq(pdev, irq_index); >> + if (virt_msir < 0) { >> + dev_err(&pdev->dev, "Cannot translate IRQ index %d\n", >> + irq_index); >> + return -EINVAL; >> + } >> + >> + err = request_irq(virt_msir, xgene_msi_isr, 0, "xgene-msi", msi); >> + if (err) { >> + dev_err(&pdev->dev, "request irq failed\n"); >> + return err; >> + } >> + >> + if (alloc_cpumask_var(&mask, GFP_KERNEL)) { >> + cpumask_setall(mask); >> + irq_set_affinity(virt_msir, mask); >> + free_cpumask_var(mask); >> + } >> + >> + msi->msi_virqs[irq_index] = virt_msir; >> + >> + return 0; >> +} >> + >> +static const struct of_device_id xgene_msi_match_table[] = { >> + {.compatible = "apm,xgene1-msi", >> + .data = xgene_msi_init_storm_settings}, >> + {}, >> +}; >> + >> +static int xgene_msi_probe(struct platform_device *pdev) >> +{ >> + struct resource *res; >> + int rc, irq_index; >> + struct device_node *np; >> + const struct of_device_id *matched_np; >> + struct xgene_msi *xgene_msi; >> + xgene_msi_initcall_t init_fn; >> + u32 nr_hw_irqs, nr_msi_vecs; >> + >> + np = of_find_matching_node_and_match(NULL, >> + xgene_msi_match_table, &matched_np); >> + if (!np) >> + return -ENODEV; >> + >> + xgene_msi = kzalloc(sizeof(struct xgene_msi), GFP_KERNEL); >> + if (!xgene_msi) { >> + dev_err(&pdev->dev, "failed to allocate X-Gene MSI data\n"); >> + return -ENOMEM; >> + } >> + >> + init_fn = (xgene_msi_initcall_t) matched_np->data; >> + rc = init_fn(xgene_msi); >> + if (rc) >> + return rc; >> + >> + platform_set_drvdata(pdev, xgene_msi); >> + >> + nr_msi_vecs = xgene_msi->settings->nr_msi_vec; >> + xgene_msi->irqhost = irq_domain_add_linear(pdev->dev.of_node, >> + nr_msi_vecs, &xgene_msi_host_ops, xgene_msi); >> + if (!xgene_msi->irqhost) { >> + dev_err(&pdev->dev, "No memory for MSI irqhost\n"); >> + rc = -ENOMEM; >> + goto error; >> + } >> + >> + res = platform_get_resource(pdev, IORESOURCE_MEM, 0); >> + xgene_msi->msi_regs = devm_ioremap_resource(&pdev->dev, res); >> + if (IS_ERR(xgene_msi->msi_regs)) { >> + dev_err(&pdev->dev, "no reg space\n"); >> + rc = -EINVAL; >> + goto error; >> + } >> + >> + xgene_msi->msi_addr_hi = upper_32_bits(res->start); >> + xgene_msi->msi_addr_lo = lower_32_bits(res->start); >> + >> + rc = xgene_msi_init_allocator(xgene_msi); >> + if (rc) { >> + dev_err(&pdev->dev, "Error allocating MSI bitmap\n"); >> + goto error; >> + } >> + >> + nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; >> + for (irq_index = 0; irq_index < nr_hw_irqs; irq_index++) { >> + rc = xgene_msi_setup_hwirq(xgene_msi, pdev, irq_index); >> + if (rc) >> + goto error; >> + } >> + >> + xgene_msi->msi_chip.dev = &pdev->dev; >> + xgene_msi->msi_chip.of_node = np; >> + xgene_msi->msi_chip.setup_irq = xgene_msi_setup_irq; >> + xgene_msi->msi_chip.teardown_irq = xgene_msi_teardown_irq; >> + >> + rc = of_pci_msi_chip_add(&xgene_msi->msi_chip); >> + if (rc) { >> + dev_err(&pdev->dev, "failed to add MSI controller chip\n"); >> + goto error; >> + } >> + >> + dev_info(&pdev->dev, "APM X-Gene PCIe MSI driver loaded\n"); >> + >> + return 0; >> +error: >> + xgene_msi_remove(pdev); >> + return rc; >> +} >> + >> +static struct platform_driver xgene_msi_driver = { >> + .driver = { >> + .name = "xgene-msi", >> + .owner = THIS_MODULE, >> + .of_match_table = xgene_msi_match_table, >> + }, >> + .probe = xgene_msi_probe, >> + .remove = xgene_msi_remove, >> +}; >> + >> +static int __init xgene_pcie_msi_init(void) >> +{ >> + return platform_driver_register(&xgene_msi_driver); >> +} >> +subsys_initcall(xgene_pcie_msi_init); >> + >> +MODULE_AUTHOR("Duc Dang <dhdang@apm.com>"); >> +MODULE_DESCRIPTION("APM X-Gene PCIe MSI driver"); >> +MODULE_LICENSE("GPL v2"); >> diff --git a/drivers/pci/host/pci-xgene.c b/drivers/pci/host/pci-xgene.c >> index ee082c0..63d58e6 100644 >> --- a/drivers/pci/host/pci-xgene.c >> +++ b/drivers/pci/host/pci-xgene.c >> @@ -468,6 +468,23 @@ static int xgene_pcie_setup(struct xgene_pcie_port *port, >> return 0; >> } >> >> +static int xgene_pcie_msi_enable(struct pci_bus *bus) >> +{ >> + struct device_node *msi_node; >> + >> + msi_node = of_parse_phandle(bus->dev.of_node, >> + "msi-parent", 0); >> + if (!msi_node) >> + return -ENODEV; >> + >> + bus->msi = of_pci_find_msi_chip_by_node(msi_node); >> + if (bus->msi) >> + bus->msi->dev = &bus->dev; >> + else >> + return -ENODEV; >> + return 0; >> +} >> + >> static int xgene_pcie_probe_bridge(struct platform_device *pdev) >> { >> struct device_node *dn = pdev->dev.of_node; >> @@ -504,6 +521,14 @@ static int xgene_pcie_probe_bridge(struct platform_device *pdev) >> if (!bus) >> return -ENOMEM; >> >> + if (IS_ENABLED(CONFIG_PCI_MSI)) { >> + ret = xgene_pcie_msi_enable(bus); >> + if (ret) { >> + dev_err(port->dev, "failed to enable MSI\n"); >> + return ret; >> + } >> + } >> + >> pci_scan_child_bus(bus); >> pci_assign_unassigned_bus_resources(bus); >> pci_bus_add_devices(bus); >> > > > -- > Jazz is not dead. It just smells funny...
On 18/03/15 18:29, Duc Dang wrote: > On Wed, Mar 18, 2015 at 11:05 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: >> On 04/03/15 19:39, Duc Dang wrote: >>> X-Gene v1 SOC supports total 2688 MSI/MSIX vectors coalesced into >>> 16 HW IRQ lines. >>> >>> Signed-off-by: Duc Dang <dhdang@apm.com> >>> Signed-off-by: Tanmay Inamdar <tinamdar@apm.com> >> >> I just had a quick look at this, and this seems to be going in the exact >> opposite direction compared to what we now have on arm64, where we move >> away from using struct msi_controller for most thing, and implement PCI >> MSI/MSIX in a generic way, using MSI domains. >> >> I suggest you have a look at how GICv2m and GICv3 ITS implement the MSI >> support. You can also have a look at what I did for the Tegra MSI >> controller in this patch: >> >> https://git.kernel.org/cgit/linux/kernel/git/maz/arm-platforms.git/commit/?h=irq/kill-msi-controller&id=83b3602fcee7972b9d549ed729b56ec28de16081 >> >> Eventually, the plan is to kill msi_controller entirely, so introducing >> new drivers that rely on it is not something I'm eager to see. > > Thanks, Marc. > > X-Gene 1 MSI is handled by separate MSI controller block, so its > driver implementation is different from GICv2m and GICv3. I will refer It will certainly be different in the sense that you won't use a stacked domain on top of the GIC. But what I want to see is the use of a generic pci_msi_domain on top of an irq_domain, just like we have on v2m and v3. Thomas has also been vocal enough about that in the past, and x86 is going down that road as well. > to what you did for Tegra MSI, but I don't see your latest changes in > 4.0-rc4. Is the change you made for Tegra MSI going to mainline soon? Not yet. As you can see in this branch, this relies on some other cleanups. But you can already convert most of your driver and put it in the shape that matches what we have for v2m and v3. Once the required cleanups are in, I'll remove the last traces of msi_controller myself if necessary. Thanks, M.
On Wed, Mar 18, 2015 at 11:52 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: > On 18/03/15 18:29, Duc Dang wrote: >> On Wed, Mar 18, 2015 at 11:05 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: >>> On 04/03/15 19:39, Duc Dang wrote: >>>> X-Gene v1 SOC supports total 2688 MSI/MSIX vectors coalesced into >>>> 16 HW IRQ lines. >>>> >>>> Signed-off-by: Duc Dang <dhdang@apm.com> >>>> Signed-off-by: Tanmay Inamdar <tinamdar@apm.com> >>> >>> I just had a quick look at this, and this seems to be going in the exact >>> opposite direction compared to what we now have on arm64, where we move >>> away from using struct msi_controller for most thing, and implement PCI >>> MSI/MSIX in a generic way, using MSI domains. >>> >>> I suggest you have a look at how GICv2m and GICv3 ITS implement the MSI >>> support. You can also have a look at what I did for the Tegra MSI >>> controller in this patch: >>> >>> https://git.kernel.org/cgit/linux/kernel/git/maz/arm-platforms.git/commit/?h=irq/kill-msi-controller&id=83b3602fcee7972b9d549ed729b56ec28de16081 >>> >>> Eventually, the plan is to kill msi_controller entirely, so introducing >>> new drivers that rely on it is not something I'm eager to see. >> >> Thanks, Marc. >> >> X-Gene 1 MSI is handled by separate MSI controller block, so its >> driver implementation is different from GICv2m and GICv3. I will refer > > It will certainly be different in the sense that you won't use a stacked > domain on top of the GIC. But what I want to see is the use of a generic > pci_msi_domain on top of an irq_domain, just like we have on v2m and v3. > Thomas has also been vocal enough about that in the past, and x86 is > going down that road as well. > >> to what you did for Tegra MSI, but I don't see your latest changes in >> 4.0-rc4. Is the change you made for Tegra MSI going to mainline soon? > > Not yet. As you can see in this branch, this relies on some other > cleanups. But you can already convert most of your driver and put it in > the shape that matches what we have for v2m and v3. Once the required > cleanups are in, I'll remove the last traces of msi_controller myself if > necessary. > Hi Marc, Bjorn, I follow Marc's suggestion and convert my X-Gene 1 MSI driver to remove msi_controller struct and use generic pci_msi_domain on top of an irq_domain. The code requires Marc's changes in irq/kill-msi-controller branch to be compiled and function correctly, so I plan to post the patch on top of Marc's tree. Please let me know if you think I should have different approach to post this patch. Another question I have is when do you plan to roll out this hierarchy irq domain implementation for MSI, as I see some of Marc's changes to kill msi_controller structure and implement MSI irq domain still does not get into Bjorn's tree. Regards, Duc Dang. > Thanks, > > M. > -- > Jazz is not dead. It just smells funny...
On Tue, 7 Apr 2015 20:56:48 +0100 Duc Dang <dhdang@apm.com> wrote: Hi Duc, > On Wed, Mar 18, 2015 at 11:52 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: > > On 18/03/15 18:29, Duc Dang wrote: > >> On Wed, Mar 18, 2015 at 11:05 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: > >>> On 04/03/15 19:39, Duc Dang wrote: > >>>> X-Gene v1 SOC supports total 2688 MSI/MSIX vectors coalesced into > >>>> 16 HW IRQ lines. > >>>> > >>>> Signed-off-by: Duc Dang <dhdang@apm.com> > >>>> Signed-off-by: Tanmay Inamdar <tinamdar@apm.com> > >>> > >>> I just had a quick look at this, and this seems to be going in the exact > >>> opposite direction compared to what we now have on arm64, where we move > >>> away from using struct msi_controller for most thing, and implement PCI > >>> MSI/MSIX in a generic way, using MSI domains. > >>> > >>> I suggest you have a look at how GICv2m and GICv3 ITS implement the MSI > >>> support. You can also have a look at what I did for the Tegra MSI > >>> controller in this patch: > >>> > >>> https://git.kernel.org/cgit/linux/kernel/git/maz/arm-platforms.git/commit/?h=irq/kill-msi-controller&id=83b3602fcee7972b9d549ed729b56ec28de16081 > >>> > >>> Eventually, the plan is to kill msi_controller entirely, so introducing > >>> new drivers that rely on it is not something I'm eager to see. > >> > >> Thanks, Marc. > >> > >> X-Gene 1 MSI is handled by separate MSI controller block, so its > >> driver implementation is different from GICv2m and GICv3. I will refer > > > > It will certainly be different in the sense that you won't use a stacked > > domain on top of the GIC. But what I want to see is the use of a generic > > pci_msi_domain on top of an irq_domain, just like we have on v2m and v3. > > Thomas has also been vocal enough about that in the past, and x86 is > > going down that road as well. > > > >> to what you did for Tegra MSI, but I don't see your latest changes in > >> 4.0-rc4. Is the change you made for Tegra MSI going to mainline soon? > > > > Not yet. As you can see in this branch, this relies on some other > > cleanups. But you can already convert most of your driver and put it in > > the shape that matches what we have for v2m and v3. Once the required > > cleanups are in, I'll remove the last traces of msi_controller myself if > > necessary. > > > > Hi Marc, Bjorn, > > I follow Marc's suggestion and convert my X-Gene 1 MSI driver to > remove msi_controller struct and use generic pci_msi_domain on top of > an irq_domain. The code requires Marc's changes in > irq/kill-msi-controller branch to be compiled and function correctly, > so I plan to post the patch on top of Marc's tree. Please let me know > if you think I should have different approach to post this patch. I don't think you should rely on this branch just yet (this is why I suggested matching what we currently have for GICv2m and v3). Keeping msi_controller is fine at the moment, even if we only use to store the MSI domain. If and when this series makes it to mainline, I'll go over the individual drivers to convert them to the new scheme, just like I did for GIC and Tegra. > Another question I have is when do you plan to roll out this hierarchy > irq domain implementation for MSI, as I see some of Marc's changes to > kill msi_controller structure and implement MSI irq domain still does > not get into Bjorn's tree. I'm still tinkering with it (I have some long standing comments from Bjorn to address), and I'm working on non-PCI MSI support at the moment. But the core MSI irq domain stuff is already in (this is what GICv2m and GICv3 ITS are using), and this should be an easy thing to convert your code. Please let me know if I can be of any help. Thanks, M.
This patch set adds MSI/MSIX termination driver support for APM X-Gene v1 SoC. APM X-Gene v1 SoC supports its own implementation of MSI, which is not compliant to GIC V2M specification for MSI Termination. There is single MSI block in X-Gene v1 SOC which serves all 5 PCIe ports. This MSI block supports 2688 MSI termination ports coalesced into 16 physical HW IRQ lines and shared across all 5 PCIe ports. v3 changes: 1. Implement MSI support using PCI MSI IRQ domain 2. Only use msi_controller to store IRQ domain v2 changes: 1. Use msi_controller structure 2. Remove arch hooks arch_teardown_msi_irqs and arch_setup_msi_irqs .../devicetree/bindings/pci/xgene-pci-msi.txt | 63 ++++ MAINTAINERS | 8 + arch/arm64/boot/dts/apm/apm-storm.dtsi | 27 ++ drivers/pci/host/Kconfig | 6 + drivers/pci/host/Makefile | 1 + drivers/pci/host/pci-xgene-msi.c | 407 +++++++++++++++++++++ drivers/pci/host/pci-xgene.c | 21 ++ 7 files changed, 533 insertions(+) create mode 100644 Documentation/devicetree/bindings/pci/xgene-pci-msi.txt create mode 100644 drivers/pci/host/pci-xgene-msi.c
Hi Marc, On Wed, Apr 8, 2015 at 12:44 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: > On Tue, 7 Apr 2015 20:56:48 +0100 > Duc Dang <dhdang@apm.com> wrote: > > Hi Duc, > >> On Wed, Mar 18, 2015 at 11:52 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: >> > On 18/03/15 18:29, Duc Dang wrote: >> >> On Wed, Mar 18, 2015 at 11:05 AM, Marc Zyngier <marc.zyngier@arm.com> wrote: >> >>> On 04/03/15 19:39, Duc Dang wrote: >> >>>> X-Gene v1 SOC supports total 2688 MSI/MSIX vectors coalesced into >> >>>> 16 HW IRQ lines. >> >>>> >> >>>> Signed-off-by: Duc Dang <dhdang@apm.com> >> >>>> Signed-off-by: Tanmay Inamdar <tinamdar@apm.com> >> >>> >> >>> I just had a quick look at this, and this seems to be going in the exact >> >>> opposite direction compared to what we now have on arm64, where we move >> >>> away from using struct msi_controller for most thing, and implement PCI >> >>> MSI/MSIX in a generic way, using MSI domains. >> >>> >> >>> I suggest you have a look at how GICv2m and GICv3 ITS implement the MSI >> >>> support. You can also have a look at what I did for the Tegra MSI >> >>> controller in this patch: >> >>> >> >>> https://git.kernel.org/cgit/linux/kernel/git/maz/arm-platforms.git/commit/?h=irq/kill-msi-controller&id=83b3602fcee7972b9d549ed729b56ec28de16081 >> >>> >> >>> Eventually, the plan is to kill msi_controller entirely, so introducing >> >>> new drivers that rely on it is not something I'm eager to see. >> >> >> >> Thanks, Marc. >> >> >> >> X-Gene 1 MSI is handled by separate MSI controller block, so its >> >> driver implementation is different from GICv2m and GICv3. I will refer >> > >> > It will certainly be different in the sense that you won't use a stacked >> > domain on top of the GIC. But what I want to see is the use of a generic >> > pci_msi_domain on top of an irq_domain, just like we have on v2m and v3. >> > Thomas has also been vocal enough about that in the past, and x86 is >> > going down that road as well. >> > >> >> to what you did for Tegra MSI, but I don't see your latest changes in >> >> 4.0-rc4. Is the change you made for Tegra MSI going to mainline soon? >> > >> > Not yet. As you can see in this branch, this relies on some other >> > cleanups. But you can already convert most of your driver and put it in >> > the shape that matches what we have for v2m and v3. Once the required >> > cleanups are in, I'll remove the last traces of msi_controller myself if >> > necessary. >> > >> >> Hi Marc, Bjorn, >> >> I follow Marc's suggestion and convert my X-Gene 1 MSI driver to >> remove msi_controller struct and use generic pci_msi_domain on top of >> an irq_domain. The code requires Marc's changes in >> irq/kill-msi-controller branch to be compiled and function correctly, >> so I plan to post the patch on top of Marc's tree. Please let me know >> if you think I should have different approach to post this patch. > > I don't think you should rely on this branch just yet (this is why I > suggested matching what we currently have for GICv2m and v3). Keeping > msi_controller is fine at the moment, even if we only use to store the > MSI domain. If and when this series makes it to mainline, I'll go over > the individual drivers to convert them to the new scheme, just like I > did for GIC and Tegra. > >> Another question I have is when do you plan to roll out this hierarchy >> irq domain implementation for MSI, as I see some of Marc's changes to >> kill msi_controller structure and implement MSI irq domain still does >> not get into Bjorn's tree. > > I'm still tinkering with it (I have some long standing comments from > Bjorn to address), and I'm working on non-PCI MSI support at the moment. > > But the core MSI irq domain stuff is already in (this is what GICv2m > and GICv3 ITS are using), and this should be an easy thing to convert > your code. > > Please let me know if I can be of any help. > Thanks for your suggestion. I converted my code using the core MSI irq domain (similar to GICv2m, still have msi_controller structure) and posted a new version (v3) of this patch. Can you please take a look when you have some time? Regards, Duc Dang. > Thanks, > > M. > -- > Without deviation from the norm, progress is not possible.
diff --git a/drivers/pci/host/Kconfig b/drivers/pci/host/Kconfig index 7b892a9..6c0f98c 100644 --- a/drivers/pci/host/Kconfig +++ b/drivers/pci/host/Kconfig @@ -84,11 +84,15 @@ config PCIE_XILINX Say 'Y' here if you want kernel to support the Xilinx AXI PCIe Host Bridge driver. +config PCI_XGENE_MSI + bool + config PCI_XGENE bool "X-Gene PCIe controller" depends on ARCH_XGENE depends on OF select PCIEPORTBUS + select PCI_XGENE_MSI if PCI_MSI help Say Y here if you want internal PCI support on APM X-Gene SoC. There are 5 internal PCIe ports available. Each port is GEN3 capable diff --git a/drivers/pci/host/Makefile b/drivers/pci/host/Makefile index e61d91c..f39bde3 100644 --- a/drivers/pci/host/Makefile +++ b/drivers/pci/host/Makefile @@ -11,5 +11,6 @@ obj-$(CONFIG_PCIE_SPEAR13XX) += pcie-spear13xx.o obj-$(CONFIG_PCI_KEYSTONE) += pci-keystone-dw.o pci-keystone.o obj-$(CONFIG_PCIE_XILINX) += pcie-xilinx.o obj-$(CONFIG_PCI_XGENE) += pci-xgene.o +obj-$(CONFIG_PCI_XGENE_MSI) += pci-xgene-msi.o obj-$(CONFIG_PCI_LAYERSCAPE) += pci-layerscape.o obj-$(CONFIG_PCI_VERSATILE) += pci-versatile.o diff --git a/drivers/pci/host/pci-xgene-msi.c b/drivers/pci/host/pci-xgene-msi.c new file mode 100644 index 0000000..e1cab39 --- /dev/null +++ b/drivers/pci/host/pci-xgene-msi.c @@ -0,0 +1,393 @@ +/* + * APM X-Gene MSI Driver + * + * Copyright (c) 2014, Applied Micro Circuits Corporation + * Author: Tanmay Inamdar <tinamdar@apm.com> + * Duc Dang <dhdang@apm.com> + * + * This program is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License as published by the + * Free Software Foundation; either version 2 of the License, or (at your + * option) any later version. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + */ +#include <linux/interrupt.h> +#include <linux/module.h> +#include <linux/msi.h> +#include <linux/of_irq.h> +#include <linux/pci.h> +#include <linux/platform_device.h> +#include <linux/of_pci.h> + +#define MSI_INDEX0 0x000000 +#define MSI_INT0 0x800000 + +struct xgene_msi_settings { + u32 index_per_group; + u32 irqs_per_index; + u32 nr_msi_vec; + u32 nr_hw_irqs; +}; + +struct xgene_msi { + struct irq_domain *irqhost; + struct msi_controller msi_chip; + struct xgene_msi_settings *settings; + u32 msi_addr_lo; + u32 msi_addr_hi; + void __iomem *msi_regs; + unsigned long *bitmap; + struct mutex bitmap_lock; + int *msi_virqs; +}; + +static inline struct xgene_msi *to_xgene_msi(struct msi_controller *msi_chip) +{ + return container_of(msi_chip, struct xgene_msi, msi_chip); +} + +struct xgene_msi_settings storm_msi_settings = { + .index_per_group = 8, + .irqs_per_index = 21, + .nr_msi_vec = 2688, + .nr_hw_irqs = 16, +}; + +typedef int (*xgene_msi_initcall_t)(struct xgene_msi *); + +static inline irq_hw_number_t virq_to_hw(unsigned int virq) +{ + struct irq_data *irq_data = irq_get_irq_data(virq); + + return WARN_ON(!irq_data) ? 0 : irq_data->hwirq; +} + +static int xgene_msi_init_storm_settings(struct xgene_msi *xgene_msi) +{ + xgene_msi->settings = &storm_msi_settings; + return 0; +} + +static struct irq_chip xgene_msi_chip = { + .name = "X-Gene-1 MSI", + .irq_enable = pci_msi_unmask_irq, + .irq_disable = pci_msi_mask_irq, + .irq_mask = pci_msi_mask_irq, + .irq_unmask = pci_msi_unmask_irq, +}; + +static int xgene_msi_host_map(struct irq_domain *h, unsigned int virq, + irq_hw_number_t hw) +{ + irq_set_chip_and_handler(virq, &xgene_msi_chip, handle_simple_irq); + irq_set_chip_data(virq, h->host_data); + set_irq_flags(irq, IRQF_VALID); + + return 0; +} + +static const struct irq_domain_ops xgene_msi_host_ops = { + .map = xgene_msi_host_map, +}; + +static int xgene_msi_alloc(struct xgene_msi *xgene_msi) +{ + u32 msi_irq_count = xgene_msi->settings->nr_msi_vec; + int msi; + + mutex_lock(&xgene_msi->bitmap_lock); + + msi = find_first_zero_bit(xgene_msi->bitmap, msi_irq_count); + if (msi < msi_irq_count) + set_bit(msi, xgene_msi->bitmap); + else + msi = -ENOSPC; + + mutex_unlock(&xgene_msi->bitmap_lock); + + return msi; +} + +static void xgene_msi_free(struct xgene_msi *xgene_msi, unsigned long irq) +{ + mutex_lock(&xgene_msi->bitmap_lock); + + if (!test_bit(irq, xgene_msi->bitmap)) + pr_err("trying to free unused MSI#%lu\n", irq); + else + clear_bit(irq, xgene_msi->bitmap); + + mutex_unlock(&xgene_msi->bitmap_lock); +} + +static int xgene_msi_init_allocator(struct xgene_msi *xgene_msi) +{ + u32 msi_irq_count = xgene_msi->settings->nr_msi_vec; + u32 hw_irq_count = xgene_msi->settings->nr_hw_irqs; + int size = BITS_TO_LONGS(msi_irq_count) * sizeof(long); + + xgene_msi->bitmap = kzalloc(size, GFP_KERNEL); + if (!xgene_msi->bitmap) + return -ENOMEM; + mutex_init(&xgene_msi->bitmap_lock); + + xgene_msi->msi_virqs = kcalloc(hw_irq_count, sizeof(int), GFP_KERNEL); + if (!xgene_msi->msi_virqs) + return -ENOMEM; + return 0; +} + +static void xgene_msi_teardown_irq(struct msi_controller *chip, + unsigned int irq) +{ + struct xgene_msi *xgene_msi = to_xgene_msi(chip); + + irq_set_msi_desc(irq, NULL); + xgene_msi_free(xgene_msi, virq_to_hw(irq)); +} + +static void xgene_compose_msi_msg(struct pci_dev *dev, int hwirq, + struct msi_msg *msg, + struct xgene_msi *xgene_msi) +{ + u32 nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; + u32 irqs_per_index = xgene_msi->settings->irqs_per_index; + u32 reg_set = hwirq / (nr_hw_irqs * irqs_per_index); + u32 group = hwirq % nr_hw_irqs; + + msg->address_hi = xgene_msi->msi_addr_hi; + msg->address_lo = xgene_msi->msi_addr_lo + + (((8 * group) + reg_set) << 16); + msg->data = (hwirq / nr_hw_irqs) % irqs_per_index; +} + +static int xgene_msi_setup_irq(struct msi_controller *chip, + struct pci_dev *pdev, struct msi_desc *desc) +{ + struct xgene_msi *xgene_msi = to_xgene_msi(chip); + struct msi_msg msg; + unsigned long virq, gic_irq; + int hwirq; + + hwirq = xgene_msi_alloc(xgene_msi); + if (hwirq < 0) { + dev_err(&pdev->dev, "failed to allocate MSI\n"); + return -ENOSPC; + } + + virq = irq_create_mapping(xgene_msi->irqhost, hwirq); + if (virq == 0) { + dev_err(&pdev->dev, "failed to map hwirq %i\n", hwirq); + return -ENOSPC; + } + + gic_irq = xgene_msi->msi_virqs[hwirq % + xgene_msi->settings->nr_hw_irqs]; + pr_debug("Mapp HWIRQ %d on GIC IRQ %lu TO VIRQ %lu\n", + hwirq, gic_irq, virq); + irq_set_msi_desc(virq, desc); + xgene_compose_msi_msg(pdev, hwirq, &msg, xgene_msi); + irq_set_handler_data(virq, (void *)gic_irq); + write_msi_msg(virq, &msg); + + return 0; +} + +static irqreturn_t xgene_msi_isr(int irq, void *data) +{ + struct xgene_msi *xgene_msi = (struct xgene_msi *) data; + unsigned int virq; + int msir_index, msir_reg, msir_val, hw_irq; + u32 intr_index, grp_select, msi_grp, processed = 0; + u32 nr_hw_irqs, irqs_per_index, index_per_group; + + msi_grp = irq - xgene_msi->msi_virqs[0]; + if (msi_grp >= xgene_msi->settings->nr_hw_irqs) { + pr_err("invalid msi received\n"); + return IRQ_NONE; + } + + nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; + irqs_per_index = xgene_msi->settings->irqs_per_index; + index_per_group = xgene_msi->settings->index_per_group; + + grp_select = readl(xgene_msi->msi_regs + MSI_INT0 + (msi_grp << 16)); + while (grp_select) { + msir_index = ffs(grp_select) - 1; + msir_reg = (msi_grp << 19) + (msir_index << 16); + msir_val = readl(xgene_msi->msi_regs + MSI_INDEX0 + msir_reg); + while (msir_val) { + intr_index = ffs(msir_val) - 1; + hw_irq = (((msir_index * irqs_per_index) + intr_index) * + nr_hw_irqs) + msi_grp; + virq = irq_find_mapping(xgene_msi->irqhost, hw_irq); + if (virq != 0) + generic_handle_irq(virq); + msir_val &= ~(1 << intr_index); + processed++; + } + grp_select &= ~(1 << msir_index); + } + + return processed > 0 ? IRQ_HANDLED : IRQ_NONE; +} + +static int xgene_msi_remove(struct platform_device *pdev) +{ + int virq, i; + struct xgene_msi *msi = platform_get_drvdata(pdev); + u32 nr_hw_irqs = msi->settings->nr_hw_irqs; + + for (i = 0; i < nr_hw_irqs; i++) { + virq = msi->msi_virqs[i]; + if (virq != 0) + free_irq(virq, msi); + } + + kfree(msi->bitmap); + msi->bitmap = NULL; + + return 0; +} + +static int xgene_msi_setup_hwirq(struct xgene_msi *msi, + struct platform_device *pdev, + int irq_index) +{ + int virt_msir; + cpumask_var_t mask; + int err; + + virt_msir = platform_get_irq(pdev, irq_index); + if (virt_msir < 0) { + dev_err(&pdev->dev, "Cannot translate IRQ index %d\n", + irq_index); + return -EINVAL; + } + + err = request_irq(virt_msir, xgene_msi_isr, 0, "xgene-msi", msi); + if (err) { + dev_err(&pdev->dev, "request irq failed\n"); + return err; + } + + if (alloc_cpumask_var(&mask, GFP_KERNEL)) { + cpumask_setall(mask); + irq_set_affinity(virt_msir, mask); + free_cpumask_var(mask); + } + + msi->msi_virqs[irq_index] = virt_msir; + + return 0; +} + +static const struct of_device_id xgene_msi_match_table[] = { + {.compatible = "apm,xgene1-msi", + .data = xgene_msi_init_storm_settings}, + {}, +}; + +static int xgene_msi_probe(struct platform_device *pdev) +{ + struct resource *res; + int rc, irq_index; + struct device_node *np; + const struct of_device_id *matched_np; + struct xgene_msi *xgene_msi; + xgene_msi_initcall_t init_fn; + u32 nr_hw_irqs, nr_msi_vecs; + + np = of_find_matching_node_and_match(NULL, + xgene_msi_match_table, &matched_np); + if (!np) + return -ENODEV; + + xgene_msi = kzalloc(sizeof(struct xgene_msi), GFP_KERNEL); + if (!xgene_msi) { + dev_err(&pdev->dev, "failed to allocate X-Gene MSI data\n"); + return -ENOMEM; + } + + init_fn = (xgene_msi_initcall_t) matched_np->data; + rc = init_fn(xgene_msi); + if (rc) + return rc; + + platform_set_drvdata(pdev, xgene_msi); + + nr_msi_vecs = xgene_msi->settings->nr_msi_vec; + xgene_msi->irqhost = irq_domain_add_linear(pdev->dev.of_node, + nr_msi_vecs, &xgene_msi_host_ops, xgene_msi); + if (!xgene_msi->irqhost) { + dev_err(&pdev->dev, "No memory for MSI irqhost\n"); + rc = -ENOMEM; + goto error; + } + + res = platform_get_resource(pdev, IORESOURCE_MEM, 0); + xgene_msi->msi_regs = devm_ioremap_resource(&pdev->dev, res); + if (IS_ERR(xgene_msi->msi_regs)) { + dev_err(&pdev->dev, "no reg space\n"); + rc = -EINVAL; + goto error; + } + + xgene_msi->msi_addr_hi = upper_32_bits(res->start); + xgene_msi->msi_addr_lo = lower_32_bits(res->start); + + rc = xgene_msi_init_allocator(xgene_msi); + if (rc) { + dev_err(&pdev->dev, "Error allocating MSI bitmap\n"); + goto error; + } + + nr_hw_irqs = xgene_msi->settings->nr_hw_irqs; + for (irq_index = 0; irq_index < nr_hw_irqs; irq_index++) { + rc = xgene_msi_setup_hwirq(xgene_msi, pdev, irq_index); + if (rc) + goto error; + } + + xgene_msi->msi_chip.dev = &pdev->dev; + xgene_msi->msi_chip.of_node = np; + xgene_msi->msi_chip.setup_irq = xgene_msi_setup_irq; + xgene_msi->msi_chip.teardown_irq = xgene_msi_teardown_irq; + + rc = of_pci_msi_chip_add(&xgene_msi->msi_chip); + if (rc) { + dev_err(&pdev->dev, "failed to add MSI controller chip\n"); + goto error; + } + + dev_info(&pdev->dev, "APM X-Gene PCIe MSI driver loaded\n"); + + return 0; +error: + xgene_msi_remove(pdev); + return rc; +} + +static struct platform_driver xgene_msi_driver = { + .driver = { + .name = "xgene-msi", + .owner = THIS_MODULE, + .of_match_table = xgene_msi_match_table, + }, + .probe = xgene_msi_probe, + .remove = xgene_msi_remove, +}; + +static int __init xgene_pcie_msi_init(void) +{ + return platform_driver_register(&xgene_msi_driver); +} +subsys_initcall(xgene_pcie_msi_init); + +MODULE_AUTHOR("Duc Dang <dhdang@apm.com>"); +MODULE_DESCRIPTION("APM X-Gene PCIe MSI driver"); +MODULE_LICENSE("GPL v2"); diff --git a/drivers/pci/host/pci-xgene.c b/drivers/pci/host/pci-xgene.c index ee082c0..63d58e6 100644 --- a/drivers/pci/host/pci-xgene.c +++ b/drivers/pci/host/pci-xgene.c @@ -468,6 +468,23 @@ static int xgene_pcie_setup(struct xgene_pcie_port *port, return 0; } +static int xgene_pcie_msi_enable(struct pci_bus *bus) +{ + struct device_node *msi_node; + + msi_node = of_parse_phandle(bus->dev.of_node, + "msi-parent", 0); + if (!msi_node) + return -ENODEV; + + bus->msi = of_pci_find_msi_chip_by_node(msi_node); + if (bus->msi) + bus->msi->dev = &bus->dev; + else + return -ENODEV; + return 0; +} + static int xgene_pcie_probe_bridge(struct platform_device *pdev) { struct device_node *dn = pdev->dev.of_node; @@ -504,6 +521,14 @@ static int xgene_pcie_probe_bridge(struct platform_device *pdev) if (!bus) return -ENOMEM; + if (IS_ENABLED(CONFIG_PCI_MSI)) { + ret = xgene_pcie_msi_enable(bus); + if (ret) { + dev_err(port->dev, "failed to enable MSI\n"); + return ret; + } + } + pci_scan_child_bus(bus); pci_assign_unassigned_bus_resources(bus); pci_bus_add_devices(bus);