Message ID | 20221203064629.1601299-10-apatel@ventanamicro.com (mailing list archive) |
---|---|
State | Superseded |
Delegated to: | Palmer Dabbelt |
Headers | show |
Series | RISC-V IPI Improvements | expand |
Context | Check | Description |
---|---|---|
conchuod/tree_selection | fail | Guessing tree name failed |
On 03/12/2022 15.46, Anup Patel wrote: > From: Marc Zyngier <maz@kernel.org> > > Now that the complexity of the AIC IPI mux has been copied into > the core code for the benefit of the riscv architecture, > shrink the AIC driver by the same amount by using that infrastructure. > > Signed-off-by: Marc Zyngier <maz@kernel.org> > --- > drivers/irqchip/Kconfig | 1 + > drivers/irqchip/irq-apple-aic.c | 161 ++------------------------------ > 2 files changed, 9 insertions(+), 153 deletions(-) > > diff --git a/drivers/irqchip/Kconfig b/drivers/irqchip/Kconfig > index 131379aa8424..0fd452b49ed9 100644 > --- a/drivers/irqchip/Kconfig > +++ b/drivers/irqchip/Kconfig > @@ -675,6 +675,7 @@ config APPLE_AIC > bool "Apple Interrupt Controller (AIC)" > depends on ARM64 > depends on ARCH_APPLE || COMPILE_TEST > + select GENERIC_IRQ_IPI_MUX > help > Support for the Apple Interrupt Controller found on Apple Silicon SoCs, > such as the M1. > diff --git a/drivers/irqchip/irq-apple-aic.c b/drivers/irqchip/irq-apple-aic.c > index 1c2813ad8bbe..1fa81fcff788 100644 > --- a/drivers/irqchip/irq-apple-aic.c > +++ b/drivers/irqchip/irq-apple-aic.c > @@ -292,7 +292,6 @@ struct aic_irq_chip { > void __iomem *base; > void __iomem *event; > struct irq_domain *hw_domain; > - struct irq_domain *ipi_domain; > struct { > cpumask_t aff; > } *fiq_aff[AIC_NR_FIQ]; > @@ -307,9 +306,6 @@ struct aic_irq_chip { > > static DEFINE_PER_CPU(uint32_t, aic_fiq_unmasked); > > -static DEFINE_PER_CPU(atomic_t, aic_vipi_flag); > -static DEFINE_PER_CPU(atomic_t, aic_vipi_enable); > - > static struct aic_irq_chip *aic_irqc; > > static void aic_handle_ipi(struct pt_regs *regs); > @@ -751,98 +747,8 @@ static void aic_ipi_send_fast(int cpu) > isb(); > } > > -static void aic_ipi_mask(struct irq_data *d) > -{ > - u32 irq_bit = BIT(irqd_to_hwirq(d)); > - > - /* No specific ordering requirements needed here. */ > - atomic_andnot(irq_bit, this_cpu_ptr(&aic_vipi_enable)); > -} > - > -static void aic_ipi_unmask(struct irq_data *d) > -{ > - struct aic_irq_chip *ic = irq_data_get_irq_chip_data(d); > - u32 irq_bit = BIT(irqd_to_hwirq(d)); > - > - atomic_or(irq_bit, this_cpu_ptr(&aic_vipi_enable)); > - > - /* > - * The atomic_or() above must complete before the atomic_read() > - * below to avoid racing aic_ipi_send_mask(). > - */ > - smp_mb__after_atomic(); > - > - /* > - * If a pending vIPI was unmasked, raise a HW IPI to ourselves. > - * No barriers needed here since this is a self-IPI. > - */ > - if (atomic_read(this_cpu_ptr(&aic_vipi_flag)) & irq_bit) { > - if (static_branch_likely(&use_fast_ipi)) > - aic_ipi_send_fast(smp_processor_id()); > - else > - aic_ic_write(ic, AIC_IPI_SEND, AIC_IPI_SEND_CPU(smp_processor_id())); > - } > -} > - > -static void aic_ipi_send_mask(struct irq_data *d, const struct cpumask *mask) > -{ > - struct aic_irq_chip *ic = irq_data_get_irq_chip_data(d); > - u32 irq_bit = BIT(irqd_to_hwirq(d)); > - u32 send = 0; > - int cpu; > - unsigned long pending; > - > - for_each_cpu(cpu, mask) { > - /* > - * This sequence is the mirror of the one in aic_ipi_unmask(); > - * see the comment there. Additionally, release semantics > - * ensure that the vIPI flag set is ordered after any shared > - * memory accesses that precede it. This therefore also pairs > - * with the atomic_fetch_andnot in aic_handle_ipi(). > - */ > - pending = atomic_fetch_or_release(irq_bit, per_cpu_ptr(&aic_vipi_flag, cpu)); > - > - /* > - * The atomic_fetch_or_release() above must complete before the > - * atomic_read() below to avoid racing aic_ipi_unmask(). > - */ > - smp_mb__after_atomic(); > - > - if (!(pending & irq_bit) && > - (atomic_read(per_cpu_ptr(&aic_vipi_enable, cpu)) & irq_bit)) { > - if (static_branch_likely(&use_fast_ipi)) > - aic_ipi_send_fast(cpu); > - else > - send |= AIC_IPI_SEND_CPU(cpu); > - } > - } > - > - /* > - * The flag writes must complete before the physical IPI is issued > - * to another CPU. This is implied by the control dependency on > - * the result of atomic_read_acquire() above, which is itself > - * already ordered after the vIPI flag write. > - */ > - if (send) > - aic_ic_write(ic, AIC_IPI_SEND, send); > -} > - > -static struct irq_chip ipi_chip = { > - .name = "AIC-IPI", > - .irq_mask = aic_ipi_mask, > - .irq_unmask = aic_ipi_unmask, > - .ipi_send_mask = aic_ipi_send_mask, > -}; > - > -/* > - * IPI IRQ domain > - */ > - > static void aic_handle_ipi(struct pt_regs *regs) > { > - int i; > - unsigned long enabled, firing; > - > /* > * Ack the IPI. We need to order this after the AIC event read, but > * that is enforced by normal MMIO ordering guarantees. > @@ -857,27 +763,7 @@ static void aic_handle_ipi(struct pt_regs *regs) > aic_ic_write(aic_irqc, AIC_IPI_ACK, AIC_IPI_OTHER); > } > > - /* > - * The mask read does not need to be ordered. Only we can change > - * our own mask anyway, so no races are possible here, as long as > - * we are properly in the interrupt handler (which is covered by > - * the barrier that is part of the top-level AIC handler's readl()). > - */ > - enabled = atomic_read(this_cpu_ptr(&aic_vipi_enable)); > - > - /* > - * Clear the IPIs we are about to handle. This pairs with the > - * atomic_fetch_or_release() in aic_ipi_send_mask(), and needs to be > - * ordered after the aic_ic_write() above (to avoid dropping vIPIs) and > - * before IPI handling code (to avoid races handling vIPIs before they > - * are signaled). The former is taken care of by the release semantics > - * of the write portion, while the latter is taken care of by the > - * acquire semantics of the read portion. > - */ > - firing = atomic_fetch_andnot(enabled, this_cpu_ptr(&aic_vipi_flag)) & enabled; > - > - for_each_set_bit(i, &firing, AIC_NR_SWIPI) > - generic_handle_domain_irq(aic_irqc->ipi_domain, i); > + ipi_mux_process(); > > /* > * No ordering needed here; at worst this just changes the timing of > @@ -887,55 +773,24 @@ static void aic_handle_ipi(struct pt_regs *regs) > aic_ic_write(aic_irqc, AIC_IPI_MASK_CLR, AIC_IPI_OTHER); > } > > -static int aic_ipi_alloc(struct irq_domain *d, unsigned int virq, > - unsigned int nr_irqs, void *args) > +static void aic_ipi_send_single(unsigned int cpu) > { > - int i; > - > - for (i = 0; i < nr_irqs; i++) { > - irq_set_percpu_devid(virq + i); > - irq_domain_set_info(d, virq + i, i, &ipi_chip, d->host_data, > - handle_percpu_devid_irq, NULL, NULL); > - } > - > - return 0; > -} > - > -static void aic_ipi_free(struct irq_domain *d, unsigned int virq, unsigned int nr_irqs) > -{ > - /* Not freeing IPIs */ > + if (static_branch_likely(&use_fast_ipi)) > + aic_ipi_send_fast(cpu); > + else > + aic_ic_write(aic_irqc, AIC_IPI_SEND, AIC_IPI_SEND_CPU(cpu)); > } > > -static const struct irq_domain_ops aic_ipi_domain_ops = { > - .alloc = aic_ipi_alloc, > - .free = aic_ipi_free, > -}; > - > static int __init aic_init_smp(struct aic_irq_chip *irqc, struct device_node *node) > { > - struct irq_domain *ipi_domain; > int base_ipi; > > - ipi_domain = irq_domain_create_linear(irqc->hw_domain->fwnode, AIC_NR_SWIPI, > - &aic_ipi_domain_ops, irqc); > - if (WARN_ON(!ipi_domain)) > - return -ENODEV; > - > - ipi_domain->flags |= IRQ_DOMAIN_FLAG_IPI_SINGLE; > - irq_domain_update_bus_token(ipi_domain, DOMAIN_BUS_IPI); > - > - base_ipi = __irq_domain_alloc_irqs(ipi_domain, -1, AIC_NR_SWIPI, > - NUMA_NO_NODE, NULL, false, NULL); > - > - if (WARN_ON(!base_ipi)) { > - irq_domain_remove(ipi_domain); > + base_ipi = ipi_mux_create(AIC_NR_SWIPI, aic_ipi_send_single); > + if (WARN_ON(base_ipi <= 0)) > return -ENODEV; > - } > > set_smp_ipi_range(base_ipi, AIC_NR_SWIPI); > > - irqc->ipi_domain = ipi_domain; > - > return 0; > } > Reviewed-by: Hector Martin <marcan@marcan.st> Tested-by: Hector Martin <marcan@marcan.st> Smoke tested on t8112, nothing exploded :) - Hector
On 05/12/2022 16.14, Hector Martin wrote: > On 03/12/2022 15.46, Anup Patel wrote: >> From: Marc Zyngier <maz@kernel.org> <snip> > > Reviewed-by: Hector Martin <marcan@marcan.st> > Tested-by: Hector Martin <marcan@marcan.st> > > Smoke tested on t8112, nothing exploded :) > > - Hector Come to think of it, rather than r-b, that (and #3) ought to be: Acked-by: Hector Martin <marcan@marcan.st> - Hector
diff --git a/drivers/irqchip/Kconfig b/drivers/irqchip/Kconfig index 131379aa8424..0fd452b49ed9 100644 --- a/drivers/irqchip/Kconfig +++ b/drivers/irqchip/Kconfig @@ -675,6 +675,7 @@ config APPLE_AIC bool "Apple Interrupt Controller (AIC)" depends on ARM64 depends on ARCH_APPLE || COMPILE_TEST + select GENERIC_IRQ_IPI_MUX help Support for the Apple Interrupt Controller found on Apple Silicon SoCs, such as the M1. diff --git a/drivers/irqchip/irq-apple-aic.c b/drivers/irqchip/irq-apple-aic.c index 1c2813ad8bbe..1fa81fcff788 100644 --- a/drivers/irqchip/irq-apple-aic.c +++ b/drivers/irqchip/irq-apple-aic.c @@ -292,7 +292,6 @@ struct aic_irq_chip { void __iomem *base; void __iomem *event; struct irq_domain *hw_domain; - struct irq_domain *ipi_domain; struct { cpumask_t aff; } *fiq_aff[AIC_NR_FIQ]; @@ -307,9 +306,6 @@ struct aic_irq_chip { static DEFINE_PER_CPU(uint32_t, aic_fiq_unmasked); -static DEFINE_PER_CPU(atomic_t, aic_vipi_flag); -static DEFINE_PER_CPU(atomic_t, aic_vipi_enable); - static struct aic_irq_chip *aic_irqc; static void aic_handle_ipi(struct pt_regs *regs); @@ -751,98 +747,8 @@ static void aic_ipi_send_fast(int cpu) isb(); } -static void aic_ipi_mask(struct irq_data *d) -{ - u32 irq_bit = BIT(irqd_to_hwirq(d)); - - /* No specific ordering requirements needed here. */ - atomic_andnot(irq_bit, this_cpu_ptr(&aic_vipi_enable)); -} - -static void aic_ipi_unmask(struct irq_data *d) -{ - struct aic_irq_chip *ic = irq_data_get_irq_chip_data(d); - u32 irq_bit = BIT(irqd_to_hwirq(d)); - - atomic_or(irq_bit, this_cpu_ptr(&aic_vipi_enable)); - - /* - * The atomic_or() above must complete before the atomic_read() - * below to avoid racing aic_ipi_send_mask(). - */ - smp_mb__after_atomic(); - - /* - * If a pending vIPI was unmasked, raise a HW IPI to ourselves. - * No barriers needed here since this is a self-IPI. - */ - if (atomic_read(this_cpu_ptr(&aic_vipi_flag)) & irq_bit) { - if (static_branch_likely(&use_fast_ipi)) - aic_ipi_send_fast(smp_processor_id()); - else - aic_ic_write(ic, AIC_IPI_SEND, AIC_IPI_SEND_CPU(smp_processor_id())); - } -} - -static void aic_ipi_send_mask(struct irq_data *d, const struct cpumask *mask) -{ - struct aic_irq_chip *ic = irq_data_get_irq_chip_data(d); - u32 irq_bit = BIT(irqd_to_hwirq(d)); - u32 send = 0; - int cpu; - unsigned long pending; - - for_each_cpu(cpu, mask) { - /* - * This sequence is the mirror of the one in aic_ipi_unmask(); - * see the comment there. Additionally, release semantics - * ensure that the vIPI flag set is ordered after any shared - * memory accesses that precede it. This therefore also pairs - * with the atomic_fetch_andnot in aic_handle_ipi(). - */ - pending = atomic_fetch_or_release(irq_bit, per_cpu_ptr(&aic_vipi_flag, cpu)); - - /* - * The atomic_fetch_or_release() above must complete before the - * atomic_read() below to avoid racing aic_ipi_unmask(). - */ - smp_mb__after_atomic(); - - if (!(pending & irq_bit) && - (atomic_read(per_cpu_ptr(&aic_vipi_enable, cpu)) & irq_bit)) { - if (static_branch_likely(&use_fast_ipi)) - aic_ipi_send_fast(cpu); - else - send |= AIC_IPI_SEND_CPU(cpu); - } - } - - /* - * The flag writes must complete before the physical IPI is issued - * to another CPU. This is implied by the control dependency on - * the result of atomic_read_acquire() above, which is itself - * already ordered after the vIPI flag write. - */ - if (send) - aic_ic_write(ic, AIC_IPI_SEND, send); -} - -static struct irq_chip ipi_chip = { - .name = "AIC-IPI", - .irq_mask = aic_ipi_mask, - .irq_unmask = aic_ipi_unmask, - .ipi_send_mask = aic_ipi_send_mask, -}; - -/* - * IPI IRQ domain - */ - static void aic_handle_ipi(struct pt_regs *regs) { - int i; - unsigned long enabled, firing; - /* * Ack the IPI. We need to order this after the AIC event read, but * that is enforced by normal MMIO ordering guarantees. @@ -857,27 +763,7 @@ static void aic_handle_ipi(struct pt_regs *regs) aic_ic_write(aic_irqc, AIC_IPI_ACK, AIC_IPI_OTHER); } - /* - * The mask read does not need to be ordered. Only we can change - * our own mask anyway, so no races are possible here, as long as - * we are properly in the interrupt handler (which is covered by - * the barrier that is part of the top-level AIC handler's readl()). - */ - enabled = atomic_read(this_cpu_ptr(&aic_vipi_enable)); - - /* - * Clear the IPIs we are about to handle. This pairs with the - * atomic_fetch_or_release() in aic_ipi_send_mask(), and needs to be - * ordered after the aic_ic_write() above (to avoid dropping vIPIs) and - * before IPI handling code (to avoid races handling vIPIs before they - * are signaled). The former is taken care of by the release semantics - * of the write portion, while the latter is taken care of by the - * acquire semantics of the read portion. - */ - firing = atomic_fetch_andnot(enabled, this_cpu_ptr(&aic_vipi_flag)) & enabled; - - for_each_set_bit(i, &firing, AIC_NR_SWIPI) - generic_handle_domain_irq(aic_irqc->ipi_domain, i); + ipi_mux_process(); /* * No ordering needed here; at worst this just changes the timing of @@ -887,55 +773,24 @@ static void aic_handle_ipi(struct pt_regs *regs) aic_ic_write(aic_irqc, AIC_IPI_MASK_CLR, AIC_IPI_OTHER); } -static int aic_ipi_alloc(struct irq_domain *d, unsigned int virq, - unsigned int nr_irqs, void *args) +static void aic_ipi_send_single(unsigned int cpu) { - int i; - - for (i = 0; i < nr_irqs; i++) { - irq_set_percpu_devid(virq + i); - irq_domain_set_info(d, virq + i, i, &ipi_chip, d->host_data, - handle_percpu_devid_irq, NULL, NULL); - } - - return 0; -} - -static void aic_ipi_free(struct irq_domain *d, unsigned int virq, unsigned int nr_irqs) -{ - /* Not freeing IPIs */ + if (static_branch_likely(&use_fast_ipi)) + aic_ipi_send_fast(cpu); + else + aic_ic_write(aic_irqc, AIC_IPI_SEND, AIC_IPI_SEND_CPU(cpu)); } -static const struct irq_domain_ops aic_ipi_domain_ops = { - .alloc = aic_ipi_alloc, - .free = aic_ipi_free, -}; - static int __init aic_init_smp(struct aic_irq_chip *irqc, struct device_node *node) { - struct irq_domain *ipi_domain; int base_ipi; - ipi_domain = irq_domain_create_linear(irqc->hw_domain->fwnode, AIC_NR_SWIPI, - &aic_ipi_domain_ops, irqc); - if (WARN_ON(!ipi_domain)) - return -ENODEV; - - ipi_domain->flags |= IRQ_DOMAIN_FLAG_IPI_SINGLE; - irq_domain_update_bus_token(ipi_domain, DOMAIN_BUS_IPI); - - base_ipi = __irq_domain_alloc_irqs(ipi_domain, -1, AIC_NR_SWIPI, - NUMA_NO_NODE, NULL, false, NULL); - - if (WARN_ON(!base_ipi)) { - irq_domain_remove(ipi_domain); + base_ipi = ipi_mux_create(AIC_NR_SWIPI, aic_ipi_send_single); + if (WARN_ON(base_ipi <= 0)) return -ENODEV; - } set_smp_ipi_range(base_ipi, AIC_NR_SWIPI); - irqc->ipi_domain = ipi_domain; - return 0; }