Message ID | 1420633698-11742-1-git-send-email-daniel.thompson@linaro.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Hi Daniel, Some minor comments below... On Wed, Jan 07, 2015 at 12:28:18PM +0000, Daniel Thompson wrote: > Some ARM platforms mux the PMU interrupt of every core into a single > SPI. On such platforms if the PMU of any core except 0 raises an interrupt > then it cannot be serviced and eventually, if you are lucky, the spurious > irq detection might forcefully disable the interrupt. > > On these SoCs it is not possible to determine which core raised the > interrupt so workaround this issue by queuing irqwork on the other > cores whenever the primary interrupt handler is unable to service the > interrupt. > > The u8500 platform has an alternative workaround that dynamically alters > the affinity of the PMU interrupt. This workaround logic is no longer > required so the original code is removed as is the hook it relied upon. > > Tested on imx6q (which has fours cores/PMUs all muxed to a single SPI). > > Signed-off-by: Daniel Thompson <daniel.thompson@linaro.org> [...] > diff --git a/arch/arm/kernel/perf_event_cpu.c b/arch/arm/kernel/perf_event_cpu.c > index dd9acc95ebc0..3d51c5f442eb 100644 > --- a/arch/arm/kernel/perf_event_cpu.c > +++ b/arch/arm/kernel/perf_event_cpu.c > @@ -59,6 +59,116 @@ int perf_num_counters(void) > } > EXPORT_SYMBOL_GPL(perf_num_counters); > > +#ifdef CONFIG_SMP > +/* > + * Workaround logic that is distributed to all cores if the PMU has only > + * a single IRQ and the CPU receiving that IRQ cannot handle it. Its > + * job is to try to service the interrupt on the current CPU. It will > + * also enable the IRQ again if all the other CPUs have already tried to > + * service it. > + */ > +static void cpu_pmu_do_percpu_work(struct irq_work *w) > +{ > + struct pmu_hw_events *hw_events = > + container_of(w, struct pmu_hw_events, work); > + struct arm_pmu *cpu_pmu = hw_events->percpu_pmu; > + > + atomic_set(&hw_events->work_ret, > + cpu_pmu->handle_irq(0, cpu_pmu)); Do you need a memory barrier here, or is that implued by enable_irq? > + if (atomic_dec_and_test(&cpu_pmu->remaining_work)) > + enable_irq(cpu_pmu->muxed_spi_workaround_irq); > +} > + > +/* > + * Called when the main interrupt handler cannot determine the source > + * of interrupt. It will deploy a workaround if we are running on an SMP > + * platform with only a single muxed SPI. > + * > + * The workaround disables the interrupt and distributes irqwork to all > + * other processors in the system. Hopefully one of them will clear the > + * interrupt... > + */ > +static irqreturn_t cpu_pmu_handle_irq_none(int irq_num, struct arm_pmu *cpu_pmu) > +{ > + irqreturn_t ret = IRQ_NONE; > + cpumask_t deploy_on_mask; > + int cpu, work_ret; > + if (irq_num != cpu_pmu->muxed_spi_workaround_irq) > + return IRQ_NONE; return ret ? > + > + disable_irq_nosync(cpu_pmu->muxed_spi_workaround_irq); > + > + cpumask_copy(&deploy_on_mask, cpu_online_mask); > + cpumask_clear_cpu(smp_processor_id(), &deploy_on_mask); > + atomic_add(cpumask_weight(&deploy_on_mask), &cpu_pmu->remaining_work); > + smp_mb__after_atomic(); What's this barrier needed for? > + > + for_each_cpu(cpu, &deploy_on_mask) { Why not for_each_online_cpu and then continue if cpu == smp_processor_id() ? I assume the race against hotplug is benign, as the interrupt will no longer be asserted to the GIC if the source CPU goes offline? > + struct pmu_hw_events *hw_events = > + per_cpu_ptr(cpu_pmu->hw_events, cpu); > + > + /* > + * The workaround code exits immediately without waiting to > + * see if the interrupt was handled by another CPU. This makes > + * it hard for us to decide between IRQ_HANDLED and IRQ_NONE. > + * However, the handler isn't shared so we don't have to worry > + * about being a good citizen, only about keeping the spurious > + * interrupt detector working. This allows us to return the > + * result of our *previous* attempt to deploy workaround. > + */ > + work_ret = atomic_read(&hw_events->work_ret); > + if (work_ret != IRQ_NONE) > + ret = work_ret; Is this actually necessary, or can we always return handled? > + > + if (!irq_work_queue_on(&hw_events->work, cpu)) > + if (atomic_dec_and_test(&cpu_pmu->remaining_work)) I'm not convinced that we can't have old work racing on the remaining work field with a subsequent interrupt. > + enable_irq(cpu_pmu->muxed_spi_workaround_irq); "This function (enable_irq) may be called from IRQ context only when desc->irq_data.chip->bus_lock and desc->chip->bus_sync_unlock are NULL !" Can we guarantee that in the general case? > + } > + > + return ret; > +} > + > +static int cpu_pmu_muxed_spi_workaround_init(struct arm_pmu *cpu_pmu) > +{ > + struct platform_device *pmu_device = cpu_pmu->plat_device; > + int cpu; > + > + for_each_possible_cpu(cpu) { > + struct pmu_hw_events *hw_events = > + per_cpu_ptr(cpu_pmu->hw_events, cpu); > + > + init_irq_work(&hw_events->work, cpu_pmu_do_percpu_work); > + atomic_set(&hw_events->work_ret, IRQ_HANDLED); > + } > + > + aomic_set(cpu_pmu->remaining_work, 0); So you didn't even build this... Will
On 08/01/15 17:30, Will Deacon wrote: > Hi Daniel, > > Some minor comments below... > > On Wed, Jan 07, 2015 at 12:28:18PM +0000, Daniel Thompson wrote: >> Some ARM platforms mux the PMU interrupt of every core into a single >> SPI. On such platforms if the PMU of any core except 0 raises an interrupt >> then it cannot be serviced and eventually, if you are lucky, the spurious >> irq detection might forcefully disable the interrupt. >> >> On these SoCs it is not possible to determine which core raised the >> interrupt so workaround this issue by queuing irqwork on the other >> cores whenever the primary interrupt handler is unable to service the >> interrupt. >> >> The u8500 platform has an alternative workaround that dynamically alters >> the affinity of the PMU interrupt. This workaround logic is no longer >> required so the original code is removed as is the hook it relied upon. >> >> Tested on imx6q (which has fours cores/PMUs all muxed to a single SPI). >> >> Signed-off-by: Daniel Thompson <daniel.thompson@linaro.org> > > [...] > >> diff --git a/arch/arm/kernel/perf_event_cpu.c b/arch/arm/kernel/perf_event_cpu.c >> index dd9acc95ebc0..3d51c5f442eb 100644 >> --- a/arch/arm/kernel/perf_event_cpu.c >> +++ b/arch/arm/kernel/perf_event_cpu.c >> @@ -59,6 +59,116 @@ int perf_num_counters(void) >> } >> EXPORT_SYMBOL_GPL(perf_num_counters); >> >> +#ifdef CONFIG_SMP >> +/* >> + * Workaround logic that is distributed to all cores if the PMU has only >> + * a single IRQ and the CPU receiving that IRQ cannot handle it. Its >> + * job is to try to service the interrupt on the current CPU. It will >> + * also enable the IRQ again if all the other CPUs have already tried to >> + * service it. >> + */ >> +static void cpu_pmu_do_percpu_work(struct irq_work *w) >> +{ >> + struct pmu_hw_events *hw_events = >> + container_of(w, struct pmu_hw_events, work); >> + struct arm_pmu *cpu_pmu = hw_events->percpu_pmu; >> + >> + atomic_set(&hw_events->work_ret, >> + cpu_pmu->handle_irq(0, cpu_pmu)); > > Do you need a memory barrier here, or is that implued by enable_irq? We are more getting away without a memory barrier... the spurious interrupt detector won't really mind if we see an out of date value (since it can tolerate getting the value wrong sometimes). I think we can moot this issue though by removing the code. See below... >> + if (atomic_dec_and_test(&cpu_pmu->remaining_work)) >> + enable_irq(cpu_pmu->muxed_spi_workaround_irq); >> +} >> + >> +/* >> + * Called when the main interrupt handler cannot determine the source >> + * of interrupt. It will deploy a workaround if we are running on an SMP >> + * platform with only a single muxed SPI. >> + * >> + * The workaround disables the interrupt and distributes irqwork to all >> + * other processors in the system. Hopefully one of them will clear the >> + * interrupt... >> + */ >> +static irqreturn_t cpu_pmu_handle_irq_none(int irq_num, struct arm_pmu *cpu_pmu) >> +{ >> + irqreturn_t ret = IRQ_NONE; >> + cpumask_t deploy_on_mask; >> + int cpu, work_ret; >> + if (irq_num != cpu_pmu->muxed_spi_workaround_irq) >> + return IRQ_NONE; > > return ret ? In general I prefer only to take return variables from variables on control paths where the return value is not constant. However this will also become moot if I remove the work_ret logic. >> + >> + disable_irq_nosync(cpu_pmu->muxed_spi_workaround_irq); >> + >> + cpumask_copy(&deploy_on_mask, cpu_online_mask); >> + cpumask_clear_cpu(smp_processor_id(), &deploy_on_mask); >> + atomic_add(cpumask_weight(&deploy_on_mask), &cpu_pmu->remaining_work); >> + smp_mb__after_atomic(); > > What's this barrier needed for? It pairs up with the implicit barrier in atomic_dec_and_test() and ensures the increment happens before the decrement (and therefore that the interrupt will be re-enabled). It can be removed providing we can rely on there being an implicit barrier in irq_work_queue_on(). Internally this function uses arch_send_call_function_single_ipi() there is definitely a barrier for that all current arm and arm64 platforms. I will remove this. >> + >> + for_each_cpu(cpu, &deploy_on_mask) { > > Why not for_each_online_cpu and then continue if cpu == smp_processor_id() ? > I assume the race against hotplug is benign, as the interrupt will no longer > be asserted to the GIC if the source CPU goes offline? If cpu_online_mask changes after we have performed the atomic_add() but before (or during) the loop then we would mis-manage the value of remaining_work might fail to re-enable the interrupt. >> + struct pmu_hw_events *hw_events = >> + per_cpu_ptr(cpu_pmu->hw_events, cpu); >> + >> + /* >> + * The workaround code exits immediately without waiting to >> + * see if the interrupt was handled by another CPU. This makes >> + * it hard for us to decide between IRQ_HANDLED and IRQ_NONE. >> + * However, the handler isn't shared so we don't have to worry >> + * about being a good citizen, only about keeping the spurious >> + * interrupt detector working. This allows us to return the >> + * result of our *previous* attempt to deploy workaround. >> + */ >> + work_ret = atomic_read(&hw_events->work_ret); >> + if (work_ret != IRQ_NONE) >> + ret = work_ret; > > Is this actually necessary, or can we always return handled? Ultimately it depends if we need the spurious interrupt detection logic to work. The work_ret approach is rather nasty (and most of your review comments are linked to it one way or another). Thus I think I'm OK to remove this altogether; spurious interrupts are very unlikely for the PMU IRQ. >> + >> + if (!irq_work_queue_on(&hw_events->work, cpu)) >> + if (atomic_dec_and_test(&cpu_pmu->remaining_work)) > > I'm not convinced that we can't have old work racing on the remaining work > field with a subsequent interrupt. I don't think that can happen. For the interrupt to be re-enabled all cores must the started executing their irq_work handlers and called atomic_dec_and_test(). Even though they may not have completed the pending flag is cleared before they are called making it safe to re-trigger them. In fact even if that were not the case the error path for irq_work_queue_on() would resolve the problem for us (at the cost of re-entering the interrupt handler during races). >> + enable_irq(cpu_pmu->muxed_spi_workaround_irq); > > "This function (enable_irq) may be called from IRQ context only when > desc->irq_data.chip->bus_lock and desc->chip->bus_sync_unlock are NULL !" > > Can we guarantee that in the general case? For the PMU I think we can. The irqchips that use these callbacks are found in one of the following directories: arch/mips drivers/base/regmap drivers/gpio drivers/mfd drivers/platform/x86 All of above would be a pretty astonishing way to hook up an intimate peripheral like the PMU. >> + } >> + >> + return ret; >> +} >> + >> +static int cpu_pmu_muxed_spi_workaround_init(struct arm_pmu *cpu_pmu) >> +{ >> + struct platform_device *pmu_device = cpu_pmu->plat_device; >> + int cpu; >> + >> + for_each_possible_cpu(cpu) { >> + struct pmu_hw_events *hw_events = >> + per_cpu_ptr(cpu_pmu->hw_events, cpu); >> + >> + init_irq_work(&hw_events->work, cpu_pmu_do_percpu_work); >> + atomic_set(&hw_events->work_ret, IRQ_HANDLED); >> + } >> + >> + aomic_set(cpu_pmu->remaining_work, 0); > > So you didn't even build this... Oh no. I built... I found that typo... I fixed... I soak tested for an hour on i.MX6 (because I had changed the function from which we deploy the workaround since v2). After all that overlooking regenerating the patch before posting it really was rather foolish. Sorry. BTW I have just done a side by side diff with what I posted and what is in my git repo. The only other difference between what I tested and what I posted was a minor whitespace change.
diff --git a/arch/arm/include/asm/pmu.h b/arch/arm/include/asm/pmu.h index b1596bd59129..295e762d5116 100644 --- a/arch/arm/include/asm/pmu.h +++ b/arch/arm/include/asm/pmu.h @@ -87,6 +87,19 @@ struct pmu_hw_events { * already have to allocate this struct per cpu. */ struct arm_pmu *percpu_pmu; + +#ifdef CONFIG_SMP + /* + * This is used to schedule workaround logic on platforms where all + * the PMUs are attached to a single SPI. + */ + struct irq_work work; + + /* + * Used to track state when deploying above workaround. + */ + atomic_t work_ret; +#endif }; struct arm_pmu { @@ -117,6 +130,10 @@ struct arm_pmu { struct platform_device *plat_device; struct pmu_hw_events __percpu *hw_events; struct notifier_block hotplug_nb; +#ifdef CONFIG_SMP + int muxed_spi_workaround_irq; + atomic_t remaining_work; +#endif }; #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu)) diff --git a/arch/arm/kernel/perf_event.c b/arch/arm/kernel/perf_event.c index f7c65adaa428..e5c537b57f94 100644 --- a/arch/arm/kernel/perf_event.c +++ b/arch/arm/kernel/perf_event.c @@ -299,8 +299,6 @@ validate_group(struct perf_event *event) static irqreturn_t armpmu_dispatch_irq(int irq, void *dev) { struct arm_pmu *armpmu; - struct platform_device *plat_device; - struct arm_pmu_platdata *plat; int ret; u64 start_clock, finish_clock; @@ -311,14 +309,9 @@ static irqreturn_t armpmu_dispatch_irq(int irq, void *dev) * dereference. */ armpmu = *(void **)dev; - plat_device = armpmu->plat_device; - plat = dev_get_platdata(&plat_device->dev); start_clock = sched_clock(); - if (plat && plat->handle_irq) - ret = plat->handle_irq(irq, armpmu, armpmu->handle_irq); - else - ret = armpmu->handle_irq(irq, armpmu); + ret = armpmu->handle_irq(irq, armpmu); finish_clock = sched_clock(); perf_sample_event_took(finish_clock - start_clock); diff --git a/arch/arm/kernel/perf_event_cpu.c b/arch/arm/kernel/perf_event_cpu.c index dd9acc95ebc0..3d51c5f442eb 100644 --- a/arch/arm/kernel/perf_event_cpu.c +++ b/arch/arm/kernel/perf_event_cpu.c @@ -59,6 +59,116 @@ int perf_num_counters(void) } EXPORT_SYMBOL_GPL(perf_num_counters); +#ifdef CONFIG_SMP +/* + * Workaround logic that is distributed to all cores if the PMU has only + * a single IRQ and the CPU receiving that IRQ cannot handle it. Its + * job is to try to service the interrupt on the current CPU. It will + * also enable the IRQ again if all the other CPUs have already tried to + * service it. + */ +static void cpu_pmu_do_percpu_work(struct irq_work *w) +{ + struct pmu_hw_events *hw_events = + container_of(w, struct pmu_hw_events, work); + struct arm_pmu *cpu_pmu = hw_events->percpu_pmu; + + atomic_set(&hw_events->work_ret, + cpu_pmu->handle_irq(0, cpu_pmu)); + + if (atomic_dec_and_test(&cpu_pmu->remaining_work)) + enable_irq(cpu_pmu->muxed_spi_workaround_irq); +} + +/* + * Called when the main interrupt handler cannot determine the source + * of interrupt. It will deploy a workaround if we are running on an SMP + * platform with only a single muxed SPI. + * + * The workaround disables the interrupt and distributes irqwork to all + * other processors in the system. Hopefully one of them will clear the + * interrupt... + */ +static irqreturn_t cpu_pmu_handle_irq_none(int irq_num, struct arm_pmu *cpu_pmu) +{ + irqreturn_t ret = IRQ_NONE; + cpumask_t deploy_on_mask; + int cpu, work_ret; + + if (irq_num != cpu_pmu->muxed_spi_workaround_irq) + return IRQ_NONE; + + disable_irq_nosync(cpu_pmu->muxed_spi_workaround_irq); + + cpumask_copy(&deploy_on_mask, cpu_online_mask); + cpumask_clear_cpu(smp_processor_id(), &deploy_on_mask); + atomic_add(cpumask_weight(&deploy_on_mask), &cpu_pmu->remaining_work); + smp_mb__after_atomic(); + + for_each_cpu(cpu, &deploy_on_mask) { + struct pmu_hw_events *hw_events = + per_cpu_ptr(cpu_pmu->hw_events, cpu); + + /* + * The workaround code exits immediately without waiting to + * see if the interrupt was handled by another CPU. This makes + * it hard for us to decide between IRQ_HANDLED and IRQ_NONE. + * However, the handler isn't shared so we don't have to worry + * about being a good citizen, only about keeping the spurious + * interrupt detector working. This allows us to return the + * result of our *previous* attempt to deploy workaround. + */ + work_ret = atomic_read(&hw_events->work_ret); + if (work_ret != IRQ_NONE) + ret = work_ret; + + if (!irq_work_queue_on(&hw_events->work, cpu)) + if (atomic_dec_and_test(&cpu_pmu->remaining_work)) + enable_irq(cpu_pmu->muxed_spi_workaround_irq); + } + + return ret; +} + +static int cpu_pmu_muxed_spi_workaround_init(struct arm_pmu *cpu_pmu) +{ + struct platform_device *pmu_device = cpu_pmu->plat_device; + int cpu; + + for_each_possible_cpu(cpu) { + struct pmu_hw_events *hw_events = + per_cpu_ptr(cpu_pmu->hw_events, cpu); + + init_irq_work(&hw_events->work, cpu_pmu_do_percpu_work); + atomic_set(&hw_events->work_ret, IRQ_HANDLED); + } + + aomic_set(cpu_pmu->remaining_work, 0); + cpu_pmu->muxed_spi_workaround_irq = platform_get_irq(pmu_device, 0); + + return 0; +} + +static void cpu_pmu_muxed_spi_workaround_term(struct arm_pmu *cpu_pmu) +{ + cpu_pmu->muxed_spi_workaround_irq = 0; +} +#else /* CONFIG_SMP */ +static int cpu_pmu_muxed_spi_workaround_init(struct arm_pmu *cpu_pmu) +{ + return 0; +} + +static void cpu_pmu_muxed_spi_workaround_term(struct arm_pmu *cpu_pmu) +{ +} + +static irqreturn_t cpu_pmu_handle_irq_none(int irq_num, struct arm_pmu *cpu_pmu) +{ + return IRQ_NONE; +} +#endif /* CONFIG_SMP */ + /* Include the PMU-specific implementations. */ #include "perf_event_xscale.c" #include "perf_event_v6.c" @@ -98,6 +208,8 @@ static void cpu_pmu_free_irq(struct arm_pmu *cpu_pmu) if (irq >= 0) free_irq(irq, per_cpu_ptr(&hw_events->percpu_pmu, i)); } + + cpu_pmu_muxed_spi_workaround_term(cpu_pmu); } } @@ -155,6 +267,16 @@ static int cpu_pmu_request_irq(struct arm_pmu *cpu_pmu, irq_handler_t handler) cpumask_set_cpu(i, &cpu_pmu->active_irqs); } + + /* + * If we are running SMP and have only one interrupt source + * then get ready to share that single irq among the cores. + */ + if (nr_cpu_ids > 1 && irqs == 1) { + err = cpu_pmu_muxed_spi_workaround_init(cpu_pmu); + if (err) + return err; + } } return 0; diff --git a/arch/arm/kernel/perf_event_v7.c b/arch/arm/kernel/perf_event_v7.c index 8993770c47de..0dd914c10803 100644 --- a/arch/arm/kernel/perf_event_v7.c +++ b/arch/arm/kernel/perf_event_v7.c @@ -792,7 +792,7 @@ static irqreturn_t armv7pmu_handle_irq(int irq_num, void *dev) * Did an overflow occur? */ if (!armv7_pmnc_has_overflowed(pmnc)) - return IRQ_NONE; + return cpu_pmu_handle_irq_none(irq_num, cpu_pmu); /* * Handle the counter(s) overflow(s) diff --git a/arch/arm/mach-ux500/cpu-db8500.c b/arch/arm/mach-ux500/cpu-db8500.c index 6f63954c8bde..917774999c5c 100644 --- a/arch/arm/mach-ux500/cpu-db8500.c +++ b/arch/arm/mach-ux500/cpu-db8500.c @@ -12,8 +12,6 @@ #include <linux/init.h> #include <linux/device.h> #include <linux/amba/bus.h> -#include <linux/interrupt.h> -#include <linux/irq.h> #include <linux/platform_device.h> #include <linux/io.h> #include <linux/mfd/abx500/ab8500.h> @@ -23,7 +21,6 @@ #include <linux/regulator/machine.h> #include <linux/random.h> -#include <asm/pmu.h> #include <asm/mach/map.h> #include "setup.h" @@ -99,30 +96,6 @@ static void __init u8500_map_io(void) iotable_init(u8500_io_desc, ARRAY_SIZE(u8500_io_desc)); } -/* - * The PMU IRQ lines of two cores are wired together into a single interrupt. - * Bounce the interrupt to the other core if it's not ours. - */ -static irqreturn_t db8500_pmu_handler(int irq, void *dev, irq_handler_t handler) -{ - irqreturn_t ret = handler(irq, dev); - int other = !smp_processor_id(); - - if (ret == IRQ_NONE && cpu_online(other)) - irq_set_affinity(irq, cpumask_of(other)); - - /* - * We should be able to get away with the amount of IRQ_NONEs we give, - * while still having the spurious IRQ detection code kick in if the - * interrupt really starts hitting spuriously. - */ - return ret; -} - -static struct arm_pmu_platdata db8500_pmu_platdata = { - .handle_irq = db8500_pmu_handler, -}; - static const char *db8500_read_soc_id(void) { void __iomem *uid = __io_address(U8500_BB_UID_BASE); @@ -143,8 +116,6 @@ static struct device * __init db8500_soc_device_init(void) } static struct of_dev_auxdata u8500_auxdata_lookup[] __initdata = { - /* Requires call-back bindings. */ - OF_DEV_AUXDATA("arm,cortex-a9-pmu", 0, "arm-pmu", &db8500_pmu_platdata), /* Requires DMA bindings. */ OF_DEV_AUXDATA("stericsson,ux500-msp-i2s", 0x80123000, "ux500-msp-i2s.0", &msp0_platform_data),
Some ARM platforms mux the PMU interrupt of every core into a single SPI. On such platforms if the PMU of any core except 0 raises an interrupt then it cannot be serviced and eventually, if you are lucky, the spurious irq detection might forcefully disable the interrupt. On these SoCs it is not possible to determine which core raised the interrupt so workaround this issue by queuing irqwork on the other cores whenever the primary interrupt handler is unable to service the interrupt. The u8500 platform has an alternative workaround that dynamically alters the affinity of the PMU interrupt. This workaround logic is no longer required so the original code is removed as is the hook it relied upon. Tested on imx6q (which has fours cores/PMUs all muxed to a single SPI). Signed-off-by: Daniel Thompson <daniel.thompson@linaro.org> --- Notes: v2 was tested on u8500 (thanks to Linus Walleij). v3 doesn't change anything conceptual but the changes were just enough for me not to preserve the Tested-By:. v3: * Removed function pointer indirection when deploying workaround code and reorganise the code accordingly (Mark Rutland). * Move the workaround state tracking into the existing percpu data structure (Mark Rutland). * Renamed cret to percpu_ret and rewrote the comment describing the purpose of this variable (Mark Rutland). * Copy the cpu_online_mask and use that to act on a consistent set of cpus throughout the workaround (Mark Rutland). * Changed "single_irq" to "muxed_spi" to more explicitly describe the problem. v2: * Fixed build problems on systems without SMP. v1: * Thanks to Lucas Stach, Russell King and Thomas Gleixner for critiquing an older, completely different way to tackle the same problem. arch/arm/include/asm/pmu.h | 17 ++++++ arch/arm/kernel/perf_event.c | 9 +-- arch/arm/kernel/perf_event_cpu.c | 122 +++++++++++++++++++++++++++++++++++++++ arch/arm/kernel/perf_event_v7.c | 2 +- arch/arm/mach-ux500/cpu-db8500.c | 29 ---------- 5 files changed, 141 insertions(+), 38 deletions(-) -- 1.9.3