Re: [PATCH v15 9/9] irqchip/apple-aic: Move over to core ipi-mux
From: Hector Martin
Date: Mon Dec 05 2022 - 02:15:11 EST
On 03/12/2022 15.46, Anup Patel wrote:
> From: Marc Zyngier <maz@xxxxxxxxxx>
>
> Now that the complexity of the AIC IPI mux has been copied into
> the core code for the benefit of the riscv architecture,
> shrink the AIC driver by the same amount by using that infrastructure.
>
> Signed-off-by: Marc Zyngier <maz@xxxxxxxxxx>
> ---
> drivers/irqchip/Kconfig | 1 +
> drivers/irqchip/irq-apple-aic.c | 161 ++------------------------------
> 2 files changed, 9 insertions(+), 153 deletions(-)
>
> diff --git a/drivers/irqchip/Kconfig b/drivers/irqchip/Kconfig
> index 131379aa8424..0fd452b49ed9 100644
> --- a/drivers/irqchip/Kconfig
> +++ b/drivers/irqchip/Kconfig
> @@ -675,6 +675,7 @@ config APPLE_AIC
> bool "Apple Interrupt Controller (AIC)"
> depends on ARM64
> depends on ARCH_APPLE || COMPILE_TEST
> + select GENERIC_IRQ_IPI_MUX
> help
> Support for the Apple Interrupt Controller found on Apple Silicon SoCs,
> such as the M1.
> diff --git a/drivers/irqchip/irq-apple-aic.c b/drivers/irqchip/irq-apple-aic.c
> index 1c2813ad8bbe..1fa81fcff788 100644
> --- a/drivers/irqchip/irq-apple-aic.c
> +++ b/drivers/irqchip/irq-apple-aic.c
> @@ -292,7 +292,6 @@ struct aic_irq_chip {
> void __iomem *base;
> void __iomem *event;
> struct irq_domain *hw_domain;
> - struct irq_domain *ipi_domain;
> struct {
> cpumask_t aff;
> } *fiq_aff[AIC_NR_FIQ];
> @@ -307,9 +306,6 @@ struct aic_irq_chip {
>
> static DEFINE_PER_CPU(uint32_t, aic_fiq_unmasked);
>
> -static DEFINE_PER_CPU(atomic_t, aic_vipi_flag);
> -static DEFINE_PER_CPU(atomic_t, aic_vipi_enable);
> -
> static struct aic_irq_chip *aic_irqc;
>
> static void aic_handle_ipi(struct pt_regs *regs);
> @@ -751,98 +747,8 @@ static void aic_ipi_send_fast(int cpu)
> isb();
> }
>
> -static void aic_ipi_mask(struct irq_data *d)
> -{
> - u32 irq_bit = BIT(irqd_to_hwirq(d));
> -
> - /* No specific ordering requirements needed here. */
> - atomic_andnot(irq_bit, this_cpu_ptr(&aic_vipi_enable));
> -}
> -
> -static void aic_ipi_unmask(struct irq_data *d)
> -{
> - struct aic_irq_chip *ic = irq_data_get_irq_chip_data(d);
> - u32 irq_bit = BIT(irqd_to_hwirq(d));
> -
> - atomic_or(irq_bit, this_cpu_ptr(&aic_vipi_enable));
> -
> - /*
> - * The atomic_or() above must complete before the atomic_read()
> - * below to avoid racing aic_ipi_send_mask().
> - */
> - smp_mb__after_atomic();
> -
> - /*
> - * If a pending vIPI was unmasked, raise a HW IPI to ourselves.
> - * No barriers needed here since this is a self-IPI.
> - */
> - if (atomic_read(this_cpu_ptr(&aic_vipi_flag)) & irq_bit) {
> - if (static_branch_likely(&use_fast_ipi))
> - aic_ipi_send_fast(smp_processor_id());
> - else
> - aic_ic_write(ic, AIC_IPI_SEND, AIC_IPI_SEND_CPU(smp_processor_id()));
> - }
> -}
> -
> -static void aic_ipi_send_mask(struct irq_data *d, const struct cpumask *mask)
> -{
> - struct aic_irq_chip *ic = irq_data_get_irq_chip_data(d);
> - u32 irq_bit = BIT(irqd_to_hwirq(d));
> - u32 send = 0;
> - int cpu;
> - unsigned long pending;
> -
> - for_each_cpu(cpu, mask) {
> - /*
> - * This sequence is the mirror of the one in aic_ipi_unmask();
> - * see the comment there. Additionally, release semantics
> - * ensure that the vIPI flag set is ordered after any shared
> - * memory accesses that precede it. This therefore also pairs
> - * with the atomic_fetch_andnot in aic_handle_ipi().
> - */
> - pending = atomic_fetch_or_release(irq_bit, per_cpu_ptr(&aic_vipi_flag, cpu));
> -
> - /*
> - * The atomic_fetch_or_release() above must complete before the
> - * atomic_read() below to avoid racing aic_ipi_unmask().
> - */
> - smp_mb__after_atomic();
> -
> - if (!(pending & irq_bit) &&
> - (atomic_read(per_cpu_ptr(&aic_vipi_enable, cpu)) & irq_bit)) {
> - if (static_branch_likely(&use_fast_ipi))
> - aic_ipi_send_fast(cpu);
> - else
> - send |= AIC_IPI_SEND_CPU(cpu);
> - }
> - }
> -
> - /*
> - * The flag writes must complete before the physical IPI is issued
> - * to another CPU. This is implied by the control dependency on
> - * the result of atomic_read_acquire() above, which is itself
> - * already ordered after the vIPI flag write.
> - */
> - if (send)
> - aic_ic_write(ic, AIC_IPI_SEND, send);
> -}
> -
> -static struct irq_chip ipi_chip = {
> - .name = "AIC-IPI",
> - .irq_mask = aic_ipi_mask,
> - .irq_unmask = aic_ipi_unmask,
> - .ipi_send_mask = aic_ipi_send_mask,
> -};
> -
> -/*
> - * IPI IRQ domain
> - */
> -
> static void aic_handle_ipi(struct pt_regs *regs)
> {
> - int i;
> - unsigned long enabled, firing;
> -
> /*
> * Ack the IPI. We need to order this after the AIC event read, but
> * that is enforced by normal MMIO ordering guarantees.
> @@ -857,27 +763,7 @@ static void aic_handle_ipi(struct pt_regs *regs)
> aic_ic_write(aic_irqc, AIC_IPI_ACK, AIC_IPI_OTHER);
> }
>
> - /*
> - * The mask read does not need to be ordered. Only we can change
> - * our own mask anyway, so no races are possible here, as long as
> - * we are properly in the interrupt handler (which is covered by
> - * the barrier that is part of the top-level AIC handler's readl()).
> - */
> - enabled = atomic_read(this_cpu_ptr(&aic_vipi_enable));
> -
> - /*
> - * Clear the IPIs we are about to handle. This pairs with the
> - * atomic_fetch_or_release() in aic_ipi_send_mask(), and needs to be
> - * ordered after the aic_ic_write() above (to avoid dropping vIPIs) and
> - * before IPI handling code (to avoid races handling vIPIs before they
> - * are signaled). The former is taken care of by the release semantics
> - * of the write portion, while the latter is taken care of by the
> - * acquire semantics of the read portion.
> - */
> - firing = atomic_fetch_andnot(enabled, this_cpu_ptr(&aic_vipi_flag)) & enabled;
> -
> - for_each_set_bit(i, &firing, AIC_NR_SWIPI)
> - generic_handle_domain_irq(aic_irqc->ipi_domain, i);
> + ipi_mux_process();
>
> /*
> * No ordering needed here; at worst this just changes the timing of
> @@ -887,55 +773,24 @@ static void aic_handle_ipi(struct pt_regs *regs)
> aic_ic_write(aic_irqc, AIC_IPI_MASK_CLR, AIC_IPI_OTHER);
> }
>
> -static int aic_ipi_alloc(struct irq_domain *d, unsigned int virq,
> - unsigned int nr_irqs, void *args)
> +static void aic_ipi_send_single(unsigned int cpu)
> {
> - int i;
> -
> - for (i = 0; i < nr_irqs; i++) {
> - irq_set_percpu_devid(virq + i);
> - irq_domain_set_info(d, virq + i, i, &ipi_chip, d->host_data,
> - handle_percpu_devid_irq, NULL, NULL);
> - }
> -
> - return 0;
> -}
> -
> -static void aic_ipi_free(struct irq_domain *d, unsigned int virq, unsigned int nr_irqs)
> -{
> - /* Not freeing IPIs */
> + if (static_branch_likely(&use_fast_ipi))
> + aic_ipi_send_fast(cpu);
> + else
> + aic_ic_write(aic_irqc, AIC_IPI_SEND, AIC_IPI_SEND_CPU(cpu));
> }
>
> -static const struct irq_domain_ops aic_ipi_domain_ops = {
> - .alloc = aic_ipi_alloc,
> - .free = aic_ipi_free,
> -};
> -
> static int __init aic_init_smp(struct aic_irq_chip *irqc, struct device_node *node)
> {
> - struct irq_domain *ipi_domain;
> int base_ipi;
>
> - ipi_domain = irq_domain_create_linear(irqc->hw_domain->fwnode, AIC_NR_SWIPI,
> - &aic_ipi_domain_ops, irqc);
> - if (WARN_ON(!ipi_domain))
> - return -ENODEV;
> -
> - ipi_domain->flags |= IRQ_DOMAIN_FLAG_IPI_SINGLE;
> - irq_domain_update_bus_token(ipi_domain, DOMAIN_BUS_IPI);
> -
> - base_ipi = __irq_domain_alloc_irqs(ipi_domain, -1, AIC_NR_SWIPI,
> - NUMA_NO_NODE, NULL, false, NULL);
> -
> - if (WARN_ON(!base_ipi)) {
> - irq_domain_remove(ipi_domain);
> + base_ipi = ipi_mux_create(AIC_NR_SWIPI, aic_ipi_send_single);
> + if (WARN_ON(base_ipi <= 0))
> return -ENODEV;
> - }
>
> set_smp_ipi_range(base_ipi, AIC_NR_SWIPI);
>
> - irqc->ipi_domain = ipi_domain;
> -
> return 0;
> }
>
Reviewed-by: Hector Martin <marcan@xxxxxxxxx>
Tested-by: Hector Martin <marcan@xxxxxxxxx>
Smoke tested on t8112, nothing exploded :)
- Hector