Re: [PATCH v5 6/6] perf: ARM DynamIQ Shared Unit PMU support
From: Mark Rutland
Date: Wed Aug 16 2017 - 10:12:15 EST
On Tue, Aug 08, 2017 at 12:37:26PM +0100, Suzuki K Poulose wrote:
> +/*
> + * struct dsu_pmu - DSU PMU descriptor
> + *
> + * @pmu_lock : Protects accesses to DSU PMU register from multiple
> + * CPUs.
> + * @hw_events : Holds the event counter state.
> + * @supported_cpus : CPUs attached to the DSU.
> + * @active_cpu : CPU to which the PMU is bound for accesses.
> + * @cpuhp_node : Node for CPU hotplug notifier link.
> + * @num_counters : Number of event counters implemented by the PMU,
> + * excluding the cycle counter.
> + * @irq : Interrupt line for counter overflow.
> + * @cpmceid_bitmap : Bitmap for the availability of architected common
> + * events (event_code < 0x40).
> + */
> +struct dsu_pmu {
> + struct pmu pmu;
> + struct device *dev;
> + raw_spinlock_t pmu_lock;
I'm in two minds about pmu_lock. The core has to take the ctx lock for
most (all?) operations, and we only allow events to be opened on a
particular CPU, so there shouldn't be concurrent accesses from other
CPUs.
We do need to disable interrupts in order to serialise against the IRQ
handler, and disabling IRQs doesn't make it clear that we're protecting
a particular resource.
Could we update the description to make it clear that the pmu_lock is
used to serialise against the IRQ handler? It also happens to protect
cross-cpu accesses, but those would be a bug.
Other PMU drivers have locks which may not be necessary; I can try to
clean those up later.
[...]
> +static struct attribute *dsu_pmu_event_attrs[] = {
> + DSU_EVENT_ATTR(cycles, 0x11),
> + DSU_EVENT_ATTR(bus_acecss, 0x19),
> + DSU_EVENT_ATTR(memory_error, 0x1a),
> + DSU_EVENT_ATTR(bus_cycles, 0x1d),
> + DSU_EVENT_ATTR(l3d_cache_allocate, 0x29),
> + DSU_EVENT_ATTR(l3d_cache_refill, 0x2a),
> + DSU_EVENT_ATTR(l3d_cache, 0x2b),
> + DSU_EVENT_ATTR(l3d_cache_wb, 0x2c),
MAX_COMMON_EVENTS seems to be 0x40, so are we just assuming the below
are implemented?
If so, why bother exposing them at all? We can't know that they're going
to work...
> + DSU_EVENT_ATTR(bus_access_rd, 0x60),
> + DSU_EVENT_ATTR(bus_access_wr, 0x61),
> + DSU_EVENT_ATTR(bus_access_shared, 0x62),
> + DSU_EVENT_ATTR(bus_access_not_shared, 0x63),
> + DSU_EVENT_ATTR(bus_access_normal, 0x64),
> + DSU_EVENT_ATTR(bus_access_periph, 0x65),
> + DSU_EVENT_ATTR(l3d_cache_rd, 0xa0),
> + DSU_EVENT_ATTR(l3d_cache_wr, 0xa1),
> + DSU_EVENT_ATTR(l3d_cache_refill_rd, 0xa2),
> + DSU_EVENT_ATTR(l3d_cache_refill_wr, 0xa3),
> + DSU_EVENT_ATTR(acp_access, 0x119),
> + DSU_EVENT_ATTR(acp_cycles, 0x11d),
> + DSU_EVENT_ATTR(acp_access_rd, 0x160),
> + DSU_EVENT_ATTR(acp_access_wr, 0x161),
> + DSU_EVENT_ATTR(pp_access, 0x219),
> + DSU_EVENT_ATTR(pp_cycles, 0x21d),
> + DSU_EVENT_ATTR(pp_access_rd, 0x260),
> + DSU_EVENT_ATTR(pp_access_wr, 0x261),
> + DSU_EVENT_ATTR(scu_snp_access, 0xc0),
> + DSU_EVENT_ATTR(scu_snp_evict, 0xc1),
> + DSU_EVENT_ATTR(scu_snp_access_cpu, 0xc2),
> + DSU_EVENT_ATTR(scu_pftch_cpu_access, 0x500),
> + DSU_EVENT_ATTR(scu_pftch_cpu_miss, 0x501),
> + DSU_EVENT_ATTR(scu_pftch_cpu_hit, 0x502),
> + DSU_EVENT_ATTR(scu_pftch_cpu_match, 0x503),
> + DSU_EVENT_ATTR(scu_pftch_cpu_kill, 0x504),
> + DSU_EVENT_ATTR(scu_stash_icn_access, 0x510),
> + DSU_EVENT_ATTR(scu_stash_icn_miss, 0x511),
> + DSU_EVENT_ATTR(scu_stash_icn_hit, 0x512),
> + DSU_EVENT_ATTR(scu_stash_icn_match, 0x513),
> + DSU_EVENT_ATTR(scu_stash_icn_kill, 0x514),
> + DSU_EVENT_ATTR(scu_hzd_address, 0xd0),
> + NULL,
> +};
> +
> +static bool dsu_pmu_event_supported(struct dsu_pmu *dsu_pmu, unsigned long evt)
> +{
> + /*
> + * DSU PMU provides a bit map for events with
> + * id < DSU_PMU_MAX_COMMON_EVENTS.
> + * Events above the range are reported as supported, as
> + * tracking the support needs per-chip tables and makes
> + * it difficult to track. If an event is not supported,
> + * it won't be counted.
> + */
> + if (evt >= DSU_PMU_MAX_COMMON_EVENTS)
> + return true;
> + /* The PMU driver doesn't support chain mode */
> + if (evt == DSU_PMU_EVT_CHAIN)
> + return false;
> + return test_bit(evt, dsu_pmu->cpmceid_bitmap);
> +}
I don't think we need to use this for event_init() (more on that below),
so I think this can be simplified.
[...]
> +static struct attribute *dsu_pmu_cpumask_attrs[] = {
> + DSU_CPUMASK_ATTR(cpumask, DSU_ACTIVE_CPU_MASK),
> + DSU_CPUMASK_ATTR(supported_cpus, DSU_SUPPORTED_CPU_MASK),
> + NULL,
> +};
Normally we only expose one mask.
Why do we need the supported cpu mask? What is the intended use-case?
[...]
> +static inline u64 dsu_pmu_read_counter(struct perf_event *event)
> +{
> + u64 val = 0;
> + unsigned long flags;
> + struct dsu_pmu *dsu_pmu = to_dsu_pmu(event->pmu);
> + int idx = event->hw.idx;
> +
> + if (WARN_ON(!cpumask_test_cpu(smp_processor_id(),
> + &dsu_pmu->supported_cpus)))
> + return 0;
It would be better to check that this is the active CPU, since reading
on another supported CPU would still be wrong.
Likewise for the check in dsu_pmu_write_counter().
[...]
> +static void dsu_pmu_enable_counter(struct dsu_pmu *dsu_pmu, int idx)
> +{
> + unsigned long flags;
> +
> + raw_spin_lock_irqsave(&dsu_pmu->pmu_lock, flags);
> + __dsu_pmu_counter_interrupt_enable(idx);
> + __dsu_pmu_enable_counter(idx);
> + raw_spin_unlock_irqrestore(&dsu_pmu->pmu_lock, flags);
> +}
> +
> +static void dsu_pmu_disable_counter(struct dsu_pmu *dsu_pmu, int idx)
> +{
> + unsigned long flags;
> +
> + raw_spin_lock_irqsave(&dsu_pmu->pmu_lock, flags);
> + __dsu_pmu_disable_counter(idx);
> + __dsu_pmu_counter_interrupt_disable(idx);
> + raw_spin_unlock_irqrestore(&dsu_pmu->pmu_lock, flags);
> +}
Do we need the disable/enable IRQs across these? AFAICT these writing to
set/clr registers, so there's no RMW race to worry about so long as the
rest of the code does likewise.
[...]
> +static inline u32 dsu_pmu_get_status(void)
> +{
> + return __dsu_pmu_get_pmovsclr();
> +}
This could do with a better name. It's not clear what status it's meant
to get.
[...]
> +static irqreturn_t dsu_pmu_handle_irq(int irq_num, void *dev)
> +{
> + int i;
> + bool handled = false;
> + struct dsu_pmu *dsu_pmu = dev;
> + struct dsu_hw_events *hw_events = &dsu_pmu->hw_events;
> + unsigned long overflow, workset;
> +
> + overflow = (unsigned long)dsu_pmu_get_status();
This cast isn't necessary.
> + bitmap_and(&workset, &overflow, hw_events->used_mask,
> + DSU_PMU_MAX_HW_CNTRS);
> +
> + if (!workset)
> + return IRQ_NONE;
> +
> + for_each_set_bit(i, &workset, DSU_PMU_MAX_HW_CNTRS) {
> + struct perf_event *event = hw_events->events[i];
> +
> + if (WARN_ON(!event))
> + continue;
Unless we explicitly clear an event's overflow flag when we remove it,
this could happen in normal operation. This should probably have a
_ONCE(), if we think we need it.
[...]
> +static int dsu_pmu_add(struct perf_event *event, int flags)
> +{
> + struct dsu_pmu *dsu_pmu = to_dsu_pmu(event->pmu);
> + struct dsu_hw_events *hw_events = &dsu_pmu->hw_events;
> + struct hw_perf_event *hwc = &event->hw;
> + int idx;
> +
> + if (!cpumask_test_cpu(smp_processor_id(), &dsu_pmu->supported_cpus))
> + return -ENOENT;
This shouldn't ever happen, and we can check against the active cpumask,
with a WARN_ON_ONCE(). We have to do this for CPU PMUs since they
support events which can migrate with tasks, but that's not the case
here.
[...]
> +static int dsu_pmu_validate_event(struct pmu *pmu,
> + struct dsu_hw_events *hw_events,
> + struct perf_event *event)
> +{
> + if (is_software_event(event))
> + return 1;
> + /* Reject groups spanning multiple HW PMUs. */
> + if (event->pmu != pmu)
> + return 0;
> + if (event->state < PERF_EVENT_STATE_OFF)
> + return 1;
> + if (event->state == PERF_EVENT_STATE_OFF && !event->attr.enable_on_exec)
> + return 1;
This doesn't make sense for a !CPU pmu, and I think you can nuke the
prior test, too, which I think doesn't make sense.
> + return dsu_pmu_get_event_idx(hw_events, event) >= 0;
> +}
This whole function could be bool.
> +/*
> + * Make sure the group of events can be scheduled at once
> + * on the PMU.
> + */
> +static int dsu_pmu_validate_group(struct perf_event *event)
> +{
> + struct perf_event *sibling, *leader = event->group_leader;
> + struct dsu_hw_events fake_hw;
> +
> + if (event->group_leader == event)
> + return 0;
> +
> + memset(fake_hw.used_mask, 0, sizeof(fake_hw.used_mask));
> + if (!dsu_pmu_validate_event(event->pmu, &fake_hw, leader))
> + return -EINVAL;
> + list_for_each_entry(sibling, &leader->sibling_list, group_entry) {
> + if (!dsu_pmu_validate_event(event->pmu, &fake_hw, sibling))
> + return -EINVAL;
> + }
> + if (!dsu_pmu_validate_event(event->pmu, &fake_hw, event))
> + return -EINVAL;
> + return 0;
> +}
The return value has the opposite polarity to dsu_pmu_validate_event(),
which is rather confusing.
[...]
> +static int dsu_pmu_event_init(struct perf_event *event)
> +{
> + struct dsu_pmu *dsu_pmu = to_dsu_pmu(event->pmu);
> +
> + if (event->attr.type != event->pmu->type)
> + return -ENOENT;
> +
> + if (!dsu_pmu_event_supported(dsu_pmu, event->attr.config))
> + return -EOPNOTSUPP;
I can see why we track this for the sysfs nodes, but why do we bother
enforcing it for event_init()? Especially given we only enforce it for a
subset of values...
We don't do that for other PMUs, and assuming that this doesn't break
the HW, I'd rather we allowed users to request raw values, even if we
think the HW won't count them.
> +
> + /* We cannot support task bound events */
> + if (event->cpu < 0) {
> + dev_dbg(dsu_pmu->pmu.dev, "Can't support per-task counters\n");
> + return -EINVAL;
> + }
We should also check (event->attach_state & PERF_ATTACH_TASK), since you
could have a per-task event with a CPU filter, and that doesn't make
sense either.
[...]
> +/**
> + * dsu_pmu_dt_get_cpus: Get the list of CPUs in the cluster.
> + */
> +static int dsu_pmu_dt_get_cpus(struct device_node *dev, cpumask_t *mask)
> +{
> + int i = 0, n, cpu;
> + struct device_node *cpu_node;
> +
> + n = of_count_phandle_with_args(dev, "cpus", NULL);
> + if (n <= 0)
> + return -ENODEV;
> + for (; i < n; i++) {
> + cpu_node = of_parse_phandle(dev, "cpus", i);
> + if (!cpu_node)
> + break;
> + cpu = of_cpu_node_to_id(cpu_node);
> + of_node_put(cpu_node);
> + if (cpu < 0)
> + break;
I believe this can happen if the kernel's nr_cpus were capped below the
number of CPUs in the system. So we need to iterate all entries of the
cpus list regardless.
> + cpumask_set_cpu(cpu, mask);
> + }
> + if (i != n)
> + return -EINVAL;
> + return 0;
> +}
[...]
> +
> +/*
> + * dsu_pmu_probe_pmu: Probe the PMU details on a CPU in the cluster.
> + */
> +static void dsu_pmu_probe_pmu(void *data)
> +{
> + struct dsu_pmu *dsu_pmu = data;
> + u64 cpmcr;
> + u32 cpmceid[2];
> +
> + if (WARN_ON(!cpumask_test_cpu(smp_processor_id(),
> + &dsu_pmu->supported_cpus)))
> + return;
I can't see how this can happen unless core kernel code is critically
broken, given that the only caller uses smp_call_function_any() with the
supported_cpus mask.
Do we need this check?
> + cpmcr = __dsu_pmu_read_pmcr();
> + dsu_pmu->num_counters = ((cpmcr >> CLUSTERPMCR_N_SHIFT) &
> + CLUSTERPMCR_N_MASK);
> + if (!dsu_pmu->num_counters)
> + return;
Is that valid? What range of values makes sense?
[...]
> + /*
> + * Find one CPU in the DSU to handle the IRQs.
> + * It is highly unlikely that we would fail
> + * to find one, given the probing has succeeded.
> + */
> + cpu = dsu_pmu_get_online_cpu(dsu_pmu);
> + if (cpu >= nr_cpu_ids)
> + return -ENODEV;
> + cpumask_set_cpu(cpu, &dsu_pmu->active_cpu);
> + rc = irq_set_affinity_hint(irq, &dsu_pmu->active_cpu);
Is setting the affinity hint strong enough?
Can the affinity be changed behind the back of this driver?
> +static int dsu_pmu_cpu_teardown(unsigned int cpu, struct hlist_node *node)
> +{
> + int dst;
> + struct dsu_pmu *dsu_pmu = hlist_entry_safe(node, struct dsu_pmu,
> + cpuhp_node);
> +
> + if (!cpumask_test_and_clear_cpu(cpu, &dsu_pmu->active_cpu))
> + return 0;
> +
> + dst = dsu_pmu_get_online_cpu_any_but(dsu_pmu, cpu);
> + if (dst < nr_cpu_ids) {
> + cpumask_set_cpu(dst, &dsu_pmu->active_cpu);
> + perf_pmu_migrate_context(&dsu_pmu->pmu, cpu, dst);
> + irq_set_affinity_hint(dsu_pmu->irq, &dsu_pmu->active_cpu);
Likewise, is this sufficient?
Thanks,
Mark.