Re: [PATCH v7 3/8] perf: imx_perf: let the driver manage the counter usage rather the user
From: Frank Li
Date: Mon Mar 18 2024 - 12:25:19 EST
On Fri, Mar 15, 2024 at 05:55:50PM +0800, Xu Yang wrote:
> In current design, the user of perf app needs to input counter ID to count
> events. However, this is not user-friendly since the user needs to lookup
> the map table to find the counter. Instead of letting the user to input
> the counter, let this driver to manage the counters in this patch.
>
> This will be implemented by:
> 1. allocate counter 0 for cycle event.
> 2. find unused counter from 1-10 for reference events.
> 3. allocate specific counter for counter-specific events.
>
> In this patch, counter attribute is removed too. To mark counter-specific
> events, counter ID will be encoded into perf_pmu_events_attr.id.
>
> Signed-off-by: Xu Yang <xu.yang_2@xxxxxxx>
Reviewed-by: Frank Li <Frank.Li@xxxxxxx>
>
> ---
> Changes in v6:
> - new patch
> Changes in v7:
> - no changes
> ---
> drivers/perf/fsl_imx9_ddr_perf.c | 168 ++++++++++++++++++-------------
> 1 file changed, 99 insertions(+), 69 deletions(-)
>
> diff --git a/drivers/perf/fsl_imx9_ddr_perf.c b/drivers/perf/fsl_imx9_ddr_perf.c
> index 4ec70775d1f0..4fdf8bcf6646 100644
> --- a/drivers/perf/fsl_imx9_ddr_perf.c
> +++ b/drivers/perf/fsl_imx9_ddr_perf.c
> @@ -41,9 +41,11 @@
>
> #define NUM_COUNTERS 11
> #define CYCLES_COUNTER 0
> +#define CYCLES_EVENT_ID 0
>
> #define CONFIG_EVENT GENMASK(7, 0)
> #define CONFIG_COUNTER GENMASK(15, 8)
> +#define CONFIG_COUNTER_OFFSET 8
>
> #define to_ddr_pmu(p) container_of(p, struct ddr_pmu, pmu)
>
> @@ -130,6 +132,8 @@ static ssize_t ddr_pmu_event_show(struct device *dev,
> return sysfs_emit(page, "event=0x%02llx\n", pmu_attr->id);
> }
>
> +#define ID(counter, id) ((counter << CONFIG_COUNTER_OFFSET) | id)
> +
> #define IMX9_DDR_PMU_EVENT_ATTR(_name, _id) \
> (&((struct perf_pmu_events_attr[]) { \
> { .attr = __ATTR(_name, 0444, ddr_pmu_event_show, NULL),\
> @@ -162,81 +166,81 @@ static struct attribute *ddr_perf_events_attrs[] = {
> IMX9_DDR_PMU_EVENT_ATTR(ddrc_pm_29, 63),
>
> /* counter1 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_0, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_1, 65),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_2, 66),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_3, 67),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_4, 68),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_5, 69),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_6, 70),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_7, 71),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_0, ID(1, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_1, ID(1, 65)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_2, ID(1, 66)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_3, ID(1, 67)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_4, ID(1, 68)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_5, ID(1, 69)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_6, ID(1, 70)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_riq_7, ID(1, 71)),
>
> /* counter2 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_0, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_1, 65),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_2, 66),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_3, 67),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_4, 68),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_5, 69),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_6, 70),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_7, 71),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_empty, 72),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pm_rd_trans_filt, 73),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_0, ID(2, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_1, ID(2, 65)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_2, ID(2, 66)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_3, ID(2, 67)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_4, ID(2, 68)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_5, ID(2, 69)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_6, ID(2, 70)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_ld_wiq_7, ID(2, 71)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_empty, ID(2, 72)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pm_rd_trans_filt, ID(2, 73)),
>
> /* counter3 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_0, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_1, 65),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_2, 66),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_3, 67),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_4, 68),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_5, 69),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_6, 70),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_7, 71),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_full, 72),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pm_wr_trans_filt, 73),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_0, ID(3, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_1, ID(3, 65)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_2, ID(3, 66)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_3, ID(3, 67)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_4, ID(3, 68)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_5, ID(3, 69)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_6, ID(3, 70)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_collision_7, ID(3, 71)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_full, ID(3, 72)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pm_wr_trans_filt, ID(3, 73)),
>
> /* counter4 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_0, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_1, 65),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_2, 66),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_3, 67),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_4, 68),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_5, 69),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_6, 70),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_7, 71),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq2_rmw, 72),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pm_rd_beat_filt, 73),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_0, ID(4, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_1, ID(4, 65)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_2, ID(4, 66)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_3, ID(4, 67)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_4, ID(4, 68)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_5, ID(4, 69)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_6, ID(4, 70)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_row_open_7, ID(4, 71)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq2_rmw, ID(4, 72)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pm_rd_beat_filt, ID(4, 73)),
>
> /* counter5 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_0, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_1, 65),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_2, 66),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_3, 67),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_4, 68),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_5, 69),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_6, 70),
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_7, 71),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq1, 72),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_0, ID(5, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_1, ID(5, 65)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_2, ID(5, 66)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_3, ID(5, 67)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_4, ID(5, 68)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_5, ID(5, 69)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_6, ID(5, 70)),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_start_7, ID(5, 71)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq1, ID(5, 72)),
>
> /* counter6 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_end_0, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq2, 72),
> + IMX9_DDR_PMU_EVENT_ATTR(ddrc_qx_valid_end_0, ID(6, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq2, ID(6, 72)),
>
> /* counter7 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_1_2_full, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_wrq0, 65),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_1_2_full, ID(7, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_wrq0, ID(7, 65)),
>
> /* counter8 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_bias_switched, 64),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_1_4_full, 65),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_bias_switched, ID(8, 64)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_1_4_full, ID(8, 65)),
>
> /* counter9 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_wrq1, 65),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_3_4_full, 66),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_wrq1, ID(9, 65)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_3_4_full, ID(9, 66)),
>
> /* counter10 specific events */
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_misc_mrk, 65),
> - IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq0, 66),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_misc_mrk, ID(10, 65)),
> + IMX9_DDR_PMU_EVENT_ATTR(eddrtq_pmon_ld_rdq0, ID(10, 66)),
> NULL,
> };
>
> @@ -245,14 +249,12 @@ static const struct attribute_group ddr_perf_events_attr_group = {
> .attrs = ddr_perf_events_attrs,
> };
>
> -PMU_FORMAT_ATTR(event, "config:0-7");
> -PMU_FORMAT_ATTR(counter, "config:8-15");
> +PMU_FORMAT_ATTR(event, "config:0-15");
> PMU_FORMAT_ATTR(axi_id, "config1:0-17");
> PMU_FORMAT_ATTR(axi_mask, "config2:0-17");
>
> static struct attribute *ddr_perf_format_attrs[] = {
> &format_attr_event.attr,
> - &format_attr_counter.attr,
> &format_attr_axi_id.attr,
> &format_attr_axi_mask.attr,
> NULL,
> @@ -366,13 +368,10 @@ static void ddr_perf_counter_local_config(struct ddr_pmu *pmu, int config,
> }
> }
>
> -static void ddr_perf_monitor_config(struct ddr_pmu *pmu, int cfg, int cfg1, int cfg2)
> +static void ddr_perf_monitor_config(struct ddr_pmu *pmu, int event,
> + int counter, int axi_id, int axi_mask)
> {
> u32 pmcfg1, pmcfg2;
> - int event, counter;
> -
> - event = FIELD_GET(CONFIG_EVENT, cfg);
> - counter = FIELD_GET(CONFIG_COUNTER, cfg);
>
> pmcfg1 = readl_relaxed(pmu->base + PMCFG1);
>
> @@ -392,12 +391,12 @@ static void ddr_perf_monitor_config(struct ddr_pmu *pmu, int cfg, int cfg1, int
> pmcfg1 &= ~PMCFG1_RD_BT_FILT_EN;
>
> pmcfg1 &= ~FIELD_PREP(PMCFG1_ID_MASK, 0x3FFFF);
> - pmcfg1 |= FIELD_PREP(PMCFG1_ID_MASK, cfg2);
> + pmcfg1 |= FIELD_PREP(PMCFG1_ID_MASK, axi_mask);
> writel(pmcfg1, pmu->base + PMCFG1);
>
> pmcfg2 = readl_relaxed(pmu->base + PMCFG2);
> pmcfg2 &= ~FIELD_PREP(PMCFG2_ID, 0x3FFFF);
> - pmcfg2 |= FIELD_PREP(PMCFG2_ID, cfg1);
> + pmcfg2 |= FIELD_PREP(PMCFG2_ID, axi_id);
> writel(pmcfg2, pmu->base + PMCFG2);
> }
>
> @@ -465,6 +464,28 @@ static void ddr_perf_event_start(struct perf_event *event, int flags)
> hwc->state = 0;
> }
>
> +static int ddr_perf_alloc_counter(struct ddr_pmu *pmu, int event, int counter)
> +{
> + int i;
> +
> + if (event == CYCLES_EVENT_ID) {
> + // Cycles counter is dedicated for cycle event.
> + if (pmu->events[CYCLES_COUNTER] == NULL)
> + return CYCLES_COUNTER;
> + } else if (counter != 0) {
> + // Counter specific event use specific counter.
> + if (pmu->events[counter] == NULL)
> + return counter;
> + } else {
> + // Auto allocate counter for referene event.
> + for (i = 1; i < NUM_COUNTERS; i++)
> + if (pmu->events[i] == NULL)
> + return i;
> + }
> +
> + return -ENOENT;
> +}
> +
> static int ddr_perf_event_add(struct perf_event *event, int flags)
> {
> struct ddr_pmu *pmu = to_ddr_pmu(event->pmu);
> @@ -472,10 +493,17 @@ static int ddr_perf_event_add(struct perf_event *event, int flags)
> int cfg = event->attr.config;
> int cfg1 = event->attr.config1;
> int cfg2 = event->attr.config2;
> - int counter;
> + int event_id, counter;
>
> + event_id = FIELD_GET(CONFIG_EVENT, cfg);
> counter = FIELD_GET(CONFIG_COUNTER, cfg);
>
> + counter = ddr_perf_alloc_counter(pmu, event_id, counter);
> + if (counter < 0) {
> + dev_dbg(pmu->dev, "There are not enough counters\n");
> + return -EOPNOTSUPP;
> + }
> +
> pmu->events[counter] = event;
> pmu->active_events++;
> hwc->idx = counter;
> @@ -485,7 +513,7 @@ static int ddr_perf_event_add(struct perf_event *event, int flags)
> ddr_perf_event_start(event, flags);
>
> /* read trans, write trans, read beat */
> - ddr_perf_monitor_config(pmu, cfg, cfg1, cfg2);
> + ddr_perf_monitor_config(pmu, event_id, counter, cfg1, cfg2);
>
> return 0;
> }
> @@ -506,9 +534,11 @@ static void ddr_perf_event_del(struct perf_event *event, int flags)
> {
> struct ddr_pmu *pmu = to_ddr_pmu(event->pmu);
> struct hw_perf_event *hwc = &event->hw;
> + int counter = hwc->idx;
>
> ddr_perf_event_stop(event, PERF_EF_UPDATE);
>
> + pmu->events[counter] = NULL;
> pmu->active_events--;
> hwc->idx = -1;
> }
> --
> 2.34.1
>