Re: [PATCH v9 46/48] perf jevents: Add collection of topdown like metrics for arm64

From: James Clark

Date: Mon Dec 15 2025 - 08:24:41 EST




On 09/12/2025 23:23, Ian Rogers wrote:
On Tue, Dec 9, 2025 at 3:31 AM James Clark <james.clark@xxxxxxxxxx> wrote:

On 02/12/2025 5:50 pm, Ian Rogers wrote:
Metrics are created using legacy, common and recommended events. As
events may be missing a TryEvent function will give None if an event
is missing. To workaround missing JSON events for cortex-a53, sysfs
encodings are used.

Signed-off-by: Ian Rogers <irogers@xxxxxxxxxx>
---
An earlier review of this patch by Leo Yan is here:
https://lore.kernel.org/lkml/8168c713-005c-4fd9-a928-66763dab746a@xxxxxxx/
Hopefully all corrections were made.
---
tools/perf/pmu-events/arm64_metrics.py | 145 ++++++++++++++++++++++++-
1 file changed, 142 insertions(+), 3 deletions(-)

[...]
+ MetricGroup("lpm_topdown_be_bound", [
+ MetricGroup("lpm_topdown_be_dtlb", [
+ Metric("lpm_topdown_be_dtlb_walks", "Dtlb walks per instruction",
+ d_ratio(dtlb_walk, ins_ret), "walk/insn"),
+ Metric("lpm_topdown_be_dtlb_walk_rate", "Dtlb walks per L1D TLB access",
+ d_ratio(dtlb_walk, l1d_tlb) if l1d_tlb else None, "100%"),
+ ]) if dtlb_walk else None,
+ MetricGroup("lpm_topdown_be_mix", [
+ Metric("lpm_topdown_be_mix_ld", "Percentage of load instructions",
+ d_ratio(ld_spec, inst_spec), "100%") if ld_spec else None,
+ Metric("lpm_topdown_be_mix_st", "Percentage of store instructions",
+ d_ratio(st_spec, inst_spec), "100%") if st_spec else None,
+ Metric("lpm_topdown_be_mix_simd", "Percentage of SIMD instructions",
+ d_ratio(ase_spec, inst_spec), "100%") if ase_spec else None,
+ Metric("lpm_topdown_be_mix_fp",
+ "Percentage of floating point instructions",
+ d_ratio(vfp_spec, inst_spec), "100%") if vfp_spec else None,
+ Metric("lpm_topdown_be_mix_dp",
+ "Percentage of data processing instructions",
+ d_ratio(dp_spec, inst_spec), "100%") if dp_spec else None,
+ Metric("lpm_topdown_be_mix_crypto",
+ "Percentage of data processing instructions",
+ d_ratio(crypto_spec, inst_spec), "100%") if crypto_spec else None,
+ Metric(
+ "lpm_topdown_be_mix_br", "Percentage of branch instructions",
+ d_ratio(br_immed_spec + br_indirect_spec + br_ret_spec,
+ inst_spec), "100%") if br_immed_spec and br_indirect_spec and br_ret_spec else None,

Hi Ian,

I've been trying to engage with the team that's publishing the metrics
in Arm [1] to see if there was any chance in getting some unity between
these new metrics and their existing json ones. The feedback from them
was that the decision to only publish metrics for certain cores is
deliberate and there is no plan to change anything. The metrics there
are well tested, known to be working, and usually contain workarounds
for specific issues. They don't want to do "Arm wide" common metrics for
existing cores as they believe it has more potential to mislead people
than help.

So this is sad, but I'll drop the patch from the series so as not to
delay things and keep carrying it in Google's tree. Just looking in
tools/perf/pmu-events/arch/arm64/arm there are 20 ARM models of which
only neoverse models (5 of the 20) have metrics. Could ARM's metric
people step up to fill the void? Models like cortex-a76 are actively
sold in the Raspberry Pi 5 and yet lack metrics.

I don't know of any plan to do so, although it's obviously possible. I think if some kind of official request came through some customer (including Google) then it could happen. But it looks like nobody is really asking for it so it wasn't done and that's how the decision was made.


I think there has to be a rule at some point of, "don't let perfect be
the enemy of good." There's no implication that ARM should maintain
these metrics and they be perfect just as there isn't an implication
that ARM should maintain the legacy metrics like
"stalled_cycles_per_instruction":
https://web.git.kernel.org/pub/scm/linux/kernel/git/perf/perf-tools-next.git/tree/tools/perf/pmu-events/arch/common/common/metrics.json?h=perf-tools-next#n49


I kind of agree, I'm not saying to definitely not add these metrics. Just making sure that you're aware of the potential for some issues and conflicts going forwards.

As you say, the bugs can always be fixed. It's just that now we have to do it in two places.

I'm guessing the cycles breakdown:
https://lore.kernel.org/lkml/20251202175043.623597-48-irogers@xxxxxxxxxx/
is okay and will keep that for ARM.


Yes this one looks fine.

I'm commenting on this "lpm_topdown_be_mix_br" as one example, that the
equivalent Arm metric "branch_percentage" excludes br_ret_spec because
br_indirect_spec also counts returns. Or on neoverse-n3 it's
"PC_WRITE_SPEC / INST_SPEC".

This is the value in upstreaming metrics like this, to bug fix. This
is what has happened with the AMD and Intel metrics. I'm happy we can
deliver more metrics to users on those CPUs.

I see that you've prefixed all the metrics so the names won't clash from
Kan's feedback [2]. But it makes me wonder if at some point some kind of
alias list could be implemented to override the generated metrics with
hand written json ones. But by that point why not just use the same
names? The Arm metric team's feedback was that there isn't really an
industry standard for naming, and that differences between architectures
would make it almost impossible to standardise anyway in their opinion.

So naming is always a challenge. One solution here is the ilist
application. When doing the legacy event reorganization I remember you
arguing that legacy events should be the norm and not the exception,
but it is because of all the model quirks precisely why that doesn't
work. By working through the quirks with cross platform metrics,
that's value to users and not misleading. If it does mislead then
that's a bug, let's fix it. Presenting users with no data isn't a fix
nor being particularly helpful.

But here we're adding duplicate metrics with different names, where the
new ones are known to have issues. It's not a great user experience IMO,
but at the same time missing old cores from the Arm metrics isn't a
great user experience either. I actually don't have a solution, other
than to say I tried to get them to consider more unified naming.

So the lpm_ metrics are on top of whatever a vendor wants to add.
There are often more than one way to compute a metric, such as memory
controller counters vs l3 cache, on Intel an lpm_ metric may use
uncore counters while a tma_ metric uses the cache. I don't know if
sticking "ARM doesn't support this" in all the ARM lpm_ metric
descriptions would mitigate your metric creators' concerns, it is

Probably not necessary to do that no, but maybe some kind of header or documentation about what "lpm_" stands for would be good. I don't know if we'd expect users to know about the subtle differences or what lpm stands for?

something implied by Linux's licensing. We do highlight metrics that
contain experimental events, such as on Intel, should be considered
similarly experimental.

I also have to say that I do still agree with Andi's old feedback [3]
that the existing json was good enough, and maybe this isn't the right
direction, although it's not very useful feedback at this point. I
thought I had replied to that thread long ago, but must not have pressed
send, sorry about that.

So having handwritten long metrics in json it's horrid, Having been
there I wouldn't want to be doing more of it. No comments, no line
breaks, huge potential for typos, peculiar rules on when commas are
allowed (so removing a line breaks parsing), .. This is why we have
make_legacy_cache.py
https://web.git.kernel.org/pub/scm/linux/kernel/git/perf/perf-tools-next.git/tree/tools/perf/pmu-events/make_legacy_cache.py?h=perf-tools-next
writing 1216 legacy cache event descriptions (7266 lines of json) vs
129 lines of python. I'm going to be on team python all day long. In
terms of the Linux build, I don't think there's a reasonable
alternative language.

Thanks,
Ian


Comments and line breaks could have been worked around by bodging the json parser or moving to yaml or something. But yes having a big formula in a one line string with conditionals wasn't that great. The biggest problem I think this fixes is the fact that it fills in missing metrics.

It does however make it harder to autogenerate using any formulas published elsewhere. Whether that's actually an issue or not I don't know because these are going to be changing much less frequently than something like the event names that are published in json (and there's fewer metrics than events as well).


[1]:
https://gitlab.arm.com/telemetry-solution/telemetry-solution/-/tree/main/data
[2]:
https://lore.kernel.org/lkml/43548903-b7c8-47c4-b1da-0258293ecbd4@xxxxxxxxxxxxxxx
[3]: https://lore.kernel.org/lkml/ZeJJyCmXO9GxpDiF@tassilo/