Hi Arnaldo,
On Sun, Jul 24, 2022, 6:34 PM Xing Zhengjun <zhengjun.xing@xxxxxxxxxxxxxxx <mailto:zhengjun.xing@xxxxxxxxxxxxxxx>> wrote:
Hi Ian,
On 7/23/2022 6:32 AM, Ian Rogers wrote:
> Use script at:
>
https://github.com/intel/event-converter-for-linux-perf/blob/master/download_and_gen.py
<https://github.com/intel/event-converter-for-linux-perf/blob/master/download_and_gen.py>
>
It is better to add the event JSON file version and TMA version for the
future track. For example, the event list is based on broadwellx JSON
file v19, the metrics are based on TMA4.4-full.
Thanks Xing,
I'll add that in v2. I'd skipped it this time as I'd been adding it to the mapfile. I'll also break apart the tremontx to snowridgex change to match yours. I also will rebase to see if that'll fix the git am issue. Apologies in advance to everyone's inboxes :-)
Thanks,
Ian
> to download and generate the latest events and metrics. Manually copy
> the broadwellx files into perf and update mapfile.csv.
>
> Tested with 'perf test':
> 10: PMU events :
> 10.1: PMU event table sanity : Ok
> 10.2: PMU event map aliases : Ok
> 10.3: Parsing of PMU event table metrics : Ok
> 10.4: Parsing of PMU event table metrics with fake PMUs : Ok
> 90: perf all metricgroups test : Ok
> 91: perf all metrics test : Skip
> 93: perf all PMU test : Ok
>
> Signed-off-by: Ian Rogers <irogers@xxxxxxxxxx
<mailto:irogers@xxxxxxxxxx>>
> ---
> .../arch/x86/broadwellx/bdx-metrics.json | 570 ++-
> .../pmu-events/arch/x86/broadwellx/cache.json | 22 +-
> .../arch/x86/broadwellx/floating-point.json | 9 +-
> .../arch/x86/broadwellx/frontend.json | 2 +-
> .../arch/x86/broadwellx/memory.json | 39 +-
> .../pmu-events/arch/x86/broadwellx/other.json | 2 +-
> .../arch/x86/broadwellx/pipeline.json | 4 +-
> .../arch/x86/broadwellx/uncore-cache.json | 3788
++++++++++++++++-
> .../x86/broadwellx/uncore-interconnect.json | 1438 ++++++-
> .../arch/x86/broadwellx/uncore-memory.json | 2849 ++++++++++++-
> .../arch/x86/broadwellx/uncore-other.json | 3252 ++++++++++++++
> .../arch/x86/broadwellx/uncore-power.json | 437 +-
> .../arch/x86/broadwellx/virtual-memory.json | 2 +-
> tools/perf/pmu-events/arch/x86/mapfile.csv | 2 +-
> 14 files changed, 12103 insertions(+), 313 deletions(-)
> create mode 100644
tools/perf/pmu-events/arch/x86/broadwellx/uncore-other.json
>
> diff --git
a/tools/perf/pmu-events/arch/x86/broadwellx/bdx-metrics.json
b/tools/perf/pmu-events/arch/x86/broadwellx/bdx-metrics.json
> index b055947c0afe..720ee7c9332d 100644
> --- a/tools/perf/pmu-events/arch/x86/broadwellx/bdx-metrics.json
> +++ b/tools/perf/pmu-events/arch/x86/broadwellx/bdx-metrics.json
> @@ -74,12 +74,6 @@
> "MetricGroup": "Branches;Fed;FetchBW",
> "MetricName": "UpTB"
> },
> - {
> - "BriefDescription": "Cycles Per Instruction (per Logical
Processor)",
> - "MetricExpr": "1 / (INST_RETIRED.ANY /
CPU_CLK_UNHALTED.THREAD)",
> - "MetricGroup": "Pipeline;Mem",
> - "MetricName": "CPI"
> - },
> {
> "BriefDescription": "Per-Logical Processor actual
clocks when the Logical Processor is active.",
> "MetricExpr": "CPU_CLK_UNHALTED.THREAD",
> @@ -130,43 +124,25 @@
> "MetricName": "FLOPc_SMT"
> },
> {
> - "BriefDescription": "Actual per-core usage of the
Floating Point execution units (regardless of the vector width)",
> + "BriefDescription": "Actual per-core usage of the
Floating Point non-X87 execution units (regardless of precision or
vector-width)",
> "MetricExpr": "( (FP_ARITH_INST_RETIRED.SCALAR_SINGLE +
FP_ARITH_INST_RETIRED.SCALAR_DOUBLE) +
(FP_ARITH_INST_RETIRED.128B_PACKED_DOUBLE +
FP_ARITH_INST_RETIRED.128B_PACKED_SINGLE +
FP_ARITH_INST_RETIRED.256B_PACKED_DOUBLE +
FP_ARITH_INST_RETIRED.256B_PACKED_SINGLE) ) / ( 2 *
CPU_CLK_UNHALTED.THREAD )",
> "MetricGroup": "Cor;Flops;HPC",
> "MetricName": "FP_Arith_Utilization",
> - "PublicDescription": "Actual per-core usage of the
Floating Point execution units (regardless of the vector width).
Values > 1 are possible due to Fused-Multiply Add (FMA) counting."
> + "PublicDescription": "Actual per-core usage of the
Floating Point non-X87 execution units (regardless of precision or
vector-width). Values > 1 are possible due to ([BDW+] Fused-Multiply
Add (FMA) counting - common; [ADL+] use all of ADD/MUL/FMA in Scalar
or 128/256-bit vectors - less common)."
> },
> {
> - "BriefDescription": "Actual per-core usage of the
Floating Point execution units (regardless of the vector width). SMT
version; use when SMT is enabled and measuring per logical CPU.",
> + "BriefDescription": "Actual per-core usage of the
Floating Point non-X87 execution units (regardless of precision or
vector-width). SMT version; use when SMT is enabled and measuring
per logical CPU.",
> "MetricExpr": "( (FP_ARITH_INST_RETIRED.SCALAR_SINGLE +
FP_ARITH_INST_RETIRED.SCALAR_DOUBLE) +
(FP_ARITH_INST_RETIRED.128B_PACKED_DOUBLE +
FP_ARITH_INST_RETIRED.128B_PACKED_SINGLE +
FP_ARITH_INST_RETIRED.256B_PACKED_DOUBLE +
FP_ARITH_INST_RETIRED.256B_PACKED_SINGLE) ) / ( 2 * ( (
CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) ) )",
> "MetricGroup": "Cor;Flops;HPC_SMT",
> "MetricName": "FP_Arith_Utilization_SMT",
> - "PublicDescription": "Actual per-core usage of the
Floating Point execution units (regardless of the vector width).
Values > 1 are possible due to Fused-Multiply Add (FMA) counting.
SMT version; use when SMT is enabled and measuring per logical CPU."
> + "PublicDescription": "Actual per-core usage of the
Floating Point non-X87 execution units (regardless of precision or
vector-width). Values > 1 are possible due to ([BDW+] Fused-Multiply
Add (FMA) counting - common; [ADL+] use all of ADD/MUL/FMA in Scalar
or 128/256-bit vectors - less common). SMT version; use when SMT is
enabled and measuring per logical CPU."
> },
> {
> - "BriefDescription": "Instruction-Level-Parallelism
(average number of uops executed when there is at least 1 uop
executed)",
> + "BriefDescription": "Instruction-Level-Parallelism
(average number of uops executed when there is execution) per-core",
> "MetricExpr": "UOPS_EXECUTED.THREAD / ((
cpu@UOPS_EXECUTED.CORE\\,cmask\\=1@ / 2 ) if #SMT_on else
UOPS_EXECUTED.CYCLES_GE_1_UOP_EXEC)",
> "MetricGroup": "Backend;Cor;Pipeline;PortsUtil",
> "MetricName": "ILP"
> },
> - {
> - "BriefDescription": "Branch Misprediction Cost: Fraction
of TMA slots wasted per non-speculative branch misprediction
(retired JEClear)",
> - "MetricExpr": " ( ((BR_MISP_RETIRED.ALL_BRANCHES / (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT )) * ((
UOPS_ISSUED.ANY - UOPS_RETIRED.RETIRE_SLOTS + 4 *
INT_MISC.RECOVERY_CYCLES ) / (4 * CPU_CLK_UNHALTED.THREAD))) + (4 *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 *
CPU_CLK_UNHALTED.THREAD)) * (BR_MISP_RETIRED.ALL_BRANCHES * (12 * (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT + BACLEARS.ANY )
/ CPU_CLK_UNHALTED.THREAD) / ( BR_MISP_RETIRED.ALL_BRANCHES +
MACHINE_CLEARS.COUNT + BACLEARS.ANY )) / #(4 *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 *
CPU_CLK_UNHALTED.THREAD)) ) * (4 * CPU_CLK_UNHALTED.THREAD) /
BR_MISP_RETIRED.ALL_BRANCHES",
> - "MetricGroup": "Bad;BrMispredicts",
> - "MetricName": "Branch_Misprediction_Cost"
> - },
> - {
> - "BriefDescription": "Branch Misprediction Cost: Fraction
of TMA slots wasted per non-speculative branch misprediction
(retired JEClear)",
> - "MetricExpr": " ( ((BR_MISP_RETIRED.ALL_BRANCHES / (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT )) * ((
UOPS_ISSUED.ANY - UOPS_RETIRED.RETIRE_SLOTS + 4 * (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) ) / (4 * ( (
CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK )
)))) + (4 * IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 * (
( CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) )))
* (BR_MISP_RETIRED.ALL_BRANCHES * (12 * (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT + BACLEARS.ANY )
/ CPU_CLK_UNHALTED.THREAD) / ( BR_MISP_RETIRED.ALL_BRANCHES +
MACHINE_CLEARS.COUNT + BACLEARS.ANY )) / #(4 *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 * ( (
CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) )))
) * (4 * ( ( CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) ))
/ BR_MISP_RETIRED.ALL_BRANCHES",
> - "MetricGroup": "Bad;BrMispredicts_SMT",
> - "MetricName": "Branch_Misprediction_Cost_SMT"
> - },
> - {
> - "BriefDescription": "Number of Instructions per
non-speculative Branch Misprediction (JEClear)",
> - "MetricExpr": "INST_RETIRED.ANY /
BR_MISP_RETIRED.ALL_BRANCHES",
> - "MetricGroup": "Bad;BadSpec;BrMispredicts",
> - "MetricName": "IpMispredict"
> - },
> {
> "BriefDescription": "Core actual clocks when any
Logical Processor is active on the Physical Core",
> "MetricExpr": "( ( CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1
+ CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) )",
> @@ -256,6 +232,18 @@
> "MetricGroup": "Summary;TmaL1",
> "MetricName": "Instructions"
> },
> + {
> + "BriefDescription": "Average number of Uops retired in
cycles where at least one uop has retired.",
> + "MetricExpr": "UOPS_RETIRED.RETIRE_SLOTS /
cpu@UOPS_RETIRED.RETIRE_SLOTS\\,cmask\\=1@",
> + "MetricGroup": "Pipeline;Ret",
> + "MetricName": "Retire"
> + },
> + {
> + "BriefDescription": "",
> + "MetricExpr": "UOPS_EXECUTED.THREAD /
cpu@UOPS_EXECUTED.THREAD\\,cmask\\=1@",
> + "MetricGroup": "Cor;Pipeline;PortsUtil;SMT",
> + "MetricName": "Execute"
> + },
> {
> "BriefDescription": "Fraction of Uops delivered by the
DSB (aka Decoded ICache; or Uop Cache)",
> "MetricExpr": "IDQ.DSB_UOPS / (( IDQ.DSB_UOPS +
LSD.UOPS + IDQ.MITE_UOPS + IDQ.MS_UOPS ) )",
> @@ -263,35 +251,34 @@
> "MetricName": "DSB_Coverage"
> },
> {
> - "BriefDescription": "Actual Average Latency for L1
data-cache miss demand load instructions (in core cycles)",
> - "MetricExpr": "L1D_PEND_MISS.PENDING / (
MEM_LOAD_UOPS_RETIRED.L1_MISS + mem_load_uops_retired.hit_lfb )",
> - "MetricGroup": "Mem;MemoryBound;MemoryLat",
> - "MetricName": "Load_Miss_Real_Latency",
> - "PublicDescription": "Actual Average Latency for L1
data-cache miss demand load instructions (in core cycles). Latency
may be overestimated for multi-load instructions - e.g. repeat strings."
> + "BriefDescription": "Number of Instructions per
non-speculative Branch Misprediction (JEClear) (lower number means
higher occurrence rate)",
> + "MetricExpr": "INST_RETIRED.ANY /
BR_MISP_RETIRED.ALL_BRANCHES",
> + "MetricGroup": "Bad;BadSpec;BrMispredicts",
> + "MetricName": "IpMispredict"
> },
> {
> - "BriefDescription": "Memory-Level-Parallelism (average
number of L1 miss demand load when there is at least one such miss.
Per-Logical Processor)",
> - "MetricExpr": "L1D_PEND_MISS.PENDING /
L1D_PEND_MISS.PENDING_CYCLES",
> - "MetricGroup": "Mem;MemoryBound;MemoryBW",
> - "MetricName": "MLP"
> + "BriefDescription": "Branch Misprediction Cost: Fraction
of TMA slots wasted per non-speculative branch misprediction
(retired JEClear)",
> + "MetricExpr": " ( ((BR_MISP_RETIRED.ALL_BRANCHES / (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT )) * ((
UOPS_ISSUED.ANY - UOPS_RETIRED.RETIRE_SLOTS + 4 *
INT_MISC.RECOVERY_CYCLES ) / (4 * CPU_CLK_UNHALTED.THREAD))) + (4 *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 *
CPU_CLK_UNHALTED.THREAD)) * (BR_MISP_RETIRED.ALL_BRANCHES * (12 * (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT + BACLEARS.ANY )
/ CPU_CLK_UNHALTED.THREAD) / ( BR_MISP_RETIRED.ALL_BRANCHES +
MACHINE_CLEARS.COUNT + BACLEARS.ANY )) / #(4 *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 *
CPU_CLK_UNHALTED.THREAD)) ) * (4 * CPU_CLK_UNHALTED.THREAD) /
BR_MISP_RETIRED.ALL_BRANCHES",
> + "MetricGroup": "Bad;BrMispredicts",
> + "MetricName": "Branch_Misprediction_Cost"
> },
> {
> - "BriefDescription": "Average data fill bandwidth to the
L1 data cache [GB / sec]",
> - "MetricExpr": "64 * L1D.REPLACEMENT / 1000000000 /
duration_time",
> - "MetricGroup": "Mem;MemoryBW",
> - "MetricName": "L1D_Cache_Fill_BW"
> + "BriefDescription": "Branch Misprediction Cost: Fraction
of TMA slots wasted per non-speculative branch misprediction
(retired JEClear)",
> + "MetricExpr": " ( ((BR_MISP_RETIRED.ALL_BRANCHES / (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT )) * ((
UOPS_ISSUED.ANY - UOPS_RETIRED.RETIRE_SLOTS + 4 * (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) ) / (4 * ( (
CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK )
)))) + (4 * IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 * (
( CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) )))
* (BR_MISP_RETIRED.ALL_BRANCHES * (12 * (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT + BACLEARS.ANY )
/ CPU_CLK_UNHALTED.THREAD) / ( BR_MISP_RETIRED.ALL_BRANCHES +
MACHINE_CLEARS.COUNT + BACLEARS.ANY )) / #(4 *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / (4 * ( (
CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) )))
) * (4 * ( ( CPU_CLK_UNHALTED.THREAD / 2 ) * ( 1 +
CPU_CLK_UNHALTED.ONE_THREAD_ACTIVE / CPU_CLK_UNHALTED.REF_XCLK ) ))
/ BR_MISP_RETIRED.ALL_BRANCHES",
> + "MetricGroup": "Bad;BrMispredicts_SMT",
> + "MetricName": "Branch_Misprediction_Cost_SMT"
> },
> {
> - "BriefDescription": "Average data fill bandwidth to the
L2 cache [GB / sec]",
> - "MetricExpr": "64 * L2_LINES_IN.ALL / 1000000000 /
duration_time",
> - "MetricGroup": "Mem;MemoryBW",
> - "MetricName": "L2_Cache_Fill_BW"
> + "BriefDescription": "Actual Average Latency for L1
data-cache miss demand load operations (in core cycles)",
> + "MetricExpr": "L1D_PEND_MISS.PENDING / (
MEM_LOAD_UOPS_RETIRED.L1_MISS + mem_load_uops_retired.hit_lfb )",
> + "MetricGroup": "Mem;MemoryBound;MemoryLat",
> + "MetricName": "Load_Miss_Real_Latency"
> },
> {
> - "BriefDescription": "Average per-core data fill
bandwidth to the L3 cache [GB / sec]",
> - "MetricExpr": "64 * LONGEST_LAT_CACHE.MISS / 1000000000
/ duration_time",
> - "MetricGroup": "Mem;MemoryBW",
> - "MetricName": "L3_Cache_Fill_BW"
> + "BriefDescription": "Memory-Level-Parallelism (average
number of L1 miss demand load when there is at least one such miss.
Per-Logical Processor)",
> + "MetricExpr": "L1D_PEND_MISS.PENDING /
L1D_PEND_MISS.PENDING_CYCLES",
> + "MetricGroup": "Mem;MemoryBound;MemoryBW",
> + "MetricName": "MLP"
> },
> {
> "BriefDescription": "L1 cache true misses per kilo
instruction for retired demand loads",
> @@ -306,13 +293,13 @@
> "MetricName": "L2MPKI"
> },
> {
> - "BriefDescription": "L2 cache misses per kilo
instruction for all request types (including speculative)",
> + "BriefDescription": "L2 cache ([RKL+] true) misses per
kilo instruction for all request types (including speculative)",
> "MetricExpr": "1000 * L2_RQSTS.MISS / INST_RETIRED.ANY",
> "MetricGroup": "Mem;CacheMisses;Offcore",
> "MetricName": "L2MPKI_All"
> },
> {
> - "BriefDescription": "L2 cache misses per kilo
instruction for all demand loads (including speculative)",
> + "BriefDescription": "L2 cache ([RKL+] true) misses per
kilo instruction for all demand loads (including speculative)",
> "MetricExpr": "1000 * L2_RQSTS.DEMAND_DATA_RD_MISS /
INST_RETIRED.ANY",
> "MetricGroup": "Mem;CacheMisses",
> "MetricName": "L2MPKI_Load"
> @@ -348,6 +335,48 @@
> "MetricGroup": "Mem;MemoryTLB_SMT",
> "MetricName": "Page_Walks_Utilization_SMT"
> },
> + {
> + "BriefDescription": "Average per-core data fill
bandwidth to the L1 data cache [GB / sec]",
> + "MetricExpr": "64 * L1D.REPLACEMENT / 1000000000 /
duration_time",
> + "MetricGroup": "Mem;MemoryBW",
> + "MetricName": "L1D_Cache_Fill_BW"
> + },
> + {
> + "BriefDescription": "Average per-core data fill
bandwidth to the L2 cache [GB / sec]",
> + "MetricExpr": "64 * L2_LINES_IN.ALL / 1000000000 /
duration_time",
> + "MetricGroup": "Mem;MemoryBW",
> + "MetricName": "L2_Cache_Fill_BW"
> + },
> + {
> + "BriefDescription": "Average per-core data fill
bandwidth to the L3 cache [GB / sec]",
> + "MetricExpr": "64 * LONGEST_LAT_CACHE.MISS / 1000000000
/ duration_time",
> + "MetricGroup": "Mem;MemoryBW",
> + "MetricName": "L3_Cache_Fill_BW"
> + },
> + {
> + "BriefDescription": "Average per-thread data fill
bandwidth to the L1 data cache [GB / sec]",
> + "MetricExpr": "(64 * L1D.REPLACEMENT / 1000000000 /
duration_time)",
> + "MetricGroup": "Mem;MemoryBW",
> + "MetricName": "L1D_Cache_Fill_BW_1T"
> + },
> + {
> + "BriefDescription": "Average per-thread data fill
bandwidth to the L2 cache [GB / sec]",
> + "MetricExpr": "(64 * L2_LINES_IN.ALL / 1000000000 /
duration_time)",
> + "MetricGroup": "Mem;MemoryBW",
> + "MetricName": "L2_Cache_Fill_BW_1T"
> + },
> + {
> + "BriefDescription": "Average per-thread data fill
bandwidth to the L3 cache [GB / sec]",
> + "MetricExpr": "(64 * LONGEST_LAT_CACHE.MISS / 1000000000
/ duration_time)",
> + "MetricGroup": "Mem;MemoryBW",
> + "MetricName": "L3_Cache_Fill_BW_1T"
> + },
> + {
> + "BriefDescription": "Average per-thread data access
bandwidth to the L3 cache [GB / sec]",
> + "MetricExpr": "0",
> + "MetricGroup": "Mem;MemoryBW;Offcore",
> + "MetricName": "L3_Cache_Access_BW_1T"
> + },
> {
> "BriefDescription": "Average CPU Utilization",
> "MetricExpr": "CPU_CLK_UNHALTED.REF_TSC / msr@tsc@",
> @@ -364,7 +393,8 @@
> "BriefDescription": "Giga Floating Point Operations Per
Second",
> "MetricExpr": "( ( 1 * (
FP_ARITH_INST_RETIRED.SCALAR_SINGLE +
FP_ARITH_INST_RETIRED.SCALAR_DOUBLE ) + 2 *
FP_ARITH_INST_RETIRED.128B_PACKED_DOUBLE + 4 * (
FP_ARITH_INST_RETIRED.128B_PACKED_SINGLE +
FP_ARITH_INST_RETIRED.256B_PACKED_DOUBLE ) + 8 *
FP_ARITH_INST_RETIRED.256B_PACKED_SINGLE ) / 1000000000 ) /
duration_time",
> "MetricGroup": "Cor;Flops;HPC",
> - "MetricName": "GFLOPs"
> + "MetricName": "GFLOPs",
> + "PublicDescription": "Giga Floating Point Operations Per
Second. Aggregate across all supported options of: FP precisions,
scalar and vector instructions, vector-width and AMX engine."
> },
> {
> "BriefDescription": "Average Frequency Utilization
relative nominal frequency",
> @@ -461,5 +491,439 @@
> "MetricExpr": "(cstate_pkg@c7\\-residency@ / msr@tsc@)
* 100",
> "MetricGroup": "Power",
> "MetricName": "C7_Pkg_Residency"
> + },
> + {
> + "BriefDescription": "CPU operating frequency (in GHz)",
> + "MetricExpr": "( CPU_CLK_UNHALTED.THREAD /
CPU_CLK_UNHALTED.REF_TSC * #SYSTEM_TSC_FREQ ) / 1000000000",
> + "MetricGroup": "",
> + "MetricName": "cpu_operating_frequency",
> + "ScaleUnit": "1GHz"
> + },
> + {
> + "BriefDescription": "Cycles per instruction retired;
indicating how much time each executed instruction took; in units of
cycles.",
> + "MetricExpr": "CPU_CLK_UNHALTED.THREAD / INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "cpi",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "The ratio of number of completed
memory load instructions to the total number completed instructions",
> + "MetricExpr": "MEM_UOPS_RETIRED.ALL_LOADS /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "loads_per_instr",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "The ratio of number of completed
memory store instructions to the total number completed instructions",
> + "MetricExpr": "MEM_UOPS_RETIRED.ALL_STORES /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "stores_per_instr",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of requests missing
L1 data cache (includes data+rfo w/ prefetches) to the total number
of completed instructions",
> + "MetricExpr": "L1D.REPLACEMENT / INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName":
"l1d_mpi_includes_data_plus_rfo_with_prefetches",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of demand load
requests hitting in L1 data cache to the total number of completed
instructions",
> + "MetricExpr": "MEM_LOAD_UOPS_RETIRED.L1_HIT /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "l1d_demand_data_read_hits_per_instr",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of code read
requests missing in L1 instruction cache (includes prefetches) to
the total number of completed instructions",
> + "MetricExpr": "L2_RQSTS.ALL_CODE_RD / INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName":
"l1_i_code_read_misses_with_prefetches_per_instr",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of completed demand
load requests hitting in L2 cache to the total number of completed
instructions",
> + "MetricExpr": "MEM_LOAD_UOPS_RETIRED.L2_HIT /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "l2_demand_data_read_hits_per_instr",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of requests missing
L2 cache (includes code+data+rfo w/ prefetches) to the total number
of completed instructions",
> + "MetricExpr": "L2_LINES_IN.ALL / INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName":
"l2_mpi_includes_code_plus_data_plus_rfo_with_prefetches",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of completed data
read request missing L2 cache to the total number of completed
instructions",
> + "MetricExpr": "MEM_LOAD_UOPS_RETIRED.L2_MISS /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "l2_demand_data_read_mpi",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of code read
request missing L2 cache to the total number of completed instructions",
> + "MetricExpr": "L2_RQSTS.CODE_RD_MISS / INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "l2_demand_code_mpi",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of data read
requests missing last level core cache (includes demand w/
prefetches) to the total number of completed instructions",
> + "MetricExpr": "(
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ +
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x192@ ) /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "llc_data_read_mpi_demand_plus_prefetch",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of code read
requests missing last level core cache (includes demand w/
prefetches) to the total number of completed instructions",
> + "MetricExpr": "(
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x181@ +
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x191@ ) /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "llc_code_read_mpi_demand_plus_prefetch",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Average latency of a last level
cache (LLC) demand and prefetch data read miss (read memory access)
in nano seconds",
> + "MetricExpr": "( 1000000000 * (
cbox@UNC_C_TOR_OCCUPANCY.MISS_OPCODE\\,filter_opc\\=0x182@ /
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ ) / (
UNC_C_CLOCKTICKS / ( source_count(UNC_C_CLOCKTICKS) * #num_packages
) ) ) * duration_time",
> + "MetricGroup": "",
> + "MetricName":
"llc_data_read_demand_plus_prefetch_miss_latency",
> + "ScaleUnit": "1ns"
> + },
> + {
> + "BriefDescription": "Average latency of a last level
cache (LLC) demand and prefetch data read miss (read memory access)
addressed to local memory in nano seconds",
> + "MetricExpr": "( 1000000000 * (
cbox@UNC_C_TOR_OCCUPANCY.MISS_LOCAL_OPCODE\\,filter_opc\\=0x182@ /
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ ) / (
UNC_C_CLOCKTICKS / ( source_count(UNC_C_CLOCKTICKS) * #num_packages
) ) ) * duration_time",
> + "MetricGroup": "",
> + "MetricName":
"llc_data_read_demand_plus_prefetch_miss_latency_for_local_requests",
> + "ScaleUnit": "1ns"
> + },
> + {
> + "BriefDescription": "Average latency of a last level
cache (LLC) demand and prefetch data read miss (read memory access)
addressed to remote memory in nano seconds",
> + "MetricExpr": "( 1000000000 * (
cbox@UNC_C_TOR_OCCUPANCY.MISS_REMOTE_OPCODE\\,filter_opc\\=0x182@ /
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ ) / (
UNC_C_CLOCKTICKS / ( source_count(UNC_C_CLOCKTICKS) * #num_packages
) ) ) * duration_time",
> + "MetricGroup": "",
> + "MetricName":
"llc_data_read_demand_plus_prefetch_miss_latency_for_remote_requests",
> + "ScaleUnit": "1ns"
> + },
> + {
> + "BriefDescription": "Ratio of number of completed page
walks (for all page sizes) caused by a code fetch to the total
number of completed instructions. This implies it missed in the ITLB
(Instruction TLB) and further levels of TLB.",
> + "MetricExpr": "ITLB_MISSES.WALK_COMPLETED /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "itlb_mpi",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of completed page
walks (for 2 megabyte and 4 megabyte page sizes) caused by a code
fetch to the total number of completed instructions. This implies it
missed in the Instruction Translation Lookaside Buffer (ITLB) and
further levels of TLB.",
> + "MetricExpr": "ITLB_MISSES.WALK_COMPLETED_2M_4M /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "itlb_large_page_mpi",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of completed page
walks (for all page sizes) caused by demand data loads to the total
number of completed instructions. This implies it missed in the DTLB
and further levels of TLB.",
> + "MetricExpr": "DTLB_LOAD_MISSES.WALK_COMPLETED /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "dtlb_load_mpi",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Ratio of number of completed page
walks (for all page sizes) caused by demand data stores to the total
number of completed instructions. This implies it missed in the DTLB
and further levels of TLB.",
> + "MetricExpr": "DTLB_STORE_MISSES.WALK_COMPLETED /
INST_RETIRED.ANY",
> + "MetricGroup": "",
> + "MetricName": "dtlb_store_mpi",
> + "ScaleUnit": "1per_instr"
> + },
> + {
> + "BriefDescription": "Memory read that miss the last
level cache (LLC) addressed to local DRAM as a percentage of total
memory read accesses, does not include LLC prefetches.",
> + "MetricExpr": "100 *
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ / (
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ +
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ )",
> + "MetricGroup": "",
> + "MetricName": "numa_percent_reads_addressed_to_local_dram",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "Memory reads that miss the last
level cache (LLC) addressed to remote DRAM as a percentage of total
memory read accesses, does not include LLC prefetches.",
> + "MetricExpr": "100 *
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ / (
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ +
cbox@UNC_C_TOR_INSERTS.MISS_OPCODE\\,filter_opc\\=0x182@ )",
> + "MetricGroup": "",
> + "MetricName": "numa_percent_reads_addressed_to_remote_dram",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "Uncore operating frequency in GHz",
> + "MetricExpr": "UNC_C_CLOCKTICKS / (
source_count(UNC_C_CLOCKTICKS) * #num_packages ) / 1000000000",
> + "MetricGroup": "",
> + "MetricName": "uncore_frequency",
> + "ScaleUnit": "1GHz"
> + },
> + {
> + "BriefDescription": "Intel(R) Quick Path Interconnect
(QPI) data transmit bandwidth (MB/sec)",
> + "MetricExpr": "( UNC_Q_TxL_FLITS_G0.DATA * 8 / 1000000)
/ duration_time",
> + "MetricGroup": "",
> + "MetricName": "qpi_data_transmit_bw_only_data",
> + "ScaleUnit": "1MB/s"
> + },
> + {
> + "BriefDescription": "DDR memory read bandwidth (MB/sec)",
> + "MetricExpr": "( UNC_M_CAS_COUNT.RD * 64 / 1000000) /
duration_time",
> + "MetricGroup": "",
> + "MetricName": "memory_bandwidth_read",
> + "ScaleUnit": "1MB/s"
> + },
> + {
> + "BriefDescription": "DDR memory write bandwidth (MB/sec)",
> + "MetricExpr": "( UNC_M_CAS_COUNT.WR * 64 / 1000000) /
duration_time",
> + "MetricGroup": "",
> + "MetricName": "memory_bandwidth_write",
> + "ScaleUnit": "1MB/s"
> + },
> + {
> + "BriefDescription": "DDR memory bandwidth (MB/sec)",
> + "MetricExpr": "(( UNC_M_CAS_COUNT.RD +
UNC_M_CAS_COUNT.WR ) * 64 / 1000000) / duration_time",
> + "MetricGroup": "",
> + "MetricName": "memory_bandwidth_total",
> + "ScaleUnit": "1MB/s"
> + },
> + {
> + "BriefDescription": "Bandwidth of IO reads that are
initiated by end device controllers that are requesting memory from
the CPU.",
> + "MetricExpr": "(
cbox@UNC_C_TOR_INSERTS.OPCODE\\,filter_opc\\=0x19e@ * 64 / 1000000)
/ duration_time",
> + "MetricGroup": "",
> + "MetricName": "io_bandwidth_read",
> + "ScaleUnit": "1MB/s"
> + },
> + {
> + "BriefDescription": "Bandwidth of IO writes that are
initiated by end device controllers that are writing memory to the
CPU.",
> + "MetricExpr": "((
cbox@UNC_C_TOR_INSERTS.OPCODE\\,filter_opc\\=0x1c8\\,filter_tid\\=0x3e@
+
cbox@UNC_C_TOR_INSERTS.OPCODE\\,filter_opc\\=0x180\\,filter_tid\\=0x3e@
) * 64 / 1000000) / duration_time",
> + "MetricGroup": "",
> + "MetricName": "io_bandwidth_write",
> + "ScaleUnit": "1MB/s"
> + },
> + {
> + "BriefDescription": "Uops delivered from decoded
instruction cache (decoded stream buffer or DSB) as a percent of
total uops delivered to Instruction Decode Queue",
> + "MetricExpr": "100 * ( IDQ.DSB_UOPS / UOPS_ISSUED.ANY )",
> + "MetricGroup": "",
> + "MetricName":
"percent_uops_delivered_from_decoded_icache_dsb",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "Uops delivered from legacy decode
pipeline (Micro-instruction Translation Engine or MITE) as a percent
of total uops delivered to Instruction Decode Queue",
> + "MetricExpr": "100 * ( IDQ.MITE_UOPS / UOPS_ISSUED.ANY )",
> + "MetricGroup": "",
> + "MetricName":
"percent_uops_delivered_from_legacy_decode_pipeline_mite",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "Uops delivered from microcode
sequencer (MS) as a percent of total uops delivered to Instruction
Decode Queue",
> + "MetricExpr": "100 * ( IDQ.MS_UOPS / UOPS_ISSUED.ANY )",
> + "MetricGroup": "",
> + "MetricName":
"percent_uops_delivered_from_microcode_sequencer_ms",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "Uops delivered from loop stream
detector(LSD) as a percent of total uops delivered to Instruction
Decode Queue",
> + "MetricExpr": "100 * ( LSD.UOPS / UOPS_ISSUED.ANY )",
> + "MetricGroup": "",
> + "MetricName":
"percent_uops_delivered_from_loop_stream_detector_lsd",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This category represents fraction
of slots where the processor's Frontend undersupplies its Backend.
Frontend denotes the first part of the processor core responsible to
fetch operations that are executed later on by the Backend part.
Within the Frontend; a branch predictor predicts the next address to
fetch; cache-lines are fetched from the memory subsystem; parsed
into instructions; and lastly decoded into micro-operations (uops).
Ideally the Frontend can issue Machine_Width uops every cycle to the
Backend. Frontend Bound denotes unutilized issue-slots when there is
no Backend stall; i.e. bubbles where Frontend delivered no uops
while Backend could have accepted them. For example; stalls due to
instruction-cache misses would be categorized under Frontend Bound.",
> + "MetricExpr": "100 * ( IDQ_UOPS_NOT_DELIVERED.CORE / ( (
4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) )",
> + "MetricGroup": "TmaL1;PGO",
> + "MetricName": "tma_frontend_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots the CPU was stalled due to Frontend latency issues. For
example; instruction-cache misses; iTLB misses or fetch stalls after
a branch misprediction are categorized under Frontend Latency. In
such cases; the Frontend eventually delivers no uops for some period.",
> + "MetricExpr": "100 * ( ( 4 ) *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) )",
> + "MetricGroup":
"Frontend;TmaL2;m_tma_frontend_bound_percent",
> + "MetricName": "tma_fetch_latency_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
cycles the CPU was stalled due to instruction cache misses.",
> + "MetricExpr": "100 * ( ICACHE.IFDATA_STALL / (
CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup":
"BigFoot;FetchLat;IcMiss;TmaL3;m_tma_fetch_latency_percent",
> + "MetricName": "tma_icache_misses_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
cycles the CPU was stalled due to Instruction TLB (ITLB) misses.",
> + "MetricExpr": "100 * ( ( 14 * ITLB_MISSES.STLB_HIT +
cpu@ITLB_MISSES.WALK_DURATION\\,cmask\\=0x1@ + 7 *
ITLB_MISSES.WALK_COMPLETED ) / ( CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup":
"BigFoot;FetchLat;MemoryTLB;TmaL3;m_tma_fetch_latency_percent",
> + "MetricName": "tma_itlb_misses_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
cycles the CPU was stalled due to Branch Resteers. Branch Resteers
estimates the Frontend delay in fetching operations from corrected
path; following all sorts of miss-predicted branches. For example;
branchy code with lots of miss-predictions might get categorized
under Branch Resteers. Note the value of this node may overlap with
its siblings.",
> + "MetricExpr": "100 * ( ( 12 ) * (
BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT + BACLEARS.ANY )
/ ( CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup": "FetchLat;TmaL3;m_tma_fetch_latency_percent",
> + "MetricName": "tma_branch_resteers_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
cycles the CPU was stalled due to switches from DSB to MITE
pipelines. The DSB (decoded i-cache) is a Uop Cache where the
front-end directly delivers Uops (micro operations) avoiding heavy
x86 decoding. The DSB pipeline has shorter latency and delivered
higher bandwidth than the MITE (legacy instruction decode pipeline).
Switching between the two pipelines can cause penalties hence this
metric measures the exposed penalty.",
> + "MetricExpr": "100 * ( DSB2MITE_SWITCHES.PENALTY_CYCLES
/ ( CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup":
"DSBmiss;FetchLat;TmaL3;m_tma_fetch_latency_percent",
> + "MetricName": "tma_dsb_switches_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
cycles CPU was stalled due to Length Changing Prefixes (LCPs). Using
proper compiler flags or Intel Compiler by default will certainly
avoid this. #Link: Optimization Guide about LCP BKMs.",
> + "MetricExpr": "100 * ( ILD_STALL.LCP / (
CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup": "FetchLat;TmaL3;m_tma_fetch_latency_percent",
> + "MetricName": "tma_lcp_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric estimates the fraction
of cycles when the CPU was stalled due to switches of uop delivery
to the Microcode Sequencer (MS). Commonly used instructions are
optimized for delivery by the DSB (decoded i-cache) or MITE (legacy
instruction decode) pipelines. Certain operations cannot be handled
natively by the execution pipeline; and must be performed by
microcode (small programs injected into the execution stream).
Switching to the MS too often can negatively impact performance. The
MS is designated to deliver long uop flows required by CISC
instructions like CPUID; or uncommon conditions like Floating Point
Assists when dealing with Denormals.",
> + "MetricExpr": "100 * ( ( 2 ) * IDQ.MS_SWITCHES / (
CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup":
"FetchLat;MicroSeq;TmaL3;m_tma_fetch_latency_percent",
> + "MetricName": "tma_ms_switches_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots the CPU was stalled due to Frontend bandwidth issues. For
example; inefficiencies at the instruction decoders; or restrictions
for caching in the DSB (decoded uops cache) are categorized under
Fetch Bandwidth. In such cases; the Frontend typically delivers
suboptimal amount of uops to the Backend.",
> + "MetricExpr": "100 * ( ( IDQ_UOPS_NOT_DELIVERED.CORE / (
( 4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) - ( ( 4 ) *
IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) )",
> + "MetricGroup":
"FetchBW;Frontend;TmaL2;m_tma_frontend_bound_percent",
> + "MetricName": "tma_fetch_bandwidth_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents Core
fraction of cycles in which CPU was likely limited due to the MITE
pipeline (the legacy decode pipeline). This pipeline is used for
code that was not pre-cached in the DSB or LSD. For example;
inefficiencies due to asymmetric decoders; use of long immediate or
LCP can manifest as MITE fetch bandwidth bottleneck.",
> + "MetricExpr": "100 * ( ( IDQ.ALL_MITE_CYCLES_ANY_UOPS -
IDQ.ALL_MITE_CYCLES_4_UOPS ) / ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 )
if #SMT_on else ( CPU_CLK_UNHALTED.THREAD ) ) / 2 )",
> + "MetricGroup":
"DSBmiss;FetchBW;TmaL3;m_tma_fetch_bandwidth_percent",
> + "MetricName": "tma_mite_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents Core
fraction of cycles in which CPU was likely limited due to DSB
(decoded uop cache) fetch pipeline. For example; inefficient
utilization of the DSB cache structure or bank conflict when reading
from it; are categorized here.",
> + "MetricExpr": "100 * ( ( IDQ.ALL_DSB_CYCLES_ANY_UOPS -
IDQ.ALL_DSB_CYCLES_4_UOPS ) / ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 )
if #SMT_on else ( CPU_CLK_UNHALTED.THREAD ) ) / 2 )",
> + "MetricGroup":
"DSB;FetchBW;TmaL3;m_tma_fetch_bandwidth_percent",
> + "MetricName": "tma_dsb_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This category represents fraction
of slots wasted due to incorrect speculations. This include slots
used to issue uops that do not eventually get retired and slots for
which the issue-pipeline was blocked due to recovery from earlier
incorrect speculation. For example; wasted work due to
miss-predicted branches are categorized under Bad Speculation
category. Incorrect data speculation followed by Memory Ordering
Nukes is another example.",
> + "MetricExpr": "100 * ( ( UOPS_ISSUED.ANY - (
UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) )",
> + "MetricGroup": "TmaL1",
> + "MetricName": "tma_bad_speculation_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots the CPU has wasted due to Branch Misprediction. These slots
are either wasted by uops fetched from an incorrectly speculated
program path; or stalls when the out-of-order part of the machine
needs to recover its state from a speculative path.",
> + "MetricExpr": "100 * ( ( BR_MISP_RETIRED.ALL_BRANCHES /
( BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT ) ) * ( (
UOPS_ISSUED.ANY - ( UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) )",
> + "MetricGroup":
"BadSpec;BrMispredicts;TmaL2;m_tma_bad_speculation_percent",
> + "MetricName": "tma_branch_mispredicts_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots the CPU has wasted due to Machine Clears. These slots are
either wasted by uops fetched prior to the clear; or stalls the
out-of-order portion of the machine needs to recover its state after
the clear. For example; this can happen due to memory ordering Nukes
(e.g. Memory Disambiguation) or Self-Modifying-Code (SMC) nukes.",
> + "MetricExpr": "100 * ( ( ( UOPS_ISSUED.ANY - (
UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) - ( ( BR_MISP_RETIRED.ALL_BRANCHES /
( BR_MISP_RETIRED.ALL_BRANCHES + MACHINE_CLEARS.COUNT ) ) * ( (
UOPS_ISSUED.ANY - ( UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) ) )",
> + "MetricGroup":
"BadSpec;MachineClears;TmaL2;m_tma_bad_speculation_percent",
> + "MetricName": "tma_machine_clears_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This category represents fraction
of slots where no uops are being delivered due to a lack of required
resources for accepting new uops in the Backend. Backend is the
portion of the processor core where the out-of-order scheduler
dispatches ready uops into their respective execution units; and
once completed these uops get retired according to program order.
For example; stalls due to data-cache misses or stalls due to the
divider unit being overloaded are both categorized under Backend
Bound. Backend Bound is further divided into two main categories:
Memory Bound and Core Bound.",
> + "MetricExpr": "100 * ( 1 - ( (
IDQ_UOPS_NOT_DELIVERED.CORE / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_ISSUED.ANY - (
UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_RETIRED.RETIRE_SLOTS ) /
( ( 4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) ) )",
> + "MetricGroup": "TmaL1",
> + "MetricName": "tma_backend_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots the Memory subsystem within the Backend was a bottleneck. Memory Bound estimates fraction of slots where pipeline is likely
stalled due to demand load or store instructions. This accounts
mainly for (1) non-completed in-flight memory demand loads which
coincides with execution units starvation; in addition to (2) cases
where stores could impose backpressure on the pipeline when many of
them get buffered at the same time (less common out of the two).",
> + "MetricExpr": "100 * ( ( ( CYCLE_ACTIVITY.STALLS_MEM_ANY
+ RESOURCE_STALLS.SB <http://RESOURCE_STALLS.SB> ) / ( (
CYCLE_ACTIVITY.STALLS_TOTAL + UOPS_EXECUTED.CYCLES_GE_1_UOP_EXEC - (
UOPS_EXECUTED.CYCLES_GE_3_UOPS_EXEC if ( ( INST_RETIRED.ANY / (
CPU_CLK_UNHALTED.THREAD ) ) > 1.8 ) else
UOPS_EXECUTED.CYCLES_GE_2_UOPS_EXEC ) - ( RS_EVENTS.EMPTY_CYCLES if
( ( ( 4 ) * IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / ( ( 4
) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) > 0.1 ) else 0 ) +
RESOURCE_STALLS.SB <http://RESOURCE_STALLS.SB> ) ) ) * ( 1 - ( (
IDQ_UOPS_NOT_DELIVERED.CORE / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_ISSUED.ANY - (
UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_RETIRED.RETIRE_SLOTS ) /
( ( 4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) ) ) )",
> + "MetricGroup": "Backend;TmaL2;m_tma_backend_bound_percent",
> + "MetricName": "tma_memory_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric estimates how often the
CPU was stalled without loads missing the L1 data cache. The L1
data cache typically has the shortest latency. However; in certain
cases like loads blocked on older stores; a load might suffer due to
high latency even though it is being satisfied by the L1. Another
example is loads who miss in the TLB. These cases are characterized
by execution unit stalls; while some non-completed demand load lives
in the machine without having that demand load missing the L1 cache.",
> + "MetricExpr": "100 * ( max( (
CYCLE_ACTIVITY.STALLS_MEM_ANY - CYCLE_ACTIVITY.STALLS_L1D_MISS ) / (
CPU_CLK_UNHALTED.THREAD ) , 0 ) )",
> + "MetricGroup":
"CacheMisses;MemoryBound;TmaL3mem;TmaL3;m_tma_memory_bound_percent",
> + "MetricName": "tma_l1_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric estimates how often the
CPU was stalled due to L2 cache accesses by loads. Avoiding cache
misses (i.e. L1 misses/L2 hits) can improve the latency and increase
performance.",
> + "MetricExpr": "100 * ( ( CYCLE_ACTIVITY.STALLS_L1D_MISS
- CYCLE_ACTIVITY.STALLS_L2_MISS ) / ( CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup":
"CacheMisses;MemoryBound;TmaL3mem;TmaL3;m_tma_memory_bound_percent",
> + "MetricName": "tma_l2_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric estimates how often the
CPU was stalled due to loads accesses to L3 cache or contended with
a sibling Core. Avoiding cache misses (i.e. L2 misses/L3 hits) can
improve the latency and increase performance.",
> + "MetricExpr": "100 * ( ( MEM_LOAD_UOPS_RETIRED.L3_HIT /
( MEM_LOAD_UOPS_RETIRED.L3_HIT + ( 7 ) *
MEM_LOAD_UOPS_RETIRED.L3_MISS ) ) * CYCLE_ACTIVITY.STALLS_L2_MISS /
( CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup":
"CacheMisses;MemoryBound;TmaL3mem;TmaL3;m_tma_memory_bound_percent",
> + "MetricName": "tma_l3_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric estimates how often the
CPU was stalled on accesses to external memory (DRAM) by loads.
Better caching can improve the latency and increase performance.",
> + "MetricExpr": "100 * ( min( ( ( 1 - (
MEM_LOAD_UOPS_RETIRED.L3_HIT / ( MEM_LOAD_UOPS_RETIRED.L3_HIT + ( 7
) * MEM_LOAD_UOPS_RETIRED.L3_MISS ) ) ) *
CYCLE_ACTIVITY.STALLS_L2_MISS / ( CPU_CLK_UNHALTED.THREAD ) ) , ( 1
) ) )",
> + "MetricGroup":
"MemoryBound;TmaL3mem;TmaL3;m_tma_memory_bound_percent",
> + "MetricName": "tma_dram_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric estimates how often CPU
was stalled due to RFO store memory accesses; RFO store issue a
read-for-ownership request before the write. Even though store
accesses do not typically stall out-of-order CPUs; there are few
cases where stores can lead to actual stalls. This metric will be
flagged should RFO stores be a bottleneck.",
> + "MetricExpr": "100 * ( RESOURCE_STALLS.SB
<http://RESOURCE_STALLS.SB> / ( CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup":
"MemoryBound;TmaL3mem;TmaL3;m_tma_memory_bound_percent",
> + "MetricName": "tma_store_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots where Core non-memory issues were of a bottleneck. Shortage
in hardware compute resources; or dependencies in software's
instructions are both categorized under Core Bound. Hence it may
indicate the machine ran out of an out-of-order resource; certain
execution units are overloaded or dependencies in program's data- or
instruction-flow are limiting the performance (e.g. FP-chained
long-latency arithmetic operations).",
> + "MetricExpr": "100 * ( ( 1 - ( (
IDQ_UOPS_NOT_DELIVERED.CORE / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_ISSUED.ANY - (
UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_RETIRED.RETIRE_SLOTS ) /
( ( 4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) ) ) - ( ( (
CYCLE_ACTIVITY.STALLS_MEM_ANY + RESOURCE_STALLS.SB
<http://RESOURCE_STALLS.SB> ) / ( ( CYCLE_ACTIVITY.STALLS_TOTAL +
UOPS_EXECUTED.CYCLES_GE_1_UOP_EXEC - (
UOPS_EXECUTED.CYCLES_GE_3_UOPS_EXEC if ( ( INST_RETIRED.ANY / (
CPU_CLK_UNHALTED.THREAD ) ) > 1.8 ) else
UOPS_EXECUTED.CYCLES_GE_2_UOPS_EXEC ) - ( RS_EVENTS.EMPTY_CYCLES if
( ( ( 4 ) * IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / ( ( 4
) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) > 0.1 ) else 0 ) +
RESOURCE_STALLS.SB <http://RESOURCE_STALLS.SB> ) ) ) * ( 1 - ( (
IDQ_UOPS_NOT_DELIVERED.CORE / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_ISSUED.ANY - (
UOPS_RETIRED.RETIRE_SLOTS ) + ( 4 ) * ( (
INT_MISC.RECOVERY_CYCLES_ANY / 2 ) if #SMT_on else
INT_MISC.RECOVERY_CYCLES ) ) / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) + ( ( UOPS_RETIRED.RETIRE_SLOTS ) /
( ( 4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) ) ) ) )",
> + "MetricGroup":
"Backend;TmaL2;Compute;m_tma_backend_bound_percent",
> + "MetricName": "tma_core_bound_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
cycles where the Divider unit was active. Divide and square root
instructions are performed by the Divider unit and can take
considerably longer latency than integer or Floating Point addition;
subtraction; or multiplication.",
> + "MetricExpr": "100 * ( ARITH.FPU_DIV_ACTIVE / ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) )",
> + "MetricGroup": "TmaL3;m_tma_core_bound_percent",
> + "MetricName": "tma_divider_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric estimates fraction of
cycles the CPU performance was potentially limited due to Core
computation issues (non divider-related). Two distinct categories
can be attributed into this metric: (1) heavy data-dependency among
contiguous instructions would manifest in this metric - such cases
are often referred to as low Instruction Level Parallelism (ILP).
(2) Contention on some hardware execution unit other than Divider.
For example; when there are too many multiply operations.",
> + "MetricExpr": "100 * ( ( ( ( CYCLE_ACTIVITY.STALLS_TOTAL
+ UOPS_EXECUTED.CYCLES_GE_1_UOP_EXEC - (
UOPS_EXECUTED.CYCLES_GE_3_UOPS_EXEC if ( ( INST_RETIRED.ANY / (
CPU_CLK_UNHALTED.THREAD ) ) > 1.8 ) else
UOPS_EXECUTED.CYCLES_GE_2_UOPS_EXEC ) - ( RS_EVENTS.EMPTY_CYCLES if
( ( ( 4 ) * IDQ_UOPS_NOT_DELIVERED.CYCLES_0_UOPS_DELIV.CORE / ( ( 4
) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) > 0.1 ) else 0 ) +
RESOURCE_STALLS.SB <http://RESOURCE_STALLS.SB> ) ) -
RESOURCE_STALLS.SB <http://RESOURCE_STALLS.SB> -
CYCLE_ACTIVITY.STALLS_MEM_ANY ) / ( CPU_CLK_UNHALTED.THREAD ) )",
> + "MetricGroup": "PortsUtil;TmaL3;m_tma_core_bound_percent",
> + "MetricName": "tma_ports_utilization_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This category represents fraction
of slots utilized by useful work i.e. issued uops that eventually
get retired. Ideally; all pipeline slots would be attributed to the
Retiring category. Retiring of 100% would indicate the maximum
Pipeline_Width throughput was achieved. Maximizing Retiring
typically increases the Instructions-per-cycle (see IPC metric).
Note that a high Retiring value does not necessary mean there is no
room for more performance. For example; Heavy-operations or
Microcode Assists are categorized under Retiring. They often
indicate suboptimal performance and can often be optimized or
avoided. ",
> + "MetricExpr": "100 * ( ( UOPS_RETIRED.RETIRE_SLOTS ) / (
( 4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) )",
> + "MetricGroup": "TmaL1",
> + "MetricName": "tma_retiring_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots where the CPU was retiring light-weight operations --
instructions that require no more than one uop (micro-operation).
This correlates with total number of instructions used by the
program. A uops-per-instruction (see UPI metric) ratio of 1 or less
should be expected for decently optimized software running on Intel
Core/Xeon products. While this often indicates efficient X86
instructions were executed; high value does not necessarily mean
better performance cannot be achieved.",
> + "MetricExpr": "100 * ( ( ( UOPS_RETIRED.RETIRE_SLOTS ) /
( ( 4 ) * ( ( CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) - ( ( ( ( UOPS_RETIRED.RETIRE_SLOTS
) / UOPS_ISSUED.ANY ) * IDQ.MS_UOPS / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) ) )",
> + "MetricGroup": "Retire;TmaL2;m_tma_retiring_percent",
> + "MetricName": "tma_light_operations_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents overall
arithmetic floating-point (FP) operations fraction the CPU has
executed (retired). Note this metric's value may exceed its parent
due to use of \"Uops\" CountDomain and FMA double-counting.",
> + "MetricExpr": "100 * ( ( INST_RETIRED.X87 * ( (
UOPS_RETIRED.RETIRE_SLOTS ) / INST_RETIRED.ANY ) / (
UOPS_RETIRED.RETIRE_SLOTS ) ) + ( (
FP_ARITH_INST_RETIRED.SCALAR_SINGLE +
FP_ARITH_INST_RETIRED.SCALAR_DOUBLE ) / ( UOPS_RETIRED.RETIRE_SLOTS
) ) + ( min( ( ( FP_ARITH_INST_RETIRED.128B_PACKED_DOUBLE +
FP_ARITH_INST_RETIRED.128B_PACKED_SINGLE +
FP_ARITH_INST_RETIRED.256B_PACKED_DOUBLE +
FP_ARITH_INST_RETIRED.256B_PACKED_SINGLE ) / (
UOPS_RETIRED.RETIRE_SLOTS ) ) , ( 1 ) ) ) )",
> + "MetricGroup": "HPC;TmaL3;m_tma_light_operations_percent",
> + "MetricName": "tma_fp_arith_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots where the CPU was retiring heavy-weight operations --
instructions that require two or more uops or microcoded sequences.
This highly-correlates with the uop length of these
instructions/sequences.",
> + "MetricExpr": "100 * ( ( ( ( UOPS_RETIRED.RETIRE_SLOTS )
/ UOPS_ISSUED.ANY ) * IDQ.MS_UOPS / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) ) )",
> + "MetricGroup": "Retire;TmaL2;m_tma_retiring_percent",
> + "MetricName": "tma_heavy_operations_percent",
> + "ScaleUnit": "1%"
> + },
> + {
> + "BriefDescription": "This metric represents fraction of
slots the CPU was retiring uops fetched by the Microcode Sequencer
(MS) unit. The MS is used for CISC instructions not supported by
the default decoders (like repeat move strings; or CPUID); or by
microcode assists used to address some operation modes (like in
Floating Point assists). These cases can often be avoided.",
> + "MetricExpr": "100 * ( ( ( UOPS_RETIRED.RETIRE_SLOTS ) /
UOPS_ISSUED.ANY ) * IDQ.MS_UOPS / ( ( 4 ) * ( (
CPU_CLK_UNHALTED.THREAD_ANY / 2 ) if #SMT_on else (
CPU_CLK_UNHALTED.THREAD ) ) ) )",
> + "MetricGroup":
"MicroSeq;TmaL3;m_tma_heavy_operations_percent",
> + "MetricName": "tma_microcode_sequencer_percent",
> + "ScaleUnit": "1%"
> }
> ]
> diff --git a/tools/perf/pmu-events/arch/x86/broadwellx/cache.json
b/tools/perf/pmu-events/arch/x86/broadwellx/cache.json
> index 127abe08362f..2efc4c0ee740 100644
> --- a/tools/perf/pmu-events/arch/x86/broadwellx/cache.json
> +++ b/tools/perf/pmu-events/arch/x86/broadwellx/cache.json
> @@ -814,9 +814,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_CODE_RD.LLC_HIT.HIT_OTHER_CORE_NO_FWD",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x04003C0244",
> + "MSRValue": "0x4003C0244",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch code
reads hit in the L3 and the snoops to sibling cores hit in either
E/S state and the line is not forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -829,7 +828,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x10003C0091",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch data
reads hit in the L3 and the snoop to one of the sibling cores hits
the line in M state and the line is forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -840,9 +838,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_DATA_RD.LLC_HIT.HIT_OTHER_CORE_NO_FWD",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x04003C0091",
> + "MSRValue": "0x4003C0091",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch data
reads hit in the L3 and the snoops to sibling cores hit in either
E/S state and the line is not forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -855,7 +852,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x10003C07F7",
> "Offcore": "1",
> - "PublicDescription": "Counts all data/code/rfo reads
(demand & prefetch) hit in the L3 and the snoop to one of the
sibling cores hits the line in M state and the line is forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -866,9 +862,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_READS.LLC_HIT.HIT_OTHER_CORE_NO_FWD",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x04003C07F7",
> + "MSRValue": "0x4003C07F7",
> "Offcore": "1",
> - "PublicDescription": "Counts all data/code/rfo reads
(demand & prefetch) hit in the L3 and the snoops to sibling cores
hit in either E/S state and the line is not forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -881,7 +876,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3F803C8FFF",
> "Offcore": "1",
> - "PublicDescription": "Counts all requests hit in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -894,7 +888,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x10003C0122",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch RFOs
hit in the L3 and the snoop to one of the sibling cores hits the
line in M state and the line is forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -905,9 +898,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_RFO.LLC_HIT.HIT_OTHER_CORE_NO_FWD",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x04003C0122",
> + "MSRValue": "0x4003C0122",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch RFOs
hit in the L3 and the snoops to sibling cores hit in either E/S
state and the line is not forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -920,7 +912,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3F803C0002",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand data writes
(RFOs) hit in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -933,7 +924,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x10003C0002",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand data writes
(RFOs) hit in the L3 and the snoop to one of the sibling cores hits
the line in M state and the line is forwarded",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -946,7 +936,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3F803C0200",
> "Offcore": "1",
> - "PublicDescription": "Counts prefetch (that bring data
to LLC only) code reads hit in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -959,7 +948,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3F803C0100",
> "Offcore": "1",
> - "PublicDescription": "Counts all prefetch (that bring
data to LLC only) RFOs hit in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -973,4 +961,4 @@
> "SampleAfterValue": "100003",
> "UMask": "0x10"
> }
> -]
> \ No newline at end of file
> +]
> diff --git
a/tools/perf/pmu-events/arch/x86/broadwellx/floating-point.json
b/tools/perf/pmu-events/arch/x86/broadwellx/floating-point.json
> index 9ad37dddb354..93bbc8600321 100644
> --- a/tools/perf/pmu-events/arch/x86/broadwellx/floating-point.json
> +++ b/tools/perf/pmu-events/arch/x86/broadwellx/floating-point.json
> @@ -5,6 +5,7 @@
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xc7",
> "EventName": "FP_ARITH_INST_RETIRED.128B_PACKED_DOUBLE",
> + "PublicDescription": "Number of SSE/AVX computational
128-bit packed double precision floating-point instructions retired;
some instructions will count twice as noted below. Each count
represents 2 computation operations, one for each element. Applies
to SSE* and AVX* packed double precision floating-point
instructions: ADD SUB HADD HSUB SUBADD MUL DIV MIN MAX SQRT DPP
FM(N)ADD/SUB. DPP and FM(N)ADD/SUB instructions count twice as they
perform 2 calculations per element. The DAZ and FTZ flags in the
MXCSR register need to be set when using these events.",
> "SampleAfterValue": "2000003",
> "UMask": "0x4"
> },
> @@ -14,6 +15,7 @@
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xc7",
> "EventName": "FP_ARITH_INST_RETIRED.128B_PACKED_SINGLE",
> + "PublicDescription": "Number of SSE/AVX computational
128-bit packed single precision floating-point instructions retired;
some instructions will count twice as noted below. Each count
represents 4 computation operations, one for each element. Applies
to SSE* and AVX* packed single precision floating-point
instructions: ADD SUB HADD HSUB SUBADD MUL DIV MIN MAX SQRT RSQRT
RCP DPP FM(N)ADD/SUB. DPP and FM(N)ADD/SUB instructions count twice
as they perform 2 calculations per element. The DAZ and FTZ flags in
the MXCSR register need to be set when using these events.",
> "SampleAfterValue": "2000003",
> "UMask": "0x8"
> },
> @@ -23,6 +25,7 @@
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xc7",
> "EventName": "FP_ARITH_INST_RETIRED.256B_PACKED_DOUBLE",
> + "PublicDescription": "Number of SSE/AVX computational
256-bit packed double precision floating-point instructions retired;
some instructions will count twice as noted below. Each count
represents 4 computation operations, one for each element. Applies
to SSE* and AVX* packed double precision floating-point
instructions: ADD SUB HADD HSUB SUBADD MUL DIV MIN MAX SQRT
FM(N)ADD/SUB. FM(N)ADD/SUB instructions count twice as they perform
2 calculations per element. The DAZ and FTZ flags in the MXCSR
register need to be set when using these events.",
> "SampleAfterValue": "2000003",
> "UMask": "0x10"
> },
> @@ -32,6 +35,7 @@
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xc7",
> "EventName": "FP_ARITH_INST_RETIRED.256B_PACKED_SINGLE",
> + "PublicDescription": "Number of SSE/AVX computational
256-bit packed single precision floating-point instructions retired;
some instructions will count twice as noted below. Each count
represents 8 computation operations, one for each element. Applies
to SSE* and AVX* packed single precision floating-point
instructions: ADD SUB HADD HSUB SUBADD MUL DIV MIN MAX SQRT RSQRT
RCP DPP FM(N)ADD/SUB. DPP and FM(N)ADD/SUB instructions count twice
as they perform 2 calculations per element. The DAZ and FTZ flags in
the MXCSR register need to be set when using these events.",
> "SampleAfterValue": "2000003",
> "UMask": "0x20"
> },
> @@ -59,6 +63,7 @@
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xc7",
> "EventName": "FP_ARITH_INST_RETIRED.SCALAR",
> + "PublicDescription": "Number of SSE/AVX computational
scalar single precision and double precision floating-point
instructions retired; some instructions will count twice as noted
below. Each count represents 1 computational operation. Applies to
SSE* and AVX* scalar single precision floating-point instructions:
ADD SUB MUL DIV MIN MAX SQRT RSQRT RCP FM(N)ADD/SUB. FM(N)ADD/SUB
instructions count twice as they perform 2 calculations per element.
The DAZ and FTZ flags in the MXCSR register need to be set when
using these events.",
> "SampleAfterValue": "2000003",
> "UMask": "0x3"
> },
> @@ -68,6 +73,7 @@
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xc7",
> "EventName": "FP_ARITH_INST_RETIRED.SCALAR_DOUBLE",
> + "PublicDescription": "Number of SSE/AVX computational
scalar double precision floating-point instructions retired; some
instructions will count twice as noted below. Each count represents
1 computational operation. Applies to SSE* and AVX* scalar double
precision floating-point instructions: ADD SUB MUL DIV MIN MAX SQRT
FM(N)ADD/SUB. FM(N)ADD/SUB instructions count twice as they perform
2 calculations per element. The DAZ and FTZ flags in the MXCSR
register need to be set when using these events.",
> "SampleAfterValue": "2000003",
> "UMask": "0x1"
> },
> @@ -77,6 +83,7 @@
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xc7",
> "EventName": "FP_ARITH_INST_RETIRED.SCALAR_SINGLE",
> + "PublicDescription": "Number of SSE/AVX computational
scalar single precision floating-point instructions retired; some
instructions will count twice as noted below. Each count represents
1 computational operation. Applies to SSE* and AVX* scalar single
precision floating-point instructions: ADD SUB MUL DIV MIN MAX SQRT
RSQRT RCP FM(N)ADD/SUB. FM(N)ADD/SUB instructions count twice as
they perform 2 calculations per element. The DAZ and FTZ flags in
the MXCSR register need to be set when using these events.",
> "SampleAfterValue": "2000003",
> "UMask": "0x2"
> },
> @@ -190,4 +197,4 @@
> "SampleAfterValue": "2000003",
> "UMask": "0x3"
> }
> -]
> \ No newline at end of file
> +]
> diff --git
a/tools/perf/pmu-events/arch/x86/broadwellx/frontend.json
b/tools/perf/pmu-events/arch/x86/broadwellx/frontend.json
> index f0bcb945ff76..37ce8034b2ed 100644
> --- a/tools/perf/pmu-events/arch/x86/broadwellx/frontend.json
> +++ b/tools/perf/pmu-events/arch/x86/broadwellx/frontend.json
> @@ -292,4 +292,4 @@
> "SampleAfterValue": "2000003",
> "UMask": "0x1"
> }
> -]
> \ No newline at end of file
> +]
> diff --git
a/tools/perf/pmu-events/arch/x86/broadwellx/memory.json
b/tools/perf/pmu-events/arch/x86/broadwellx/memory.json
> index cce993b197e3..545f61f691b9 100644
> --- a/tools/perf/pmu-events/arch/x86/broadwellx/memory.json
> +++ b/tools/perf/pmu-events/arch/x86/broadwellx/memory.json
> @@ -247,7 +247,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3FBFC00244",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch code
reads miss in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -258,9 +257,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_CODE_RD.LLC_MISS.LOCAL_DRAM",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x0604000244",
> + "MSRValue": "0x604000244",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch code
reads miss the L3 and the data is returned from local dram",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -273,7 +271,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3FBFC00091",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch data
reads miss in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -284,9 +281,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_DATA_RD.LLC_MISS.LOCAL_DRAM",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x0604000091",
> + "MSRValue": "0x604000091",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch data
reads miss the L3 and the data is returned from local dram",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -297,9 +293,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_DATA_RD.LLC_MISS.REMOTE_DRAM",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x063BC00091",
> + "MSRValue": "0x63BC00091",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch data
reads miss the L3 and the data is returned from remote dram",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -312,7 +307,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x103FC00091",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch data
reads miss the L3 and the modified data is transferred from remote
cache",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -323,9 +317,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_DATA_RD.LLC_MISS.REMOTE_HIT_FORWARD",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x087FC00091",
> + "MSRValue": "0x87FC00091",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch data
reads miss the L3 and clean or shared data is transferred from
remote cache",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -338,20 +331,18 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3FBFC007F7",
> "Offcore": "1",
> - "PublicDescription": "Counts all data/code/rfo reads
(demand & prefetch) miss in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> {
> - "BriefDescription": "Counts all data/code/rfo reads
(demand & prefetch)miss the L3 and the data is returned from local
dram",
> + "BriefDescription": "Counts all data/code/rfo reads
(demand & prefetch) miss the L3 and the data is returned from local
dram",
> "Counter": "0,1,2,3",
> "CounterHTOff": "0,1,2,3",
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_READS.LLC_MISS.LOCAL_DRAM",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x06040007F7",
> + "MSRValue": "0x6040007F7",
> "Offcore": "1",
> - "PublicDescription": "Counts all data/code/rfo reads
(demand & prefetch)miss the L3 and the data is returned from local
dram",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -362,9 +353,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_READS.LLC_MISS.REMOTE_DRAM",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x063BC007F7",
> + "MSRValue": "0x63BC007F7",
> "Offcore": "1",
> - "PublicDescription": "Counts all data/code/rfo reads
(demand & prefetch) miss the L3 and the data is returned from remote
dram",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -377,7 +367,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x103FC007F7",
> "Offcore": "1",
> - "PublicDescription": "Counts all data/code/rfo reads
(demand & prefetch) miss the L3 and the modified data is transferred
from remote cache",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -388,9 +377,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_READS.LLC_MISS.REMOTE_HIT_FORWARD",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x087FC007F7",
> + "MSRValue": "0x87FC007F7",
> "Offcore": "1",
> - "PublicDescription": "Counts all data/code/rfo reads
(demand & prefetch) miss the L3 and clean or shared data is
transferred from remote cache",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -403,7 +391,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3FBFC08FFF",
> "Offcore": "1",
> - "PublicDescription": "Counts all requests miss in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -416,7 +403,6 @@
> "MSRIndex": "0x1a6,0x1a7",
> "MSRValue": "0x3FBFC00122",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & prefetch RFOs
miss in the L3",
> "SampleAfterValue": "100003",
> "UMask": "0x1"
> },
> @@ -427,9 +413,8 @@
> "EventCode": "0xB7, 0xBB",
> "EventName":
"OFFCORE_RESPONSE.ALL_RFO.LLC_MISS.LOCAL_DRAM",
> "MSRIndex": "0x1a6,0x1a7",
> - "MSRValue": "0x0604000122",
> + "MSRValue": "0x604000122",
> "Offcore": "1",
> - "PublicDescription": "Counts all demand & pref