Hi Jeremy,
Sorry, I don't have your previous patch emails to reply on right patch
context.
So commenting on top of this patch.
AFAIU, the PPTT v5 patches still rely on CLIDR_EL1 register to know the type
of
Caches enabled/available on the platform. With PPTT, it should not rely on
architecture
registers. There can be platforms which can report cache availability in
PPTT but not in
architecture registers.
The following code snippet shows usage of CLIDR_EL1
In arch/arm64/kernel/cacheinfo.c
static inline enum cache_type get_cache_type(int level)
{
u64 clidr;
if (level > MAX_CACHE_LEVEL)
return CACHE_TYPE_NOCACHE;
clidr = read_sysreg(clidr_el1);
return CLIDR_CTYPE(clidr, level);
}
static int __populate_cache_leaves(unsigned int cpu)
{
unsigned int level, idx;
enum cache_type type;
struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
struct cacheinfo *this_leaf = this_cpu_ci->info_list;
for (idx = 0, level = 1; level <= this_cpu_ci->num_levels &&
idx < this_cpu_ci->num_leaves; idx++, level++) {
type = get_cache_type(level);
if (type == CACHE_TYPE_SEPARATE) {
ci_leaf_init(this_leaf++, CACHE_TYPE_DATA, level);
ci_leaf_init(this_leaf++, CACHE_TYPE_INST, level);
} else {
ci_leaf_init(this_leaf++, type, level);
}
}
return 0;
}
In populate_cache_leaves() the cache type is read from CLIDR_EL1 register.
If CLIDR_EL1 reports CACHE_TYPE_NOCACHE for a particular level then sysfs
entry
/sys/devices/system/cpu/cpu0/index<n>/type is not created and hence
userspace tools
like lstopo will not report this cache level.
Regards
Vijay
-----Original Message-----[mailto:linux-arm-kernel-bounces@xxxxxxxxxxxxxxxxxxx]
From: linux-arm-kernel
On Behalf Of Rafael J. Wysocki<linux-acpi@xxxxxxxxxxxxxxx>;
Sent: Thursday, December 14, 2017 4:40 AM
To: Jeremy Linton <jeremy.linton@xxxxxxx>
Cc: Mark Rutland <mark.rutland@xxxxxxx>; Jonathan.Zhang@xxxxxxxxxx;
Jayachandran.Nair@xxxxxxxxxx; Lorenzo Pieralisi
<lorenzo.pieralisi@xxxxxxx>; Catalin Marinas <catalin.marinas@xxxxxxx>;
Rafael J. Wysocki <rafael@xxxxxxxxxx>; jhugo@xxxxxxxxxxxxxx; Will Deacon
<will.deacon@xxxxxxx>; Linux PM <linux-pm@xxxxxxxxxxxxxxx>; Rafael J.
Wysocki <rjw@xxxxxxxxxxxxx>; Greg Kroah-Hartman
<gregkh@xxxxxxxxxxxxxxxxxxx>; Linux Kernel Mailing List <linux-
kernel@xxxxxxxxxxxxxxx>; ACPI Devel Maling List
Viresh Kumar <viresh.kumar@xxxxxxxxxx>; Hanjun Guo
<hanjun.guo@xxxxxxxxxx>; Al Stone <ahs3@xxxxxxxxxx>; Sudeep Holla
<sudeep.holla@xxxxxxx>; austinwc@xxxxxxxxxxxxxx;
wangxiongfeng2@xxxxxxxxxx; linux-arm-kernel@xxxxxxxxxxxxxxxxxxx; Len
Brown <lenb@xxxxxxxxxx>
Subject: Re: [PATCH v5 6/9] ACPI/PPTT: Add topology parsing code
On Thu, Dec 14, 2017 at 12:06 AM, Jeremy Linton <jeremy.linton@xxxxxxx>
wrote:
Hi,
On 12/13/2017 04:28 PM, Rafael J. Wysocki wrote:
On Wed, Dec 13, 2017 at 6:38 PM, Lorenzo Pieralisi
<lorenzo.pieralisi@xxxxxxx> wrote:
On Tue, Dec 12, 2017 at 10:13:08AM -0600, Jeremy Linton wrote:
Hi,
First, thanks for taking a look at this.
On 12/11/2017 07:12 PM, Rafael J. Wysocki wrote:
On Friday, December 1, 2017 11:23:27 PM CET Jeremy Linton wrote:
The PPTT can be used to determine the groupings of CPU's at given
levels in the system. Lets add a few routines to the PPTT parsing
code to return a unique id for each unique level in the processor
hierarchy. This can then be matched to build
thread/core/cluster/die/package/etc mappings for each processing
element in the system.
Signed-off-by: Jeremy Linton <jeremy.linton@xxxxxxx>
Why can't this be folded into patch [2/9]?
It can, and I will be happy squash it.
It was requested that the topology portion of the parser be split
out back in v3.
https://www.spinics.net/lists/linux-acpi/msg78487.html
I asked to split cache/topology since I am not familiar with cache
code and Sudeep - who looks after the cache code - won't be able to
review this series in time for v4.16.
OK, so why do we need it in 4.16?
I think its more case of as soon as possible. That is because there
are machines where the topology is completely incorrect due to
assumptions the kernel makes based on registers that aren't defined
for that purpose (say describing which cores are in a physical socket,
or LLC's attached to interconnects or memory controllers).
This incorrect topology information is reported to things like the
kernel scheduler, which then makes poor scheduling decisions resulting
in sub-optimal system performance.
This patchset (and ACPI 6.2) clears up a lot of those problems.
As long as the ACPI tables are as expected that is, I suppose?
Anyway, fair enough, but I don't want to rush it in.
Thanks,
Rafael
_______________________________________________
linux-arm-kernel mailing list
linux-arm-kernel@xxxxxxxxxxxxxxxxxxx
http://lists.infradead.org/mailman/listinfo/linux-arm-kernel