Hi Peter,
On 25/03/25 17:39, Peter Zijlstra wrote:
Hi all,
One of the many things on the eternal todo list has been finishing the
below hackery.
It is an attempt at modelling cache affinity -- and while the patch
really only targets LLC, it could very well be extended to also apply to
clusters (L2). Specifically any case of multiple cache domains inside a
node.
Anyway, I wrote this about a year ago, and I mentioned this at the
recent OSPM conf where Gautham and Prateek expressed interest in playing
with this code.
So here goes, very rough and largely unproven code ahead :-)
It applies to current tip/master, but I know it will fail the __percpu
validation that sits in -next, although that shouldn't be terribly hard
to fix up.
As is, it only computes a CPU inside the LLC that has the highest recent
runtime, this CPU is then used in the wake-up path to steer towards this
LLC and in task_hot() to limit migrations away from it.
More elaborate things could be done, notably there is an XXX in there
somewhere about finding the best LLC inside a NODE (interaction with
NUMA_BALANCING).
Tested the patch on a 12-core, 96-thread Power10 system using a real-life
workload, DayTrader.
Here is a summary of the runs:
Users | Instances | Throughput vs Base | Avg Resp. Time vs Base
--------------------------------------------------------------
30 | 1 | -25.3% | +50%
60 | 1 | -25.1% | +50%
30 | 3 | -22.8% | +33%
As of now, the patch negatively impacts performance both in terms of
throughput and latency.
I will conduct more extensive testing with both microbenchmarks and
real-life workloads.
Thanks,
Madadi Vineeth Reddy