Re: [RFC PATCH 3/3] idle: store the idle state index in the structrq

From: Daniel Lezcano
Date: Fri Jan 31 2014 - 05:44:38 EST


On 01/30/2014 10:02 PM, Nicolas Pitre wrote:
On Thu, 30 Jan 2014, Lorenzo Pieralisi wrote:

On Thu, Jan 30, 2014 at 05:25:27PM +0000, Daniel Lezcano wrote:
On 01/30/2014 05:35 PM, Peter Zijlstra wrote:
On Thu, Jan 30, 2014 at 05:27:54PM +0100, Daniel Lezcano wrote:
IIRC, Alex Shi sent a patchset to improve the choosing of the idlest cpu and
the exit_latency was needed.

Right. However if we have a 'natural' order in the state array the index
itself might often be sufficient to find the least idle state, in this
specific case the absolute exit latency doesn't matter, all we want is
the lowest one.

Indeed. It could be simple as that. I feel we may need more informations
in the future but comparing the indexes could be a nice simple and
efficient solution.

As long as we take into account that some states might require multiple
CPUs to be idle in order to be entered, fine by me. But we should
certainly avoid waking up a CPU in a cluster that is in eg C2 (all CPUs in
C2, so cluster in C2) when there are CPUs in C3 in other clusters with
some CPUs running in those clusters, because there C3 means "CPU in C3, not
cluster in C3". Overall what I am saying is that what you are doing
makes perfect sense but we have to take the above into account.

Some states have CPU and cluster (or we can call it package) components,
and that's true on ARM and other architectures too, to the best of my
knowledge.

The notion of cluster or package maps pretty naturally onto scheduling
domains. And the search for an idle CPU to wake up should avoid a
scheduling domain with a load of zero (which is obviously a prerequisite
for a power save mode to be applied to the cluster level) if there exist
idle CPUs in another domain already which load is not zero (all other
considerations being equal). Hence your concern would be addressed
without any particular issue even if the individual CPU idle state index
is not exactly in sync with reality because of other hardware related
constraints.

The other solution consists in making the index dynamic. That means
letting backend idle drivers change it i.e. when the last man in a
cluster goes idle it could update the index for all the other CPUs in
the cluster. There is no locking needed as the scheduler is only
consuming this info, and the scheduler getting it wrong on rare
occasions is not a big deal either. But that looks pretty ugly as at
least 2 levels of abstractions would be breached in this case.

Yes, I agree it would break the level of abstractions and I don't think it is worth to take into account this for now.

Let's consider the following status:

1. there are archs where the cluster dependency is handled by the firmware and where the 'intermediate' idle state to wait for the cpu sync is hidden because of the level of abstraction of such firmware.
This is the case for x86 arch and ARM platform with PSCI which represent most of the hardware.

2. there are archs where the cluster dependency is handled by the cpuidle couple idle state and where the cpumask (stored in the idle state structure) gives us this dependency which is a very small part of the hardware and where most of the boards at EOL (omap4, tegra2).

3. there are archs where the cluster dependency is built from the device tree and where a mapping for the cluster topology is discussed.

4. there are archs where the cluster dependency is reflected by the usage of the multiple cpuidle driver support (big.Little).

Having the index stored in the struct rq is a good first step to integrate the cpuidle with the scheduler even if we don't have an accurate result at the beginning.


--
<http://www.linaro.org/> Linaro.org â Open source software for ARM SoCs

Follow Linaro: <http://www.facebook.com/pages/Linaro> Facebook |
<http://twitter.com/#!/linaroorg> Twitter |
<http://www.linaro.org/linaro-blog/> Blog

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/