Re: [PATCH 00/10] x86: Reduce memory and intra-node effects withlarge count NR_CPUs

From: Mike Travis
Date: Mon Jan 14 2008 - 13:01:04 EST



Ingo Molnar wrote:
> * Andi Kleen <ak@xxxxxxx> wrote:
>
>>> i.e. we've got ~22K bloat per CPU - which is not bad, but because
>>> it's a static component, it hurts smaller boxes. For distributors to
>>> enable CONFIG_NR_CPU=1024 by default i guess that bloat has to drop
>>> below 1-2K per CPU :-/ [that would still mean 1-2MB total bloat but
>>> that's much more acceptable than 23MB]
>> Even 1-2MB overhead would be too much for distributors I think.
>> Ideally there must be near zero overhead for possible CPUs (and I see
>> no principle reason why this is not possible) Worst case a low few
>> hundred KBs, but even that would be much.
>
> i think this patchset already gives a net win, by moving stuff from
> NR_CPUS arrays into per_cpu area. (Travis please confirm that this is
> indeed what the numbers show)
>
> The (total-)size of the per-cpu area(s) grows linearly with the number
> of CPUs, so we'll have the expected near-zero overhead on 4-8-16-32 CPUs
> and the expected larger total overhead on 1024 CPUs.
>
> Ingo

Yes, and it's just the first step. Ideally, there is *no* extra memory
used by specifying NR_CPUS = <whatever> and all the extra memory only
comes into play when they are "possible/probable". This means that almost
all of the data needs to be in the percpu area (and compact that as much
as possible) or in the initdata section and discarded after use.

And Andi is right, the distributors will not default the NR_CPUS to a large
value unless there is zero or very little overhead. And since so much
depends on using standard configurations (certifications, etc.) we cannot
depend on a special build.

Thanks,
Mike

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/