Re: [PATCH 0/2] Migrate data off physical pages with corrected memory errors (Version 7)
From: Russ Anderson
Date: Sun Jul 20 2008 - 13:39:28 EST
On Sat, Jul 19, 2008 at 12:37:11PM +0200, Andi Kleen wrote:
> Russ Anderson <rja@xxxxxxx> writes:
> > [PATCH 0/2] Migrate data off physical pages with corrected memory errors (Version 7)
> FWIW I discussed this with some hardware people and the general
> opinion was that it was way too aggressive to disable a page on the
> first corrected error like this patchkit currently does.
Part of the "fun" of memory error decision making is that memory hardware
can fail in different ways based on design, manufacturing process, running
conditions (ie temperature), etc. So the right answer for one type
of memory hardware may be the wrong answer for another type. That is
why the decision making part of the migration code is implemented
as a kernel loadable module. That way distros/vendors can use
a module appropriate for the specific hardware.
The patch has a module for IA64, based on experience on IA64 hardware.
It is a first step, to get the basic functionality in the kernel.
The module can be enhanced for different failure modes and hardware
Note also the functionality to return pages that have been marked
bad. This allows the pages to be freed if the module is too aggressive.
> The corrected bit error could be caused by a temporary condition
> e.g. in the DIMM link, and does not necessarily mean that part of the
> DIMM is really going bad. Permanently disabling would only be
> justified if you saw repeated corrected errors over a long time from
> the same DIMM.
That is true in some cases. We have extensive experience with Altix
hardware where corrected errors quickly degrade to uncorrected errors.
> There are also some potential scenarios where being so aggressive
> could hurt, e.g. if you have a low rate of random corrected events
> spread randomly all over your memory (e.g. with a flakey DIMM
> connection) after a long enough uptime you could lose significant parts
> of your memory even though the DIMM is actually still ok.
That is a function of system size. The fewer DIMMs in the system the
greater that could be a issue. Altix systems tend to have many DIMMs
(~20,000 in one customer system). So disabling the memory on a
DIMM with a flaky connector is a small percentage of overall memory.
On a large NUMA machine the flaky DIMM connector would only effect
memory on one node.
> Also the other issue that if the DIMM is going bad then it's likely
> larger areas than just the lines making up this page. So you
> would still risk uncorrected errors anyways because disabling
> the page would only cover a small subset of the affected area.
Sure. A common failure mode is that a row/column on a DRAM goes
bad, which effects a range of addresses. I have a DIMM on one
of my test machines which behaves that way. It was valuable for
testing the code because several meg worth of pages get migrated.
It is a good stress test for the migration code.
A good enhancement would be to migrate all the data off a DRAM and/or
DIMM when a threshold is exceeded. That would take knowledge of the
physical memory to memory map layout.
> If you really wanted to do this you probably should hook it up
> to mcelog's (or the IA64 equivalent) DIMM database
Is there an IA64 equivalent? I've looked at the x86_64 mcelog,
but have not found a IA64 version.
> and then
> control it from user space with suitable large thresholds
> and DIMM specific knowledge. But it's unlikely it can be really
> done nicely in a way that is isolated from very specific
> knowledge about the underlying memory configuration.
Agreed. An interface to export the physical memory configuration
(from ACPI tables?) would be useful.
Russ Anderson, OS RAS/Partitioning Project Lead
SGI - Silicon Graphics Inc rja@xxxxxxx
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/