On Wed, Aug 26, 2009 at 07:58:40AM -0400, Ric Wheeler wrote:Drive in raid 5 failed; hot spare was available (no idea about
UPS). System apparently locked up trying to talk to the failed drive,
or maybe admin just was not patient enough, so he just powercycled the
array. He lost the array.
So while most people will not agressively powercycle the RAID array,
drive failure still provokes little tested error paths, and getting
unclean shutdown is quite easy in such case.
Then what we need to document is do not power cycle an array during a
rebuild, right?
Well, the softwar raid layer could be improved so that it implements
scrubbing by default (i.e., have the md package install a cron job to
implement a periodict scrub pass automatically). The MD code could
also regularly check to make sure the hot spare is OK; the other
possibility is that hot spare, which hadn't been used in a long time,
had silently failed.
In the end, there are cascading failures that will defeat any data
protection scheme, but that does not mean that the value of that scheme
is zero. We need to be get more people to use RAID (including MD5) and
try to enhance it as we go. Just using a single disk is not a good
thing...
Yep; the solution is to improve the storage devices. It is *not* to
encourage people to think RAID is not worth it, or that somehow ext2
is better than ext3 because it runs fsck's all the time at boot up.
That's just crazy talk.
- Ted