Re: [patch] ext2/3: document conditions when reliable operation ispossible

From: Ric Wheeler
Date: Thu Aug 27 2009 - 07:44:29 EST


On 08/26/2009 11:53 PM, Rob Landley wrote:
On Tuesday 25 August 2009 18:40:50 Ric Wheeler wrote:
Repeat experiment until you get up to something like google scale or the
other papers on failures in national labs in the US and then we can have an
informed discussion.
On google scale anvil lightning can fry your machine out of a clear sky.

However, there are still a few non-enterprise users out there, and knowing
that specific usage patterns don't behave like they expect might be useful to
them.


You are missing the broader point of both papers. They (and people like me when back at EMC) look at large numbers of machines and try to fix what actually breaks when run in the real world and causes data loss. The motherboards, S-ATA controllers, disk types are the same class of parts that I have in my desktop box today.

The advantage of google, national labs, etc is that they have large numbers of systems and can draw conclusions that are meaningful to our broad user base.

Specifically, in using S-ATA drives (just like ours, maybe slightly more reliable) they see up to 7% of those drives fail each year. All users have "soft" drive failures like single remapped sectors.

These errors happen extremely commonly and are what RAID deals with well.

What does not happen commonly is that during the RAID rebuild (kicked off only after a drive is kicked out), you push the power button or have a second failure (power outage).

We will have more users loose data if they decide to use ext2 instead of ext3 and use only single disk storage.

We have real numbers that show that is true. Injecting double faults into a system that handles single faults is frankly not that interesting.

You can get better protection from these double faults if you move to "cloud" like storage configs where each box is fault tolerant, but you also spread your data over multiple boxes in multiple locations.

Regards,

Ric

I can promise you that hot unplugging and replugging a S-ATA drive will
also lose you data if you are actively writing to it (ext2, 3,
whatever).
I can promise you that running S-ATA drive will also lose you data,
even if you are not actively writing to it. Just wait 10 years; so
what is your point?
I lost a s-ata drive 24 hours after installing it in a new box. If I had
MD5 RAID5, I would not have lost any.

My point is that you fail to take into account the rate of failures of a
given configuration and the probability of data loss given those rates.
Actually, that's _exactly_ what he's talking about.

When writing to a degraded raid or a flash disk, journaling is essentially
useless. If you get a power failure, kernel panic, somebody tripping over a
USB cable, and so on, your filesystem will not be protected by journaling.
Your data won't be trashed _every_ time, but the likelihood is much greater
than experience with journaling in other contexts would suggest.

Worse, the journaling may be counterproductive by _hiding_ many errors that
fsck would promptly detect, so when the error is detected it may not be
associated with the event that caused it. It also may not be noticed until
good backups of the data have been overwritten or otherwise cycled out.

You seem to be arguing that Linux is no longer used anywhere but the
enterprise, so issues affecting USB flash keys or cheap software-only RAID
aren't worth documenting?

Rob

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/