Re: Poor Software RAID-0 performance with 126.96.36.199
From: Lennart Sorensen
Date: Mon Nov 21 2005 - 15:47:34 EST
On Mon, Nov 21, 2005 at 09:31:14PM +0100, Lars Roland wrote:
> I have created a stripe across two 500Gb disks located on separate IDE
> channels using:
> mdadm -Cv /dev/md0 -c32 -n2 -l0 /dev/hdb /dev/hdd
Does -l0 equal stripe or linear? The mdadm man page doesn't seem clear
o that to me.
If it defaults to linear, then you shouldn't expect any performance gain
since that would just stick one drive after the other (no striping).
Try explicitly statping -l stripe instead of -l 0.
> the performance is awful on both kernel 188.8.131.52 and 184.108.40.206 (even
> with hdparm and blockdev tuning), both bonnie++ and hdparm (included
> below) shows a single disk operating faster than the stripe:
> dkstorage01:~# hdparm -t /dev/md0
> Timing buffered disk reads: 182 MB in 3.01 seconds = 60.47 MB/sec
> dkstorage02:~# hdparm -t /dev/hdc1
> Timing buffered disk reads: 184 MB in 3.02 seconds = 60.93 MB/sec
How about at least testing one of the drives involved in the raid,
although I assume they are identical in your case given the numbers.
Did you test this with other kernel versions (older ones) to see if it
was better in the past?
Any idea where the ide controller is connected? If it is PCI the whole
bus only has 133MB/s to give on many systems (some have more of course),
so maybe 60M/s is quite good.
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/