2.6.18.1 + grsecurity JFS failed with dbAllocNext: Corrupt dmap page
From: lkml-2006i-ticket@xxxxxxxxxx
Date: Tue Oct 31 2006 - 03:52:26 EST
Hello,
on 2.6.18.1 + grsecurity I totally lost an JFS partition.
I was doing regular work on freshly installed debian stable (on
olerd/rather tested hardware that worked 24/24 7/7)
I installed 2.6.18.1 + grsecurity, started memtest and stress in the
background, enabled swap with encryption, and played a bit with SysRq -
using it to dump CPU state and tasks list (playing around).
After a while I got tons of errors like
ERROR: (device hda3): DT_GETPAGE: dtree page corrupt
ERROR: (device hda3): dbAllocNext: Corrupt dmap page
about 10 per second. I rebooted (computer hanged)
After reset the partition used - hda3 (as /, it was the only rw mounted
partition) was gone, unable to repair "since both master and secondary
superblocks are damaged".
If anyone is interested I saved first 100 MB of this ~6 GB partition
using dd, it looks totally different in hexdump then other partition (at
least the begin) so it suggests that indeed superblock and other data
was totally shreded / filled with garbage.
The config it Althon 1700 with 512 RAM, kernel 2.6.18.1 + grsecurity.
Config: in 250 Hz timer frequency, with Voluntary Kernel Preemption,
with no privilaged ioports, possible with some additional kernel hackng
options (like vm debugging),
otherwise config was like the 2.6.18 (but not .1) that run perfectly on
other machines.
The failure occured during heavy load and stress testing while playing
with SysRq at same time (but only non-destructive things like -p -t -m)
The smartctrl shows 2 UDMA errors for this device (but they might be
very old).
Possible that this is hardware fault, but letting now perhaps anyone had
simmilar problems...
ERROR: (device hda3): DT_GETPAGE: dtree page corrupt
ERROR: (device hda3): dbAllocNext: Corrupt dmap page
--
LimCore C++ Software Architect / Team Lead
---> oo Linux programs
limcore
software
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/