archive mirror
 help / color / mirror / Atom feed
* RAID1/5, SMP 2.4.18-27.7-xsmp, EXT3 and file system corruption
@ 2003-04-27  6:20 David Redman
  0 siblings, 0 replies; only message in thread
From: David Redman @ 2003-04-27  6:20 UTC (permalink / raw)
  To: linux-kernel

First off, I am not subscribed to the list, I used to be but not any 
more; please reply directly to me: or to

So onto the problem: I've been seeing data corruption on my RAID 
metadrives for some time (months); over reboots they almost never come 
up clean, and a umount and fsck -f always ends up finding damage.
These are large EXT3 FSs - 120GB.

This is one of the failing configs:
	/dev/md0 4x40GB IBM drives RAID5
	/dev/md1 2x120GB IBM drives RAID1
I have also tried:
	/dev/md0 4x40 RAID5
	/dev/md1 /dev/md0 + 1x120 RAID1
and I see corruption on both metadisks in both configurations.

The arrays get consistently damaged; I get duplicate blocks, corrupted 
inodes, and directory holes.
An example from dmesg:
	Apr 25 04:19:35 nasbox kernel: EXT3-fs error (device md(9,2)):
	ext3_readdir: dir dectory #7603330 contains a hole at offset
This is from the RAID5 array..

My machine is nothing special Dual P3@700Mhz 384Mb DRAM, P6DBE 
super-micro motherboard. the IDE controllers running the discs are all 
PROMISE IDE PDC20262 controllers, the drives are all using UltraDMA 
cables and all drives are the master on thier own controllers (Ie 3 
cards, each with 2 drives). Root/swap is on the motherboard IDE. My 
kernel is a stock Redhat7.3; A home-brew also gets corrupted FSs.

Anyone else seen this kind of corruption? I've searched google and 
noticed a few other postings, but I thought I'd ask here.

Since I broke the mirror and just mounted a single drive (ie no 
/dev/md1, just using a raw drive /dev/hde1) I have not noticed any 
corruption on the filesystem, the other drive /dev/md0 (RAID5) is still 
getting corrupted.

I have also tried running a mirror on the single 120GB drive in degraded 
mode (ie one disk missing) and it also gets corruption. At this point I 
have to assume that my problem is /dev/md related, and because I dont 
see anyone else complaining I am assuming it is somehow tied to SMP.

I'm going to boot a non-SMP kernel and see if that changes the behavior; 
problem is it takes a horribly long time to fsck the arrays when/if they 
go bad.. sigh.

Thanks for any input;


^ permalink raw reply	[flat|nested] only message in thread

only message in thread, other threads:[~2003-04-27  6:08 UTC | newest]

Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2003-04-27  6:20 RAID1/5, SMP 2.4.18-27.7-xsmp, EXT3 and file system corruption David Redman

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).