RAID 5 with 6 disks - one parity, no hot spares.
Description of Problem:
I had two cases of the following scenario:
a. A disk fails. During the failure a hiccup on SCSI bus causes a bad block
written to another disk.
b. Array marks failed disk as bad and continues in degraded mode.
c. In a very short time array encounters a bad block and marks the second disk
d. At this point system is pretty much dead because array is not responding.
e. Recovery involves fixing bad blocks by writing some data into them and
direct editing raid volume superblocks to include (semi-) mistakenly marked as
bad disk back in the array.
Seems like raid code should be smart enough to understand that its running in
degraded mode and no disk should be removed from array.
If array is in degraded mode it shell remain up and report read errors when
encounters bad blocks in one or more of the disks.
When I tried to run
raidhotremove /dev/md0 /dev/sdb1
I found following in /var/log/messages: (see attachment).
Created attachment 47320 [details]
extract from the /var/log/messages
Thanks for the bug report. However, Red Hat no longer maintains this version of
the product. Please upgrade to the latest version and open a new bug if the problem
The Fedora Legacy project (http://fedoralegacy.org/) maintains some older releases,
and if you believe this bug is interesting to them, please report the problem in
the bug tracker at: http://bugzilla.fedora.us/