Howdy folks, i've got a strange thing happening to my raid array and thought i would see what you guys thought. If this has already been covered somewhere else, please forgive, but as far as i know, i haven't read anything quite like this on the forums.
I'm running sme 7.3 on a raid 5 disk array (adaptec Ultra 2 wide raid controller) with 3 18 gig drives sda1, sda2 and sda3. Since i have built this machine, every so often, and not in any pattern that i can tell, sda2 gets a failure message and is removed from the raid array.
Of course, the raid array goes into degraded mode, but continues to function.
here's the strange part.
EVERY time it does this, i test the drive that is says has failed, and the drive functions perfectly!!
All i have to do is run the mdadm -a command again for it to rebuild the raid using the same disk it says is bad, and it goes back to normal, until out of the blue, it decides that the drive has failed and removes it again.
It has done this a total of 4 times in the 1.5 yrs i've had this machine running.
Anyone have a clue as to what might be behind this?
I'll post whatever logs you request, just make sure you include how to get to them because i don't know where to find everything.
Thanks guys!!
SME ROCKS by the way!!!
BigPoppa