I recently set up a new IBM server for a client, and installed 5.6, followed by the most recent updates. Subsequent to this, the second disk in the software RAID (2 x 40 GB drives, different models, primary master and secondary slave) became inaccessible.
It appeared that the second drive had failed, so I had it replaced with another Maxtor drive. In trying to rebuild the RAID array, I used Darrell May's how-to, and found that I had no idea of the drive geometry to use in setting the partitions. I therefore backed up to desktop and re-installed 5.6. Before I applied the updates, I checked the RAID status and saw that it was sync'ing OK. I then installed the latest SME Server upgrades, and following the reboot, I found that the second drive was no longer active.
dmesg output shows:
md: Autodetecting RAID arrays.
[events: 00000008]
[events: 00000008]
[events: 00000008]
md: autorun ...
md: considering hda1 ...
md: adding hda1 ...
md: created md0
md: bind
md: running:
md: hda1's event counter: 00000008
md0: former device [dev 16:41] is unavailable, removing from array!
md: RAID level 1 does not need chunksize! Continuing anyway.
md0: max total readahead window set to 508k
md0: 1 data-disks, max readahead per data-disk: 508k
raid1: device hda1 operational as mirror 0
raid1: md0, not all disks are operational -- trying to recover array
raid1: raid set md0 active with 1 out of 2 mirrors
md: updating md0 RAID superblock on device
md: hda1 [events: 00000009]<6>(write) hda1's sb offset: 104320
md: recovery thread got woken up ...
md0: no spare disk to reconstruct array! -- continuing in degraded mode
md: recovery thread finished ...
md: considering hda2 ...
md: adding hda2 ...
md: created md1
md: bind
md: running:
md: hda2's event counter: 00000008
md1: former device [dev 16:42] is unavailable, removing from array!
md: md1: raid array is not clean -- starting background reconstruction
md: RAID level 1 does not need chunksize! Continuing anyway.
md1: max total readahead window set to 508k
md1: 1 data-disks, max readahead per data-disk: 508k
raid1: device hda2 operational as mirror 0
raid1: md1, not all disks are operational -- trying to recover array
raid1: raid set md1 active with 1 out of 2 mirrors
md: updating md1 RAID superblock on device
md: hda2 [events: 00000009]<6>(write) hda2's sb offset: 38708544
md: recovery thread got woken up ...
md1: no spare disk to reconstruct array! -- continuing in degraded mode
md0: no spare disk to reconstruct array! -- continuing in degraded mode
md: recovery thread finished ...
md: considering hda3 ...
md: adding hda3 ...
md: created md2
md: bind
md: running:
md: hda3's event counter: 00000008
md2: former device [dev 16:43] is unavailable, removing from array!
md: RAID level 1 does not need chunksize! Continuing anyway.
md2: max total readahead window set to 508k
md2: 1 data-disks, max readahead per data-disk: 508k
raid1: device hda3 operational as mirror 0
raid1: md2, not all disks are operational -- trying to recover array
raid1: raid set md2 active with 1 out of 2 mirrors
md: updating md2 RAID superblock on device
md: hda3 [events: 00000009]<6>(write) hda3's sb offset: 264960
md: recovery thread got woken up ...
md2: no spare disk to reconstruct array! -- continuing in degraded mode
md1: no spare disk to reconstruct array! -- continuing in degraded mode
md0: no spare disk to reconstruct array! -- continuing in degraded mode
md: recovery thread finished ...
md: ... autorun DONE.
The only other change I made was to replace the shipped tg3.0 Broadcom Ethernet driver with IBM's own driver.
I am very suspicious that what initially appeared to be a hardware fault is actually somehow related to the 5.6U5 upgrade.
I'd appreciate any thoughts that would assist in resolving this.
Thanks,
Des Dougan