On Fri, Nov 11, 2005 at 02:25:35PM -0800, Rick Stevens wrote: > On Fri, 2005-11-11 at 14:16 -0800, Danny Howard wrote: > > Sweet! I can "fail" a disk and remove it thus: > > mdadm --fail /dev/md0 /dev/sdb1 > > mdadm --fail /dev/md1 /dev/sdb2 > > mdadm --fail /dev/md2 /dev/sdb3 > > [ ... physically remove disk, system is fine ... ] > > [ ... put the disk back in, system is fine ... ] > > mdadm --remove /dev/md0 /dev/sdb1 > > mdadm --add /dev/md0 /dev/sdb1 > > mdadm --remove /dev/md1 /dev/sdb2 > > mdadm --add /dev/md1 /dev/sdb2 > > mdadm --remove /dev/md2 /dev/sdb3 > > mdadm --add /dev/md2 /dev/sdb3 > > [ ... md2 does a rebuild, but /boot and <swap> are fine -- nice! ... ] > > The RAID should go into degraded mode and continue to run. If you > pulled the disk out while the system was powered up AND the system isn't > hot swap-compatible (and most built-in SATA stuff isn't), then you've > confused the SCSI bus badly and I'd be amazed if it worked at all after > that. Your error messages indicate that's the case here. Rick, Given support of the OS, the drive is hot-swap. This works fine with FreeBSD, and this works fine with Linux if I tell md that the disk is failed. If the disk fails without me telling md, then md will react badly. > If, however, the SATA drives were in a hot swap-compatible enclosure and > you see the same problem, then something else is wrong and we'd need to > look at that a bit more closely. What do you suggest? Sincerely, -danny -- http://dannyman.toldme.com/