Re: Failed Drives in Software RAID

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, 2005-01-18 at 13:10 -0800, Shane Archer wrote:
> Hi all,
> 
> I am trying to set up a software RAID5 on FC3, and having a heckuva time.
> 
> The computer is configured like this:
> 
> /boot and / are on a drive that is plugged into the motherboard's SATA0 port.
> Three 160GB WD drives are plugged into a Promise SATA150TX2 controller.
> The motherboard is an Intel SE7520AF2 server board.
> 
> The problem is, whenever I go to create the array, one drive always gets 
> reported as failed. I have done repeated surface scans on the drives (using 
> both fsck and badblocks) and I am quite certain the drives are alright. On 
> top of that, when the array starts building, the system eventually locks 
> up, without fail.
> 
> So the question becomes, where might I best troubleshoot this? Could it be 
> a software problem? A hardware problem? Has anybody had problems with the 
> Promise card in FC3?
> 
> Here's the mdadm command I am using:
> 
> # mdadm -Cv /dev/md0 -l5 -n3 -c128 /dev/sdb /dev/sdc /dev/sdd
> 
> And here's the log after running it:
> 
> Jan 17 21:57:47 xtreme kernel: md: bind<sdb>
> Jan 17 21:57:47 xtreme kernel: md: bind<sdc>
> Jan 17 21:57:47 xtreme kernel: md: bind<sdd>
> Jan 17 21:57:47 xtreme kernel: raid5: automatically using best checksumming 
> function: pIII_sse
> Jan 17 21:57:47 xtreme kernel:    pIII_sse  :  2896.000 MB/sec
> Jan 17 21:57:47 xtreme kernel: raid5: using function: pIII_sse (2896.000 
> MB/sec)
> Jan 17 21:57:47 xtreme kernel: md: raid5 personality registered as nr 4
> Jan 17 21:57:47 xtreme kernel: raid5: device sdc operational as raid disk 1
> Jan 17 21:57:47 xtreme kernel: raid5: device sdb operational as raid disk 0
> Jan 17 21:57:47 xtreme kernel: raid5: allocated 3162kB for md0
> Jan 17 21:57:47 xtreme kernel: raid5: raid level 5 set md0 active with 2 
> out of 3 devices, algorithm 2
> Jan 17 21:57:47 xtreme kernel: RAID5 conf printout:
> Jan 17 21:57:47 xtreme kernel:  --- rd:3 wd:2 fd:1
> Jan 17 21:57:47 xtreme kernel:  disk 0, o:1, dev:sdb
> Jan 17 21:57:47 xtreme kernel:  disk 1, o:1, dev:sdc
> Jan 17 21:57:47 xtreme kernel: RAID5 conf printout:
> Jan 17 21:57:47 xtreme kernel:  --- rd:3 wd:2 fd:1
> Jan 17 21:57:47 xtreme kernel:  disk 0, o:1, dev:sdb
> Jan 17 21:57:47 xtreme kernel:  disk 1, o:1, dev:sdc
> Jan 17 21:57:47 xtreme kernel:  disk 2, o:1, dev:sdd
> Jan 17 21:57:47 xtreme kernel: .<6>md: syncing RAID array md0
> Jan 17 21:57:47 xtreme kernel: md: minimum _guaranteed_ reconstruction 
> speed: 1000 KB/sec/disc.
> Jan 17 21:57:47 xtreme kernel: md: minimum _guaranteed_ reconstruction 
> speed: 1000 KB/sec/disc.
> Jan 17 21:57:47 xtreme kernel: md: using maximum available idle IO bandwith 
> (but not more than 200000 KB/sec)
> for reconstruction.
> Jan 17 21:57:47 xtreme kernel: md: using 128k window, over a total of 
> 156249856 blocks.
> 
> And here's the output from mdadm -E afterwards:
> 
> # mdadm -E /dev/sdb
> /dev/sdb:
>            Magic : a92b4efc
>          Version : 00.90.01
>             UUID : 43e05773:9c581b8a:23c8cdb0:d71025ac
>    Creation Time : Mon Jan 17 21:57:47 2005
>       Raid Level : raid5
>      Device Size : 156249856 (149.01 GiB 159.100 GB)
>     Raid Devices : 3
>    Total Devices : 3
> Preferred Minor : 0
> 
>      Update Time : Mon Jan 17 21:57:47 2005
>            State : clean
>   Active Devices : 2
> Working Devices : 3
>   Failed Devices : 1
>    Spare Devices : 1
>         Checksum : 11683037 - correct
>           Events : 0.1
> 
>           Layout : left-symmetric
>       Chunk Size : 128K
> 
>        Number   Major   Minor   RaidDevice State
> this     0       8       16        0      active sync   /dev/sdb
>     0     0       8       16        0      active sync   /dev/sdb
>     1     1       8       32        1      active sync   /dev/sdc
>     2     2       0        0        2      faulty removed
>     3     3       8       48        3      spare   /dev/sdd
> 
>  From the output above, it would appear that /dev/sdd is the dead drive. 
> However, if I unplug that drive, reboot the system, and run the same mdadm 
> -Cv command with only two drives, it still reports 1 working, 1 failed. 
> So...hardware problem?
> 
> Thanks for any help,
> 
> Shane


I'm pretty sure what you are seeing here is a normal part of the array
initializing. If you run mdadm --detail /dev/md0 does it show the array
is rebuilding?

If so, then that error should go away once the rebuild is complete.

Daniel


[Index of Archives]     [Current Fedora Users]     [Fedora Desktop]     [Fedora SELinux]     [Yosemite News]     [Yosemite Photos]     [KDE Users]     [Fedora Tools]     [Fedora Docs]

  Powered by Linux