On Fri, 2006-07-28 at 18:03 -0500, Dan Carl wrote:
I know its a raid 0 is a stripe.
Its my swap partition.
Why would I need fault tolerance on my swap.
From your first post below:
Now I can reach the drive via fdisk but I have made more
problems now (no swap now)
and I'm not sure the steps to rebuild.
That is a good reason to make sure that vital disk partitions are not
made critically weak. When striping across 3 drives the failure
probability is made 3X as likely and any single failure toasts the
entire device.
Since swap can use multiple partitions the likelyhood of failure and
total loss of swap space can be reduced by simply defining multiple swap
partitions without using striping.
Anyway,
I did what Sam suggested.
md0 is fine, md1 doesn't exist
mdadm -Q -D /dev/md2
it yeilded
/dev/md2:
Version : 00.90.01
Creation Time : Mon Feb 14 06:42:28 2005
Raid Level : raid5
Array Size : 34812416 (33.20 GiB 35.65 GB)
Device Size : 17406208 (16.60 GiB 17.82 GB)
Raid Devices : 3
Total Devices : 2
Preferred Minor : 2
Persistence : Superblock is persistent
Update Time : Fri Jul 28 17:56:25 2006
State : clean, degraded
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0
Layout : left-symmetric
Chunk Size : 256K
Number Major Minor RaidDevice State
0 8 3 0 active sync /dev/sda3
1 8 19 1 active sync /dev/sdb3
2 0 0 -1 removed
UUID : b4b161bc:2953b117:9c13c568:47693baa
Events : 0.31307539
So mdadm -a needs to be used to add the 3rd device back to md2. Sam's
instructions were clear on that. For more information and education use
the man page for mdadm.
MANAGE MODE
Usage: mdadm device options... devices...
This usage will allow individual devices in an array to be
failed,
removed or added. It is possible to perform multiple
operations with
one command. For example:
mdadm /dev/md0 -f /dev/hda1 -r /dev/hda1 -a /dev/hda1
will firstly mark /dev/hda1 as faulty in /dev/md0 and will then
remove
it from the array and finally add it back in as a spare.
However only
one md array can be affected by a single command.
I would do the following that you have not already stated done.
1. create the partition(s) on your new /dev/hdc
2. use mdadm as follows to add it to md2
mdadm /dev/md2 -a /dev/sdc3
note that I assume your partitions are created and numbered as you have
already stated.
What if the next step this is my mail server and I really don't have the
time to reload it.
I have my fstab, partition, mdstat, infomation.
I ran this command sfdisk -d > sdb-parts.dump before a added the new
drive.
Will any of this help?
fdisk -l will list the partition information for each drive including
start and end cylinders such as this.
[root@raptor pgsql]# fdisk -l /dev/hda
Disk /dev/hda: 250.0 GB, 250059350016 bytes
255 heads, 63 sectors/track, 30401 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot Start End Blocks Id System
/dev/hda1 * 1 21 168651 83 Linux
/dev/hda2 22 532 4104607+ 83 Linux
/dev/hda3 533 721 1518142+ 82 Linux
swap
/dev/hda4 722 30401 238404600 5 Extended
/dev/hda5 722 1359 5124703+ 83 Linux
....
From that you can get not only the size of each partition, but the
actual cylinders used and can recreate the table on the new drive
appropriately with fdisk.
Like I said before the only raid/partition experience I have is at
initial
installation.
----- Original Message -----
From: "Jeff Vian" <jvian10@xxxxxxxxxxx>
To: "For users of Fedora Core releases" <fedora-list@xxxxxxxxxx>
Sent: Friday, July 28, 2006 5:44 PM
Subject: Re: software raid drive failed, please provide step bu step
torebuild
> On Fri, 2006-07-28 at 16:29 -0500, Dan Carl wrote:
> > I have/had a software raid running and sdc drive failed.
> > I got a replacement drive today and installed it.
> > My only experience with set partitions and raids in during initail
setup.
> > I could not fdisk the new drive because i guess it wasn't reconized
> > so I
> > rebooted.
> > Now I can reach the drive via fdisk but I have made more problems now
(no
> > swap now)
> > and I'm not sure the steps to rebuild.
> > Background:
> > I have a FC3 with a software raid.
> > I have 3 SCSI 18gb hard drives
> > If I recall this how I set it up
> > md0 /boot 100MB raid 1 sda, sdb and sdc as spare
> > md1 /swp 768MB raid 0 sda, sdb, sdc
> This toasted your /swp partition.
> Raid 0 is striping, and a single failure toasts the entire device.
>
> You would have been ahead with a non-raid swap, and had 3 separate
> partitions, one on each device, for swap. Failure of one would not
> have
> toasted all.
>
>
> > md2 / ext3 33GB raid 5 sda, sdb, sdc
> >
> > Can someone please help?
> >
>
> --
> fedora-list mailing list
> fedora-list@xxxxxxxxxx
> To unsubscribe: https://www.redhat.com/mailman/listinfo/fedora-list
--
fedora-list mailing list
fedora-list@xxxxxxxxxx
To unsubscribe: https://www.redhat.com/mailman/listinfo/fedora-list