Re: 2.6.17-mm4

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 29/06/2006 8:36 p.m., Andrew Morton wrote:
ftp://ftp.kernel.org/pub/linux/kernel/people/akpm/patches/2.6/2.6.17/2.6.17-mm4/

- The RAID patches have been dropped due to testing failures in -mm3.

Things are still not perfect on the raid front, but better than -mm3.  The raid
oopsing on boot is gone, but my RAID-1 arrays seem to be coming up degraded for no obvious reason. The system was shut down fine before booting -mm3 and comes up OK after reverting to the older -mm releases (eg -mm1) and the supposedly faulty partitions readded back into the array.

[root@tornado ~]# cat /proc/mdstat
Personalities : [raid1]
md1 : active raid1 sdc3[2](F) sda3[0]
      4891712 blocks [2/1] [U_]
      bitmap: 13/150 pages [52KB], 16KB chunk

md2 : active raid1 sdc5[2](F) sda5[0]
      4891648 blocks [2/1] [U_]
      bitmap: 36/150 pages [144KB], 16KB chunk

md3 : active raid1 sdc6[2](F) sda6[0]
      104320 blocks [2/1] [U_]
      bitmap: 1/13 pages [4KB], 4KB chunk

md4 : active raid1 sdc7[1] sda7[2](F)
      497856 blocks [2/1] [_U]
      bitmap: 11/61 pages [44KB], 4KB chunk

md6 : active raid1 sdc10[1] sda10[0]
      20008832 blocks [2/2] [UU]

md5 : active raid1 sdc11[2](F) sda11[0]
      20008832 blocks [2/1] [U_]

md0 : active raid1 sdc2[2](F) sda2[0]
      24410688 blocks [2/1] [U_]
      bitmap: 9/187 pages [36KB], 64KB chunk

unused devices: <none>
[root@tornado ~]#

From dmesg:

md: Autodetecting RAID arrays.
md: autorun ...
md: considering sdc11 ...
md:  adding sdc11 ...
md: sdc10 has different UUID to sdc11
md: sdc7 has different UUID to sdc11
md: sdc6 has different UUID to sdc11
md: sdc5 has different UUID to sdc11
md: sdc3 has different UUID to sdc11
md: sdc2 has different UUID to sdc11
md:  adding sda11 ...
md: sda10 has different UUID to sdc11
md: sda7 has different UUID to sdc11
md: sda6 has different UUID to sdc11
md: sda5 has different UUID to sdc11
md: sda3 has different UUID to sdc11
md: sda2 has different UUID to sdc11
md: created md5
md: bind<sda11>
md: bind<sdc11>
md: running: <sdc11><sda11>
raid1: raid set md5 active with 2 out of 2 mirrors
md: considering sdc10 ...
md:  adding sdc10 ...
md: sdc7 has different UUID to sdc10
md: sdc6 has different UUID to sdc10
md: sdc5 has different UUID to sdc10
md: sdc3 has different UUID to sdc10
md: sdc2 has different UUID to sdc10
md:  adding sda10 ...
md: sda7 has different UUID to sdc10
md: sda6 has different UUID to sdc10
md: sda5 has different UUID to sdc10
md: sda3 has different UUID to sdc10
md: sda2 has different UUID to sdc10
md: created md6
md: bind<sda10>
md: bind<sdc10>
md: running: <sdc10><sda10>
raid1: raid set md6 active with 2 out of 2 mirrors
md: considering sdc7 ...
md:  adding sdc7 ...
md: sdc6 has different UUID to sdc7
md: sdc5 has different UUID to sdc7
md: sdc3 has different UUID to sdc7
md: sdc2 has different UUID to sdc7
md:  adding sda7 ...
md: sda6 has different UUID to sdc7
md: sda5 has different UUID to sdc7
md: sda3 has different UUID to sdc7
md: sda2 has different UUID to sdc7
md: created md4
md: bind<sda7>
md: bind<sdc7>
md: running: <sdc7><sda7>
raid1: raid set md4 active with 2 out of 2 mirrors
md4: bitmap initialized from disk: read 4/4 pages, set 7 bits, status: 0
created bitmap (61 pages) for device md4
raid1: Disk failure on sda7, disabling device.
        Operation continuing on 1 devices
md: considering sdc6 ...
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:1, o:0, dev:sda7
 disk 1, wo:0, o:1, dev:sdc7
md:  adding sdc6 ...
md: sdc5 has different UUID to sdc6
md: sdc3 has different UUID to sdc6
md: sdc2 has different UUID to sdc6
md:  adding sda6 ...
md: sda5 has different UUID to sdc6
RAID1 conf printout:
 --- wd:1 rd:2
 disk 1, wo:0, o:1, dev:sdc7
md: sda3 has different UUID to sdc6
md: sda2 has different UUID to sdc6
md: created md3
md: bind<sda6>
md: bind<sdc6>
md: running: <sdc6><sda6>
raid1: raid set md3 active with 2 out of 2 mirrors
md3: bitmap initialized from disk: read 1/1 pages, set 2 bits, status: 0
created bitmap (13 pages) for device md3
raid1: Disk failure on sdc6, disabling device.
        Operation continuing on 1 devices
md: considering sdc5 ...
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda6
 disk 1, wo:1, o:0, dev:sdc6
md:  adding sdc5 ...
md: sdc3 has different UUID to sdc5
md: sdc2 has different UUID to sdc5
md:  adding sda5 ...
md: sda3 has different UUID to sdc5
md: sda2 has different UUID to sdc5
md: created md2
md: bind<sda5>
md: bind<sdc5>
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda6
md: running: <sdc5><sda5>
raid1: raid set md2 active with 2 out of 2 mirrors
md2: bitmap initialized from disk: read 10/10 pages, set 80 bits, status: 0
created bitmap (150 pages) for device md2
raid1: Disk failure on sdc5, disabling device.
        Operation continuing on 1 devices
md: considering sdc3 ...
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda5
 disk 1, wo:1, o:0, dev:sdc5
md:  adding sdc3 ...
md: sdc2 has different UUID to sdc3
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda5
md:  adding sda3 ...
md: sda2 has different UUID to sdc3
md: created md1
md: bind<sda3>
md: bind<sdc3>
md: running: <sdc3><sda3>
raid1: raid set md1 active with 2 out of 2 mirrors
md1: bitmap initialized from disk: read 10/10 pages, set 55 bits, status: 0
created bitmap (150 pages) for device md1
raid1: Disk failure on sdc3, disabling device.
        Operation continuing on 1 devices
md: considering sdc2 ...
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda3
 disk 1, wo:1, o:0, dev:sdc3
md:  adding sdc2 ...
md:  adding sda2 ...
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda3
md: created md0
md: bind<sda2>
md: bind<sdc2>
md: running: <sdc2><sda2>
raid1: raid set md0 active with 2 out of 2 mirrors
md0: bitmap initialized from disk: read 12/12 pages, set 37 bits, status: 0
created bitmap (187 pages) for device md0
raid1: Disk failure on sdc2, disabling device.
        Operation continuing on 1 devices
md: ... autorun DONE.
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda2
 disk 1, wo:1, o:0, dev:sdc2
RAID1 conf printout:
 --- wd:1 rd:2
 disk 0, wo:0, o:1, dev:sda2
kjournald starting.  Commit interval 5 seconds
EXT3-fs: mounted filesystem with ordered data mode.


Here's one of the arrays:

[root@tornado ~]# mdadm --examine /dev/sda5
/dev/sda5:
          Magic : a92b4efc
        Version : 00.90.01
           UUID : d18ac80d:63c7bd9d:8cbb8566:6a378891
  Creation Time : Sun Sep  4 18:25:28 2005
     Raid Level : raid1
    Device Size : 4891648 (4.67 GiB 5.01 GB)
     Array Size : 4891648 (4.67 GiB 5.01 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 2

    Update Time : Thu Jun 29 23:34:42 2006
          State : clean
Internal Bitmap : present
 Active Devices : 1
Working Devices : 1
 Failed Devices : 1
  Spare Devices : 0
       Checksum : 5f63b753 - correct
         Events : 0.16052320


      Number   Major   Minor   RaidDevice State
this     0       8        5        0      active sync   /dev/sda5

   0     0       8        5        0      active sync   /dev/sda5
   1     1       0        0        1      faulty removed
[root@tornado ~]# mdadm --examine /dev/sdc5
/dev/sdc5:
          Magic : a92b4efc
        Version : 00.90.01
           UUID : d18ac80d:63c7bd9d:8cbb8566:6a378891
  Creation Time : Sun Sep  4 18:25:28 2005
     Raid Level : raid1
    Device Size : 4891648 (4.67 GiB 5.01 GB)
     Array Size : 4891648 (4.67 GiB 5.01 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 2

    Update Time : Thu Jun 29 23:08:32 2006
          State : clean
Internal Bitmap : present
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 5f63b148 - correct
         Events : 0.16052318


      Number   Major   Minor   RaidDevice State
this     1       8       37        1      active sync   /dev/sdc5

   0     0       8        5        0      active sync   /dev/sda5
   1     1       8       37        1      active sync   /dev/sdc5
[root@tornado ~]#

Reuben

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux