linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* upgrade to jessie/newer kernel and mdadm problems
@ 2015-05-01  9:40 Jérôme Tytgat
  2015-05-04 12:42 ` Phil Turmel
  0 siblings, 1 reply; 3+ messages in thread
From: Jérôme Tytgat @ 2015-05-01  9:40 UTC (permalink / raw)
  To: linux-raid

Hello list,

Sorry for the long post, but I wanted to be as much informative as I can be.
Forgive my lack of knowlegde in mdadm, I know how to create it using the 
debian installer and few things to get information but that's all.
Forgive also my english, I'm not a native in this language.

My system has been installed in 2007, and I've upgraded it several times 
until this week to Debian Jessie (the latest version).

So, I've upgraded my system to jessie today (only partially with apt-get 
upgrade + kernel upgrade) and I faced with a problem with my RAID 1 soft.

I have two disk (/dev/sdb and /dev/sdc) which are members of the raid array
There's 10 partitions on these disks, each one is a array (ie sdb1 and 
sdc1).
All of these form my raids array : md0 to md9

Today , one of my md partition was missing.
Before upgrade I had partitions from md0 to md9, after reboot I'm 
missing md9.

my mdadm.conf before reboot looked like that:
 >-------------------------------------------------------------------------------------------------------------<

    # mdadm.conf
    #
    # Please refer to mdadm.conf(5) for information about this file.
    #

    # by default, scan all partitions (/proc/partitions) for MD superblocks.
    # alternatively, specify devices to scan, using wildcards if desired.
    DEVICE partitions

    # auto-create devices with Debian standard permissions
    CREATE owner=root group=disk mode=0660 auto=yes

    # automatically tag new arrays as belonging to the local system
    HOMEHOST <system>

    # instruct the monitoring daemon where to send mail alerts
    MAILADDR root

    # definitions of existing MD arrays
    ARRAY /dev/md0 level=raid1 num-devices=2
    UUID=350e253f:863b7b04:b1617c47:b213a024
    ARRAY /dev/md1 level=raid1 num-devices=2
    UUID=086e68ed:3607317f:60b56e23:6bae62bc
    ARRAY /dev/md2 level=raid1 num-devices=2
    UUID=0f6e3ed5:aeee975a:c3647deb:763d68ce
    ARRAY /dev/md3 level=raid1 num-devices=2
    UUID=6b560fe8:f24d6f2e:8942bd3a:1903abbc
    ARRAY /dev/md4 level=raid1 num-devices=2
    UUID=462dab96:b7ca2a17:7c4aebf1:d4d7ec3b
    ARRAY /dev/md5 level=raid1 num-devices=2
    UUID=ea511351:3abc7b12:4c81e838:93dbd21a
    ARRAY /dev/md6 level=raid1 num-devices=2
    UUID=da0d76c6:91422584:dc3d6162:37ced53b
    ARRAY /dev/md7 level=raid1 num-devices=2
    UUID=387c831c:8a6d05e3:b649696c:0870b930
    ARRAY /dev/md8 level=raid1 num-devices=2
    UUID=b07c4ab4:39d0ba53:9913afa9:fd9cc323
    ARRAY /dev/md9 level=raid1 num-devices=2
    UUID=36c4edd0:a0492cc9:0cd2fce0:2745e358

    # This file was auto-generated on Wed, 07 Feb 2007 17:15:33 +0000
    # by mkconf $Id: mkconf 261 2006-11-09 13:32:35Z madduck $

 >-------------------------------------------------------------------------------------------------------------<

After reboot and after a mdadm -Es, I got this:
 >-------------------------------------------------------------------------------------------------------------<

    # mdadm.conf
    #
    # Please refer to mdadm.conf(5) for information about this file.
    #

    # by default, scan all partitions (/proc/partitions) for MD superblocks.
    # alternatively, specify devices to scan, using wildcards if desired.
    DEVICE partitions

    # auto-create devices with Debian standard permissions
    CREATE owner=root group=disk mode=0660 auto=yes

    # automatically tag new arrays as belonging to the local system
    HOMEHOST <system>

    # instruct the monitoring daemon where to send mail alerts
    MAILADDR root

    # definitions of existing MD arrays
    ARRAY metadata=imsm UUID=d9cfa6d9:2a715e4f:1fbc2095:be342429
    ARRAY /dev/md/raidlin container=d9cfa6d9:2a715e4f:1fbc2095:be342429
    member=0 UUID=91449a9d:9242bfe9:d99bceb0:a59f9314
    ARRAY /dev/md0 UUID=350e253f:863b7b04:b1617c47:b213a024
    ARRAY /dev/md1 UUID=086e68ed:3607317f:60b56e23:6bae62bc
    ARRAY /dev/md2 UUID=0f6e3ed5:aeee975a:c3647deb:763d68ce
    ARRAY /dev/md3 UUID=6b560fe8:f24d6f2e:8942bd3a:1903abbc
    ARRAY /dev/md4 UUID=462dab96:b7ca2a17:7c4aebf1:d4d7ec3b
    ARRAY /dev/md5 UUID=ea511351:3abc7b12:4c81e838:93dbd21a
    ARRAY /dev/md6 UUID=da0d76c6:91422584:dc3d6162:37ced53b
    ARRAY /dev/md7 UUID=387c831c:8a6d05e3:b649696c:0870b930
    ARRAY /dev/md8 UUID=b07c4ab4:39d0ba53:9913afa9:fd9cc323

 >-------------------------------------------------------------------------------------------------------------<

This is what is looking my fdisk extract and /proc/mdstat:
 >-------------------------------------------------------------------------------------------------------------<

    # fdisk -l /dev/sdb

    Disk /dev/sdb: 250.1 GB, 250059350016 bytes
    255 heads, 63 sectors/track, 30401 cylinders, total 488397168 sectors
    Units = sectors of 1 * 512 = 512 bytes
    Sector size (logical/physical): 512 bytes / 512 bytes
    I/O size (minimum/optimal): 512 bytes / 512 bytes
    Disk identifier: 0x0001edee

        Device Boot      Start         End      Blocks   Id  System
    /dev/sdb1   *          63      979964      489951   fd  Linux raid
    autodetect
    /dev/sdb2          979965    10747484     4883760   fd  Linux raid
    autodetect
    /dev/sdb3        10747485    14651279     1951897+  fd  Linux raid
    autodetect
    /dev/sdb4        14651280   488392064   236870392+   5  Extended
    /dev/sdb5        14651343    24418799     4883728+  fd  Linux raid
    autodetect
    /dev/sdb6        24418863    43953839     9767488+  fd  Linux raid
    autodetect
    /dev/sdb7        43953903    53721359     4883728+  fd  Linux raid
    autodetect
    /dev/sdb8        53721423    63488879     4883728+  fd  Linux raid
    autodetect
    /dev/sdb9        63488943    73256399     4883728+  fd  Linux raid
    autodetect
    /dev/sdb10       73256463    83023919     4883728+  fd  Linux raid
    autodetect
    /dev/sdb11       83023983   488392064   202684041   fd  Linux raid
    autodetect

 >-------------------------------------------------------------------------------------------------------------<

    # fdisk -l /dev/sdc

    Disk /dev/sdc: 250.1 GB, 250059350016 bytes
    255 heads, 63 sectors/track, 30401 cylinders, total 488397168 sectors
    Units = sectors of 1 * 512 = 512 bytes
    Sector size (logical/physical): 512 bytes / 512 bytes
    I/O size (minimum/optimal): 512 bytes / 512 bytes
    Disk identifier: 0x000c352c

        Device Boot      Start         End      Blocks   Id  System
    /dev/sdc1   *          63      979964      489951   fd  Linux raid
    autodetect
    /dev/sdc2          979965    10747484     4883760   fd  Linux raid
    autodetect
    /dev/sdc3        10747485    14651279     1951897+  fd  Linux raid
    autodetect
    /dev/sdc4        14651280   488392064   236870392+   5  Extended
    /dev/sdc5        14651343    24418799     4883728+  fd  Linux raid
    autodetect
    /dev/sdc6        24418863    43953839     9767488+  fd  Linux raid
    autodetect
    /dev/sdc7        43953903    53721359     4883728+  fd  Linux raid
    autodetect
    /dev/sdc8        53721423    63488879     4883728+  fd  Linux raid
    autodetect
    /dev/sdc9        63488943    73256399     4883728+  fd  Linux raid
    autodetect
    /dev/sdc10       73256463    83023919     4883728+  fd  Linux raid
    autodetect
    /dev/sdc11       83023983   488392064   202684041   fd  Linux raid
    autodetect

 >-------------------------------------------------------------------------------------------------------------<

    # cat /proc/mdstat
    Personalities : [raid1]
    md8 : active raid1 md126p10[0]
           4883648 blocks [2/1] [U_]

    md7 : active raid1 md126p9[0]
           4883648 blocks [2/1] [U_]

    md6 : active raid1 md126p8[0]
           4883648 blocks [2/1] [U_]

    md5 : active raid1 md126p7[0]
           4883648 blocks [2/1] [U_]

    md4 : active raid1 md126p6[0]
           9767424 blocks [2/1] [U_]

    md3 : active raid1 md126p5[0]
           4883648 blocks [2/1] [U_]

    md2 : active (auto-read-only) raid1 md126p3[0]
           1951808 blocks [2/1] [U_]

    md1 : active raid1 md126p2[0]
           4883648 blocks [2/1] [U_]

    md0 : active raid1 md126p1[0]
           489856 blocks [2/1] [U_]

    md126 : active raid1 sdc[1] sdb[0]
           244195328 blocks super external:/md127/0 [2/2] [UU]

    md127 : inactive sdc[1](S) sdb[0](S)
           6306 blocks super external:imsm

    unused devices: <none>

 >-------------------------------------------------------------------------------------------------------------<


I'm not an expert at md at all, so I called a friend and we were able to 
find my lost md9 partition, it was md126p11.

This is what we saw in dmesg:
 >-------------------------------------------------------------------------------------------------------------<

    [    1.593297] md: bind<sdb>
    [    1.598582] md: bind<sdc>
    [    1.599902] md: bind<sdb>
    [    1.600045] md: bind<sdc>
    [    1.606550] md: raid1 personality registered for level 1
    [    1.607396] md/raid1:md126: active with 2 out of 2 mirrors
    [    1.607433] md126: detected capacity change from 0 to 250056015872
    [    1.632396] md: md126 switched to read-write mode.
    [    1.669910]  md126: p1 p2 p3 p4 < p5 p6 p7 p8 p9 p10 p11 >
    [    1.670501] md126: p11 size 405368082 extends beyond EOD, truncated
    [    4.100396] md: md0 stopped.
    [    4.100919] md: bind<md126p1>
    [    4.101708] md/raid1:md0: active with 1 out of 2 mirrors
    [    4.101734] md0: detected capacity change from 0 to 501612544
    [    4.102179]  md0: unknown partition table
    [    4.224625] md: md1 stopped.
    [    4.225109] md: bind<md126p2>
    [    4.225886] md/raid1:md1: active with 1 out of 2 mirrors
    [    4.225911] md1: detected capacity change from 0 to 5000855552
    [    4.226658]  md1: unknown partition table
    [    4.420746] md: md2 stopped.
    [    4.421441] md: bind<md126p3>
    [    4.422216] md/raid1:md2: active with 1 out of 2 mirrors
    [    4.422241] md2: detected capacity change from 0 to 1998651392
    [    4.422677]  md2: unknown partition table
    [    4.595729] md: md3 stopped.
    [    4.596410] md: bind<md126p5>
    [    4.597189] md/raid1:md3: active with 1 out of 2 mirrors
    [    4.597215] md3: detected capacity change from 0 to 5000855552
    [    4.597638]  md3: unknown partition table
    [    4.668224] md: md4 stopped.
    [    4.668693] md: bind<md126p6>
    [    4.669446] md/raid1:md4: active with 1 out of 2 mirrors
    [    4.669474] md4: detected capacity change from 0 to 10001842176
    [    4.669909]  md4: unknown partition table
    [    4.783732] md: md5 stopped.
    [    4.784236] md: bind<md126p7>
    [    4.785024] md/raid1:md5: active with 1 out of 2 mirrors
    [    4.785049] md5: detected capacity change from 0 to 5000855552
    [    4.785479]  md5: unknown partition table
    [    4.970769] md: md6 stopped.
    [    4.971366] md: bind<md126p8>
    [    4.972129] md/raid1:md6: active with 1 out of 2 mirrors
    [    4.972158] md6: detected capacity change from 0 to 5000855552
    [    4.972594]  md6: unknown partition table
    [    5.137394] md: md7 stopped.
    [    5.138011] md: bind<md126p9>
    [    5.138754] md/raid1:md7: active with 1 out of 2 mirrors
    [    5.138779] md7: detected capacity change from 0 to 5000855552
    [    5.139232]  md7: unknown partition table
    [    5.329093] md: md8 stopped.
    [    5.330228] md: bind<md126p10>
    [    5.330977] md/raid1:md8: active with 1 out of 2 mirrors
    [    5.331003] md8: detected capacity change from 0 to 5000855552
    [    5.350896]  md8: unknown partition table

 >-------------------------------------------------------------------------------------------------------------<

note the line "[ 1.670501] md126: p11 size 405368082 extends beyond EOD, 
truncated"

After force mounting /dev/md126p11, I noted these messages in dmesg:
 >-------------------------------------------------------------------------------------------------------------<

    [ 4361.248369] md126p11: rw=32, want=405367936, limit=405366673
    [ 4361.248382] XFS (md126p11): Mounting V4 Filesystem
    [ 4361.657842] XFS (md126p11): Ending clean mount
    [ 4609.295281] md126p11: rw=32, want=405367936, limit=405366673
    [ 4609.295300] XFS (md126p11): Mounting V4 Filesystem
    [ 4609.513109] XFS (md126p11): Ending clean mount

 >-------------------------------------------------------------------------------------------------------------<


and in daemon.log (lines correspond to the reboot after the upgrade, the 
time I lost my md9):
 >-------------------------------------------------------------------------------------------------------------<

    Apr 29 20:26:26 shax mdadm[4735]: DeviceDisappeared event detected
    on md device /dev/md9
    Apr 29 20:26:29 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md8
    Apr 29 20:26:30 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md7
    Apr 29 20:26:31 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md6
    Apr 29 20:26:33 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md5
    Apr 29 20:26:34 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md4
    Apr 29 20:26:35 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md3
    Apr 29 20:26:35 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md2
    Apr 29 20:26:36 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md1
    Apr 29 20:26:37 shax mdadm[4735]: DegradedArray event detected on md
    device /dev/md0
    Apr 29 20:26:37 shax mdadm[4735]: NewArray event detected on md
    device /dev/md127
    Apr 29 20:26:37 shax mdadm[4735]: NewArray event detected on md
    device /dev/md126

 >-------------------------------------------------------------------------------------------------------------<

output of fdisk -l /dev/md126 :

 >-------------------------------------------------------------------------------------------------------------<

    fdisk -l /dev/md126

    Disk /dev/md126: 250.1 GB, 250056015872 bytes
    255 heads, 63 sectors/track, 30400 cylinders, total 488390656 sectors
    Units = sectors of 1 * 512 = 512 bytes
    Sector size (logical/physical): 512 bytes / 512 bytes
    I/O size (minimum/optimal): 512 bytes / 512 bytes
    Disk identifier: 0x000c352c

           Device Boot      Start         End      Blocks   Id System
    /dev/md126p1   *          63      979964      489951   fd  Linux
    raid autodetect
    /dev/md126p2          979965    10747484     4883760   fd  Linux
    raid autodetect
    /dev/md126p3        10747485    14651279     1951897+  fd  Linux
    raid autodetect
    /dev/md126p4        14651280   488392064   236870392+   5 Extended
    /dev/md126p5        14651343    24418799     4883728+  fd  Linux
    raid autodetect
    /dev/md126p6        24418863    43953839     9767488+  fd  Linux
    raid autodetect
    /dev/md126p7        43953903    53721359     4883728+  fd  Linux
    raid autodetect
    /dev/md126p8        53721423    63488879     4883728+  fd  Linux
    raid autodetect
    /dev/md126p9        63488943    73256399     4883728+  fd  Linux
    raid autodetect
    /dev/md126p10       73256463    83023919     4883728+  fd  Linux
    raid autodetect
    /dev/md126p11       83023983   488392064   202684041   fd  Linux
    raid autodetect

 >-------------------------------------------------------------------------------------------------------------<

mdadm -E /dev/sdbX ou mdadm -E /dev/sdcX doesn't give anything :
 >-------------------------------------------------------------------------------------------------------------<

    # mdadm -E /dev/sdb1
    mdadm: cannot open /dev/sdb1: No such device or address

    # mdadm -E /dev/sdb2
    mdadm: cannot open /dev/sdb2: No such device or address

    # mdadm -E /dev/sdb3
    mdadm: cannot open /dev/sdb3: No such device or address

    # mdadm -E /dev/sdb4
    mdadm: cannot open /dev/sdb4: No such device or address

    # mdadm -E /dev/sdb5
    mdadm: cannot open /dev/sdb5: No such device or address

    # mdadm -E /dev/sdb6
    mdadm: cannot open /dev/sdb6: No such device or address

    # mdadm -E /dev/sdb7
    mdadm: cannot open /dev/sdb7: No such device or address

    # mdadm -E /dev/sdb8
    mdadm: cannot open /dev/sdb8: No such device or address

    # mdadm -E /dev/sdb9
    mdadm: cannot open /dev/sdb9: No such device or address

    # mdadm -E /dev/sdb10
    mdadm: cannot open /dev/sdb10: No such device or address

    # mdadm -E /dev/sdb11
    mdadm: cannot open /dev/sdb11: No such device or address

 >-------------------------------------------------------------------------------------------------------------<

mdadm -E /dev/md126pX do give something (except for md126p11):
 >-------------------------------------------------------------------------------------------------------------<


    # mdadm -E /dev/md126p1
    /dev/md126p1:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : 350e253f:863b7b04:b1617c47:b213a024
       Creation Time : Mon May  7 20:29:35 2007
          Raid Level : raid1
       Used Dev Size : 489856 (478.46 MiB 501.61 MB)
          Array Size : 489856 (478.46 MiB 501.61 MB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 0

         Update Time : Thu Apr 30 08:00:08 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : 6372ca75 - correct
              Events : 1769


           Number   Major   Minor   RaidDevice State
    this     0     259        0        0      active sync
      /dev/md/raidlin_0p1

        0     0     259        0        0      active sync
      /dev/md/raidlin_0p1
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p2
    /dev/md126p2:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : 086e68ed:3607317f:60b56e23:6bae62bc
       Creation Time : Mon May  7 20:29:41 2007
          Raid Level : raid1
       Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
          Array Size : 4883648 (4.66 GiB 5.00 GB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 1

         Update Time : Thu Apr 30 20:57:21 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : 4fd14bc9 - correct
              Events : 3980


           Number   Major   Minor   RaidDevice State
    this     0     259        1        0      active sync
      /dev/md/raidlin_0p2

        0     0     259        1        0      active sync
      /dev/md/raidlin_0p2
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p3
    /dev/md126p3:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : 0f6e3ed5:aeee975a:c3647deb:763d68ce
       Creation Time : Mon May  7 20:29:48 2007
          Raid Level : raid1
       Used Dev Size : 1951808 (1906.38 MiB 1998.65 MB)
          Array Size : 1951808 (1906.38 MiB 1998.65 MB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 2

         Update Time : Thu Apr 30 18:27:58 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : 3cc9ac24 - correct
              Events : 1703


           Number   Major   Minor   RaidDevice State
    this     0     259        2        0      active sync
      /dev/md/raidlin_0p3

        0     0     259        2        0      active sync
      /dev/md/raidlin_0p3
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p4
    /dev/md126p4:
        MBR Magic : aa55
    Partition[0] :      9767457 sectors at           63 (type fd)
    Partition[1] :     19535040 sectors at      9767520 (type 05)

    # mdadm -E /dev/md126p5
    /dev/md126p5:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : 6b560fe8:f24d6f2e:8942bd3a:1903abbc
       Creation Time : Mon May  7 20:29:55 2007
          Raid Level : raid1
       Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
          Array Size : 4883648 (4.66 GiB 5.00 GB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 3

         Update Time : Thu Apr 30 20:57:18 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : 44e1e6e6 - correct
              Events : 7856


           Number   Major   Minor   RaidDevice State
    this     0     259        4        0      active sync
      /dev/md/raidlin_0p5

        0     0     259        4        0      active sync
      /dev/md/raidlin_0p5
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p6
    /dev/md126p6:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : 462dab96:b7ca2a17:7c4aebf1:d4d7ec3b
       Creation Time : Mon May  7 20:30:02 2007
          Raid Level : raid1
       Used Dev Size : 9767424 (9.31 GiB 10.00 GB)
          Array Size : 9767424 (9.31 GiB 10.00 GB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 4

         Update Time : Thu Apr 30 20:57:14 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : 945d0c07 - correct
              Events : 2999


           Number   Major   Minor   RaidDevice State
    this     0     259        5        0      active sync
      /dev/md/raidlin_0p6

        0     0     259        5        0      active sync
      /dev/md/raidlin_0p6
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p7
    /dev/md126p7:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : ea511351:3abc7b12:4c81e838:93dbd21a
       Creation Time : Mon May  7 20:30:09 2007
          Raid Level : raid1
       Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
          Array Size : 4883648 (4.66 GiB 5.00 GB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 5

         Update Time : Thu Apr 30 20:57:18 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : 4a634da7 - correct
              Events : 8626


           Number   Major   Minor   RaidDevice State
    this     0     259        6        0      active sync
      /dev/md/raidlin_0p7

        0     0     259        6        0      active sync
      /dev/md/raidlin_0p7
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p8
    /dev/md126p8:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : da0d76c6:91422584:dc3d6162:37ced53b
       Creation Time : Mon May  7 20:30:15 2007
          Raid Level : raid1
       Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
          Array Size : 4883648 (4.66 GiB 5.00 GB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 6

         Update Time : Thu Apr 30 20:57:31 2015
               State : active
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : c4540a0c - correct
              Events : 30082


           Number   Major   Minor   RaidDevice State
    this     0     259        7        0      active sync
      /dev/md/raidlin_0p8

        0     0     259        7        0      active sync
      /dev/md/raidlin_0p8
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p9
    /dev/md126p9:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : 387c831c:8a6d05e3:b649696c:0870b930
       Creation Time : Mon May  7 20:30:21 2007
          Raid Level : raid1
       Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
          Array Size : 4883648 (4.66 GiB 5.00 GB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 7

         Update Time : Thu Apr 30 20:57:10 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : c69b931a - correct
              Events : 4852


           Number   Major   Minor   RaidDevice State
    this     0     259        8        0      active sync
      /dev/md/raidlin_0p9

        0     0     259        8        0      active sync
      /dev/md/raidlin_0p9
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p10
    /dev/md126p10:
               Magic : a92b4efc
             Version : 0.90.00
                UUID : b07c4ab4:39d0ba53:9913afa9:fd9cc323
       Creation Time : Mon May  7 20:30:28 2007
          Raid Level : raid1
       Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
          Array Size : 4883648 (4.66 GiB 5.00 GB)
        Raid Devices : 2
       Total Devices : 1
    Preferred Minor : 8

         Update Time : Thu Apr 30 20:57:28 2015
               State : clean
      Active Devices : 1
    Working Devices : 1
      Failed Devices : 1
       Spare Devices : 0
            Checksum : c5f5d015 - correct
              Events : 19271


           Number   Major   Minor   RaidDevice State
    this     0     259        9        0      active sync
      /dev/md/raidlin_0p10

        0     0     259        9        0      active sync
      /dev/md/raidlin_0p10
        1     1       0        0        1      faulty removed

    # mdadm -E /dev/md126p11
    mdadm: No md superblock detected on /dev/md126p11. 

 >-------------------------------------------------------------------------------------------------------------<

Some more commands:
 >-------------------------------------------------------------------------------------------------------------<

    # mdadm -D /dev/md126
    /dev/md126:
           Container : /dev/md/imsm0, member 0
          Raid Level : raid1
          Array Size : 244195328 (232.88 GiB 250.06 GB)
       Used Dev Size : 244195328 (232.88 GiB 250.06 GB)
        Raid Devices : 2
       Total Devices : 2

               State : active
      Active Devices : 2
    Working Devices : 2
      Failed Devices : 0
       Spare Devices : 0


                UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
         Number   Major   Minor   RaidDevice State
            1       8       32        0      active sync   /dev/sdc
            0       8       16        1      active sync   /dev/sdb

 >-------------------------------------------------------------------------------------------------------------<

    # mdadm -E /dev/sdb
    mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
    mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
    /dev/sdb:
               Magic : Intel Raid ISM Cfg Sig.
             Version : 1.1.00
         Orig Family : 26b5a9e0
              Family : 26b5a9e0
          Generation : 00004db7
          Attributes : All supported
                UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
            Checksum : 261d2aed correct
         MPB Sectors : 1
               Disks : 2
        RAID Devices : 1

       Disk01 Serial : VFC100R10BE79D
               State : active
                  Id : 00010000
         Usable Size : 488390862 (232.88 GiB 250.06 GB)

    [raidlin]:
                UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
          RAID Level : 1
             Members : 2
               Slots : [UU]
         Failed disk : none
           This Slot : 1
          Array Size : 488390656 (232.88 GiB 250.06 GB)
        Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
       Sector Offset : 0
         Num Stripes : 1907776
          Chunk Size : 64 KiB
            Reserved : 0
       Migrate State : idle
           Map State : normal
         Dirty State : dirty

       Disk00 Serial : VFC100R10BRKMD
               State : active
                  Id : 00000000
         Usable Size : 488390862 (232.88 GiB 250.06 GB)

 >-------------------------------------------------------------------------------------------------------------<

    # mdadm -E /dev/sdc
    mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
    mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
    /dev/sdc:
               Magic : Intel Raid ISM Cfg Sig.
             Version : 1.1.00
         Orig Family : 26b5a9e0
              Family : 26b5a9e0
          Generation : 00004dbc
          Attributes : All supported
                UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
            Checksum : 261c2af2 correct
         MPB Sectors : 1
               Disks : 2
        RAID Devices : 1

       Disk00 Serial : VFC100R10BRKMD
               State : active
                  Id : 00000000
         Usable Size : 488390862 (232.88 GiB 250.06 GB)

    [raidlin]:
                UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
          RAID Level : 1
             Members : 2
               Slots : [UU]
         Failed disk : none
           This Slot : 0
          Array Size : 488390656 (232.88 GiB 250.06 GB)
        Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
       Sector Offset : 0
         Num Stripes : 1907776
          Chunk Size : 64 KiB
            Reserved : 0
       Migrate State : idle
           Map State : normal
         Dirty State : clean

       Disk01 Serial : VFC100R10BE79D
               State : active
                  Id : 00010000
         Usable Size : 488390862 (232.88 GiB 250.06 GB)

 >-------------------------------------------------------------------------------------------------------------<

I see something wicked in there and I don't know how to correct it...
I see at least two problems :
- mdadm is not able to detect md9
- my array does not seems fully operationa

Can you help ?

If you need more logs / command output, tell me which one.
I hope I'm not too much confusing...

kernel : Linux shax 3.16.0-4-686-pae #1 SMP Debian 3.16.7-ckt9-3~deb8u1 
(2015-04-24) i686 GNU/Linux





^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: upgrade to jessie/newer kernel and mdadm problems
  2015-05-01  9:40 upgrade to jessie/newer kernel and mdadm problems Jérôme Tytgat
@ 2015-05-04 12:42 ` Phil Turmel
  2015-05-04 13:07   ` Jérôme Tytgat
  0 siblings, 1 reply; 3+ messages in thread
From: Phil Turmel @ 2015-05-04 12:42 UTC (permalink / raw)
  To: Jérôme Tytgat, linux-raid

Good morning Jérôme,

On 05/01/2015 05:40 AM, Jérôme Tytgat wrote:
> Hello list,
> 
> Sorry for the long post, but I wanted to be as much informative as I can
> be.

I was leaving your case for people who know IMSM to pipe up, as I don't
have any experience with it.  But the silence is deafening :-(

> Forgive my lack of knowlegde in mdadm, I know how to create it using the
> debian installer and few things to get information but that's all.
> Forgive also my english, I'm not a native in this language.

It's ok.  Your report was thorough.

> My system has been installed in 2007, and I've upgraded it several times
> until this week to Debian Jessie (the latest version).

That explains the use of v0.90 metadata.

> So, I've upgraded my system to jessie today (only partially with apt-get
> upgrade + kernel upgrade) and I faced with a problem with my RAID 1 soft.
> 
> I have two disk (/dev/sdb and /dev/sdc) which are members of the raid array
> There's 10 partitions on these disks, each one is a array (ie sdb1 and
> sdc1).
> All of these form my raids array : md0 to md9

Understood.

> Today , one of my md partition was missing.
> Before upgrade I had partitions from md0 to md9, after reboot I'm
> missing md9.
> 
> my mdadm.conf before reboot looked like that:

[trim /]

Very good.  Although the level= and num-devices= clauses aren't
necessary, and sometimes troublesome.

> After reboot and after a mdadm -Es, I got this:

[trim /]

>    # definitions of existing MD arrays
>    ARRAY metadata=imsm UUID=d9cfa6d9:2a715e4f:1fbc2095:be342429
>    ARRAY /dev/md/raidlin container=d9cfa6d9:2a715e4f:1fbc2095:be342429
>    member=0 UUID=91449a9d:9242bfe9:d99bceb0:a59f9314

So here your system misidentified your drives as members of an
Intel-based MD-compatible hardware raid, then found raid members inside it.

This is almost certainly a side effect of using v0.90 metadata.  It has
always had a design problem distinguishing between a raid partition at
the end of a disk and a raid occupying an entire device.  It's one of
the reasons that metadata was deprecated long ago.

If you haven't used the system in this weird state, fixing it should be
relatively simple:

1) use mdadm --stop on all the arrays, in numerical order.  /proc/mdstat
should then be empty.

2) manually assemble your arrays, one by one, using the
--update=metadata clause to converted them to v1.0 metadata.

3) If md9 refuses to assemble (possibly damaged by the usage as IMSM),
re-create it with metadata v1.0.

4) Replace your mdadm.conf with a new scan, then update your initramfs.

However, if you've been using the system in this degraded state, you
will need to do the manual assembly with only the good partitions, then
add the other partitions to rebuild each.

Hope this helps.

Phil
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: upgrade to jessie/newer kernel and mdadm problems
  2015-05-04 12:42 ` Phil Turmel
@ 2015-05-04 13:07   ` Jérôme Tytgat
  0 siblings, 0 replies; 3+ messages in thread
From: Jérôme Tytgat @ 2015-05-04 13:07 UTC (permalink / raw)
  To: linux-raid

> I was leaving your case for people who know IMSM to pipe up, as I don't
> have any experience with it.  But the silence is deafening :-(

That's OK, a good guy (PascalHambourg) in the french debian forum was 
able to help me a lot.


> However, if you've been using the system in this degraded state, you
> will need to do the manual assembly with only the good partitions, then
> add the other partitions to rebuild each.

Yes, it was in use, but this what we done :

1. restoring the original mdadm.conf

2. modifying the DEVICE lines with this : DEVICE /dev/sdb?* /dev/sdc?*

3. updating initram: update-initramfs -u (got some errors but we ignored 
them, however I made two initrd to be failsafe)

4. rebooted, md126 was gone and md9 back. However all arrays had a 
partition marked as fail

5. rebuilded each partiton with mdadm /dev/mdX --add /dev/sdcX or mdadm 
/dev/mdX --add /dev/sdbX accordingly (sometimes, the failed partition 
was on sdb and sometimes on sdc, this scarried me as I thought I would 
loose everything if one drive failed).

6. rebuild was ok, I needed to remove the superblock on /dev/sdb and 
/dev/sdc because it looked like this (it contains Intel RAID data but my 
disk are softraid and we thought that was one origin of the problem):


<--------------------------------------------------------------------->
# mdadm -E /dev/sdb
mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
/dev/sdb:
           Magic : Intel Raid ISM Cfg Sig.
         Version : 1.1.00
     Orig Family : 26b5a9e0
          Family : 26b5a9e0
      Generation : 00004db7
      Attributes : All supported
            UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
        Checksum : 261d2aed correct
     MPB Sectors : 1
           Disks : 2
    RAID Devices : 1

   Disk01 Serial : VFC100R10BE79D
           State : active
              Id : 00010000
     Usable Size : 488390862 (232.88 GiB 250.06 GB)

[raidlin]:
            UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
      RAID Level : 1
         Members : 2
           Slots : [UU]
     Failed disk : none
       This Slot : 1
      Array Size : 488390656 (232.88 GiB 250.06 GB)
    Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
   Sector Offset : 0
     Num Stripes : 1907776
      Chunk Size : 64 KiB
        Reserved : 0
   Migrate State : idle
       Map State : normal
     Dirty State : dirty

   Disk00 Serial : VFC100R10BRKMD
           State : active
              Id : 00000000
     Usable Size : 488390862 (232.88 GiB 250.06 GB)
<--------------------------------------------------------------------->
# mdadm -E /dev/sdc
mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
/dev/sdc:
           Magic : Intel Raid ISM Cfg Sig.
         Version : 1.1.00
     Orig Family : 26b5a9e0
          Family : 26b5a9e0
      Generation : 00004dbc
      Attributes : All supported
            UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
        Checksum : 261c2af2 correct
     MPB Sectors : 1
           Disks : 2
    RAID Devices : 1

   Disk00 Serial : VFC100R10BRKMD
           State : active
              Id : 00000000
     Usable Size : 488390862 (232.88 GiB 250.06 GB)

[raidlin]:
            UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
      RAID Level : 1
         Members : 2
           Slots : [UU]
     Failed disk : none
       This Slot : 0
      Array Size : 488390656 (232.88 GiB 250.06 GB)
    Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
   Sector Offset : 0
     Num Stripes : 1907776
      Chunk Size : 64 KiB
        Reserved : 0
   Migrate State : idle
       Map State : normal
     Dirty State : clean

   Disk01 Serial : VFC100R10BE79D
           State : active
              Id : 00010000
     Usable Size : 488390862 (232.88 GiB 250.06 GB)
<--------------------------------------------------------------------->

7. So I rebooted into initramfs shell by editing GRUB command line to 
add "break" at the end of kernel line

8. stopped the raid (they was mounted accordingly to /proc/mdstat) : 
mdadm --stop --scan

9. removed the superblock on /dev/sdb and /dev/sdc : mdadm 
--zero-superblock --metadata=imsm /dev/sdb ; mdadm --zero-superblock 
--metadata=imsm /dev/sdc

This is what they look now :
<--------------------------------------------------------------------->
# mdadm -E /dev/sdb
/dev/sdb:
    MBR Magic : aa55
Partition[0] :       979902 sectors at           63 (type fd)
Partition[1] :      9767520 sectors at       979965 (type fd)
Partition[2] :      3903795 sectors at     10747485 (type fd)
Partition[3] :    473740785 sectors at     14651280 (type 05)
<--------------------------------------------------------------------->
# mdadm -E /dev/sdc
/dev/sdc:
    MBR Magic : aa55
Partition[0] :       979902 sectors at           63 (type fd)
Partition[1] :      9767520 sectors at       979965 (type fd)
Partition[2] :      3903795 sectors at     10747485 (type fd)
Partition[3] :    473740785 sectors at     14651280 (type 05)
<--------------------------------------------------------------------->

10. rebooted, changed back DEVICE line in mdadm.conf to DEVICE 
partitions

The array looks like OK now.

Anyway, Should I upgrade to superblock 1.0 (or 1.2) ? If so, can I use 
your method do it in initramfs shell (because my system is live with 
active raid arrays) ?


Full thread there (in french, sorry): 
https://www.debian-fr.org/mise-a-jour-vers-jessie-et-mdadm-t51945.html

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2015-05-04 13:07 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-05-01  9:40 upgrade to jessie/newer kernel and mdadm problems Jérôme Tytgat
2015-05-04 12:42 ` Phil Turmel
2015-05-04 13:07   ` Jérôme Tytgat

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).