linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* RAID5 assembles with wrong array size
@ 2009-04-24 14:34 Joel Means
  0 siblings, 0 replies; only message in thread
From: Joel Means @ 2009-04-24 14:34 UTC (permalink / raw)
  To: linux-raid

I seem to be having major issues with my primary storage array.  I am
running Debian Lenny with a 2.6.29 kernel, compiled myself, and mdadm
version 2.6.7.2.  I have a seven disk RAID5.  My problem is that when
I assemble my array, the array size shown is too small.  Here is what
I get from mdadm -D /dev/md0:

/dev/md0:
        Version : 00.90
  Creation Time : Thu Apr 23 21:25:14 2009
     Raid Level : raid5
     Array Size : 782819968 (746.56 GiB 801.61 GB)
  Used Dev Size : 488383936 (465.76 GiB 500.11 GB)
   Raid Devices : 7
  Total Devices : 7
Preferred Minor : 0
    Persistence : Superblock is persistent

    Update Time : Thu Apr 23 21:25:14 2009
          State : clean, degraded
 Active Devices : 6
Working Devices : 7
 Failed Devices : 0
  Spare Devices : 1

         Layout : left-symmetric
     Chunk Size : 64K

           UUID : 16b9b60d:ab7eb6b3:4bb6d167:
00581514 (local to host localhost)
         Events : 0.1

    Number   Major   Minor   RaidDevice State
       0       8       33        0      active sync   /dev/sdc1
       1       8       97        1      active sync   /dev/sdg1
       2       8       65        2      active sync   /dev/sde1
       3       8       49        3      active sync   /dev/sdd1
       4       8        1        4      active sync   /dev/sda1
       5       8       81        5      active sync   /dev/sdf1
       6       0        0        6      removed

       7       8       17        -      spare   /dev/sdb1


Using mdadm -E on each of the drives gives the correct info:

/dev/sda1:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : 16b9b60d:ab7eb6b3:4bb6d167:00581514 (local to host localhost)
  Creation Time : Thu Apr 23 21:25:14 2009
     Raid Level : raid5
  Used Dev Size : 488383936 (465.76 GiB 500.11 GB)
     Array Size : 2930303616 (2794.56 GiB 3000.63 GB)
   Raid Devices : 7
  Total Devices : 8
Preferred Minor : 0

    Update Time : Thu Apr 23 21:25:14 2009
          State : clean
 Active Devices : 6
Working Devices : 7
 Failed Devices : 1
  Spare Devices : 1
       Checksum : 68722cb1 - correct
         Events : 1

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     4       8        1        4      active sync   /dev/sda1

   0     0       8       33        0      active sync   /dev/sdc1
   1     1       8       97        1      active sync   /dev/sdg1
   2     2       8       65        2      active sync   /dev/sde1
   3     3       8       49        3      active sync   /dev/sdd1
   4     4       8        1        4      active sync   /dev/sda1
   5     5       8       81        5      active sync   /dev/sdf1
   6     6       0        0        6      faulty
   7     7       8       17        7      spare   /dev/sdb1


Note the difference in Array Size.  I can remove and re-add /dev/sdb1,
and the array will start rebuilding, but the info from 'mdadm -E'
doesn't change.   Stopping and re-assembling the array puts the
following into /var/log/kern.log:

Apr 24 07:52:22 meansnet kernel: [40612.345814] md: md0 stopped.
Apr 24 07:52:22 meansnet kernel: [40612.345855] md: unbind<sdc1>
Apr 24 07:52:22 meansnet kernel: [40612.364018] md: export_rdev(sdc1)
Apr 24 07:52:22 meansnet kernel: [40612.364079] md: unbind<sdb1>
Apr 24 07:52:22 meansnet kernel: [40612.380013] md: export_rdev(sdb1)
Apr 24 07:52:22 meansnet kernel: [40612.380037] md: unbind<sdf1>
Apr 24 07:52:22 meansnet kernel: [40612.396013] md: export_rdev(sdf1)
Apr 24 07:52:22 meansnet kernel: [40612.396034] md: unbind<sda1>
Apr 24 07:52:22 meansnet kernel: [40612.408021] md: export_rdev(sda1)
Apr 24 07:52:22 meansnet kernel: [40612.408062] md: unbind<sdd1>
Apr 24 07:52:22 meansnet kernel: [40612.420021] md: export_rdev(sdd1)
Apr 24 07:52:22 meansnet kernel: [40612.420042] md: unbind<sde1>
Apr 24 07:52:22 meansnet kernel: [40612.432012] md: export_rdev(sde1)
Apr 24 07:52:22 meansnet kernel: [40612.432033] md: unbind<sdg1>
Apr 24 07:52:22 meansnet kernel: [40612.444013] md: export_rdev(sdg1)
Apr 24 07:52:45 meansnet kernel: [40632.630953] md: md0 stopped.
Apr 24 07:52:54 meansnet kernel: [40632.865163] md: bind<sdg1>
Apr 24 07:53:02 meansnet kernel: [40632.865333] md: bind<sde1>
Apr 24 07:53:10 meansnet kernel: [40632.865460] md: bind<sdd1>
Apr 24 07:53:10 meansnet kernel: [40632.871509] md: bind<sda1>
Apr 24 07:53:10 meansnet kernel: [40632.871659] md: bind<sdf1>
Apr 24 07:53:10 meansnet kernel: [40632.878380] md: bind<sdb1>
Apr 24 07:53:10 meansnet kernel: [40632.878509] md: bind<sdc1>
Apr 24 07:53:10 meansnet kernel: [40632.916237] raid5: device sdc1
operational as raid disk 0
Apr 24 07:53:10 meansnet kernel: [40632.916255] raid5: device sdf1
operational as raid disk 5
Apr 24 07:53:10 meansnet kernel: [40632.916271] raid5: device sda1
operational as raid disk 4
Apr 24 07:53:10 meansnet kernel: [40632.916288] raid5: device sdd1
operational as raid disk 3
Apr 24 07:53:10 meansnet kernel: [40632.916305] raid5: device sde1
operational as raid disk 2
Apr 24 07:53:10 meansnet kernel: [40632.916322] raid5: device sdg1
operational as raid disk 1
Apr 24 07:53:10 meansnet kernel: [40632.916897] raid5: allocated 7330kB for md0
Apr 24 07:53:10 meansnet kernel: [40632.916913] raid5: raid level 5
set md0 active with 6 out of 7 devices, algorithm 2
Apr 24 07:53:10 meansnet kernel: [40632.916940] RAID5 conf printout:
Apr 24 07:53:10 meansnet kernel: [40632.916954]  --- rd:7 wd:6
Apr 24 07:53:10 meansnet kernel: [40632.916967]  disk 0, o:1, dev:sdc1
Apr 24 07:53:10 meansnet kernel: [40632.916982]  disk 1, o:1, dev:sdg1
Apr 24 07:53:10 meansnet kernel: [40632.916996]  disk 2, o:1, dev:sde1
Apr 24 07:53:10 meansnet kernel: [40632.917019]  disk 3, o:1, dev:sdd1
Apr 24 07:53:10 meansnet kernel: [40632.917034]  disk 4, o:1, dev:sda1
Apr 24 07:53:10 meansnet kernel: [40632.917048]  disk 5, o:1, dev:sdf1
Apr 24 07:53:10 meansnet kernel: [40632.917483]  md0: unknown partition table



Running 'fsck.jfs' on /dev/md0 gives me an error about corrupt
superblocks, so I don't know if my data is hosed or not.  This was
working fine for over a year.   It could be the upgrade to the new
kernel that did this, but trying to revert to the older kernel gave me
several other issues.  Does anyone have any thoughts on what might be
done to fix this?
Thanks,

Joel

^ permalink raw reply	[flat|nested] only message in thread

only message in thread, other threads:[~2009-04-24 14:34 UTC | newest]

Thread overview: (only message) (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-04-24 14:34 RAID5 assembles with wrong array size Joel Means

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).