linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Dave Jiang <dave.jiang@intel.com>
To: chris <tknchris@gmail.com>
Cc: Linux-RAID <linux-raid@vger.kernel.org>
Subject: Re: Recovery/Access of imsm raid via mdadm?
Date: Thu, 10 Jan 2013 10:09:56 -0700	[thread overview]
Message-ID: <50EEF5E4.8070509@intel.com> (raw)
In-Reply-To: <CAKnNFz96pxExLDpoX1Yi+m0ASb-pV9eUVwZ7xmNQE9LnQHWT_g@mail.gmail.com>

On 01/10/2013 09:23 AM, chris wrote:
> Hello,
>
> I have a machine which was running a imsm raid volume, where the
> motherboard failed and I do not have access to another system with
> imsm. I remember noticing some time ago that mdadm could recognize
> these arrays, so I decided to try recovery in a spare machine with the
> disks from the array.
>
> I guess my questions are:
> Is this the right forum for help with this?
> Am I even going down a feasible path here or is this array dependent
> on the HBA in some way?
> If it is possible any ideas of anything else I can do to debug this further?

Typically mdadm probes the OROM and look for platform details. But you 
can try overriding with:
export IMSM_NO_PLATFORM=1

See if that works for you.

> The original array was a raid 5 of 4x2TB sata disks
>
> When I examine the first disk, things look good:
>
> mdadm --examine /dev/sdb
> mdadm: /dev/sdb is not attached to Intel(R) RAID controller.
> mdadm: /dev/sdb is not attached to Intel(R) RAID controller.
> /dev/sdb:
>            Magic : Intel Raid ISM Cfg Sig.
>          Version : 1.3.00
>      Orig Family : 226cc5df
>           Family : 226cc5df
>       Generation : 000019dc
>       Attributes : All supported
>             UUID : b67ea029:aaea7da2:2540c1eb:ebe98af1
>         Checksum : 651263bf correct
>      MPB Sectors : 2
>            Disks : 4
>     RAID Devices : 1
>
>    Disk02 Serial : Z1E1RPA9
>            State : active
>               Id : 00030000
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
> [Volume0]:
>             UUID : 51a415ba:dc9c8cd7:5b3ea8de:465b4630
>       RAID Level : 5
>          Members : 4
>            Slots : [__U_]
>      Failed disk : 0
>        This Slot : 2
>       Array Size : 11721072640 (5589.04 GiB 6001.19 GB)
>     Per Dev Size : 3907024648 (1863.01 GiB 2000.40 GB)
>    Sector Offset : 0
>      Num Stripes : 15261814
>       Chunk Size : 128 KiB
>         Reserved : 0
>    Migrate State : idle
>        Map State : failed
>      Dirty State : clean
>
>    Disk00 Serial : Z1E1AKPH:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
>    Disk01 Serial : Z24091Q5:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
>    Disk03 Serial : Z1E19E4K:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
>
> When I try to scan for arrays I get this:
> # mdadm --examine --scan
> HBAs of devices does not match (null) != (null)
> ARRAY metadata=imsm UUID=b67ea029:aaea7da2:2540c1eb:ebe98af1
> ARRAY /dev/md/Volume0 container=b67ea029:aaea7da2:2540c1eb:ebe98af1
> member=0 UUID=51a415ba:dc9c8cd7:5b3ea8de:465b4630
> ARRAY metadata=imsm UUID=b67ea029:aaea7da2:2540c1eb:ebe98af1
> ARRAY /dev/md/Volume0 container=b67ea029:aaea7da2:2540c1eb:ebe98af1
> member=0 UUID=51a415ba:dc9c8cd7:5b3ea8de:465b4630
>
> My first concern is the warning that the HBA is missing, the whole
> reason I am going at it this way is because I don't have the HBA.
> Second concern is duplicate detection of the same array.
>
> If i try to run # mdadm -As:
> mdadm: No arrays found in config file or automatically
>
>   I also tried adding the output from --examine --scan to
> /etc/mdadm/mdadm.conf but after trying that I now get blank output:
>
> # mdadm --assemble /dev/md/Volume0
> #
> # mdadm --assemble --scan
> #
>
> full examine of all disks involved:
>
> /dev/sdb:
>            Magic : Intel Raid ISM Cfg Sig.
>          Version : 1.3.00
>      Orig Family : 226cc5df
>           Family : 226cc5df
>       Generation : 000019dc
>       Attributes : All supported
>             UUID : b67ea029:aaea7da2:2540c1eb:ebe98af1
>         Checksum : 651263bf correct
>      MPB Sectors : 2
>            Disks : 4
>     RAID Devices : 1
>
>    Disk02 Serial : Z1E1RPA9
>            State : active
>               Id : 00030000
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
> [Volume0]:
>             UUID : 51a415ba:dc9c8cd7:5b3ea8de:465b4630
>       RAID Level : 5
>          Members : 4
>            Slots : [__U_]
>      Failed disk : 0
>        This Slot : 2
>       Array Size : 11721072640 (5589.04 GiB 6001.19 GB)
>     Per Dev Size : 3907024648 (1863.01 GiB 2000.40 GB)
>    Sector Offset : 0
>      Num Stripes : 15261814
>       Chunk Size : 128 KiB
>         Reserved : 0
>    Migrate State : idle
>        Map State : failed
>      Dirty State : clean
>
>    Disk00 Serial : Z1E1AKPH:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
>    Disk01 Serial : Z24091Q5:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
>    Disk03 Serial : Z1E19E4K:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
> /dev/sdc:
>     MBR Magic : aa55
> Partition[0] :   3907027057 sectors at           63 (type 42)
> /dev/sdd:
>            Magic : Intel Raid ISM Cfg Sig.
>          Version : 1.3.00
>      Orig Family : 226cc5df
>           Family : 226cc5df
>       Generation : 000019d9
>       Attributes : All supported
>             UUID : b67ea029:aaea7da2:2540c1eb:ebe98af1
>         Checksum : 641438ba correct
>      MPB Sectors : 2
>            Disks : 4
>     RAID Devices : 1
>
>    Disk03 Serial : Z1E19E4K
>            State : active
>               Id : 00020000
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
> [Volume0]:
>             UUID : 51a415ba:dc9c8cd7:5b3ea8de:465b4630
>       RAID Level : 5
>          Members : 4
>            Slots : [__UU]
>      Failed disk : 0
>        This Slot : 3
>       Array Size : 11721072640 (5589.04 GiB 6001.19 GB)
>     Per Dev Size : 3907024648 (1863.01 GiB 2000.40 GB)
>    Sector Offset : 0
>      Num Stripes : 15261814
>       Chunk Size : 128 KiB
>         Reserved : 0
>    Migrate State : idle
>        Map State : failed
>      Dirty State : clean
>
>    Disk00 Serial : Z1E1AKPH:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
>    Disk01 Serial : Z24091Q5:0
>            State : active
>               Id : ffffffff
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
>
>    Disk02 Serial : Z1E1RPA9
>            State : active
>               Id : 00030000
>      Usable Size : 3907024648 (1863.01 GiB 2000.40 GB)
> /dev/sde:
>     MBR Magic : aa55
> Partition[0] :   4294967295 sectors at            1 (type ee)
>
> # dpkg -l | grep mdadm
> ii  mdadm                                                       3.2.5-1+b1
>
> thanks
> chris
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


  reply	other threads:[~2013-01-10 17:09 UTC|newest]

Thread overview: 18+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2013-01-10 16:23 Recovery/Access of imsm raid via mdadm? chris
2013-01-10 17:09 ` Dave Jiang [this message]
2013-01-10 20:19   ` chris
2013-01-11  1:42     ` Dan Williams
2013-01-11 17:53       ` chris
2013-01-13 19:00         ` chris
2013-01-13 21:05           ` Dan Williams
2013-01-14  0:56             ` chris
2013-01-14 12:36               ` Dorau, Lukasz
2013-01-14 14:10               ` Dorau, Lukasz
2013-01-14 14:24               ` Dorau, Lukasz
2013-01-14 15:25                 ` chris
2013-01-15 10:25                   ` Dorau, Lukasz
2013-01-16 16:49                     ` chris
2013-01-16 16:53                       ` chris
2013-01-16 22:47                       ` Dan Williams
2013-01-17 15:12                         ` Charles Polisher
2013-01-17 16:07                         ` chris

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=50EEF5E4.8070509@intel.com \
    --to=dave.jiang@intel.com \
    --cc=linux-raid@vger.kernel.org \
    --cc=tknchris@gmail.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).