linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* Recovering RAID set after OS disk failed
@ 2014-06-02  5:37 Davide Guarisco
  2014-06-02 12:36 ` Kővári Péter
  0 siblings, 1 reply; 7+ messages in thread
From: Davide Guarisco @ 2014-06-02  5:37 UTC (permalink / raw)
  To: linux-raid

[-- Attachment #1: Type: text/plain, Size: 992 bytes --]

Hello and sorry if this is not the place for a relatively “newbie” question.

Five years ago I built a NAS box running on Ubuntu server then-current version. The NAS has 4 SATA drives (the RAID set) and a PATA system drive (for the OS). Now the system drive has failed (click of death). I believe the that RAID data is still safe and sound, but my question is how to proceed in such a scenario. (I do have a backup but it’s fairly old). I do not remember exactly how I created the RAID set, other than being with mdadm and probably RAID 5. 

I have now replaced the failed PATA drive with a new 32 GB PATA SSD, and installed Ubuntu Server 14.04. The system is up and running and it sees the four SATA drives. I have installed Webmin and I am ready to recover the RAID set. 


I read the Wiki and it suggests running a “permutation” Perl script. Is this a reasonable thing to do? 
I could not find much information on a case like this in a Goole search, so any help appreciated. 



[-- Attachment #2: Message signed with OpenPGP using GPGMail --]
[-- Type: application/pgp-signature, Size: 881 bytes --]

^ permalink raw reply	[flat|nested] 7+ messages in thread

* RE: Recovering RAID set after OS disk failed
  2014-06-02  5:37 Recovering RAID set after OS disk failed Davide Guarisco
@ 2014-06-02 12:36 ` Kővári Péter
  2014-06-03  5:49   ` Davide Guarisco
  0 siblings, 1 reply; 7+ messages in thread
From: Kővári Péter @ 2014-06-02 12:36 UTC (permalink / raw)
  To: linux-raid

Hi Davide,

Open / ssh a console on your NAS box, and issue the following command and send us the results: 
$ cat /proc/mdstat

Please also issue the following commands

mdadm --examine /dev/sdX[Y]

Where X is one of the raid drive's name and Y is the partition number, if you created the raid set on partitions. (If not, then leave the number.) So, for example (assuming that your OS drive is /dev/sda, so your raid drives are /dev/sdb, /dev/sdc and so on) issue the following commands:

$ mdadm --examine /dev/sdb
or
$ mdadm --examine /dev/sdb1

and so on for all 4 drives. And send back the results.

p.s.
Before everything else,  you might try  auto assembling th eset by:
$ mdadm -v --assemble --scan

It might assemble your raid set for you successfully out of the box.  (If not, send here the output.)
If this assembles your set successfully, then you just need to save your config in /etc/mdam/mdadm.conf, do an initramfs update and you are good to go.
So to save the config issue:
$ mdadm --examine --scan >> /etc/mdadm/mdadm.conf

then update initramfs so th eset will auto assmble on next boot:
$ update-initramfs -k all -u

Best regards,
Peter
---------------------------------------------------------------------------------------------------------------
: peter@kovari.priv.hu
: pkovari@gmail.com
:  www.kovari.priv.hu

-----Original Message-----
From: linux-raid-owner@vger.kernel.org [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Davide Guarisco
Sent: Monday, June 2, 2014 7:38 AM
To: linux-raid@vger.kernel.org
Subject: Recovering RAID set after OS disk failed

Hello and sorry if this is not the place for a relatively “newbie” question.

Five years ago I built a NAS box running on Ubuntu server then-current version. The NAS has 4 SATA drives (the RAID set) and a PATA system drive (for the OS). Now the system drive has failed (click of death). I believe the that RAID data is still safe and sound, but my question is how to proceed in such a scenario. (I do have a backup but it’s fairly old). I do not remember exactly how I created the RAID set, other than being with mdadm and probably RAID 5. 

I have now replaced the failed PATA drive with a new 32 GB PATA SSD, and installed Ubuntu Server 14.04. The system is up and running and it sees the four SATA drives. I have installed Webmin and I am ready to recover the RAID set. 


I read the Wiki and it suggests running a “permutation” Perl script. Is this a reasonable thing to do? 
I could not find much information on a case like this in a Goole search, so any help appreciated. 




--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Recovering RAID set after OS disk failed
  2014-06-02 12:36 ` Kővári Péter
@ 2014-06-03  5:49   ` Davide Guarisco
  2014-06-03  6:05     ` Eyal Lebedinsky
  0 siblings, 1 reply; 7+ messages in thread
From: Davide Guarisco @ 2014-06-03  5:49 UTC (permalink / raw)
  To: linux-raid

[-- Attachment #1: Type: text/plain, Size: 5652 bytes --]

Peter, thanks for your help. Below are the answers.


On Jun 2, 2014, at 05:36, Kővári Péter <peter@kovari.priv.hu> wrote:

> Hi Davide,
> 
> Open / ssh a console on your NAS box, and issue the following command and send us the results: 
> $ cat /proc/mdstat


Personalities :
unused devices: <none>



> 
> Please also issue the following commands
> 
> mdadm --examine /dev/sdX[Y]

$ mdadm —examine /dev/sdb
mdadm: cannot open /dev/sdb: Permission denied


$ sudo mdadm —examine /dev/sdb
/dev/sdb:
	MBR Magic : aa55
Partition[0] : 1953520002 sectors at       63 (type fd)


$ sudo mdadm —examine /dev/sdb1
/dev/sdb1:
          Magic : a92b4efc
        Version : 0.90.00
           UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
  Creation Time : Tue Mar  3 23:27:50 2009
     Raid Level : raid5
  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
     Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 0

    Update Time : Wed May 28 21:52:54 2014
          State : clean
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 2d5185d8 - correct
         Events : 46

         Layout : left-symmetric
     Chunk Size : 128K

      Number   Major   Minor   RaidDevice State
this     0       8       17        0      active sync   /dev/sdb1

   0     0       8       17        0      active sync   /dev/sdb1
   1     1       8       33        1      active sync   /dev/sdc1
   2     2       8       49        2      active sync   /dev/sdd1
   3     3       8       65        3      active sync   /dev/sde1


> mdadm --examine /dev/sdc1

/dev/sdc1:
          Magic : a92b4efc
        Version : 0.90.00
           UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
  Creation Time : Tue Mar  3 23:27:50 2009
     Raid Level : raid5
  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
     Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 0

    Update Time : Wed May 28 21:52:54 2014
          State : clean
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 2d5185ea - correct
         Events : 46

         Layout : left-symmetric
     Chunk Size : 128K

      Number   Major   Minor   RaidDevice State
this     1       8       33        1      active sync   /dev/sdc1

   0     0       8       17        0      active sync   /dev/sdb1
   1     1       8       33        1      active sync   /dev/sdc1
   2     2       8       49        2      active sync   /dev/sdd1
   3     3       8       65        3      active sync   /dev/sde1


…etc. So it seems to me that we are OK, with the RAID 5 set setup on /dev/sdb1, /dev/sdc1, /dev/sdd1, /dev/sde1.

> 
> Where X is one of the raid drive's name and Y is the partition number, if you created the raid set on partitions. (If not, then leave the number.) So, for example (assuming that your OS drive is /dev/sda, so your raid drives are /dev/sdb, /dev/sdc and so on) issue the following commands:
> 
> $ mdadm --examine /dev/sdb
> or
> $ mdadm --examine /dev/sdb1
> 
> and so on for all 4 drives. And send back the results.
> 
> p.s.
> Before everything else,  you might try  auto assembling th eset by:
> $ mdadm -v --assemble —scan

Trying this holding my breath….

> mdadm -v --assemble --scan

mdadm: looking for devices for /dev/md0
mdadm: no recogniseable superblock on /dev/dm-1
mdadm: no recogniseable superblock on /dev/dm-0
mdadm: no RAID superblock on /dev/sde
mdadm: no RAID superblock on /dev/sdd
mdadm: no RAID superblock on /dev/sdb
mdadm: no RAID superblock on /dev/sdc
mdadm: no RAID superblock on /dev/sda5
mdadm: no RAID superblock on /dev/sda2
mdadm: no RAID superblock on /dev/sda1
mdadm: no RAID superblock on /dev/sda
mdadm: /dev/sde1 is identified as a member of /dev/md0, slot 3.
mdadm: /dev/sdd1 is identified as a member of /dev/md0, slot 2.
mdadm: /dev/sdb1 is identified as a member of /dev/md0, slot 0.
mdadm: /dev/sdc1 is identified as a member of /dev/md0, slot 1.
mdadm: added /dev/sdc1 to /dev/md0 as 1
mdadm: added /dev/sdd1 to /dev/md0 as 2
mdadm: added /dev/sde1 to /dev/md0 as 3
mdadm: added /dev/sdb1 to /dev/md0 as 0
mdadm: /dev/md0 has been started with 4 drives.


OK, this seems successful as well. My RAID is /dev/md0. 




> 
> It might assemble your raid set for you successfully out of the box.  (If not, send here the output.)
> If this assembles your set successfully, then you just need to save your config in /etc/mdam/mdadm.conf, do an initramfs update and you are good to go.
> So to save the config issue:
> $ mdadm --examine --scan >> /etc/mdadm/mdadm.conf

> cat /etc/mdadm/mdadm.conf

ARRAY /dev/md0 UUID=f8a943c7:2ffa13d0:9770de34:eca2e81c



> 
> then update initramfs so th eset will auto assmble on next boot:
> $ update-initramfs -k all -u

> update-initramfs -k all -u

update-initramfs: Generating /boot/initrd.img-3.13.0-24-generic



But now:

> sudo fdisk -l

Disk /dev/md0: 3000.6 GB, 3000606523392 bytes
2 heads, 4 sectors/track, 732569952 cylinders, total 5860559616 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 131072 bytes / 393216 bytes
Disk identifier: 0x00000000

Disk /dev/md0 doesn’t contain a valid partition table



How do I fix this and how to I gain access to /dev/md0?

Thanks,
	Davide







[-- Attachment #2: Message signed with OpenPGP using GPGMail --]
[-- Type: application/pgp-signature, Size: 881 bytes --]

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Recovering RAID set after OS disk failed
  2014-06-03  5:49   ` Davide Guarisco
@ 2014-06-03  6:05     ` Eyal Lebedinsky
  2014-06-04  5:00       ` Davide Guarisco
  0 siblings, 1 reply; 7+ messages in thread
From: Eyal Lebedinsky @ 2014-06-03  6:05 UTC (permalink / raw)
  To: linux-raid

Davide,

Do you expect a partition table or do you use the whole disk as the fs (or whatever
higher layers you have)?

What does your /etc/fstab entry say?

Did you try a simple
	sudo mount /dev/md0

You do not say if you rebooted (to let initrd) do its thing.

Eyal

On 06/03/14 15:49, Davide Guarisco wrote:
> Peter, thanks for your help. Below are the answers.
>
>
> On Jun 2, 2014, at 05:36, Kővári Péter <peter@kovari.priv.hu> wrote:
>
>> Hi Davide,
>>
>> Open / ssh a console on your NAS box, and issue the following command and send us the results:
>> $ cat /proc/mdstat
>
>
> Personalities :
> unused devices: <none>
>
>
>
>>
>> Please also issue the following commands
>>
>> mdadm --examine /dev/sdX[Y]
>
> $ mdadm —examine /dev/sdb
> mdadm: cannot open /dev/sdb: Permission denied
>
>
> $ sudo mdadm —examine /dev/sdb
> /dev/sdb:
> 	MBR Magic : aa55
> Partition[0] : 1953520002 sectors at       63 (type fd)
>
>
> $ sudo mdadm —examine /dev/sdb1
> /dev/sdb1:
>            Magic : a92b4efc
>          Version : 0.90.00
>             UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
>    Creation Time : Tue Mar  3 23:27:50 2009
>       Raid Level : raid5
>    Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>       Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>     Raid Devices : 4
>    Total Devices : 4
> Preferred Minor : 0
>
>      Update Time : Wed May 28 21:52:54 2014
>            State : clean
>   Active Devices : 4
> Working Devices : 4
>   Failed Devices : 0
>    Spare Devices : 0
>         Checksum : 2d5185d8 - correct
>           Events : 46
>
>           Layout : left-symmetric
>       Chunk Size : 128K
>
>        Number   Major   Minor   RaidDevice State
> this     0       8       17        0      active sync   /dev/sdb1
>
>     0     0       8       17        0      active sync   /dev/sdb1
>     1     1       8       33        1      active sync   /dev/sdc1
>     2     2       8       49        2      active sync   /dev/sdd1
>     3     3       8       65        3      active sync   /dev/sde1
>
>
>> mdadm --examine /dev/sdc1
>
> /dev/sdc1:
>            Magic : a92b4efc
>          Version : 0.90.00
>             UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
>    Creation Time : Tue Mar  3 23:27:50 2009
>       Raid Level : raid5
>    Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>       Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>     Raid Devices : 4
>    Total Devices : 4
> Preferred Minor : 0
>
>      Update Time : Wed May 28 21:52:54 2014
>            State : clean
>   Active Devices : 4
> Working Devices : 4
>   Failed Devices : 0
>    Spare Devices : 0
>         Checksum : 2d5185ea - correct
>           Events : 46
>
>           Layout : left-symmetric
>       Chunk Size : 128K
>
>        Number   Major   Minor   RaidDevice State
> this     1       8       33        1      active sync   /dev/sdc1
>
>     0     0       8       17        0      active sync   /dev/sdb1
>     1     1       8       33        1      active sync   /dev/sdc1
>     2     2       8       49        2      active sync   /dev/sdd1
>     3     3       8       65        3      active sync   /dev/sde1
>
>
> …etc. So it seems to me that we are OK, with the RAID 5 set setup on /dev/sdb1, /dev/sdc1, /dev/sdd1, /dev/sde1.
>
>>
>> Where X is one of the raid drive's name and Y is the partition number, if you created the raid set on partitions. (If not, then leave the number.) So, for example (assuming that your OS drive is /dev/sda, so your raid drives are /dev/sdb, /dev/sdc and so on) issue the following commands:
>>
>> $ mdadm --examine /dev/sdb
>> or
>> $ mdadm --examine /dev/sdb1
>>
>> and so on for all 4 drives. And send back the results.
>>
>> p.s.
>> Before everything else,  you might try  auto assembling th eset by:
>> $ mdadm -v --assemble —scan
>
> Trying this holding my breath….
>
>> mdadm -v --assemble --scan
>
> mdadm: looking for devices for /dev/md0
> mdadm: no recogniseable superblock on /dev/dm-1
> mdadm: no recogniseable superblock on /dev/dm-0
> mdadm: no RAID superblock on /dev/sde
> mdadm: no RAID superblock on /dev/sdd
> mdadm: no RAID superblock on /dev/sdb
> mdadm: no RAID superblock on /dev/sdc
> mdadm: no RAID superblock on /dev/sda5
> mdadm: no RAID superblock on /dev/sda2
> mdadm: no RAID superblock on /dev/sda1
> mdadm: no RAID superblock on /dev/sda
> mdadm: /dev/sde1 is identified as a member of /dev/md0, slot 3.
> mdadm: /dev/sdd1 is identified as a member of /dev/md0, slot 2.
> mdadm: /dev/sdb1 is identified as a member of /dev/md0, slot 0.
> mdadm: /dev/sdc1 is identified as a member of /dev/md0, slot 1.
> mdadm: added /dev/sdc1 to /dev/md0 as 1
> mdadm: added /dev/sdd1 to /dev/md0 as 2
> mdadm: added /dev/sde1 to /dev/md0 as 3
> mdadm: added /dev/sdb1 to /dev/md0 as 0
> mdadm: /dev/md0 has been started with 4 drives.
>
>
> OK, this seems successful as well. My RAID is /dev/md0.
>
>
>
>
>>
>> It might assemble your raid set for you successfully out of the box.  (If not, send here the output.)
>> If this assembles your set successfully, then you just need to save your config in /etc/mdam/mdadm.conf, do an initramfs update and you are good to go.
>> So to save the config issue:
>> $ mdadm --examine --scan >> /etc/mdadm/mdadm.conf
>
>> cat /etc/mdadm/mdadm.conf
>
> ARRAY /dev/md0 UUID=f8a943c7:2ffa13d0:9770de34:eca2e81c
>
>
>
>>
>> then update initramfs so th eset will auto assmble on next boot:
>> $ update-initramfs -k all -u
>
>> update-initramfs -k all -u
>
> update-initramfs: Generating /boot/initrd.img-3.13.0-24-generic
>
>
>
> But now:
>
>> sudo fdisk -l
>
> Disk /dev/md0: 3000.6 GB, 3000606523392 bytes
> 2 heads, 4 sectors/track, 732569952 cylinders, total 5860559616 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 131072 bytes / 393216 bytes
> Disk identifier: 0x00000000
>
> Disk /dev/md0 doesn’t contain a valid partition table
>
>
>
> How do I fix this and how to I gain access to /dev/md0?
>
> Thanks,
> 	Davide
>
>
>
>
>
>

-- 
Eyal Lebedinsky (eyal@eyal.emu.id.au)
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Recovering RAID set after OS disk failed
  2014-06-03  6:05     ` Eyal Lebedinsky
@ 2014-06-04  5:00       ` Davide Guarisco
  2014-06-04  5:20         ` Eyal Lebedinsky
  0 siblings, 1 reply; 7+ messages in thread
From: Davide Guarisco @ 2014-06-04  5:00 UTC (permalink / raw)
  To: linux-raid

[-- Attachment #1: Type: text/plain, Size: 11547 bytes --]


On Jun 2, 2014, at 23:05, Eyal Lebedinsky <eyal@eyal.emu.id.au> wrote:

> Davide,
> 
> Do you expect a partition table or do you use the whole disk as the fs (or whatever
> higher layers you have)?


I do not remember. It’s possible I originally setup two partitions, one for Time Machine and one as shared storage. 


> 
> What does your /etc/fstab entry say?


fstab is all new, so it says:

davide@gecko:/etc$ cat fstab
# /etc/fstab: static file system information.
#
# Use 'blkid' to print the universally unique identifier for a
# device; this may be used with UUID= as a more robust way to name devices
# that works even if disks are added and removed. See fstab(5).
#
# <file system> <mount point>   <type>  <options>       <dump>  <pass>
/dev/mapper/gecko--vg-root /               ext4    errors=remount-ro 0       1
# /boot was on /dev/sda1 during installation
UUID=89a63dd3-93fb-4a37-acd5-34f6790ef3e7 /boot           ext2    defaults        0       2
/dev/mapper/gecko--vg-swap_1 none            swap    sw              0       0




> 
> Did you try a simple
> 	sudo mount /dev/md0
> 
> You do not say if you rebooted (to let initrd) do its thing.
> 
> Eyal


All right, I will reboot now…

After reboot:

davide@gecko:~$ sudo fdisk -l
[sudo] password for davide: 
no talloc stackframe at ../source3/param/loadparm.c:4864, leaking memory

Disk /dev/sda: 31.9 GB, 31937527808 bytes
255 heads, 63 sectors/track, 3882 cylinders, total 62377984 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x000384c2

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1   *        2048      499711      248832   83  Linux
/dev/sda2          501758    62375935    30937089    5  Extended
/dev/sda5          501760    62375935    30937088   8e  Linux LVM

Disk /dev/sdb: 1000.2 GB, 1000204886016 bytes
255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0xe3494e66

   Device Boot      Start         End      Blocks   Id  System
/dev/sdb1              63  1953520064   976760001   fd  Linux raid autodetect

Disk /dev/sdc: 1000.2 GB, 1000204886016 bytes
255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x6b47e429

   Device Boot      Start         End      Blocks   Id  System
/dev/sdc1              63  1953520064   976760001   fd  Linux raid autodetect

Disk /dev/sde: 1000.2 GB, 1000204886016 bytes
255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0xfdfcb365

   Device Boot      Start         End      Blocks   Id  System
/dev/sde1              63  1953520064   976760001   fd  Linux raid autodetect

Disk /dev/sdd: 1000.2 GB, 1000204886016 bytes
255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x6303429f

   Device Boot      Start         End      Blocks   Id  System
/dev/sdd1              63  1953520064   976760001   fd  Linux raid autodetect

Disk /dev/md0: 3000.6 GB, 3000606523392 bytes
2 heads, 4 sectors/track, 732569952 cylinders, total 5860559616 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 131072 bytes / 393216 bytes
Disk identifier: 0x00000000

Disk /dev/md0 doesn't contain a valid partition table

Disk /dev/mapper/gecko--vg-root: 29.5 GB, 29536288768 bytes
255 heads, 63 sectors/track, 3590 cylinders, total 57688064 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000

Disk /dev/mapper/gecko--vg-root doesn't contain a valid partition table

Disk /dev/mapper/gecko--vg-swap_1: 2139 MB, 2139095040 bytes
255 heads, 63 sectors/track, 260 cylinders, total 4177920 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000

Disk /dev/mapper/gecko--vg-swap_1 doesn't contain a valid partition table



Now trying to mount md0:

davide@gecko:~$ sudo mount /dev/md0
mount: can't find /dev/md0 in /etc/fstab or /etc/mtab


OK, I am lost now. 

Davide












> 
> On 06/03/14 15:49, Davide Guarisco wrote:
>> Peter, thanks for your help. Below are the answers.
>> 
>> 
>> On Jun 2, 2014, at 05:36, Kővári Péter <peter@kovari.priv.hu> wrote:
>> 
>>> Hi Davide,
>>> 
>>> Open / ssh a console on your NAS box, and issue the following command and send us the results:
>>> $ cat /proc/mdstat
>> 
>> 
>> Personalities :
>> unused devices: <none>
>> 
>> 
>> 
>>> 
>>> Please also issue the following commands
>>> 
>>> mdadm --examine /dev/sdX[Y]
>> 
>> $ mdadm —examine /dev/sdb
>> mdadm: cannot open /dev/sdb: Permission denied
>> 
>> 
>> $ sudo mdadm —examine /dev/sdb
>> /dev/sdb:
>> 	MBR Magic : aa55
>> Partition[0] : 1953520002 sectors at       63 (type fd)
>> 
>> 
>> $ sudo mdadm —examine /dev/sdb1
>> /dev/sdb1:
>>           Magic : a92b4efc
>>         Version : 0.90.00
>>            UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
>>   Creation Time : Tue Mar  3 23:27:50 2009
>>      Raid Level : raid5
>>   Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>      Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>    Raid Devices : 4
>>   Total Devices : 4
>> Preferred Minor : 0
>> 
>>     Update Time : Wed May 28 21:52:54 2014
>>           State : clean
>>  Active Devices : 4
>> Working Devices : 4
>>  Failed Devices : 0
>>   Spare Devices : 0
>>        Checksum : 2d5185d8 - correct
>>          Events : 46
>> 
>>          Layout : left-symmetric
>>      Chunk Size : 128K
>> 
>>       Number   Major   Minor   RaidDevice State
>> this     0       8       17        0      active sync   /dev/sdb1
>> 
>>    0     0       8       17        0      active sync   /dev/sdb1
>>    1     1       8       33        1      active sync   /dev/sdc1
>>    2     2       8       49        2      active sync   /dev/sdd1
>>    3     3       8       65        3      active sync   /dev/sde1
>> 
>> 
>>> mdadm --examine /dev/sdc1
>> 
>> /dev/sdc1:
>>           Magic : a92b4efc
>>         Version : 0.90.00
>>            UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
>>   Creation Time : Tue Mar  3 23:27:50 2009
>>      Raid Level : raid5
>>   Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>      Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>    Raid Devices : 4
>>   Total Devices : 4
>> Preferred Minor : 0
>> 
>>     Update Time : Wed May 28 21:52:54 2014
>>           State : clean
>>  Active Devices : 4
>> Working Devices : 4
>>  Failed Devices : 0
>>   Spare Devices : 0
>>        Checksum : 2d5185ea - correct
>>          Events : 46
>> 
>>          Layout : left-symmetric
>>      Chunk Size : 128K
>> 
>>       Number   Major   Minor   RaidDevice State
>> this     1       8       33        1      active sync   /dev/sdc1
>> 
>>    0     0       8       17        0      active sync   /dev/sdb1
>>    1     1       8       33        1      active sync   /dev/sdc1
>>    2     2       8       49        2      active sync   /dev/sdd1
>>    3     3       8       65        3      active sync   /dev/sde1
>> 
>> 
>> …etc. So it seems to me that we are OK, with the RAID 5 set setup on /dev/sdb1, /dev/sdc1, /dev/sdd1, /dev/sde1.
>> 
>>> 
>>> Where X is one of the raid drive's name and Y is the partition number, if you created the raid set on partitions. (If not, then leave the number.) So, for example (assuming that your OS drive is /dev/sda, so your raid drives are /dev/sdb, /dev/sdc and so on) issue the following commands:
>>> 
>>> $ mdadm --examine /dev/sdb
>>> or
>>> $ mdadm --examine /dev/sdb1
>>> 
>>> and so on for all 4 drives. And send back the results.
>>> 
>>> p.s.
>>> Before everything else,  you might try  auto assembling th eset by:
>>> $ mdadm -v --assemble —scan
>> 
>> Trying this holding my breath….
>> 
>>> mdadm -v --assemble --scan
>> 
>> mdadm: looking for devices for /dev/md0
>> mdadm: no recogniseable superblock on /dev/dm-1
>> mdadm: no recogniseable superblock on /dev/dm-0
>> mdadm: no RAID superblock on /dev/sde
>> mdadm: no RAID superblock on /dev/sdd
>> mdadm: no RAID superblock on /dev/sdb
>> mdadm: no RAID superblock on /dev/sdc
>> mdadm: no RAID superblock on /dev/sda5
>> mdadm: no RAID superblock on /dev/sda2
>> mdadm: no RAID superblock on /dev/sda1
>> mdadm: no RAID superblock on /dev/sda
>> mdadm: /dev/sde1 is identified as a member of /dev/md0, slot 3.
>> mdadm: /dev/sdd1 is identified as a member of /dev/md0, slot 2.
>> mdadm: /dev/sdb1 is identified as a member of /dev/md0, slot 0.
>> mdadm: /dev/sdc1 is identified as a member of /dev/md0, slot 1.
>> mdadm: added /dev/sdc1 to /dev/md0 as 1
>> mdadm: added /dev/sdd1 to /dev/md0 as 2
>> mdadm: added /dev/sde1 to /dev/md0 as 3
>> mdadm: added /dev/sdb1 to /dev/md0 as 0
>> mdadm: /dev/md0 has been started with 4 drives.
>> 
>> 
>> OK, this seems successful as well. My RAID is /dev/md0.
>> 
>> 
>> 
>> 
>>> 
>>> It might assemble your raid set for you successfully out of the box.  (If not, send here the output.)
>>> If this assembles your set successfully, then you just need to save your config in /etc/mdam/mdadm.conf, do an initramfs update and you are good to go.
>>> So to save the config issue:
>>> $ mdadm --examine --scan >> /etc/mdadm/mdadm.conf
>> 
>>> cat /etc/mdadm/mdadm.conf
>> 
>> ARRAY /dev/md0 UUID=f8a943c7:2ffa13d0:9770de34:eca2e81c
>> 
>> 
>> 
>>> 
>>> then update initramfs so th eset will auto assmble on next boot:
>>> $ update-initramfs -k all -u
>> 
>>> update-initramfs -k all -u
>> 
>> update-initramfs: Generating /boot/initrd.img-3.13.0-24-generic
>> 
>> 
>> 
>> But now:
>> 
>>> sudo fdisk -l
>> 
>> Disk /dev/md0: 3000.6 GB, 3000606523392 bytes
>> 2 heads, 4 sectors/track, 732569952 cylinders, total 5860559616 sectors
>> Units = sectors of 1 * 512 = 512 bytes
>> Sector size (logical/physical): 512 bytes / 512 bytes
>> I/O size (minimum/optimal): 131072 bytes / 393216 bytes
>> Disk identifier: 0x00000000
>> 
>> Disk /dev/md0 doesn’t contain a valid partition table
>> 
>> 
>> 
>> How do I fix this and how to I gain access to /dev/md0?
>> 
>> Thanks,
>> 	Davide
>> 
>> 
>> 
>> 
>> 
>> 
> 
> -- 
> Eyal Lebedinsky (eyal@eyal.emu.id.au)
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html






[-- Attachment #2: Message signed with OpenPGP using GPGMail --]
[-- Type: application/pgp-signature, Size: 881 bytes --]

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Recovering RAID set after OS disk failed
  2014-06-04  5:00       ` Davide Guarisco
@ 2014-06-04  5:20         ` Eyal Lebedinsky
  2014-06-08  4:57           ` Davide Guarisco
  0 siblings, 1 reply; 7+ messages in thread
From: Eyal Lebedinsky @ 2014-06-04  5:20 UTC (permalink / raw)
  To: linux-raid



On 06/04/14 15:00, Davide Guarisco wrote:
>
> On Jun 2, 2014, at 23:05, Eyal Lebedinsky <eyal@eyal.emu.id.au> wrote:
>
>> Davide,
>>
>> Do you expect a partition table or do you use the whole disk as the fs (or whatever
>> higher layers you have)?
>
>
> I do not remember. It’s possible I originally setup two partitions, one for Time Machine and one as shared storage.
>
>
>>
>> What does your /etc/fstab entry say?
>
>
> fstab is all new, so it says:
>
> davide@gecko:/etc$ cat fstab
> # /etc/fstab: static file system information.
> #
> # Use 'blkid' to print the universally unique identifier for a
> # device; this may be used with UUID= as a more robust way to name devices
> # that works even if disks are added and removed. See fstab(5).
> #
> # <file system> <mount point>   <type>  <options>       <dump>  <pass>
> /dev/mapper/gecko--vg-root /               ext4    errors=remount-ro 0       1
> # /boot was on /dev/sda1 during installation
> UUID=89a63dd3-93fb-4a37-acd5-34f6790ef3e7 /boot           ext2    defaults        0       2
> /dev/mapper/gecko--vg-swap_1 none            swap    sw              0       0
>
>
>
>
>>
>> Did you try a simple
>> 	sudo mount /dev/md0
>>
>> You do not say if you rebooted (to let initrd) do its thing.
>>
>> Eyal
>
>
> All right, I will reboot now…
>
> After reboot:
>
> davide@gecko:~$ sudo fdisk -l
> [sudo] password for davide:
> no talloc stackframe at ../source3/param/loadparm.c:4864, leaking memory
>
> Disk /dev/sda: 31.9 GB, 31937527808 bytes
> 255 heads, 63 sectors/track, 3882 cylinders, total 62377984 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x000384c2
>
>     Device Boot      Start         End      Blocks   Id  System
> /dev/sda1   *        2048      499711      248832   83  Linux
> /dev/sda2          501758    62375935    30937089    5  Extended
> /dev/sda5          501760    62375935    30937088   8e  Linux LVM
>
> Disk /dev/sdb: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0xe3494e66
>
>     Device Boot      Start         End      Blocks   Id  System
> /dev/sdb1              63  1953520064   976760001   fd  Linux raid autodetect
>
> Disk /dev/sdc: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x6b47e429
>
>     Device Boot      Start         End      Blocks   Id  System
> /dev/sdc1              63  1953520064   976760001   fd  Linux raid autodetect
>
> Disk /dev/sde: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0xfdfcb365
>
>     Device Boot      Start         End      Blocks   Id  System
> /dev/sde1              63  1953520064   976760001   fd  Linux raid autodetect
>
> Disk /dev/sdd: 1000.2 GB, 1000204886016 bytes
> 255 heads, 63 sectors/track, 121601 cylinders, total 1953525168 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x6303429f
>
>     Device Boot      Start         End      Blocks   Id  System
> /dev/sdd1              63  1953520064   976760001   fd  Linux raid autodetect
>
> Disk /dev/md0: 3000.6 GB, 3000606523392 bytes
> 2 heads, 4 sectors/track, 732569952 cylinders, total 5860559616 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 131072 bytes / 393216 bytes
> Disk identifier: 0x00000000
>
> Disk /dev/md0 doesn't contain a valid partition table
>
> Disk /dev/mapper/gecko--vg-root: 29.5 GB, 29536288768 bytes
> 255 heads, 63 sectors/track, 3590 cylinders, total 57688064 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
>
> Disk /dev/mapper/gecko--vg-root doesn't contain a valid partition table
>
> Disk /dev/mapper/gecko--vg-swap_1: 2139 MB, 2139095040 bytes
> 255 heads, 63 sectors/track, 260 cylinders, total 4177920 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
>
> Disk /dev/mapper/gecko--vg-swap_1 doesn't contain a valid partition table
>
>
>
> Now trying to mount md0:
>
> davide@gecko:~$ sudo mount /dev/md0
> mount: can't find /dev/md0 in /etc/fstab or /etc/mtab
>
>
> OK, I am lost now.
>
> Davide

So you do not have an fstab entry, no problem. File systems usually mount just fine
with automatic detection (replace mountPoint with the directory where it usually mounts).
	sudo mount -o ro /dev/md0 /mountPoint
[if it looks good then you can remove the readonly option '-o ro']

Or you can check the array directly:
	sudo file -s /dev/md0

For me I get:

$ sudo file -s /dev/md0
/dev/md0: Linux rev 1.0 ext4 filesystem data, UUID=1db56f55-de4f-435e-80ed-e525f07d30df, volume name "data1" (needs journal recovery) (extents) (64bit) (large files) (huge files)

If there is no fs showing then you probably did have some partitions. Or it did
not assemble correctly (hope not).

cheers

>>
>> On 06/03/14 15:49, Davide Guarisco wrote:
>>> Peter, thanks for your help. Below are the answers.
>>>
>>>
>>> On Jun 2, 2014, at 05:36, Kővári Péter <peter@kovari.priv.hu> wrote:
>>>
>>>> Hi Davide,
>>>>
>>>> Open / ssh a console on your NAS box, and issue the following command and send us the results:
>>>> $ cat /proc/mdstat
>>>
>>>
>>> Personalities :
>>> unused devices: <none>
>>>
>>>
>>>
>>>>
>>>> Please also issue the following commands
>>>>
>>>> mdadm --examine /dev/sdX[Y]
>>>
>>> $ mdadm —examine /dev/sdb
>>> mdadm: cannot open /dev/sdb: Permission denied
>>>
>>>
>>> $ sudo mdadm —examine /dev/sdb
>>> /dev/sdb:
>>> 	MBR Magic : aa55
>>> Partition[0] : 1953520002 sectors at       63 (type fd)
>>>
>>>
>>> $ sudo mdadm —examine /dev/sdb1
>>> /dev/sdb1:
>>>            Magic : a92b4efc
>>>          Version : 0.90.00
>>>             UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
>>>    Creation Time : Tue Mar  3 23:27:50 2009
>>>       Raid Level : raid5
>>>    Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>>       Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>>     Raid Devices : 4
>>>    Total Devices : 4
>>> Preferred Minor : 0
>>>
>>>      Update Time : Wed May 28 21:52:54 2014
>>>            State : clean
>>>   Active Devices : 4
>>> Working Devices : 4
>>>   Failed Devices : 0
>>>    Spare Devices : 0
>>>         Checksum : 2d5185d8 - correct
>>>           Events : 46
>>>
>>>           Layout : left-symmetric
>>>       Chunk Size : 128K
>>>
>>>        Number   Major   Minor   RaidDevice State
>>> this     0       8       17        0      active sync   /dev/sdb1
>>>
>>>     0     0       8       17        0      active sync   /dev/sdb1
>>>     1     1       8       33        1      active sync   /dev/sdc1
>>>     2     2       8       49        2      active sync   /dev/sdd1
>>>     3     3       8       65        3      active sync   /dev/sde1
>>>
>>>
>>>> mdadm --examine /dev/sdc1
>>>
>>> /dev/sdc1:
>>>            Magic : a92b4efc
>>>          Version : 0.90.00
>>>             UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
>>>    Creation Time : Tue Mar  3 23:27:50 2009
>>>       Raid Level : raid5
>>>    Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>>>       Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
>>>     Raid Devices : 4
>>>    Total Devices : 4
>>> Preferred Minor : 0
>>>
>>>      Update Time : Wed May 28 21:52:54 2014
>>>            State : clean
>>>   Active Devices : 4
>>> Working Devices : 4
>>>   Failed Devices : 0
>>>    Spare Devices : 0
>>>         Checksum : 2d5185ea - correct
>>>           Events : 46
>>>
>>>           Layout : left-symmetric
>>>       Chunk Size : 128K
>>>
>>>        Number   Major   Minor   RaidDevice State
>>> this     1       8       33        1      active sync   /dev/sdc1
>>>
>>>     0     0       8       17        0      active sync   /dev/sdb1
>>>     1     1       8       33        1      active sync   /dev/sdc1
>>>     2     2       8       49        2      active sync   /dev/sdd1
>>>     3     3       8       65        3      active sync   /dev/sde1
>>>
>>>
>>> …etc. So it seems to me that we are OK, with the RAID 5 set setup on /dev/sdb1, /dev/sdc1, /dev/sdd1, /dev/sde1.
>>>
>>>>
>>>> Where X is one of the raid drive's name and Y is the partition number, if you created the raid set on partitions. (If not, then leave the number.) So, for example (assuming that your OS drive is /dev/sda, so your raid drives are /dev/sdb, /dev/sdc and so on) issue the following commands:
>>>>
>>>> $ mdadm --examine /dev/sdb
>>>> or
>>>> $ mdadm --examine /dev/sdb1
>>>>
>>>> and so on for all 4 drives. And send back the results.
>>>>
>>>> p.s.
>>>> Before everything else,  you might try  auto assembling th eset by:
>>>> $ mdadm -v --assemble —scan
>>>
>>> Trying this holding my breath….
>>>
>>>> mdadm -v --assemble --scan
>>>
>>> mdadm: looking for devices for /dev/md0
>>> mdadm: no recogniseable superblock on /dev/dm-1
>>> mdadm: no recogniseable superblock on /dev/dm-0
>>> mdadm: no RAID superblock on /dev/sde
>>> mdadm: no RAID superblock on /dev/sdd
>>> mdadm: no RAID superblock on /dev/sdb
>>> mdadm: no RAID superblock on /dev/sdc
>>> mdadm: no RAID superblock on /dev/sda5
>>> mdadm: no RAID superblock on /dev/sda2
>>> mdadm: no RAID superblock on /dev/sda1
>>> mdadm: no RAID superblock on /dev/sda
>>> mdadm: /dev/sde1 is identified as a member of /dev/md0, slot 3.
>>> mdadm: /dev/sdd1 is identified as a member of /dev/md0, slot 2.
>>> mdadm: /dev/sdb1 is identified as a member of /dev/md0, slot 0.
>>> mdadm: /dev/sdc1 is identified as a member of /dev/md0, slot 1.
>>> mdadm: added /dev/sdc1 to /dev/md0 as 1
>>> mdadm: added /dev/sdd1 to /dev/md0 as 2
>>> mdadm: added /dev/sde1 to /dev/md0 as 3
>>> mdadm: added /dev/sdb1 to /dev/md0 as 0
>>> mdadm: /dev/md0 has been started with 4 drives.
>>>
>>>
>>> OK, this seems successful as well. My RAID is /dev/md0.
>>>
>>>
>>>
>>>
>>>>
>>>> It might assemble your raid set for you successfully out of the box.  (If not, send here the output.)
>>>> If this assembles your set successfully, then you just need to save your config in /etc/mdam/mdadm.conf, do an initramfs update and you are good to go.
>>>> So to save the config issue:
>>>> $ mdadm --examine --scan >> /etc/mdadm/mdadm.conf
>>>
>>>> cat /etc/mdadm/mdadm.conf
>>>
>>> ARRAY /dev/md0 UUID=f8a943c7:2ffa13d0:9770de34:eca2e81c
>>>
>>>
>>>
>>>>
>>>> then update initramfs so th eset will auto assmble on next boot:
>>>> $ update-initramfs -k all -u
>>>
>>>> update-initramfs -k all -u
>>>
>>> update-initramfs: Generating /boot/initrd.img-3.13.0-24-generic
>>>
>>>
>>>
>>> But now:
>>>
>>>> sudo fdisk -l
>>>
>>> Disk /dev/md0: 3000.6 GB, 3000606523392 bytes
>>> 2 heads, 4 sectors/track, 732569952 cylinders, total 5860559616 sectors
>>> Units = sectors of 1 * 512 = 512 bytes
>>> Sector size (logical/physical): 512 bytes / 512 bytes
>>> I/O size (minimum/optimal): 131072 bytes / 393216 bytes
>>> Disk identifier: 0x00000000
>>>
>>> Disk /dev/md0 doesn’t contain a valid partition table
>>>
>>>
>>>
>>> How do I fix this and how to I gain access to /dev/md0?
>>>
>>> Thanks,
>>> 	Davide
>>
>> --
>> Eyal Lebedinsky (eyal@eyal.emu.id.au)


-- 
Eyal Lebedinsky (eyal@eyal.emu.id.au)
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Recovering RAID set after OS disk failed
  2014-06-04  5:20         ` Eyal Lebedinsky
@ 2014-06-08  4:57           ` Davide Guarisco
  0 siblings, 0 replies; 7+ messages in thread
From: Davide Guarisco @ 2014-06-08  4:57 UTC (permalink / raw)
  To: linux-raid

[-- Attachment #1: Type: text/plain, Size: 2361 bytes --]


On Jun 3, 2014, at 22:20, Eyal Lebedinsky <eyal@eyal.emu.id.au> wrote:

> 
> 
> On 06/04/14 15:00, Davide Guarisco wrote:
>> 
>> On Jun 2, 2014, at 23:05, Eyal Lebedinsky <eyal@eyal.emu.id.au> wrote:
>> 
>> […]
>> 
>> OK, I am lost now.
>> 
>> Davide
> 
> So you do not have an fstab entry, no problem. File systems usually mount just fine
> with automatic detection (replace mountPoint with the directory where it usually mounts).
> 	sudo mount -o ro /dev/md0 /mountPoint
> [if it looks good then you can remove the readonly option '-o ro']
> 
> Or you can check the array directly:
> 	sudo file -s /dev/md0
> 
> For me I get:
> 
> $ sudo file -s /dev/md0
> /dev/md0: Linux rev 1.0 ext4 filesystem data, UUID=1db56f55-de4f-435e-80ed-e525f07d30df, volume name "data1" (needs journal recovery) (extents) (64bit) (large files) (huge files)
> 
> If there is no fs showing then you probably did have some partitions. Or it did
> not assemble correctly (hope not).
> 
> cheers


I was busy for the last few days….

So, yes, it is working !!!!

davide@gecko:~$ sudo mdadm --detail /dev/md0
/dev/md0:
        Version : 0.90
  Creation Time : Tue Mar  3 23:27:50 2009
     Raid Level : raid5
     Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 0
    Persistence : Superblock is persistent

    Update Time : Wed May 28 21:52:54 2014
          State : clean 
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 128K

           UUID : f8a943c7:2ffa13d0:9770de34:eca2e81c (local to host gecko)
         Events : 0.46

    Number   Major   Minor   RaidDevice State
       0       8       17        0      active sync   /dev/sdb1
       1       8       33        1      active sync   /dev/sdc1
       2       8       49        2      active sync   /dev/sdd1
       3       8       65        3      active sync   /dev/sde1


And of course, I need to mount it….

davide@gecko:~$ sudo mkdir /gecko
davide@gecko:~$ sudo mount /dev/md0 /gecko

cd /gecko, ls shows me all my files!

Now all is left to do is to save the fstab entry and install netatalk.

Thanks a lot to all who helped me out!

Davide


[-- Attachment #2: Message signed with OpenPGP using GPGMail --]
[-- Type: application/pgp-signature, Size: 881 bytes --]

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2014-06-08  4:57 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2014-06-02  5:37 Recovering RAID set after OS disk failed Davide Guarisco
2014-06-02 12:36 ` Kővári Péter
2014-06-03  5:49   ` Davide Guarisco
2014-06-03  6:05     ` Eyal Lebedinsky
2014-06-04  5:00       ` Davide Guarisco
2014-06-04  5:20         ` Eyal Lebedinsky
2014-06-08  4:57           ` Davide Guarisco

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).