From: "Sunpyo Hong" <sunpyo.hong@amac.com>
To: 'Robin Hill' <robin@robinhill.me.uk>, linux-raid@vger.kernel.org
Subject: RE: Raid 5 Issue, cannot recognize EXT3 File system.
Date: Thu, 17 Sep 2009 17:54:29 -0400 [thread overview]
Message-ID: <200909171754312.SM06100@trainer> (raw)
In-Reply-To:
I am also able to see the filesystem of the NAS, however not the data
partition. EX:
I can see the filesystem on:
#ls /media/disk (which is the mount of the filesystem in the NAS) This is
also considered to be my /dev/md0, /dev/md2 is the data partition
-----Original Message-----
From: Sunpyo Hong [mailto:sunpyo.hong@amac.com]
Sent: Thursday, September 17, 2009 5:47 PM
To: 'Robin Hill'; 'linux-raid@vger.kernel.org'
Subject: RE: Raid 5 Issue, cannot recognize EXT3 File system.
First off, lemme tell you the initial problem. I had a WD ShareSpace that
had one of the disk go bad. They sent a replacement and it was suppose to
rebuild on its own, however after the build, the array went bad and it was
no longer able to see any of the files.
I downloaded and tested the drives using windows data recovery tools I saw
that the ext3 was Linux FS and that using these tools would not help in the
recovery. However through the tools I was able to see and recover some of
the files, but the files themselves were usable. I confirmed with WD that
ext3 was in fact the FS and took steps to recover the data. These are the
steps I took in order for me to assemble the raid.
Right now I have 3/4 drives with the data. I did #mdadm --assemble --scan,
which let me assemble the raid. However at this point I was not able to see
any of the files or mount the drive to the mount point it was once at. I
have also tried #mdadm --create with the array in the right order /w the
missing disk.
Initially the --assemble --scan assembled the array /dev/md2 with the disks
in the wrong order. I know because I physically saw where the disks were in
relation to the disk order and wrote down the disk order on every HD.
Here's everything I could find in terms of information that you asked for.
It's a lot.
#dmesg
[ 339.440187] raid5: device sdb4 operational as raid disk 1
[ 339.440189] raid5: device sdd4 operational as raid disk 3
[ 339.440192] raid5: device sdc4 operational as raid disk 2
[ 339.440610] raid5: allocated 4219kB for md2
[ 339.440612] raid5: raid level 5 set md2 active with 3 out of 4 devices,
algorithm 2
[ 339.440615] RAID5 conf printout:
[ 339.440617] --- rd:4 wd:3
[ 339.440619] disk 1, o:1, dev:sdb4
[ 339.440620] disk 2, o:1, dev:sdc4
[ 339.440622] disk 3, o:1, dev:sdd4
[ 339.440817] md2: unknown partition table
[ 538.840033] kjournald starting. Commit interval 5 seconds
[ 538.891844] EXT3 FS on md0, internal journal
[ 538.891849] EXT3-fs: mounted filesystem with ordered data mode.
[ 581.585031] VFS: Can't find ext4 filesystem on dev md2.
[ 587.056825] VFS: Can't find ext3 filesystem on dev md2.
#fdisk -l
WARNING: GPT (GUID Partition Table) detected on '/dev/sda'! The util fdisk
doesn't support GPT. Use GNU Parted.
Disk /dev/sda: 2000.3 GB, 2000398934016 bytes
255 heads, 63 sectors/track, 243201 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0x00000000
Device Boot Start End Blocks Id System
/dev/sda1 1 243202 1953514583+ ee GPT
Disk /dev/sdb: 2000.3 GB, 2000398934016 bytes
255 heads, 63 sectors/track, 243201 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0xdd07e5e3
Device Boot Start End Blocks Id System
/dev/sdb1 1 26 208844+ fd Linux raid
autodetect
/dev/sdb2 27 156 1044225 fd Linux raid
autodetect
/dev/sdb3 157 182 208845 fd Linux raid
autodetect
/dev/sdb4 183 243201 1952050117+ fd Linux raid
autodetect
Disk /dev/sdc: 2000.3 GB, 2000398934016 bytes
255 heads, 63 sectors/track, 243201 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0xdd07e5e4
Device Boot Start End Blocks Id System
/dev/sdc1 1 26 208844+ fd Linux raid
autodetect
/dev/sdc2 27 156 1044225 fd Linux raid
autodetect
/dev/sdc3 157 182 208845 fd Linux raid
autodetect
/dev/sdc4 183 243201 1952050117+ fd Linux raid
autodetect
Disk /dev/sdd: 2000.3 GB, 2000398934016 bytes
255 heads, 63 sectors/track, 243201 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0xdd07e5e2
Device Boot Start End Blocks Id System
/dev/sdd1 1 26 208844+ fd Linux raid
autodetect
/dev/sdd2 27 156 1044225 fd Linux raid
autodetect
/dev/sdd3 157 182 208845 fd Linux raid
autodetect
/dev/sdd4 183 243201 1952050117+ fd Linux raid
autodetect
Disk /dev/md0: 213 MB, 213778432 bytes
2 heads, 4 sectors/track, 52192 cylinders
Units = cylinders of 8 * 512 = 4096 bytes
Disk identifier: 0x00000000
Disk /dev/md0 doesn't contain a valid partition table
Disk /dev/md2: 5996.6 GB, 5996697747456 bytes
2 heads, 4 sectors/track, 1464037536 cylinders
Units = cylinders of 8 * 512 = 4096 bytes
Disk identifier: 0x00000000
Disk /dev/md2 doesn't contain a valid partition table
#cat /proc/mdstat
Personalities : [raid1] [raid6] [raid5] [raid4]
md2 : active raid5 sdb4[1] sdd4[3] sdc4[2]
5856150144 blocks level 5, 64k chunk, algorithm 2 [4/3] [_UUU]
md0 : active raid1 sdd1[0] sdb1[2] sdc1[1]
208768 blocks [4/3] [UUU_]
#cat /etc/fstab
aufs / aufs rw 0 0
tmpfs /tmp tmpfs nosuid,nodev 0 0
/dev/sda2 swap swap defaults 0 0
/dev/sdb2 swap swap defaults 0 0
/dev/sdc2 swap swap defaults 0 0
/dev/sdd2 swap swap defaults 0 0
#mount -t ext3 /dev/md2 /media/disk/shares
mount: wrong fs type, bad option, bad superblock on /dev/md2,
missing codepage or helper program, or other error
In some cases useful info is found in syslog - try
dmesg | tail or so
#mdadm -Ds -v
ARRAY /dev/md0 level=raid1 num-devices=4 metadata=00.90
UUID=15e54255:f58be7ca:7f4a592f:038fedf2
devices=/dev/sdd1,/dev/sdc1,/dev/sdb1
ARRAY /dev/md2 level=raid5 num-devices=4 metadata=00.90
UUID=0b23d5e1:f5a27618:e368bf24:bd0fce41
devices=/dev/sdb4,/dev/sdc4,/dev/sdd4
#mdadm -Es -v
ARRAY /dev/md0 level=raid1 num-devices=4
UUID=15e54255:f58be7ca:7f4a592f:038fedf2
devices=/dev/sdd1,/dev/sdc1,/dev/sdb1
ARRAY /dev/md1 level=raid1 num-devices=4
UUID=57cd5e76:0d56f114:50bd5336:4477d020
devices=/dev/sdd2,/dev/sdc2,/dev/sdb2
ARRAY /dev/md2 level=raid5 num-devices=4
UUID=0b23d5e1:f5a27618:e368bf24:bd0fce41
devices=/dev/sdd4,/dev/sdc4,/dev/sdb4
-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Robin Hill
Sent: Thursday, September 17, 2009 5:02 PM
To: linux-raid@vger.kernel.org
Subject: Re: Raid 5 Issue, cannot recognize EXT3 File system.
On Thu Sep 17, 2009 at 04:20:16PM -0400, Sunpyo Hong wrote:
> I've contacted just about everyone that knows a thing about RAID5, but no
> one is really able to help me. Anyhow I've read up a lot on RAID5 arrays
and
> how to properly assemble them. However I've run into a problem with a NAS
> system from WD that I just can't seem to figure out.
>
> I have a ¾ disks in the array, 1 went down and is out of commission. I've
> been able to assemble my array through mdadm using --assemble --scan.
However
> I cannot access the array due to the fact that the array cannot read a
> filesystem. Everytime I try to mount I get mount: wrong fs type
etc. I
> know that the FS is an ext3 FS. However I cannot seem to get this
> thing going. I was wondering if anyone could point me in the right
> direction with this. I can't seem to find anyone that is capable of
> solving this. I would appreciate any help. Thanks!
>
What's the output of 'cat /proc/mdstat' after you assemble the array?
And what exact error (and dmesg output) do you get when trying to mount
it as ext3?
Cheers,
Robin
--
___
( ' } | Robin Hill <robin@robinhill.me.uk> |
/ / ) | Little Jim says .... |
// !! | "He fallen in de water !!" |
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
next reply other threads:[~2009-09-17 21:54 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-09-17 21:54 Sunpyo Hong [this message]
2009-09-17 22:09 ` Raid 5 Issue, cannot recognize EXT3 File system Majed B.
2009-09-21 15:32 ` Sunpyo Hong
2009-09-21 15:56 ` Robin Hill
2009-09-21 16:14 ` Sunpyo Hong
2009-09-22 4:33 ` NeilBrown
2009-09-22 15:15 ` Sunpyo Hong
2009-09-22 15:23 ` Majed B.
2009-09-22 18:42 ` Sunpyo Hong
2009-09-23 0:14 ` Majed B.
2009-09-23 0:56 ` Guy Watkins
2009-09-23 13:56 ` Sunpyo Hong
2009-09-23 14:42 ` John Robinson
2009-09-23 15:14 ` Robin Hill
2009-09-23 15:50 ` Sunpyo Hong
2009-09-25 16:35 ` Sunpyo Hong
-- strict thread matches above, loose matches on Subject: below --
2009-09-17 20:20 Sunpyo Hong
2009-09-17 21:01 ` Robin Hill
2009-09-17 21:26 ` Majed B.
2009-09-17 21:46 ` Sunpyo Hong
2009-09-18 8:13 ` Robin Hill
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=200909171754312.SM06100@trainer \
--to=sunpyo.hong@amac.com \
--cc=linux-raid@vger.kernel.org \
--cc=robin@robinhill.me.uk \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).