linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* large filesystem corruptions
@ 2010-03-12 20:01 Kapetanakis Giannis
  2010-03-12 20:35 ` Ric Wheeler
  0 siblings, 1 reply; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-12 20:01 UTC (permalink / raw)
  To: linux-raid

Hi,

The last few days I'm trying to setup a large filesystem
14TB and is always getting corrupted...

This is my latest non-working setup. Sorry for long message
but I want to make clear my  actual setup

2.6.18-164.11.1.el5PAE (x86)
4GB RAM

CONFIG_EFI_VARS=y
CONFIG_EFI=y

/dev/sdb - 6 x 1.5TB SATA drives in hardware RAID 5 (256 chunk size)
/dev/sdc - 6 x 1.5TB SATA drives in hardware RAID 5 (256 chunk size)

-- Both hardware raids are GPT labeled

Model: Adaptec ARRAY01 (scsi)
Disk /dev/sdb: 7489GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt

Number  Start   End     Size    File system  Name     Flags
  1      17.4kB  7489GB  7489GB               ARRAY01

Model: Adaptec ARRAY02 (scsi)
Disk /dev/sdc: 7489GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt

Number  Start   End     Size    File system  Name     Flags
  1      17.4kB  7489GB  7489GB               ARRAY02

-- /dev/md_d0 software raid0 on /dev/sdb1 and /dev/sdc1
-- (doing on /dev/sdb and /dev/sdc also corrupted)
-- this raid0 is also GPT labeled (256 chunk size)

mdadm --create /dev/md_d0 -a p1 -c 256 -l 0 -n 2 /dev/sdb1 /dev/sdc1

   md_d0 : active raid0 sdb1[0] sdc1[1]
       14627614208 blocks 256k chunks

Model: Unknown (unknown)
Disk /dev/md_d0: 15.0TB
Sector size (logical/physical): 512B/512B
Partition Table: gpt

Number  Start   End     Size    File system  Name     Flags
  1      17.4kB  15.0TB  15.0TB               primary


-- LVM2 on top of /dev/md_d0p1

  --- Physical volume ---
   PV Name               /dev/md_d0p1
   VG Name               vgshare
   PV Size               13.62 TB / not usable 3.47 MB
   PE Size (KByte)       4096

   --- Volume group ---
   VG Name               vgshare
   System ID
   Format                lvm2
   VG Size               13.62 TB
   PE Size               4.00 MB

   --- Logical volume ---
   LV Name                /dev/vgshare/share
   VG Name                vgshare
   LV UUID                Aoj27F-kf4U-i6XE-eNWg-hMLX-MS1h-s3oArp
   LV Write Access        read/write
   LV Status              available
   # open                 0
   LV Size                7.00 TB
   Current LE             1835008
   Segments               1
   Allocation             inherit
   Read ahead sectors     auto
   - currently set to     2048
   Block device           253:4

-- /dev/vgshare/share is ext4 formated with
mkfs.ext4 -b 4096 -E stride=64,stripe-width=128 /dev/md_d0p1

All well so far. And I start write data...

Then I create a new 2TB test lv
and format it.

/share_7TB is lost, /test_2TB is there....

Mar 12 21:13:28 server kernel: EXT4-fs error (device dm-4): 
ext4_mb_generate_buddy: EXT4-fs: group 0: 32768 blocks in bitmap, 3248 in gd
Mar 12 21:13:30 server kernel: EXT4-fs error (device dm-4): 
ext4_mb_generate_buddy: EXT4-fs: group 1648: 24544 blocks in bitmap, 153 
in gd
Mar 12 21:13:31 server kernel: attempt to access beyond end of device
Mar 12 21:13:31 server kernel: dm-4: rw=2, want=15493450520, 
limit=15032385536
Mar 12 21:13:31 server kernel: attempt to access beyond end of device
--snip
Mar 12 21:17:49 server kernel: EXT4-fs error (device dm-4): 
ext4_mb_release_inode_pa: free 1802, pa_free 1458
Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 93430033 blocks 705745 
reqs (482734 success)
Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 327298 extents scanned, 
241152 goal hits, 219206 2^N hits, 0 breaks, 0 lost
Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 9561 generated and it 
took 2012010656
Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 85047925 preallocated, 
30759591 discarded
Mar 12 21:18:09 server kernel: EXT4-fs: ext4_check_descriptors: Inode 
table for group 0 not in group (block 1936681314)!
Mar 12 21:18:09 server kernel: EXT4-fs: group descriptors corrupted!

GFS instead of ext4 also corrupted.

ext4 on top of /dev/md0 (non partitioned) on top of /dev/sdb /dev/sdc 
(without GPT) also corrupted.

I want to use software raid0 on top of the two hardware radi5
for better performance.

I understood that labeling GPT would solve this problem.
Is it x86 problem? Something fishy with my setup probably
but can't figure it out.

thanks and sorry for long message,
but I can't find a way to get this mirror server up
on it's feet again after this upgrade....

regards,

Giannis







^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-12 20:01 large filesystem corruptions Kapetanakis Giannis
@ 2010-03-12 20:35 ` Ric Wheeler
  2010-03-13  0:29   ` Kapetanakis Giannis
  0 siblings, 1 reply; 12+ messages in thread
From: Ric Wheeler @ 2010-03-12 20:35 UTC (permalink / raw)
  To: Kapetanakis Giannis; +Cc: linux-raid

On 03/12/2010 03:01 PM, Kapetanakis Giannis wrote:
> Hi,
>
> The last few days I'm trying to setup a large filesystem
> 14TB and is always getting corrupted...
>
> This is my latest non-working setup. Sorry for long message
> but I want to make clear my  actual setup
>
> 2.6.18-164.11.1.el5PAE (x86)
> 4GB RAM
>
> CONFIG_EFI_VARS=y
> CONFIG_EFI=y
>
> /dev/sdb - 6 x 1.5TB SATA drives in hardware RAID 5 (256 chunk size)
> /dev/sdc - 6 x 1.5TB SATA drives in hardware RAID 5 (256 chunk size)
>
> -- Both hardware raids are GPT labeled
>
> Model: Adaptec ARRAY01 (scsi)
> Disk /dev/sdb: 7489GB
> Sector size (logical/physical): 512B/512B
> Partition Table: gpt
>
> Number  Start   End     Size    File system  Name     Flags
>  1      17.4kB  7489GB  7489GB               ARRAY01
>
> Model: Adaptec ARRAY02 (scsi)
> Disk /dev/sdc: 7489GB
> Sector size (logical/physical): 512B/512B
> Partition Table: gpt
>
> Number  Start   End     Size    File system  Name     Flags
>  1      17.4kB  7489GB  7489GB               ARRAY02
>
> -- /dev/md_d0 software raid0 on /dev/sdb1 and /dev/sdc1
> -- (doing on /dev/sdb and /dev/sdc also corrupted)
> -- this raid0 is also GPT labeled (256 chunk size)
>
> mdadm --create /dev/md_d0 -a p1 -c 256 -l 0 -n 2 /dev/sdb1 /dev/sdc1
>
>   md_d0 : active raid0 sdb1[0] sdc1[1]
>       14627614208 blocks 256k chunks
>
> Model: Unknown (unknown)
> Disk /dev/md_d0: 15.0TB
> Sector size (logical/physical): 512B/512B
> Partition Table: gpt
>
> Number  Start   End     Size    File system  Name     Flags
>  1      17.4kB  15.0TB  15.0TB               primary
>
>
> -- LVM2 on top of /dev/md_d0p1
>
>  --- Physical volume ---
>   PV Name               /dev/md_d0p1
>   VG Name               vgshare
>   PV Size               13.62 TB / not usable 3.47 MB
>   PE Size (KByte)       4096
>
>   --- Volume group ---
>   VG Name               vgshare
>   System ID
>   Format                lvm2
>   VG Size               13.62 TB
>   PE Size               4.00 MB
>
>   --- Logical volume ---
>   LV Name                /dev/vgshare/share
>   VG Name                vgshare
>   LV UUID                Aoj27F-kf4U-i6XE-eNWg-hMLX-MS1h-s3oArp
>   LV Write Access        read/write
>   LV Status              available
>   # open                 0
>   LV Size                7.00 TB
>   Current LE             1835008
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     2048
>   Block device           253:4
>
> -- /dev/vgshare/share is ext4 formated with
> mkfs.ext4 -b 4096 -E stride=64,stripe-width=128 /dev/md_d0p1
>
> All well so far. And I start write data...
>
> Then I create a new 2TB test lv
> and format it.
>
> /share_7TB is lost, /test_2TB is there....
>
> Mar 12 21:13:28 server kernel: EXT4-fs error (device dm-4): 
> ext4_mb_generate_buddy: EXT4-fs: group 0: 32768 blocks in bitmap, 3248 
> in gd
> Mar 12 21:13:30 server kernel: EXT4-fs error (device dm-4): 
> ext4_mb_generate_buddy: EXT4-fs: group 1648: 24544 blocks in bitmap, 
> 153 in gd
> Mar 12 21:13:31 server kernel: attempt to access beyond end of device
> Mar 12 21:13:31 server kernel: dm-4: rw=2, want=15493450520, 
> limit=15032385536
> Mar 12 21:13:31 server kernel: attempt to access beyond end of device
> --snip
> Mar 12 21:17:49 server kernel: EXT4-fs error (device dm-4): 
> ext4_mb_release_inode_pa: free 1802, pa_free 1458
> Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 93430033 blocks 
> 705745 reqs (482734 success)
> Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 327298 extents 
> scanned, 241152 goal hits, 219206 2^N hits, 0 breaks, 0 lost
> Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 9561 generated and it 
> took 2012010656
> Mar 12 21:17:49 server kernel: EXT4-fs: mballoc: 85047925 
> preallocated, 30759591 discarded
> Mar 12 21:18:09 server kernel: EXT4-fs: ext4_check_descriptors: Inode 
> table for group 0 not in group (block 1936681314)!
> Mar 12 21:18:09 server kernel: EXT4-fs: group descriptors corrupted!
>
> GFS instead of ext4 also corrupted.
>
> ext4 on top of /dev/md0 (non partitioned) on top of /dev/sdb /dev/sdc 
> (without GPT) also corrupted.
>
> I want to use software raid0 on top of the two hardware radi5
> for better performance.
>
> I understood that labeling GPT would solve this problem.
> Is it x86 problem? Something fishy with my setup probably
> but can't figure it out.
>
> thanks and sorry for long message,
> but I can't find a way to get this mirror server up
> on it's feet again after this upgrade....
>
> regards,
>
> Giannis
>

This is probably an issue with the early version of ext4 you are using - 
note that the support for ext4 > 16TB is still gated by some work done 
up in the tools chain.

Have you tried xfs?

regards,

Ric


^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-12 20:35 ` Ric Wheeler
@ 2010-03-13  0:29   ` Kapetanakis Giannis
  2010-03-13  0:55     ` Kapetanakis Giannis
  2010-03-13 23:45     ` Bill Davidsen
  0 siblings, 2 replies; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-13  0:29 UTC (permalink / raw)
  To: Ric Wheeler; +Cc: linux-raid

On 12/03/10 22:35, Ric Wheeler wrote:

> This is probably an issue with the early version of ext4 you are using -
> note that the support for ext4 > 16TB is still gated by some work done
> up in the tools chain.
>
> Have you tried xfs?
>
> regards,
>
> Ric

Thanks for answering,

My filesystem would be 15 TB < 16 TB.
GFS also crashed and burned so are you sure that this
this is a problem with ext4?
Would Fedora and newer kernel be better?

On my tests the crashing filesystem was 7TB.
When I added a second fs of 2 TB on lvm, so total > 8TB
I had the crash.

I did a new test now and didn't use GFT partitions
but the whole physical/logical drives

sdb -
     | ---> md0 ---> LVM ---> ext4 filesystems
sdc -

all sdb, sdc, md0 are gpt labeled without gpt partitions
inside. No crash so far but without any data written.

Maybe the gpt partitions did the bad thing?
Can md0 use large gpt drives with no partitions?
can lvm2 use large raid device with no partition pv?

I could try XFS but i'm not familiar with it,
so I wouldn't know the optimized values for
such a large fs

regards,

Giannis


^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13  0:29   ` Kapetanakis Giannis
@ 2010-03-13  0:55     ` Kapetanakis Giannis
  2010-03-13  1:58       ` Michael Evans
  2010-03-13 23:45     ` Bill Davidsen
  1 sibling, 1 reply; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-13  0:55 UTC (permalink / raw)
  To: Ric Wheeler; +Cc: linux-raid

On 13/03/10 02:29, Kapetanakis Giannis wrote:
> I did a new test now and didn't use GFT partitions
> but the whole physical/logical drives
>
> sdb -
> | ---> md0 ---> LVM ---> ext4 filesystems
> sdc -
>
> all sdb, sdc, md0 are gpt labeled without gpt partitions
> inside. No crash so far but without any data written.
>
> Maybe the gpt partitions did the bad thing?
> Can md0 use large gpt drives with no partitions?
> can lvm2 use large raid device with no partition pv?

crashed and burned also:

Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4): 
ext4_mb_generate_buddy: EXT4-fs: group 48: 24544 blocks in bitmap, 2016 
in gd
Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4): 
mb_free_blocks: double-free of inode 12's block 1583104(bit 10240 in 
group 48)
Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4): 
mb_free_blocks: double-free of inode 12's block 1583105(bit 10241 in 
group 48)
--snip

so gpt partitions was not a problem.

Next in list: XFS

    682  2:47    mkfs.xfs -f /dev/vgshare/share
    684  2:47    mount /dev/vgshare/share /share/
    686  2:47    mkfs.xfs -f /dev/vgshare/test
    687  2:47    mount /dev/vgshare/test /test/
    689  2:47    cd /share/
    691  2:48    dd if=/dev/zero of=papaki bs=4096

Mar 13 02:47:23 server kernel: Filesystem "dm-4": Disabling barriers, 
not supported by the underlying device
Mar 13 02:47:23 server kernel: XFS mounting filesystem dm-4
Mar 13 02:47:48 server kernel: Filesystem "dm-5": Disabling barriers, 
not supported by the underlying device
Mar 13 02:47:48 server kernel: XFS mounting filesystem dm-5
Mar 13 02:48:05 server kernel: Filesystem "dm-4": XFS internal error 
xfs_trans_cancel at line 1138 of file 
/home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_trans.c. 
  Caller 0xf90e0bbc
Mar 13 02:48:05 server kernel:  [<f90d85fe>] xfs_trans_cancel+0x4d/0xd6 
[xfs]
Mar 13 02:48:05 server kernel:  [<f90e0bbc>] xfs_create+0x4ec/0x525 [xfs]
Mar 13 02:48:05 server kernel:  [<f90e0bbc>] xfs_create+0x4ec/0x525 [xfs]
Mar 13 02:48:05 server kernel:  [<f90e88f4>] xfs_vn_mknod+0x19c/0x380 [xfs]
Mar 13 02:48:05 server kernel:  [<c04760e9>] __getblk+0x30/0x27a
Mar 13 02:48:05 server kernel:  [<f8852ac7>] 
do_get_write_access+0x441/0x46e [jbd]
Mar 13 02:48:05 server kernel:  [<f8889502>] 
__ext3_get_inode_loc+0x109/0x2d5 [ext3]
Mar 13 02:48:05 server kernel:  [<c045a7aa>] 
get_page_from_freelist+0x96/0x370
Mar 13 02:48:05 server kernel:  [<f90b6827>] xfs_dir_lookup+0x91/0xff [xfs]
Mar 13 02:48:05 server kernel:  [<f90c3c51>] xfs_iunlock+0x51/0x6d [xfs]
Mar 13 02:48:05 server kernel:  [<c04824f0>] __link_path_walk+0xc62/0xd33
Mar 13 02:48:05 server kernel:  [<c0480b43>] vfs_create+0xc8/0x12f
Mar 13 02:48:05 server kernel:  [<c04834ef>] open_namei+0x16a/0x5fb
Mar 13 02:48:05 server kernel:  [<c0472a92>] __dentry_open+0xea/0x1ab
Mar 13 02:48:05 server kernel:  [<c0472be2>] do_filp_open+0x1c/0x31
Mar 13 02:48:05 server kernel:  [<c0472c35>] do_sys_open+0x3e/0xae
Mar 13 02:48:05 server kernel:  [<c0472cd2>] sys_open+0x16/0x18
Mar 13 02:48:05 server kernel:  [<c0404f17>] syscall_call+0x7/0xb
Mar 13 02:48:05 server kernel:  =======================
Mar 13 02:48:05 server kernel: xfs_force_shutdown(dm-4,0x8) called from 
line 1139 of file 
/home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_trans.c. 
  Return address = 0xf90eb6c4
Mar 13 02:48:05 server kernel: Filesystem "dm-4": Corruption of 
in-memory data detected.  Shutting down filesystem: dm-4
Mar 13 02:48:05 server kernel: Please umount the filesystem, and rectify 
the problem(s)
Mar 13 02:48:45 server kernel: xfs_force_shutdown(dm-4,0x1) called from 
line 424 of file 
/home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_rw.c. 
Return address = 0xf90eb6c4
Mar 13 02:48:45 server kernel: xfs_force_shutdown(dm-4,0x1) called from 
line 424 of file 
/home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_rw.c. 
Return address = 0xf90eb6c4

xfs_check /dev/vgshare/share
XFS: Log inconsistent (didn't find previous header)
XFS: failed to find log head
ERROR: cannot find log head/tail, run xfs_repair

xfs_repair /dev/vgshare/share
Phase 1 - find and verify superblock...
bad primary superblock - filesystem mkfs-in-progress bit set !!!

attempting to find secondary superblock...
...................................

I stopped it, can't wait to search 7TB to find the secondary 
superblock...probably won't find anything

/test works

So are we sure it's the fs?
Something else is fishy...

regards,

Giannis

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13  0:55     ` Kapetanakis Giannis
@ 2010-03-13  1:58       ` Michael Evans
  2010-03-13  8:12         ` Kapetanakis Giannis
                           ` (2 more replies)
  0 siblings, 3 replies; 12+ messages in thread
From: Michael Evans @ 2010-03-13  1:58 UTC (permalink / raw)
  To: Kapetanakis Giannis; +Cc: Ric Wheeler, linux-raid

On Fri, Mar 12, 2010 at 4:55 PM, Kapetanakis Giannis
<bilias@edu.physics.uoc.gr> wrote:
> On 13/03/10 02:29, Kapetanakis Giannis wrote:
>>
>> I did a new test now and didn't use GFT partitions
>> but the whole physical/logical drives
>>
>> sdb -
>> | ---> md0 ---> LVM ---> ext4 filesystems
>> sdc -
>>
>> all sdb, sdc, md0 are gpt labeled without gpt partitions
>> inside. No crash so far but without any data written.
>>
>> Maybe the gpt partitions did the bad thing?
>> Can md0 use large gpt drives with no partitions?
>> can lvm2 use large raid device with no partition pv?
>
> crashed and burned also:
>
> Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4):
> ext4_mb_generate_buddy: EXT4-fs: group 48: 24544 blocks in bitmap, 2016 in
> gd
> Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4): mb_free_blocks:
> double-free of inode 12's block 1583104(bit 10240 in group 48)
> Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4): mb_free_blocks:
> double-free of inode 12's block 1583105(bit 10241 in group 48)
> --snip
>
> so gpt partitions was not a problem.
>
> Next in list: XFS
>
>   682  2:47    mkfs.xfs -f /dev/vgshare/share
>   684  2:47    mount /dev/vgshare/share /share/
>   686  2:47    mkfs.xfs -f /dev/vgshare/test
>   687  2:47    mount /dev/vgshare/test /test/
>   689  2:47    cd /share/
>   691  2:48    dd if=/dev/zero of=papaki bs=4096
>
> Mar 13 02:47:23 server kernel: Filesystem "dm-4": Disabling barriers, not
> supported by the underlying device
> Mar 13 02:47:23 server kernel: XFS mounting filesystem dm-4
> Mar 13 02:47:48 server kernel: Filesystem "dm-5": Disabling barriers, not
> supported by the underlying device
> Mar 13 02:47:48 server kernel: XFS mounting filesystem dm-5
> Mar 13 02:48:05 server kernel: Filesystem "dm-4": XFS internal error
> xfs_trans_cancel at line 1138 of file
> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_trans.c.
>  Caller 0xf90e0bbc
> Mar 13 02:48:05 server kernel:  [<f90d85fe>] xfs_trans_cancel+0x4d/0xd6
> [xfs]
> Mar 13 02:48:05 server kernel:  [<f90e0bbc>] xfs_create+0x4ec/0x525 [xfs]
> Mar 13 02:48:05 server kernel:  [<f90e0bbc>] xfs_create+0x4ec/0x525 [xfs]
> Mar 13 02:48:05 server kernel:  [<f90e88f4>] xfs_vn_mknod+0x19c/0x380 [xfs]
> Mar 13 02:48:05 server kernel:  [<c04760e9>] __getblk+0x30/0x27a
> Mar 13 02:48:05 server kernel:  [<f8852ac7>] do_get_write_access+0x441/0x46e
> [jbd]
> Mar 13 02:48:05 server kernel:  [<f8889502>]
> __ext3_get_inode_loc+0x109/0x2d5 [ext3]
> Mar 13 02:48:05 server kernel:  [<c045a7aa>]
> get_page_from_freelist+0x96/0x370
> Mar 13 02:48:05 server kernel:  [<f90b6827>] xfs_dir_lookup+0x91/0xff [xfs]
> Mar 13 02:48:05 server kernel:  [<f90c3c51>] xfs_iunlock+0x51/0x6d [xfs]
> Mar 13 02:48:05 server kernel:  [<c04824f0>] __link_path_walk+0xc62/0xd33
> Mar 13 02:48:05 server kernel:  [<c0480b43>] vfs_create+0xc8/0x12f
> Mar 13 02:48:05 server kernel:  [<c04834ef>] open_namei+0x16a/0x5fb
> Mar 13 02:48:05 server kernel:  [<c0472a92>] __dentry_open+0xea/0x1ab
> Mar 13 02:48:05 server kernel:  [<c0472be2>] do_filp_open+0x1c/0x31
> Mar 13 02:48:05 server kernel:  [<c0472c35>] do_sys_open+0x3e/0xae
> Mar 13 02:48:05 server kernel:  [<c0472cd2>] sys_open+0x16/0x18
> Mar 13 02:48:05 server kernel:  [<c0404f17>] syscall_call+0x7/0xb
> Mar 13 02:48:05 server kernel:  =======================
> Mar 13 02:48:05 server kernel: xfs_force_shutdown(dm-4,0x8) called from line
> 1139 of file
> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_trans.c.
>  Return address = 0xf90eb6c4
> Mar 13 02:48:05 server kernel: Filesystem "dm-4": Corruption of in-memory
> data detected.  Shutting down filesystem: dm-4
> Mar 13 02:48:05 server kernel: Please umount the filesystem, and rectify the
> problem(s)
> Mar 13 02:48:45 server kernel: xfs_force_shutdown(dm-4,0x1) called from line
> 424 of file
> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_rw.c. Return
> address = 0xf90eb6c4
> Mar 13 02:48:45 server kernel: xfs_force_shutdown(dm-4,0x1) called from line
> 424 of file
> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_rw.c. Return
> address = 0xf90eb6c4
>
> xfs_check /dev/vgshare/share
> XFS: Log inconsistent (didn't find previous header)
> XFS: failed to find log head
> ERROR: cannot find log head/tail, run xfs_repair
>
> xfs_repair /dev/vgshare/share
> Phase 1 - find and verify superblock...
> bad primary superblock - filesystem mkfs-in-progress bit set !!!
>
> attempting to find secondary superblock...
> ...................................
>
> I stopped it, can't wait to search 7TB to find the secondary
> superblock...probably won't find anything
>
> /test works
>
> So are we sure it's the fs?
> Something else is fishy...
>
> regards,
>
> Giannis
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

This is a really basic thing, but do you have the x86 support for very
large block devices (I can't remember what the option is, since I've
been running 64 bits on any system that even remotely came close to
needing it anyway) enabled in the config as well?

Here's a hit from google, CONFIG_LBD http://cateee.net/lkddb/web-lkddb/LBD.html

Enable block devices of size 2TB and larger.

Since you're using a device >2TB in size, I will assume you are using
one of the three 'version 1' superblock types.  Either at the end 1.0,
beginning 1.1 or 4kb in from the beginning.

Please provide the full output of mdadm -Dvvs

You can use any block device as a member of an md array.  However if
you are going 'whole drive' then it would be a very good idea to erase
the existing partition table structure prior to putting a raid
superblock on the device.  This way there is no confusion about if the
device has partitions or is in fact a raid member.  Similarly when
transitioning back the other way ensuring that the old metadata for
the array is erased is also a good idea.

The kernel you're running seems to be ... exceptionally old and
heavily patched.  I have no way of knowing if the many, many, patches
that fixed numerous issues over the /years/ since it's release have
been included.  Please make sure you have the most recent release from
your vendor and ask them for support in parallel.
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13  1:58       ` Michael Evans
@ 2010-03-13  8:12         ` Kapetanakis Giannis
  2010-03-13  9:25         ` Kapetanakis Giannis
  2010-03-13 13:07         ` Ric Wheeler
  2 siblings, 0 replies; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-13  8:12 UTC (permalink / raw)
  To: Michael Evans; +Cc: linux-raid

On 13/03/10 03:58, Michael Evans wrote:

> This is a really basic thing, but do you have the x86 support for very
> large block devices (I can't remember what the option is, since I've
> been running 64 bits on any system that even remotely came close to
> needing it anyway) enabled in the config as well?
>
> Here's a hit from google, CONFIG_LBD http://cateee.net/lkddb/web-lkddb/LBD.html
>
> Enable block devices of size 2TB and larger.

Yes I have LBD support
grep LBD /boot/config-2.6.18-164.11.1.el5PAE
CONFIG_LBD=y


> Since you're using a device>2TB in size, I will assume you are using
> one of the three 'version 1' superblock types.  Either at the end 1.0,
> beginning 1.1 or 4kb in from the beginning.
>
> Please provide the full output of mdadm -Dvvs

If you mean metadata then I'm at default -> 0.90
Is this the problem? I 've seen in the manual that
2 TB is the limit for raid 1 and above

"0, 0.90, default:   Use the original 0.90  format  superblock.   This 
format limits  arrays  to 28 component devices and limits component 
devices of levels 1 and greater to 2 terabytes."


[root@server ~]# mdadm -Dvvs
mdadm: bad uuid: UUID=324587ca:484d94c7:f06cbaee:5b63cd3
/dev/md0:
         Version : 0.90
   Creation Time : Sat Mar 13 02:00:23 2010
      Raid Level : raid0
      Array Size : 14627614208 (13949.98 GiB 14978.68 GB)
    Raid Devices : 2
   Total Devices : 2
Preferred Minor : 0
     Persistence : Superblock is persistent

     Update Time : Sat Mar 13 02:00:23 2010
           State : clean
  Active Devices : 2
Working Devices : 2
  Failed Devices : 0
   Spare Devices : 0

      Chunk Size : 256K

            UUID : 324587ca:484d94c7:f06cbaee:5b63cd37
          Events : 0.1

     Number   Major   Minor   RaidDevice State
        0       8       16        0      active sync   /dev/sdb
        1       8       32        1      active sync   /dev/sdc



> You can use any block device as a member of an md array.  However if
> you are going 'whole drive' then it would be a very good idea to erase
> the existing partition table structure prior to putting a raid
> superblock on the device.  This way there is no confusion about if the
> device has partitions or is in fact a raid member.  Similarly when
> transitioning back the other way ensuring that the old metadata for
> the array is erased is also a good idea.

I have erased prior creating the gpt and the raid devide
dd if=/dev/zero of=/dev/sdb bs=512 count=64
dd if=/dev/zero of=/dev/sdc bs=512 count=64

(I accidentally erased my boot disk also but I managed to recover)

> The kernel you're running seems to be ... exceptionally old and
> heavily patched.  I have no way of knowing if the many, many, patches
> that fixed numerous issues over the /years/ since it's release have
> been included.  Please make sure you have the most recent release from
> your vendor and ask them for support in parallel.

This is centos 5.4 stock kernel. So this is redhat 5.4 stock kernel
They say they support all these...

thanks
Giannis

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13  1:58       ` Michael Evans
  2010-03-13  8:12         ` Kapetanakis Giannis
@ 2010-03-13  9:25         ` Kapetanakis Giannis
  2010-03-13 13:07         ` Ric Wheeler
  2 siblings, 0 replies; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-13  9:25 UTC (permalink / raw)
  To: Michael Evans; +Cc: linux-raid

On 13/03/10 03:58, Michael Evans wrote:

> Since you're using a device>2TB in size, I will assume you are using
> one of the three 'version 1' superblock types.  Either at the end 1.0,
> beginning 1.1 or 4kb in from the beginning.
>
> Please provide the full output of mdadm -Dvvs

I did a test with 1.1 metadata version.
Crashed again when I deleted the file from the large FS.

I did a mistake in my previous mail.
Kernel is not stock. It's the latest kernel from centos/redhat 5.4
System up2date.

I will open a ticket in Centos and Redhat bugzilla
and see how it goes.

Giannis
ps. Probably install fedora 12 64bit on Monday


^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13  1:58       ` Michael Evans
  2010-03-13  8:12         ` Kapetanakis Giannis
  2010-03-13  9:25         ` Kapetanakis Giannis
@ 2010-03-13 13:07         ` Ric Wheeler
  2010-03-13 13:19           ` Kapetanakis Giannis
  2 siblings, 1 reply; 12+ messages in thread
From: Ric Wheeler @ 2010-03-13 13:07 UTC (permalink / raw)
  To: Michael Evans, Kapetanakis Giannis; +Cc: linux-raid

On 03/12/2010 08:58 PM, Michael Evans wrote:
> On Fri, Mar 12, 2010 at 4:55 PM, Kapetanakis Giannis
> <bilias@edu.physics.uoc.gr>  wrote:
>    
>> On 13/03/10 02:29, Kapetanakis Giannis wrote:
>>      
>>> I did a new test now and didn't use GFT partitions
>>> but the whole physical/logical drives
>>>
>>> sdb -
>>> | --->  md0 --->  LVM --->  ext4 filesystems
>>> sdc -
>>>
>>> all sdb, sdc, md0 are gpt labeled without gpt partitions
>>> inside. No crash so far but without any data written.
>>>
>>> Maybe the gpt partitions did the bad thing?
>>> Can md0 use large gpt drives with no partitions?
>>> can lvm2 use large raid device with no partition pv?
>>>        
>> crashed and burned also:
>>
>> Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4):
>> ext4_mb_generate_buddy: EXT4-fs: group 48: 24544 blocks in bitmap, 2016 in
>> gd
>> Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4): mb_free_blocks:
>> double-free of inode 12's block 1583104(bit 10240 in group 48)
>> Mar 13 02:40:28 server kernel: EXT4-fs error (device dm-4): mb_free_blocks:
>> double-free of inode 12's block 1583105(bit 10241 in group 48)
>> --snip
>>
>> so gpt partitions was not a problem.
>>
>> Next in list: XFS
>>
>>    682  2:47    mkfs.xfs -f /dev/vgshare/share
>>    684  2:47    mount /dev/vgshare/share /share/
>>    686  2:47    mkfs.xfs -f /dev/vgshare/test
>>    687  2:47    mount /dev/vgshare/test /test/
>>    689  2:47    cd /share/
>>    691  2:48    dd if=/dev/zero of=papaki bs=4096
>>
>> Mar 13 02:47:23 server kernel: Filesystem "dm-4": Disabling barriers, not
>> supported by the underlying device
>> Mar 13 02:47:23 server kernel: XFS mounting filesystem dm-4
>> Mar 13 02:47:48 server kernel: Filesystem "dm-5": Disabling barriers, not
>> supported by the underlying device
>> Mar 13 02:47:48 server kernel: XFS mounting filesystem dm-5
>> Mar 13 02:48:05 server kernel: Filesystem "dm-4": XFS internal error
>> xfs_trans_cancel at line 1138 of file
>> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_trans.c.
>>   Caller 0xf90e0bbc
>> Mar 13 02:48:05 server kernel:  [<f90d85fe>] xfs_trans_cancel+0x4d/0xd6
>> [xfs]
>> Mar 13 02:48:05 server kernel:  [<f90e0bbc>] xfs_create+0x4ec/0x525 [xfs]
>> Mar 13 02:48:05 server kernel:  [<f90e0bbc>] xfs_create+0x4ec/0x525 [xfs]
>> Mar 13 02:48:05 server kernel:  [<f90e88f4>] xfs_vn_mknod+0x19c/0x380 [xfs]
>> Mar 13 02:48:05 server kernel:  [<c04760e9>] __getblk+0x30/0x27a
>> Mar 13 02:48:05 server kernel:  [<f8852ac7>] do_get_write_access+0x441/0x46e
>> [jbd]
>> Mar 13 02:48:05 server kernel:  [<f8889502>]
>> __ext3_get_inode_loc+0x109/0x2d5 [ext3]
>> Mar 13 02:48:05 server kernel:  [<c045a7aa>]
>> get_page_from_freelist+0x96/0x370
>> Mar 13 02:48:05 server kernel:  [<f90b6827>] xfs_dir_lookup+0x91/0xff [xfs]
>> Mar 13 02:48:05 server kernel:  [<f90c3c51>] xfs_iunlock+0x51/0x6d [xfs]
>> Mar 13 02:48:05 server kernel:  [<c04824f0>] __link_path_walk+0xc62/0xd33
>> Mar 13 02:48:05 server kernel:  [<c0480b43>] vfs_create+0xc8/0x12f
>> Mar 13 02:48:05 server kernel:  [<c04834ef>] open_namei+0x16a/0x5fb
>> Mar 13 02:48:05 server kernel:  [<c0472a92>] __dentry_open+0xea/0x1ab
>> Mar 13 02:48:05 server kernel:  [<c0472be2>] do_filp_open+0x1c/0x31
>> Mar 13 02:48:05 server kernel:  [<c0472c35>] do_sys_open+0x3e/0xae
>> Mar 13 02:48:05 server kernel:  [<c0472cd2>] sys_open+0x16/0x18
>> Mar 13 02:48:05 server kernel:  [<c0404f17>] syscall_call+0x7/0xb
>> Mar 13 02:48:05 server kernel:  =======================
>> Mar 13 02:48:05 server kernel: xfs_force_shutdown(dm-4,0x8) called from line
>> 1139 of file
>> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_trans.c.
>>   Return address = 0xf90eb6c4
>> Mar 13 02:48:05 server kernel: Filesystem "dm-4": Corruption of in-memory
>> data detected.  Shutting down filesystem: dm-4
>> Mar 13 02:48:05 server kernel: Please umount the filesystem, and rectify the
>> problem(s)
>> Mar 13 02:48:45 server kernel: xfs_force_shutdown(dm-4,0x1) called from line
>> 424 of file
>> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_rw.c. Return
>> address = 0xf90eb6c4
>> Mar 13 02:48:45 server kernel: xfs_force_shutdown(dm-4,0x1) called from line
>> 424 of file
>> /home/buildsvn/rpmbuild/BUILD/xfs-kmod-0.4/_kmod_build_PAE/xfs_rw.c. Return
>> address = 0xf90eb6c4
>>
>> xfs_check /dev/vgshare/share
>> XFS: Log inconsistent (didn't find previous header)
>> XFS: failed to find log head
>> ERROR: cannot find log head/tail, run xfs_repair
>>
>> xfs_repair /dev/vgshare/share
>> Phase 1 - find and verify superblock...
>> bad primary superblock - filesystem mkfs-in-progress bit set !!!
>>
>> attempting to find secondary superblock...
>> ...................................
>>
>> I stopped it, can't wait to search 7TB to find the secondary
>> superblock...probably won't find anything
>>
>> /test works
>>
>> So are we sure it's the fs?
>> Something else is fishy...
>>
>> regards,
>>
>> Giannis
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>      
> This is a really basic thing, but do you have the x86 support for very
> large block devices (I can't remember what the option is, since I've
> been running 64 bits on any system that even remotely came close to
> needing it anyway) enabled in the config as well?
>
> Here's a hit from google, CONFIG_LBD http://cateee.net/lkddb/web-lkddb/LBD.html
>
> Enable block devices of size 2TB and larger.
>
> Since you're using a device>2TB in size, I will assume you are using
> one of the three 'version 1' superblock types.  Either at the end 1.0,
> beginning 1.1 or 4kb in from the beginning.
>
> Please provide the full output of mdadm -Dvvs
>
> You can use any block device as a member of an md array.  However if
> you are going 'whole drive' then it would be a very good idea to erase
> the existing partition table structure prior to putting a raid
> superblock on the device.  This way there is no confusion about if the
> device has partitions or is in fact a raid member.  Similarly when
> transitioning back the other way ensuring that the old metadata for
> the array is erased is also a good idea.
>
> The kernel you're running seems to be ... exceptionally old and
> heavily patched.  I have no way of knowing if the many, many, patches
> that fixed numerous issues over the /years/ since it's release have
> been included.  Please make sure you have the most recent release from
> your vendor and ask them for support in parallel.
>    

I would agree that it would be key to try this on a newer kernel & on a 
64 bit box. If you have an issue with a specific vendor release, you 
should open a ticket/bugzilla with that vendor so they can help you 
figure this out.

ric


^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13 13:07         ` Ric Wheeler
@ 2010-03-13 13:19           ` Kapetanakis Giannis
  2010-03-16 10:02             ` Kapetanakis Giannis
  0 siblings, 1 reply; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-13 13:19 UTC (permalink / raw)
  To: Ric Wheeler; +Cc: linux-raid

On 13/03/10 15:07, Ric Wheeler wrote:
> I would agree that it would be key to try this on a newer kernel & on 
> a 64 bit box. If you have an issue with a specific vendor release, you 
> should open a ticket/bugzilla with that vendor so they can help you 
> figure this out.
>
> ric

I will try Fedora 12/64bit on Monday when I have physical access.

Right now all tests have failed.
OS in Centos 5.4 x86 PAE so I opened a ticket both in Centos and Redhat 
bugzilla.
I don't have any contract with redhat but I hope they will give a try on 
this.

https://bugzilla.redhat.com/show_bug.cgi?id=573185
http://bugs.centos.org/view.php?id=4239

My feelings are that it has to do with the software raid layer and not
the filesystem (ext4, xfs, gfs all died).
The filesystem seems to not get the appropriate physical/logical
quandaries from the raid layer.

Unless the x86 kernel is not capable of addressing
large filesystems as it is documented (16TB as you said).

best regards,

Giannis


^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13  0:29   ` Kapetanakis Giannis
  2010-03-13  0:55     ` Kapetanakis Giannis
@ 2010-03-13 23:45     ` Bill Davidsen
  2010-03-14  3:26       ` Kapetanakis Giannis
  1 sibling, 1 reply; 12+ messages in thread
From: Bill Davidsen @ 2010-03-13 23:45 UTC (permalink / raw)
  To: Kapetanakis Giannis; +Cc: Ric Wheeler, linux-raid

Kapetanakis Giannis wrote:
> On 12/03/10 22:35, Ric Wheeler wrote:
>
>> This is probably an issue with the early version of ext4 you are using -
>> note that the support for ext4 > 16TB is still gated by some work done
>> up in the tools chain.
>>
>> Have you tried xfs?
>>
>> regards,
>>
>> Ric
>
> Thanks for answering,
>
> My filesystem would be 15 TB < 16 TB.
> GFS also crashed and burned so are you sure that this
> this is a problem with ext4?
> Would Fedora and newer kernel be better?
>
> On my tests the crashing filesystem was 7TB.
> When I added a second fs of 2 TB on lvm, so total > 8TB
> I had the crash.
>
> I did a new test now and didn't use GFT partitions
> but the whole physical/logical drives
>
> sdb -
>     | ---> md0 ---> LVM ---> ext4 filesystems
> sdc -
>
> all sdb, sdc, md0 are gpt labeled without gpt partitions
> inside. No crash so far but without any data written.
>
> Maybe the gpt partitions did the bad thing?
> Can md0 use large gpt drives with no partitions?
> can lvm2 use large raid device with no partition pv?

I did not notice the LVM on your first post, it's far enough down that I 
missed it. I would be suspicious that the problem is in LVM rather than 
md, but if you're going for a new install soon it's probably not worth 
checking.

If you have video issues with FC12 you may need to run the vesa driver 
as a quick solution. There have been learning experiences in the video 
drivers in recent kernels, not just Fedora.

-- 
Bill Davidsen <davidsen@tmr.com>
  "We can't solve today's problems by using the same thinking we
   used in creating them." - Einstein


^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13 23:45     ` Bill Davidsen
@ 2010-03-14  3:26       ` Kapetanakis Giannis
  0 siblings, 0 replies; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-14  3:26 UTC (permalink / raw)
  To: linux-raid

On 14/03/10 01:45, Bill Davidsen wrote:

> I did not notice the LVM on your first post, it's far enough down that I
> missed it. I would be suspicious that the problem is in LVM rather than
> md, but if you're going for a new install soon it's probably not worth
> checking.
>
> If you have video issues with FC12 you may need to run the vesa driver
> as a quick solution. There have been learning experiences in the video
> drivers in recent kernels, not just Fedora.

No it's not the LVM layer.
I have crashes without LVM as well.

gpt label is also corrupted, so my guesses go
to raid layer (that's why i posted here in the first place)
or maybe the driver-adaptec.

Giannis


^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: large filesystem corruptions
  2010-03-13 13:19           ` Kapetanakis Giannis
@ 2010-03-16 10:02             ` Kapetanakis Giannis
  0 siblings, 0 replies; 12+ messages in thread
From: Kapetanakis Giannis @ 2010-03-16 10:02 UTC (permalink / raw)
  To: linux-raid; +Cc: Ric Wheeler, davidsen

On 13/03/10 15:19, Kapetanakis Giannis wrote:

> I will try Fedora 12/64bit on Monday when I have physical access.
>
> Right now all tests have failed.
> OS in Centos 5.4 x86 PAE so I opened a ticket both in Centos and Redhat
> bugzilla.
> I don't have any contract with redhat but I hope they will give a try on
> this.
>
> https://bugzilla.redhat.com/show_bug.cgi?id=573185
> http://bugs.centos.org/view.php?id=4239
>
> My feelings are that it has to do with the software raid layer and not
> the filesystem (ext4, xfs, gfs all died).
> The filesystem seems to not get the appropriate physical/logical
> quandaries from the raid layer.
>
> Unless the x86 kernel is not capable of addressing
> large filesystems as it is documented (16TB as you said).

I thought I should report here as well.

I've installed Centos 5.4 x64 and all going well so far.
No corruptions at all.

My setup is as before

sdb (hard raid5)
  |
  |------> md0 (soft raid0)----> LVM ----> ext4
  |
sdc (hard raid5)

The only strange thing I've noticed is that pvcreate
is messing with the gpt label on md0. Don't know of that
is normal.

https://bugzilla.redhat.com/show_bug.cgi?id=573185#c4

#fdisk -l /dev/md0

WARNING: GPT (GUID Partition Table) detected on '/dev/md0'! The util fdisk
doesn't support GPT. Use GNU Parted.


Disk /dev/md0: 14978.6 GB, 14978676948992 bytes
2 heads, 4 sectors/track, -638063744 cylinders
Units = cylinders of 8 * 512 = 4096 bytes

Disk /dev/md0 doesn't contain a valid partition table

# parted /dev/md0
GNU Parted 1.8.1
Using /dev/md0
Welcome to GNU Parted! Type 'help' to view a list of commands.
(parted) p
Error: Unable to open /dev/md0 - unrecognised disk label.

thanks all

Giannis

^ permalink raw reply	[flat|nested] 12+ messages in thread

end of thread, other threads:[~2010-03-16 10:02 UTC | newest]

Thread overview: 12+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2010-03-12 20:01 large filesystem corruptions Kapetanakis Giannis
2010-03-12 20:35 ` Ric Wheeler
2010-03-13  0:29   ` Kapetanakis Giannis
2010-03-13  0:55     ` Kapetanakis Giannis
2010-03-13  1:58       ` Michael Evans
2010-03-13  8:12         ` Kapetanakis Giannis
2010-03-13  9:25         ` Kapetanakis Giannis
2010-03-13 13:07         ` Ric Wheeler
2010-03-13 13:19           ` Kapetanakis Giannis
2010-03-16 10:02             ` Kapetanakis Giannis
2010-03-13 23:45     ` Bill Davidsen
2010-03-14  3:26       ` Kapetanakis Giannis

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).