* raid50 and 9TB volumes @ 2007-07-16 12:42 Raz 2007-07-16 13:01 ` David Chinner 0 siblings, 1 reply; 15+ messages in thread From: Raz @ 2007-07-16 12:42 UTC (permalink / raw) To: linux-xfs Hello I found that using xfs over raid50, ( two raid5's 8 disks each and raid 0 over them ) crashes the file system when the file system is ~ 9TB. crashing is easy: we simply create few hundred of files, then erase them in bulk. the same test passes in 6.4TB filesystems. this bug happens in 2.6.22 as well as 2.6.17.7. thank you . 4391322.839000] Filesystem "md3": XFS internal error xfs_alloc_read_agf at line 2176 of file fs/xfs/xfs_alloc.c. Caller 0xc10d31ea [4391322.863000] <c10d36e9> xfs_alloc_read_agf+0x199/0x220 <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 [4391322.882000] <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 [4391322.901000] <c1114294> xfs_iext_remove+0x64/0x90 <c10e1020> xfs_bmap_add_extent_delay_real+0x12a0/0x16d0 [4391322.921000] <c10d0fce> xfs_alloc_ag_vextent+0xae/0x140 <c10d3a77> xfs_alloc_vextent+0x307/0x5b0 [4391322.939000] <c10e49fb> xfs_bmap_btalloc+0x41b/0x980 <c1114a46> xfs_iext_bno_to_ext+0x126/0x1d0 [4391322.958000] <c1049845> get_page_from_freelist+0x75/0xa0 <c10e9355> xfs_bmapi+0x1495/0x18d0 [4391322.975000] <c1114a46> xfs_iext_bno_to_ext+0x126/0x1d0 <c10e698c> xfs_bmap_search_multi_extents+0xfc/0x110 [4391322.995000] <c1117a07> xfs_iomap_write_allocate+0x327/0x620 <f8871195> release_stripe+0x35/0x60 [raid5] [4391323.015000] <c11164a0> xfs_iomap+0x440/0x570 <c113979b> xfs_map_blocks+0x5b/0xa0 [4391323.031000] <c113aa3a> xfs_page_state_convert+0x46a/0x7a0 <c1044d7b> find_get_pages_tag+0x7b/0x90 [4391323.049000] <c113add9> xfs_vm_writepage+0x69/0x100 <c108df58> mpage_writepages+0x218/0x3f0 [4391323.067000] <c113ad70> xfs_vm_writepage+0x0/0x100 <c104b614> do_writepages+0x54/0x60 [4391323.083000] <c108be86> __sync_single_inode+0x66/0x1f0 <c108c098> __writeback_single_inode+0x88/0x1b0 [4391323.102000] <c1017fd7> find_busiest_group+0x287/0x2f0 <c108c3a7> sync_sb_inodes+0x1e7/0x300 [4391323.120000] <c104bef0> pdflush+0x0/0x50 <c108c595> writeback_inodes+0xd5/0xf0 [4391323.135000] <c104b3ac> wb_kupdate+0xbc/0x130 <c1085cc0> mark_mounts_for_expiry+0x0/0x180 [4391323.152000] <c104be0a> __pdflush+0xca/0x1b0 <c104bf2f> pdflush+0x3f/0x50 [4391323.166000] <c104b2f0> wb_kupdate+0x0/0x130 <c10330f7> kthread+0xb7/0xc0 [4391323.181000] <c1033040> kthread+0x0/0xc0 <c10011ed> kernel_thread_helper+0x5/0x18 -- Raz ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-16 12:42 raid50 and 9TB volumes Raz @ 2007-07-16 13:01 ` David Chinner 2007-07-16 13:57 ` Raz [not found] ` <5d96567b0707160653m5951fac9v5a56bb4c92174d63@mail.gmail.com> 0 siblings, 2 replies; 15+ messages in thread From: David Chinner @ 2007-07-16 13:01 UTC (permalink / raw) To: Raz; +Cc: linux-xfs On Mon, Jul 16, 2007 at 03:42:28PM +0300, Raz wrote: > Hello > I found that using xfs over raid50, ( two raid5's 8 disks each and > raid 0 over them ) crashes the file system when the file system is ~ > 9TB. crashing is easy: we simply create few hundred of files, then > erase them in bulk. the same test passes in 6.4TB filesystems. > this bug happens in 2.6.22 as well as 2.6.17.7. > thank you . > > 4391322.839000] Filesystem "md3": XFS internal error > xfs_alloc_read_agf at line 2176 of file fs/xfs/xfs_alloc.c. Caller > 0xc10d31ea > [4391322.863000] <c10d36e9> xfs_alloc_read_agf+0x199/0x220 > <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 Judging by the kernel addresses (<c10d36e9>) you're running an i386 kernel right? Which means there's probably a wrapping issue at 8TB somewhere in the code which has caused an AGF header to be trashed somewhere lower down in the filesystem. what does /proc/partitions say? I.e. does the kernel see the whole 9TB of space? What does xfs_repair tell you about the corruption? (assuming it doesn't OOM, which is a good chance if you really are on i386). Cheers, Dave. -- Dave Chinner Principal Engineer SGI Australian Software Group ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-16 13:01 ` David Chinner @ 2007-07-16 13:57 ` Raz 2007-07-16 15:24 ` Eric Sandeen [not found] ` <5d96567b0707160653m5951fac9v5a56bb4c92174d63@mail.gmail.com> 1 sibling, 1 reply; 15+ messages in thread From: Raz @ 2007-07-16 13:57 UTC (permalink / raw) To: linux-xfs On 7/16/07, David Chinner <dgc@sgi.com> wrote: > On Mon, Jul 16, 2007 at 03:42:28PM +0300, Raz wrote: > > Hello > > I found that using xfs over raid50, ( two raid5's 8 disks each and > > raid 0 over them ) crashes the file system when the file system is ~ > > 9TB. crashing is easy: we simply create few hundred of files, then > > erase them in bulk. the same test passes in 6.4TB filesystems. > > this bug happens in 2.6.22 as well as 2.6.17.7. > > thank you . > > > > 4391322.839000] Filesystem "md3": XFS internal error > > xfs_alloc_read_agf at line 2176 of file fs/xfs/xfs_alloc.c. Caller > > 0xc10d31ea > > [4391322.863000] <c10d36e9> xfs_alloc_read_agf+0x199/0x220 > > <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 > > Judging by the kernel addresses (<c10d36e9>) you're running > an i386 kernel right? Which means there's probably a wrapping > issue at 8TB somewhere in the code which has caused an AGF > header to be trashed somewhere lower down in the filesystem. > what does /proc/partitions say? I.e. does the kernel see > the whole 9TB of space? > > What does xfs_repair tell you about the corruption? (assuming > it doesn't OOM, which is a good chance if you really are on > i386). > > Cheers, > > Dave. > -- > Dave Chinner > Principal Engineer > SGI Australian Software Group > Well you are right. /proc/partitions says: .... 8 241 488384001 sdp1 9 1 3404964864 md1 9 2 3418684416 md2 9 3 6823647232 md3 while xfs formats md3 as 9 TB. If i am using LBD , what is the biggest size I can use on i386 ? many thanks raz -- Raz ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-16 13:57 ` Raz @ 2007-07-16 15:24 ` Eric Sandeen 0 siblings, 0 replies; 15+ messages in thread From: Eric Sandeen @ 2007-07-16 15:24 UTC (permalink / raw) To: Raz; +Cc: linux-xfs Raz wrote: > Well you are right. /proc/partitions says: > .... > 8 241 488384001 sdp1 > 9 1 3404964864 md1 > 9 2 3418684416 md2 > 9 3 6823647232 md3 > > while xfs formats md3 as 9 TB. > If i am using LBD , what is the biggest size I can use on i386 ? With LBD on, you *should* be able to get to 16TB (2^32 * 4096) in general, assuming that everything in your IO path is clean. (The 16TB limit is due to page cache addressing on x86). -Eric ^ permalink raw reply [flat|nested] 15+ messages in thread
[parent not found: <5d96567b0707160653m5951fac9v5a56bb4c92174d63@mail.gmail.com>]
* Re: raid50 and 9TB volumes [not found] ` <5d96567b0707160653m5951fac9v5a56bb4c92174d63@mail.gmail.com> @ 2007-07-16 22:18 ` David Chinner 2007-07-16 23:56 ` Neil Brown 0 siblings, 1 reply; 15+ messages in thread From: David Chinner @ 2007-07-16 22:18 UTC (permalink / raw) To: Raz; +Cc: neilb, dgc, xfs-oss On Mon, Jul 16, 2007 at 04:53:22PM +0300, Raz wrote: > On 7/16/07, David Chinner <dgc@sgi.com> wrote: > >On Mon, Jul 16, 2007 at 03:42:28PM +0300, Raz wrote: > >> Hello > >> I found that using xfs over raid50, ( two raid5's 8 disks each and > >> raid 0 over them ) crashes the file system when the file system is ~ > >> 9TB. crashing is easy: we simply create few hundred of files, then > >> erase them in bulk. the same test passes in 6.4TB filesystems. > >> this bug happens in 2.6.22 as well as 2.6.17.7. > >> thank you . > >> > >> 4391322.839000] Filesystem "md3": XFS internal error > >> xfs_alloc_read_agf at line 2176 of file fs/xfs/xfs_alloc.c. Caller > >> 0xc10d31ea > >> [4391322.863000] <c10d36e9> xfs_alloc_read_agf+0x199/0x220 > >> <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 > > > >Judging by the kernel addresses (<c10d36e9>) you're running > >an i386 kernel right? Which means there's probably a wrapping > >issue at 8TB somewhere in the code which has caused an AGF > > what is AGF ? An AGF is a "Allocation Group Freespace" structure that holds the free space indexes that the allocator uses. The AGF holds the root of the btrees used to find space, so if the AGF is trashed, you're in big trouble. :/ > >header to be trashed somewhere lower down in the filesystem. > >what does /proc/partitions say? I.e. does the kernel see > >the whole 9TB of space? > > > >What does xfs_repair tell you about the corruption? (assuming > >it doesn't OOM, which is a good chance if you really are on > >i386). > > Well you are right. /proc/partitions says: > .... > 8 241 488384001 sdp1 > 9 1 3404964864 md1 > 9 2 3418684416 md2 > 9 3 6823647232 md3 > > while xfs formats md3 as 9 TB. > If i am using LBD , what is the biggest size I can use on i386 ? Supposedly 16TB. 32bit x 4k page size = 16TB. Given that the size is not being reported correctly, I'd say that this is probably not an XFS issue. The next thing to check is how large an MD device you can create correctly. Neil, do you know of any problems with > 8TB md devices on i386? Cheers, Dave. > > many thanks > raz > > > > > -- > Raz -- Dave Chinner Principal Engineer SGI Australian Software Group ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-16 22:18 ` David Chinner @ 2007-07-16 23:56 ` Neil Brown 2007-07-17 0:12 ` David Chinner 0 siblings, 1 reply; 15+ messages in thread From: Neil Brown @ 2007-07-16 23:56 UTC (permalink / raw) To: David Chinner; +Cc: Raz, xfs-oss On Tuesday July 17, dgc@sgi.com wrote: > On Mon, Jul 16, 2007 at 04:53:22PM +0300, Raz wrote: > > > > Well you are right. /proc/partitions says: > > .... > > 8 241 488384001 sdp1 > > 9 1 3404964864 md1 > > 9 2 3418684416 md2 > > 9 3 6823647232 md3 > > > > while xfs formats md3 as 9 TB. > > If i am using LBD , what is the biggest size I can use on i386 ? > > Supposedly 16TB. 32bit x 4k page size = 16TB. Given that the size is > not being reported correctly, I'd say that this is probably not an > XFS issue. The next thing to check is how large an MD device you > can create correctly. > > Neil, do you know of any problems with > 8TB md devices on i386? Should work, but the amount of testing has been limited, and bugs have existed. Each component of a raid5 is limited to 2^32 K by the metadata, so that is 4TB. At 490GB, you are well under that. There should be no problem with a 3TB raid5, providing LBD has been selected. raid0 over 3TB devices should also be fine. There was a bug fixed in May this year that caused problem with md/raid0 was used over components larger than 4TB on a 32bit host, but that shouldn't affect you and it does suggest that someone had success with a very large raid0 once this bug was fixed. If XFS is given a 6.8TB devices and formats it as 9TB, then I would be looking at mkfs.xfs(??). NeilBrown ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-16 23:56 ` Neil Brown @ 2007-07-17 0:12 ` David Chinner 2007-07-17 0:54 ` Neil Brown 0 siblings, 1 reply; 15+ messages in thread From: David Chinner @ 2007-07-17 0:12 UTC (permalink / raw) To: Neil Brown; +Cc: David Chinner, Raz, xfs-oss On Tue, Jul 17, 2007 at 09:56:25AM +1000, Neil Brown wrote: > On Tuesday July 17, dgc@sgi.com wrote: > > On Mon, Jul 16, 2007 at 04:53:22PM +0300, Raz wrote: > > > > > > Well you are right. /proc/partitions says: > > > .... > > > 8 241 488384001 sdp1 > > > 9 1 3404964864 md1 > > > 9 2 3418684416 md2 > > > 9 3 6823647232 md3 > > > > > > while xfs formats md3 as 9 TB. > > > If i am using LBD , what is the biggest size I can use on i386 ? > > > > Supposedly 16TB. 32bit x 4k page size = 16TB. Given that the size is > > not being reported correctly, I'd say that this is probably not an > > XFS issue. The next thing to check is how large an MD device you > > can create correctly. > > > > Neil, do you know of any problems with > 8TB md devices on i386? > > Should work, but the amount of testing has been limited, and bugs > have existed. > > Each component of a raid5 is limited to 2^32 K by the metadata, so > that is 4TB. At 490GB, you are well under that. > There should be no problem with a 3TB raid5, providing LBD has been > selected. > > raid0 over 3TB devices should also be fine. There was a bug fixed in > May this year that caused problem with md/raid0 was used over > components larger than 4TB on a 32bit host, but that shouldn't affect > you and it does suggest that someone had success with a very large > raid0 once this bug was fixed. > > If XFS is given a 6.8TB devices and formats it as 9TB, then I would be > looking at mkfs.xfs(??). mkfs.xfs tries to read the last block of the device that it is given and proceeds only if that read is successful. IOWs, mkfs.xfs has been told the size of the device is 9TB, it's successfully read from offset 9TB, so the device must be at least 9TB. However, internal to the kernel there appears to be some kind of wrapping bug and typically that shows up with /proc/partition showing an incosistent size of the partition compared to other utilities. We've come across this problem repeatedly over the past few years with exactly these symptoms (the end of the FS overwriting the front of the FS), which was why it was the first question I asked. It has always been a block layer or partition problem and they always show up on i386 with filesystems larger than 2TB. FWIW, what a partitioning tool (if any) is being used here and what does it think the size of the partitions are?. Cheers, Dave. -- Dave Chinner Principal Engineer SGI Australian Software Group ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-17 0:12 ` David Chinner @ 2007-07-17 0:54 ` Neil Brown 2007-07-17 0:58 ` David Chinner 0 siblings, 1 reply; 15+ messages in thread From: Neil Brown @ 2007-07-17 0:54 UTC (permalink / raw) To: David Chinner; +Cc: Raz, xfs-oss On Tuesday July 17, dgc@sgi.com wrote: > On Tue, Jul 17, 2007 at 09:56:25AM +1000, Neil Brown wrote: > > On Tuesday July 17, dgc@sgi.com wrote: > > > On Mon, Jul 16, 2007 at 04:53:22PM +0300, Raz wrote: > > > > > > > > Well you are right. /proc/partitions says: > > > > .... > > > > 8 241 488384001 sdp1 > > > > 9 1 3404964864 md1 > > > > 9 2 3418684416 md2 > > > > 9 3 6823647232 md3 > > > > > > > > while xfs formats md3 as 9 TB. .. > > > > If XFS is given a 6.8TB devices and formats it as 9TB, then I would be > > looking at mkfs.xfs(??). > > mkfs.xfs tries to read the last block of the device that it is given > and proceeds only if that read is successful. IOWs, mkfs.xfs has been > told the size of the device is 9TB, it's successfully read from offset > 9TB, so the device must be at least 9TB. Odd. Given that the drives are 490GB, and there are 8 in a raid5 array, the raid5 arrays are really under 3.5GB. And two of them is less than 7GB. So there definitely are not 9TB worth of bytes.. mkfs.xfs uses the BLKGETSIZE64 ioctl which returns bdev->bi_inode->i_size, where as /proc/partitions uses get_capacity which uses disk->capacity, so there is some room for them to return different values... Except that on open, it calls bd_set_size(bdev, (loff_t)get_capacity(disk)<<9); which makes sure the two have the same value. I cannot see where the size difference comes from. What does /sbin/blockdev --getsize64 report for each of the different devices, as compared to what /proc/partitions reports? NeilBrown > > However, internal to the kernel there appears to be some kind of > wrapping bug and typically that shows up with /proc/partition > showing an incosistent size of the partition compared to other > utilities. ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-17 0:54 ` Neil Brown @ 2007-07-17 0:58 ` David Chinner 2007-07-23 6:09 ` Raz 0 siblings, 1 reply; 15+ messages in thread From: David Chinner @ 2007-07-17 0:58 UTC (permalink / raw) To: Neil Brown; +Cc: David Chinner, Raz, xfs-oss On Tue, Jul 17, 2007 at 10:54:36AM +1000, Neil Brown wrote: > On Tuesday July 17, dgc@sgi.com wrote: > > On Tue, Jul 17, 2007 at 09:56:25AM +1000, Neil Brown wrote: > > > On Tuesday July 17, dgc@sgi.com wrote: > > > > On Mon, Jul 16, 2007 at 04:53:22PM +0300, Raz wrote: > > > > > > > > > > Well you are right. /proc/partitions says: > > > > > .... > > > > > 8 241 488384001 sdp1 > > > > > 9 1 3404964864 md1 > > > > > 9 2 3418684416 md2 > > > > > 9 3 6823647232 md3 > > > > > > > > > > while xfs formats md3 as 9 TB. > .. > > > > > > If XFS is given a 6.8TB devices and formats it as 9TB, then I would be > > > looking at mkfs.xfs(??). > > > > mkfs.xfs tries to read the last block of the device that it is given > > and proceeds only if that read is successful. IOWs, mkfs.xfs has been > > told the size of the device is 9TB, it's successfully read from offset > > 9TB, so the device must be at least 9TB. > > Odd. > Given that the drives are 490GB, and there are 8 in a raid5 array, > the raid5 arrays are really under 3.5GB. And two of them is less than > 7GB. So there definitely are not 9TB worth of bytes.. > > mkfs.xfs uses the BLKGETSIZE64 ioctl which returns > bdev->bi_inode->i_size, where as /proc/partitions uses get_capacity > which uses disk->capacity, so there is some room for them to return > different values... Except that on open, it calls > bd_set_size(bdev, (loff_t)get_capacity(disk)<<9); > which makes sure the two have the same value. > > I cannot see where the size difference comes from. > What does > /sbin/blockdev --getsize64 > report for each of the different devices, as compared to what > /proc/partitions reports? And add to that the output of `xfs_growfs -n <mntpt>` so we can see what XFS really thinks the size of the filesystem is. Cheers, Dave. -- Dave Chinner Principal Engineer SGI Australian Software Group ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-17 0:58 ` David Chinner @ 2007-07-23 6:09 ` Raz 2007-07-24 1:01 ` David Chinner 0 siblings, 1 reply; 15+ messages in thread From: Raz @ 2007-07-23 6:09 UTC (permalink / raw) To: David Chinner; +Cc: xfs-oss On 7/17/07, David Chinner <dgc@sgi.com> wrote: > On Tue, Jul 17, 2007 at 10:54:36AM +1000, Neil Brown wrote: > > On Tuesday July 17, dgc@sgi.com wrote: > > > On Tue, Jul 17, 2007 at 09:56:25AM +1000, Neil Brown wrote: > > > > On Tuesday July 17, dgc@sgi.com wrote: > > > > > On Mon, Jul 16, 2007 at 04:53:22PM +0300, Raz wrote: > > > > > > > > > > > > Well you are right. /proc/partitions says: > > > > > > .... > > > > > > 8 241 488384001 sdp1 > > > > > > 9 1 3404964864 md1 > > > > > > 9 2 3418684416 md2 > > > > > > 9 3 6823647232 md3 > > > > > > > > > > > > while xfs formats md3 as 9 TB. > > .. > > > > > > > > If XFS is given a 6.8TB devices and formats it as 9TB, then I would be > > > > looking at mkfs.xfs(??). > > > > > > mkfs.xfs tries to read the last block of the device that it is given > > > and proceeds only if that read is successful. IOWs, mkfs.xfs has been > > > told the size of the device is 9TB, it's successfully read from offset > > > 9TB, so the device must be at least 9TB. > > > > Odd. > > Given that the drives are 490GB, and there are 8 in a raid5 array, > > the raid5 arrays are really under 3.5GB. And two of them is less than > > 7GB. So there definitely are not 9TB worth of bytes.. > > > > mkfs.xfs uses the BLKGETSIZE64 ioctl which returns > > bdev->bi_inode->i_size, where as /proc/partitions uses get_capacity > > which uses disk->capacity, so there is some room for them to return > > different values... Except that on open, it calls > > bd_set_size(bdev, (loff_t)get_capacity(disk)<<9); > > which makes sure the two have the same value. > > > > I cannot see where the size difference comes from. > > What does > > /sbin/blockdev --getsize64 > > report for each of the different devices, as compared to what > > /proc/partitions reports? > And add to that the output of `xfs_growfs -n <mntpt>` so we can > see what XFS really thinks the size of the filesystem is. > > Cheers, > > Dave. > -- > Dave Chinner > Principal Engineer > SGI Australian Software Group > My QA to re-installed the system. same kernel, different results. now, /proc/paritions reports : 9 1 5114281984 md1 9 2 5128001536 md2 9 3 10242281472 md3 blockdev --getsize64 /dev/md3 10488096227328 but xfs keeps on crashing. when formatting it ot 6.3 TB we're OK. when letting xfs's mkfs choose the /dev/hda1 243M 155M 76M 68% / /dev/hda1 243M 155M 76M 68% / /dev/md0 1.9G 35M 1.8G 2% /d0 /dev/md3 6.3T 5.7T 593G 91% /d1 when formatting to 6.4 TB: xfs_growfs -n ( or xfs_info ) reports: meta-data=/dev/md3 isize=256 agcount=33, agsize=52428544 blks = sectsz=512 attr=0 data = bsize=4096 blocks=1677721600, imaxpct=25 = sunit=256 swidth=512 blks, unwritten=1 naming = version 2 bsize=4096 log = internal bsize=4096 blocks=32768, version=1 = sectsz=512 sunit=0 blks realtime =none extsz=2097152 blocks=0, rtextents=0 when formatting without any size argument, xfs_growfs reports: meta-data=/dev/md3 isize=256 agcount=33, agsize=80017664 blks = sectsz=512 attr=0 data = bsize=4096 blocks=2560570368, imaxpct=25 = sunit=256 swidth=512 blks, unwritten=1 naming = version 2 bsize=4096 log = internal bsize=4096 blocks=32768, version=1 = sectsz=512 sunit=0 blks realtime =none extsz=2097152 blocks=0, rtextents=0 in this case , xfs crashes again. [4613896.794000] <c10d36e9> xfs_alloc_read_agf+0x199/0x220 <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 [4613896.794000] <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 <c10d31ea> xfs_alloc_fix_freelist+0x41a/0x4b0 [4613896.794000] <c1006ebc> timer_interrupt+0x6c/0xa0 <c1042e64> __do_IRQ+0xc4/0x120 [4613896.795000] <c100595e> do_IRQ+0x1e/0x30 <c1003aba> common_interrupt+0x1a/0x20 [4613896.795000] <c10d3a77> xfs_alloc_vextent+0x307/0x5b0 <c10e49fb> xfs_bmap_btalloc+0x41b/0x980 [4613896.795000] <c1114a46> xfs_iext_bno_to_ext+0x126/0x1d0 <c10283b6> update_wall_time_one_tick+0x6/0x80 [4613896.795000] <c102846a> update_wall_time+0xa/0x40 <c10e9355> xfs_bmapi+0x1495/0x18d0 [4613896.795000] <c1114a46> xfs_iext_bno_to_ext+0x126/0x1d0 <c10e698c> xfs_bmap_search_multi_extents+0xfc/0x110 [4613896.795000] <c1117a07> xfs_iomap_write_allocate+0x327/0x620 <c104807c> mempool_free+0x4c/0xa0 [4613896.795000] <c104807c> mempool_free+0x4c/0xa0 <c106b3f8> bio_fs_destructor+0x18/0x20 [4613896.795000] <c11164a0> xfs_iomap+0x440/0x570 <c113979b> xfs_map_blocks+0x5b/0xa0 [4613896.795000] <c113aa3a> xfs_page_state_convert+0x46a/0x7a0 <c1044d7b> find_get_pages_tag+0x7b/0x90 [4613896.795000] <c113add9> xfs_vm_writepage+0x69/0x100 <c108df58> mpage_writepages+0x218/0x3f0 [4613896.795000] <c113ad70> xfs_vm_writepage+0x0/0x100 <c104b614> do_writepages+0x54/0x60 [4613896.795000] <c108be86> __sync_single_inode+0x66/0x1f0 <c108c098> __writeback_single_inode+0x88/0x1b0 [4613896.795000] <c1028120> del_timer_sync+0x10/0x20 <c1298ee0> schedule_timeout+0x60/0xb0 [4613896.795000] <c10288a0> process_timeout+0x0/0x10 <c108c3a7> sync_sb_inodes+0x1e7/0x300 [4613896.795000] <c108c595> writeback_inodes+0xd5/0xf0 <c104afc2> balance_dirty_pages+0xd2/0x190 [4613896.795000] <c106a07c> generic_commit_write+0x7c/0xa0 <c1046950> generic_file_buffered_write+0x310/0x6b0 [4613896.795000] <c1082b7d> file_update_time+0x5d/0xe0 <c114366a> xfs_write+0xc0a/0xe00 [4613896.795000] <c1156eef> __bitmap_weight+0x5f/0x80 <c1141f47> xfs_read+0x1a7/0x370 [4613896.795000] <c113dc9f> xfs_file_aio_write+0x8f/0xa0 <c1065a71> do_sync_write+0xd1/0x120 [4613896.795000] <c1033650> autoremove_wake_function+0x0/0x60 <c1065b88> vfs_write+0xc8/0x190 [4613896.795000] <c1065d21> sys_write+0x51/0x80 <c10030ef> syscall_call+0x7/0xb [root@video1 eyal_kaufer]$ -- Raz ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-23 6:09 ` Raz @ 2007-07-24 1:01 ` David Chinner 2007-08-07 9:20 ` Raz 0 siblings, 1 reply; 15+ messages in thread From: David Chinner @ 2007-07-24 1:01 UTC (permalink / raw) To: Raz; +Cc: David Chinner, xfs-oss On Mon, Jul 23, 2007 at 09:09:03AM +0300, Raz wrote: > My QA to re-installed the system. same kernel, different results. now, > /proc/paritions > reports : > 9 1 5114281984 md1 > 9 2 5128001536 md2 > 9 3 10242281472 md3 > > blockdev --getsize64 /dev/md3 > 10488096227328 > > but xfs keeps on crashing. when formatting it ot 6.3 TB we're OK. when > letting xfs's mkfs choose the So at 6.3TB everything is ok. At what point does it start having problems? 6.4TB, 6.8TB, 8TB, 9TB? I know Neil pointed out that you shouldn't have 10TB but closer to 7TB - is this true? Cheers, Dave. -- Dave Chinner Principal Engineer SGI Australian Software Group ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-07-24 1:01 ` David Chinner @ 2007-08-07 9:20 ` Raz 2007-09-03 14:24 ` Raz 0 siblings, 1 reply; 15+ messages in thread From: Raz @ 2007-08-07 9:20 UTC (permalink / raw) To: David Chinner; +Cc: linux-xfs On 7/24/07, David Chinner <dgc@sgi.com> wrote: > On Mon, Jul 23, 2007 at 09:09:03AM +0300, Raz wrote: > > My QA to re-installed the system. same kernel, different results. now, > > /proc/paritions > > reports : > > 9 1 5114281984 md1 > > 9 2 5128001536 md2 > > 9 3 10242281472 md3 > > > > blockdev --getsize64 /dev/md3 > > 10488096227328 > > > > but xfs keeps on crashing. when formatting it ot 6.3 TB we're OK. when > > letting xfs's mkfs choose the > > So at 6.3TB everything is ok. At what point does it start having > problems? 6.4TB, 6.8TB, 8TB, 9TB? over 8 TB. we checked several times. in 8.5 it crashes. > I know Neil pointed out that you shouldn't have 10TB but closer to > 7TB - is this true? the drives are of 750 GB each. > Cheers, > > Dave. > -- > Dave Chinner > Principal Engineer > SGI Australian Software Group > -- Raz ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-08-07 9:20 ` Raz @ 2007-09-03 14:24 ` Raz 2007-09-03 18:55 ` Christian Kujau 0 siblings, 1 reply; 15+ messages in thread From: Raz @ 2007-09-03 14:24 UTC (permalink / raw) To: David Chinner; +Cc: linux-xfs Dave hello. What is the curret status of this problem ? If you recall , xfs in 32bit over 10 TB md device ( raid50 in this case) sees only 8TB ( and no more). The disks I am using are 750GB hitachi. kernel is 2.6.17. thank you raz On 8/7/07, Raz <raziebe@gmail.com> wrote: > On 7/24/07, David Chinner <dgc@sgi.com> wrote: > > On Mon, Jul 23, 2007 at 09:09:03AM +0300, Raz wrote: > > > My QA to re-installed the system. same kernel, different results. now, > > > /proc/paritions > > > reports : > > > 9 1 5114281984 md1 > > > 9 2 5128001536 md2 > > > 9 3 10242281472 md3 > > > > > > blockdev --getsize64 /dev/md3 > > > 10488096227328 > > > > > > but xfs keeps on crashing. when formatting it ot 6.3 TB we're OK. when > > > letting xfs's mkfs choose the > > > > So at 6.3TB everything is ok. At what point does it start having > > problems? 6.4TB, 6.8TB, 8TB, 9TB? > over 8 TB. we checked several times. in 8.5 it crashes. > > I know Neil pointed out that you shouldn't have 10TB but closer to > > 7TB - is this true? > the drives are of 750 GB each. > > Cheers, > > > > Dave. > > -- > > Dave Chinner > > Principal Engineer > > SGI Australian Software Group > > > > > -- > Raz > -- Raz ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-09-03 14:24 ` Raz @ 2007-09-03 18:55 ` Christian Kujau 2007-09-04 2:50 ` Eric Sandeen 0 siblings, 1 reply; 15+ messages in thread From: Christian Kujau @ 2007-09-03 18:55 UTC (permalink / raw) To: linux-xfs On Mon, 3 Sep 2007, Raz wrote: > What is the curret status of this problem ? If you recall , xfs in 32bit > over 10 TB md device ( raid50 in this case) sees only 8TB ( and no more). > The disks I am using are 750GB hitachi. > kernel is 2.6.17. dunno about this particular issue, but you meant "kernel 2.6.17.7 or higher", right? If not: http://oss.sgi.com/projects/xfs/faq.html#dir2 -- BOFH excuse #374: It's the InterNIC's fault. ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: raid50 and 9TB volumes 2007-09-03 18:55 ` Christian Kujau @ 2007-09-04 2:50 ` Eric Sandeen 0 siblings, 0 replies; 15+ messages in thread From: Eric Sandeen @ 2007-09-04 2:50 UTC (permalink / raw) To: Christian Kujau; +Cc: linux-xfs Christian Kujau wrote: > On Mon, 3 Sep 2007, Raz wrote: >> What is the curret status of this problem ? If you recall , xfs in 32bit >> over 10 TB md device ( raid50 in this case) sees only 8TB ( and no more). >> The disks I am using are 750GB hitachi. >> kernel is 2.6.17. > > dunno about this particular issue, but you meant "kernel 2.6.17.7 or > higher", right? If not: http://oss.sgi.com/projects/xfs/faq.html#dir2 That shouldn't be affected by volume size. Raz, are you certain that the MD volume is in good shape at this point, after Neil's questions? If it's something you can test on, I'd suggest getting lmdd and writing a pattern directly to the block device, spanning the 8T point, then go read it back & double check that all is well. XFS certainly has been tested at 8T and above; if I had to bet on it, I'd bet at a problem in another layer, or the configuration. -Eric ^ permalink raw reply [flat|nested] 15+ messages in thread
end of thread, other threads:[~2007-09-04 2:50 UTC | newest]
Thread overview: 15+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2007-07-16 12:42 raid50 and 9TB volumes Raz
2007-07-16 13:01 ` David Chinner
2007-07-16 13:57 ` Raz
2007-07-16 15:24 ` Eric Sandeen
[not found] ` <5d96567b0707160653m5951fac9v5a56bb4c92174d63@mail.gmail.com>
2007-07-16 22:18 ` David Chinner
2007-07-16 23:56 ` Neil Brown
2007-07-17 0:12 ` David Chinner
2007-07-17 0:54 ` Neil Brown
2007-07-17 0:58 ` David Chinner
2007-07-23 6:09 ` Raz
2007-07-24 1:01 ` David Chinner
2007-08-07 9:20 ` Raz
2007-09-03 14:24 ` Raz
2007-09-03 18:55 ` Christian Kujau
2007-09-04 2:50 ` Eric Sandeen
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox