* Incorrect Free Space / xfs_growfs on RAID5 Volume ?
@ 2009-05-26 9:51 Svavar Örn Eysteinsson
[not found] ` <4A1BE48F.9020107@dermichi.com>
0 siblings, 1 reply; 5+ messages in thread
From: Svavar Örn Eysteinsson @ 2009-05-26 9:51 UTC (permalink / raw)
To: xfs
Hi all.
I’m hoping that someone can help me out here, regarding growing a XFS
filesystem
on a Adaptec RAID controller with RAID5 setup.
First of all, this is my setup :
Fedora Core 6.
Kernel : 2.6.27
(stock Fedora Core 6 xfsprogs)
Adaptec SATA Raid Controller 2820SA ( 8 port )
For 2 years now, I had running a RAID5 setup with 4x 500GB SATA disks
as one logical drive. That drive would have been /dev/sdb.
I made an xfs filesystem out of it and it gave me about 1.8 TB of
usable data space.
No problems so far, nothing.
I recently added the 5th disks in the RAID5 setup, and reconfigured
(Online Expansion)
with the Adaptec Storage Manager.
So the Adaptec SM tells me that I have 2.274 TB of space.
Parity space is : 465.626 GB and the Stripe size is 256K.
Well, the RAID controller setup is done. So I headed out to the fdisk
section.
I deleted the /dev/sdb1 partition and made a new one right ahead.
(see below)
Partition number (1-4): 1
First cylinder (1-303916, default 1): 1
Last cylinder or +size or +sizeM or +sizeK (1-267349, default 267349):
267349
Disk /dev/sdb: 2499.7 GB, 2499794698240 bytes
255 heads, 63 sectors/track, 303916 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot Start End Blocks Id System
/dev/sdb1 1 267349 2147480811 83 Linux
I issued a “w” command to write to the partition table and exited fdisk.
Mounted the /dev/sdb1 partition as /raid-data.
Did grow the XFS filesystem with xfs_growfs /raid-data
Now the strange part. When I issue “df -h” command it shows much smaller
disk space added then it should have.
(see below)
Filesystem Size Used Avail Use% Mounted on
/dev/sda3 269G 181G 75G 71% /
/dev/sda1 244M 20M 211M 9% /boot
tmpfs 1013M 0 1013M 0% /dev/shm
/dev/sdb1 2.0T 1.9T 191G 91% /raid-data
Using a “non -h” df command shows me :
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/sda3 281636376 189165420 77933740 71% /
/dev/sda1 248895 20331 215714 9% /boot
tmpfs 1036736 0 1036736 0% /dev/shm
/dev/sdb1 2147349736 1948095012 199254724 91% /raid-data
Any ideas? This is not acceptable as I inserted 500GB as additional
space
to the RAID5 group and don’t have all of it. :(
So I checked the fragmentation with xfs_db and it told me
that the volume was having a 36.7% fragmentation.
I was going to issue a xfs_fsr but the system couldn’t find that
command.
Great.
I removed the stock xfsprogs rpm version, downloaded the newest
xfsprogs source (xfsprogs-3.0.1). Build’it, installed it and issued
the xfs_fsr on the volume.
Nothing changed on the free space thing, It’s totally the same.
Filesystem Size Used Avail Use% Mounted on
/dev/sda3 269G 181G 75G 71% /
/dev/sda1 244M 20M 211M 9% /boot
tmpfs 1013M 0 1013M 0% /dev/shm
/dev/sdb1 2.0T 1.9T 191G 91% /raid-data
I then decided to use the xfs_growfs with the newest version(that is
the version I
downloaded, configured and installed) on the volume.
Nothing happens, as the xfs_growfs says :
# xfs_growfs -d /raid-data
meta-data=/dev/sdb1 isize=256 agcount=36,
agsize=15257482 blks
= sectsz=512 attr=0
data = bsize=4096 blocks=536870202,
imaxpct=25
= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0
log =internal bsize=4096 blocks=32768, version=1
= sectsz=512 sunit=0 blks, lazy-
count=0
realtime =none extsz=65536 blocks=0, rtextents=0
data size unchanged, skipping
So I really need some advice or help about this situation.
Did I do anything wrong ?
Is the metadata, and or log data on the xfs volume taking all the
remain space
so I only get about 191GB free for data after the insert of 500GB disk ?
Thanks all.
Best regards,
Svavar - Reykjavik / Iceland
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
^ permalink raw reply [flat|nested] 5+ messages in thread[parent not found: <4A1BE48F.9020107@dermichi.com>]
* Re: Incorrect Free Space / xfs_growfs on RAID5 Volume ? [not found] ` <4A1BE48F.9020107@dermichi.com> @ 2009-05-27 9:26 ` Svavar Örn Eysteinsson 2009-05-27 16:29 ` Eric Sandeen 2009-05-28 9:10 ` Michael Monnerie 0 siblings, 2 replies; 5+ messages in thread From: Svavar Örn Eysteinsson @ 2009-05-27 9:26 UTC (permalink / raw) To: xfs Hi. I read on http://www.carltonbale.com/2007/05/how-to-break-the-2tb-2-terabyte-file-system-limit/ that if your kernel is compiled with CONFIG_LBD You can break the 2tb limit. Any facts on that ? **** Breaking 2TB Option 2 - Use Linux with CONFIG_LBD enabled. Most Linux file systems are capable of partitions larger than 2 TB, as long as the Linux kernel itself is. (See this comparison of Linux file systems.) Most Linux distributions now have kernels compiled with CONFIG_LBD enabled (Ubuntu 6.10 does, for example.) As long as the kernel is configured/compiled properly, it is straight-forward to create a single 4TB EXT3 (or similar) partition. * To summarize: 1 RAID array of five 1TB Drives -> 1 RAID level 5 Volume Set that is 4TB -> 1 EXT3 (or similar) Linux partition that is 4TB. **** .... Is this maby out of my scope/setup ? Is there a simple way for me to check if my kernel has this option compiled in ? I'm running Fedora Core 6 with 2.6.27.7 #1 SMP Tue Nov 25 11:50:10 GMT 2008 i686 i686 i386 GNU/Linux. And the FINAL question.... Is there any way for me to alter the raid volume, partitions to GPT or just format the /dev/sdb without loosing any data ? Maby it's just not possible without backup up data, and restore'ing ? Thanks allot guys.. Best regards, Svavar - Reykjavik - Iceland On 26.5.2009, at 12:46, Michael Weissenbacher wrote: > Hi Svavar! >> Now the strange part. When I issue “df -h” command it shows much >> smaller >> disk space added then it should have. > > You have run into the 2TB limit for a DOS Paritition Table. You must > use GPT (GUID Partition Table) to overcome the limit. You can't use > fdisk for that since it has no GPT support. An alternative would be > parted [1]. I'm not sure how this can be done without data loss. An > alternative would be to not use partitions at all and create the XFS > directly on /dev/sdb. > This is not really an XFS issue but an partitioning issue. > > [1] http://www.gnu.org/software/parted/index.shtml > > hth, > Michael _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: Incorrect Free Space / xfs_growfs on RAID5 Volume ? 2009-05-27 9:26 ` Svavar Örn Eysteinsson @ 2009-05-27 16:29 ` Eric Sandeen 2009-05-28 9:10 ` Michael Monnerie 1 sibling, 0 replies; 5+ messages in thread From: Eric Sandeen @ 2009-05-27 16:29 UTC (permalink / raw) To: Svavar Örn Eysteinsson; +Cc: xfs Svavar Örn Eysteinsson wrote: > Hi. > > I read on http://www.carltonbale.com/2007/05/how-to-break-the-2tb-2-terabyte-file-system-limit/ > that if your kernel is compiled with CONFIG_LBD You can break the 2tb > limit. Any facts on that ? This lets the kernel track block devices which are more than 2^32 sectors large, but it does not change the fact that an msdos partition table cannot have a partition this large. > **** > Breaking 2TB Option 2 - Use Linux with CONFIG_LBD enabled. Most Linux > file systems are capable of partitions larger than 2 TB, as long as > the Linux kernel itself is. (See this comparison of Linux file > systems.) Most Linux distributions now have kernels compiled with > CONFIG_LBD enabled (Ubuntu 6.10 does, for example.) As long as the > kernel is configured/compiled properly, it is straight-forward to > create a single 4TB EXT3 (or similar) partition. > > * To summarize: 1 RAID array of five 1TB Drives -> 1 RAID level 5 > Volume Set that is 4TB -> 1 EXT3 (or similar) Linux partition that is > 4TB. > **** > > .... Is this maby out of my scope/setup ? > > > Is there a simple way for me to check if my kernel has this option > compiled in ? > I'm running Fedora Core 6 with 2.6.27.7 #1 SMP Tue Nov 25 11:50:10 > GMT 2008 i686 i686 i386 GNU/Linux. That's pretty old.... but it probably has it on. I don't remember if FC6 had config-* in /boot; if so you could just check there. Otherwise grab the src.rpm and work from there ... > > And the FINAL question.... Is there any way for me to alter the raid > volume, partitions to GPT or just format the /dev/sdb without loosing > any data ? > Maby it's just not possible without backup up data, and restore'ing ? It is probably possible to put a new GPT table in place of the DOS table, but you have to be careful. The idea is that you need a GPT table with a partition starting at exactly the same place, and with the end in the correct (larger) place ... and the GPT table must all fit before the first sector of the first partition. With care, this usually works. -Eric > > Thanks allot guys.. > > > Best regards, > > Svavar - Reykjavik - Iceland > > > > On 26.5.2009, at 12:46, Michael Weissenbacher wrote: > >> Hi Svavar! >>> Now the strange part. When I issue “df -h” command it shows much >>> smaller >>> disk space added then it should have. >> You have run into the 2TB limit for a DOS Paritition Table. You must >> use GPT (GUID Partition Table) to overcome the limit. You can't use >> fdisk for that since it has no GPT support. An alternative would be >> parted [1]. I'm not sure how this can be done without data loss. An >> alternative would be to not use partitions at all and create the XFS >> directly on /dev/sdb. >> This is not really an XFS issue but an partitioning issue. >> >> [1] http://www.gnu.org/software/parted/index.shtml >> >> hth, >> Michael > > > _______________________________________________ > xfs mailing list > xfs@oss.sgi.com > http://oss.sgi.com/mailman/listinfo/xfs > _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: Incorrect Free Space / xfs_growfs on RAID5 Volume ? 2009-05-27 9:26 ` Svavar Örn Eysteinsson 2009-05-27 16:29 ` Eric Sandeen @ 2009-05-28 9:10 ` Michael Monnerie 1 sibling, 0 replies; 5+ messages in thread From: Michael Monnerie @ 2009-05-28 9:10 UTC (permalink / raw) To: xfs [-- Attachment #1.1: Type: text/plain, Size: 947 bytes --] On Mittwoch 27 Mai 2009 Svavar Örn Eysteinsson wrote: > Is there a simple way for me to check if my kernel has this option > compiled in ? gzip -cd /proc/config.gz | grep CONFIG_LBD if your kernel provides /proc/config.gz > And the FINAL question.... Is there any way for me to alter the raid > volume, partitions to GPT or just format the /dev/sdb without > loosing any data ? > Maby it's just not possible without backup up data, and restore'ing ? I googled this once, and found the anser "no". If you find a solution, please post it, I'd be very interested in it. mfg zmi -- // Michael Monnerie, Ing.BSc ----- http://it-management.at // Tel: 0660 / 415 65 31 .network.your.ideas. // PGP Key: "curl -s http://zmi.at/zmi.asc | gpg --import" // Fingerprint: AC19 F9D5 36ED CD8A EF38 500E CE14 91F7 1C12 09B4 // Keyserver: wwwkeys.eu.pgp.net Key-ID: 1C1209B4 [-- Attachment #1.2: This is a digitally signed message part. --] [-- Type: application/pgp-signature, Size: 197 bytes --] [-- Attachment #2: Type: text/plain, Size: 121 bytes --] _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: Incorrect Free Space / xfs_growfs on RAID5 Volume ? @ 2009-05-26 16:58 Richard Ems 0 siblings, 0 replies; 5+ messages in thread From: Richard Ems @ 2009-05-26 16:58 UTC (permalink / raw) To: xfs, svavar Hi! You have hit the 2 TB partition limit. You will need to change your partitioning to GPT, see http://en.wikipedia.org/wiki/GUID_Partition_Table or search for "2TB partition limit" on google. Backup your data first! Regards, Richard -- Richard Ems mail: Richard.Ems@Cape-Horn-Eng.com Cape Horn Engineering S.L. C/ Dr. J.J. Dómine 1, 5º piso 46011 Valencia Tel : +34 96 3242923 / Fax 924 http://www.cape-horn-eng.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs ^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2009-05-28 9:10 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-05-26 9:51 Incorrect Free Space / xfs_growfs on RAID5 Volume ? Svavar Örn Eysteinsson
[not found] ` <4A1BE48F.9020107@dermichi.com>
2009-05-27 9:26 ` Svavar Örn Eysteinsson
2009-05-27 16:29 ` Eric Sandeen
2009-05-28 9:10 ` Michael Monnerie
-- strict thread matches above, loose matches on Subject: below --
2009-05-26 16:58 Richard Ems
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox