From: Eric Sandeen <sandeen@sandeen.net>
To: Martin Steigerwald <Martin@lichtvoll.de>
Cc: Manny <dermaniac@gmail.com>, xfs@oss.sgi.com
Subject: Re: Insane file system overhead on large volume
Date: Sat, 28 Jan 2012 09:35:14 -0600 [thread overview]
Message-ID: <4F2415B2.3080605@sandeen.net> (raw)
In-Reply-To: <201201281555.22179.Martin@lichtvoll.de>
On 1/28/12 8:55 AM, Martin Steigerwald wrote:
> Am Freitag, 27. Januar 2012 schrieb Eric Sandeen:
>> On 1/27/12 1:50 AM, Manny wrote:
>>> Hi there,
>>>
>>> I'm not sure if this is intended behavior, but I was a bit stumped
>>> when I formatted a 30TB volume (12x3TB minus 2x3TB for parity in RAID
>>> 6) with XFS and noticed that there were only 22 TB left. I just
>>> called mkfs.xfs with default parameters - except for swith and sunit
>>> which match the RAID setup.
>>>
>>> Is it normal that I lost 8TB just for the file system? That's almost
>>> 30% of the volume. Should I set the block size higher? Or should I
>>> increase the number of allocation groups? Would that make a
>>> difference? Whats the preferred method for handling such large
>>> volumes?
>>
>> If it was 12x3TB I imagine you're confusing TB with TiB, so
>> perhaps your 30T is really only 27TiB to start with.
>>
>> Anyway, fs metadata should not eat much space:
>>
>> # mkfs.xfs -dfile,name=fsfile,size=30t
>> # ls -lh fsfile
>> -rw-r--r-- 1 root root 30T Jan 27 12:18 fsfile
>> # mount -o loop fsfile mnt/
>> # df -h mnt
>> Filesystem Size Used Avail Use% Mounted on
>> /tmp/fsfile 30T 5.0M 30T 1% /tmp/mnt
>>
>> So Christoph's question was a good one; where are you getting
>> your sizes?
To solve your original problem, can you answer the above question?
Adding your actual raid config output (/proc/mdstat maybe) would help
too.
> An academic question:
>
> Why is it that I get
>
> merkaba:/tmp> mkfs.xfs -dfile,name=fsfile,size=30t
> meta-data=fsfile isize=256 agcount=30, agsize=268435455
> blks
> = sectsz=512 attr=2, projid32bit=0
> data = bsize=4096 blocks=8053063650, imaxpct=5
> = sunit=0 swidth=0 blks
> naming =version 2 bsize=4096 ascii-ci=0
> log =Internes Protokoll bsize=4096 blocks=521728, version=2
> = sectsz=512 sunit=0 blks, lazy-count=1
> realtime =keine extsz=4096 blocks=0, rtextents=0
>
> merkaba:/tmp> mount -o loop fsfile /mnt/zeit
> merkaba:/tmp> df -hT /mnt/zeit
> Dateisystem Typ Größe Benutzt Verf. Verw% Eingehängt auf
> /dev/loop0 xfs 30T 33M 30T 1% /mnt/zeit
> merkaba:/tmp> LANG=C df -hT /mnt/zeit
> Filesystem Type Size Used Avail Use% Mounted on
> /dev/loop0 xfs 30T 33M 30T 1% /mnt/zeit
>
>
> 33MiB used on first mount instead of 5?
Not sure offhand, differences in xfsprogs version mkfs defaults perhaps.
...
> Hmmm, but creating the file on Ext4 does not work:
ext4 is not designed to handle very large files, so anything
above 16T will fail.
> fallocate instead of sparse file?
no, you just ran into file offset limits on ext4.
> And on BTRFS as well as XFS it appears to try to create a 30T file for
> real, i.e. by writing data - I stopped it before it could do too much
> harm.
Why do you say that it appears to create a 30T file for real? It
should not...
> Where did you create that hugish XFS file?
On XFS. Of course. :)
> Ciao,
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
next prev parent reply other threads:[~2012-01-28 15:35 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-01-27 7:50 Insane file system overhead on large volume Manny
2012-01-27 10:44 ` Christoph Hellwig
2012-01-27 19:15 ` Manny
2012-01-27 18:21 ` Eric Sandeen
2012-01-28 14:55 ` Martin Steigerwald
2012-01-28 15:35 ` Eric Sandeen [this message]
2012-01-28 16:05 ` Christoph Hellwig
2012-01-28 16:07 ` Eric Sandeen
2012-01-28 16:23 ` Martin Steigerwald
2012-01-29 22:18 ` Dave Chinner
2012-01-27 19:08 ` Stan Hoeppner
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4F2415B2.3080605@sandeen.net \
--to=sandeen@sandeen.net \
--cc=Martin@lichtvoll.de \
--cc=dermaniac@gmail.com \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox