* Re: e2fsck max blocks for huge non-extent file
[not found] <CAE2LqHL6uY=Sq2+aVtW-Lkbu9mvjFkaNqLaDA8Bkpmvx9AjHBg@mail.gmail.com>
@ 2025-01-13 16:33 ` Theodore Ts'o
2025-01-13 18:35 ` Darrick J. Wong
0 siblings, 1 reply; 4+ messages in thread
From: Theodore Ts'o @ 2025-01-13 16:33 UTC (permalink / raw)
To: Catalin Patulea; +Cc: linux-ext4, Kazuya Mio
On Mon, Jan 13, 2025 at 12:49:19AM -0500, Catalin Patulea wrote:
>
> I have an ext3 filesystem on which I manually enabled huge_file
> (files >2 TB) using tune2fs; then created a 3 TB file (backup image
> of another disk). Now, I am running e2fsck and it reports errors:
Hmm, it looks like this has been broken for a while. I've done a
quick look, and it appears this has been the case since e2fsprogs
1.28 and this commit:
commit da307041e75bdf3b24c1eb43132a4f9d8a1b3844
Author: Theodore Ts'o <tytso@mit.edu>
Date: Tue May 21 21:19:14 2002 -0400
Check for inodes which are too big (either too many blocks, or
would cause i_size to be too big), and offer to truncate the inode.
Remove old bogus i_size checks.
Add test case which tests e2fsck's handling of large sparse files.
Older e2fsck with the old(er) bogus i_size checks didn't handle
this correctly.
I think no one noticed since trying to support files this large on a
non-extent file is so inefficient and painful that in practice anyone
trying to use files this large would be using ext4, and not a really
ancient ext3 file system.
The fix might be as simple as this, but I haven't had a chance to test
it and do appropriate regression tests....
diff --git a/e2fsck/pass1.c b/e2fsck/pass1.c
index eb73922d3..e460a75f4 100644
--- a/e2fsck/pass1.c
+++ b/e2fsck/pass1.c
@@ -3842,7 +3842,7 @@ static int process_block(ext2_filsys fs,
problem = PR_1_TOOBIG_DIR;
if (p->is_dir && p->num_blocks + 1 >= p->max_blocks)
problem = PR_1_TOOBIG_DIR;
- if (p->is_reg && p->num_blocks + 1 >= p->max_blocks)
+ if (p->is_reg && p->num_blocks + 1 >= 1U << 31)
problem = PR_1_TOOBIG_REG;
if (!p->is_dir && !p->is_reg && blockcnt > 0)
problem = PR_1_TOOBIG_SYMLINK;
- Ted
^ permalink raw reply related [flat|nested] 4+ messages in thread
* Re: e2fsck max blocks for huge non-extent file
2025-01-13 16:33 ` e2fsck max blocks for huge non-extent file Theodore Ts'o
@ 2025-01-13 18:35 ` Darrick J. Wong
2025-01-13 19:26 ` Theodore Ts'o
0 siblings, 1 reply; 4+ messages in thread
From: Darrick J. Wong @ 2025-01-13 18:35 UTC (permalink / raw)
To: Theodore Ts'o; +Cc: Catalin Patulea, linux-ext4, Kazuya Mio
On Mon, Jan 13, 2025 at 11:33:45AM -0500, Theodore Ts'o wrote:
> On Mon, Jan 13, 2025 at 12:49:19AM -0500, Catalin Patulea wrote:
> >
> > I have an ext3 filesystem on which I manually enabled huge_file
> > (files >2 TB) using tune2fs; then created a 3 TB file (backup image
> > of another disk). Now, I am running e2fsck and it reports errors:
>
> Hmm, it looks like this has been broken for a while. I've done a
> quick look, and it appears this has been the case since e2fsprogs
> 1.28 and this commit:
>
> commit da307041e75bdf3b24c1eb43132a4f9d8a1b3844
> Author: Theodore Ts'o <tytso@mit.edu>
> Date: Tue May 21 21:19:14 2002 -0400
>
> Check for inodes which are too big (either too many blocks, or
> would cause i_size to be too big), and offer to truncate the inode.
> Remove old bogus i_size checks.
>
> Add test case which tests e2fsck's handling of large sparse files.
> Older e2fsck with the old(er) bogus i_size checks didn't handle
> this correctly.
>
> I think no one noticed since trying to support files this large on a
> non-extent file is so inefficient and painful that in practice anyone
> trying to use files this large would be using ext4, and not a really
> ancient ext3 file system.
>
> The fix might be as simple as this, but I haven't had a chance to test
> it and do appropriate regression tests....
>
> diff --git a/e2fsck/pass1.c b/e2fsck/pass1.c
> index eb73922d3..e460a75f4 100644
> --- a/e2fsck/pass1.c
> +++ b/e2fsck/pass1.c
> @@ -3842,7 +3842,7 @@ static int process_block(ext2_filsys fs,
> problem = PR_1_TOOBIG_DIR;
> if (p->is_dir && p->num_blocks + 1 >= p->max_blocks)
> problem = PR_1_TOOBIG_DIR;
> - if (p->is_reg && p->num_blocks + 1 >= p->max_blocks)
> + if (p->is_reg && p->num_blocks + 1 >= 1U << 31)
Hmm -- num_blocks is ... the number of "extent records", right? And on
a !extents file, each block mapped by an {in,}direct block counts as a
separate "extent record", right?
In that case, I think (1U<<31) isn't quite right, because the very large
file could have an ACL block, or (shudder) a "hurd translator block".
So that's (1U<<31) + 2 for !extents files.
For extents files, shouldn't this be (1U<<48) + 2? Since you /could/
create a horrifingly large extent tree with a hojillion little
fragments, right? Even if it took a million years to create such a
monster? :)
--D
> problem = PR_1_TOOBIG_REG;
> if (!p->is_dir && !p->is_reg && blockcnt > 0)
> problem = PR_1_TOOBIG_SYMLINK;
>
>
> - Ted
>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: e2fsck max blocks for huge non-extent file
2025-01-13 18:35 ` Darrick J. Wong
@ 2025-01-13 19:26 ` Theodore Ts'o
2025-01-17 3:26 ` Catalin Patulea
0 siblings, 1 reply; 4+ messages in thread
From: Theodore Ts'o @ 2025-01-13 19:26 UTC (permalink / raw)
To: Darrick J. Wong; +Cc: Catalin Patulea, linux-ext4, Kazuya Mio
On Mon, Jan 13, 2025 at 10:35:17AM -0800, Darrick J. Wong wrote:
>
> Hmm -- num_blocks is ... the number of "extent records", right? And on
> a !extents file, each block mapped by an {in,}direct block counts as a
> separate "extent record", right?
>
> In that case, I think (1U<<31) isn't quite right, because the very large
> file could have an ACL block, or (shudder) a "hurd translator block".
> So that's (1U<<31) + 2 for !extents files.
>
> For extents files, shouldn't this be (1U<<48) + 2? Since you /could/
> create a horrifingly large extent tree with a hojillion little
> fragments, right? Even if it took a million years to create such a
> monster? :)
The code paths in question are only used for indirect mapped files.
The logic for handling extent-mapped files is check_blocks_extents()
in modern versions of e2fsprogs, which is why Catalin was only seeing
this for an ext3 file systems that had huge_file enabled.
You're right though that we shouldn't be using num_blocks at all for
testing for regular files or directory files that are too big, since
num_blocks include blocks for extended attribute blocks, the
ind/dind/tind blocks, etc. We do care about num_blocks being too big
for the !huge_file case since for !huge_file file systems i_blocks is
denominated in 512 byte units, and is only 32-bits wide. So in that
case, we *do* care about the size of the file including metadata
blocks being no more than 2TiB.
- Ted
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: e2fsck max blocks for huge non-extent file
2025-01-13 19:26 ` Theodore Ts'o
@ 2025-01-17 3:26 ` Catalin Patulea
0 siblings, 0 replies; 4+ messages in thread
From: Catalin Patulea @ 2025-01-17 3:26 UTC (permalink / raw)
To: Theodore Ts'o; +Cc: Darrick J. Wong, linux-ext4, Kazuya Mio
On Mon, Jan 13, 2025 at 11:33 AM Theodore Ts'o <tytso@mit.edu> wrote:
> The fix might be as simple as this, but I haven't had a chance to test
> it and do appropriate regression tests....
>
> diff --git a/e2fsck/pass1.c b/e2fsck/pass1.c
> index eb73922d3..e460a75f4 100644
> --- a/e2fsck/pass1.c
> +++ b/e2fsck/pass1.c
> @@ -3842,7 +3842,7 @@ static int process_block(ext2_filsys fs,
> problem = PR_1_TOOBIG_DIR;
> if (p->is_dir && p->num_blocks + 1 >= p->max_blocks)
> problem = PR_1_TOOBIG_DIR;
> - if (p->is_reg && p->num_blocks + 1 >= p->max_blocks)
> + if (p->is_reg && p->num_blocks + 1 >= 1U << 31)
> problem = PR_1_TOOBIG_REG;
> if (!p->is_dir && !p->is_reg && blockcnt > 0)
> problem = PR_1_TOOBIG_SYMLINK;
I can confirm that with this patch, e2fsck passes on the test image
created as shown in my original email (dd if=/dev/zero ...). I also
confirm 'make check' passes (390 tests succeeded).
Do you have any thoughts on what a practical regression test would
look like? My repro instructions require 2.1 TB of physical disk space
and root access, which I am guessing is out of the question. For my
local tests I have been using 'qemu-nbd' and QCOW2 images to reduce
the disk space requirements, but it still requires root and ~30 minute
runtime, which still seems impractical.
> ind/dind/tind blocks, etc. We do care about num_blocks being too big
> for the !huge_file case since for !huge_file file systems i_blocks is
> denominated in 512 byte units, and is only 32-bits wide. So in that
> case, we *do* care about the size of the file including metadata
> blocks being no more than 2TiB.
In the proposed patch, "p->num_blocks + 1 >= 1U << 31", that's 2^31
512-byte blocks, would that limit file size to 1 TB?
> You're right though that we shouldn't be using num_blocks at all for
> testing for regular files or directory files that are too big, since
> num_blocks include blocks for extended attribute blocks, the
> ind/dind/tind blocks, etc. We do care about num_blocks being too big
> for the !huge_file case since for !huge_file file systems i_blocks is
> denominated in 512 byte units, and is only 32-bits wide. So in that
> case, we *do* care about the size of the file including metadata
> blocks being no more than 2TiB.
>
> - Ted
>
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2025-01-17 3:27 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
[not found] <CAE2LqHL6uY=Sq2+aVtW-Lkbu9mvjFkaNqLaDA8Bkpmvx9AjHBg@mail.gmail.com>
2025-01-13 16:33 ` e2fsck max blocks for huge non-extent file Theodore Ts'o
2025-01-13 18:35 ` Darrick J. Wong
2025-01-13 19:26 ` Theodore Ts'o
2025-01-17 3:26 ` Catalin Patulea
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).