From: Brian Foster <bfoster@redhat.com>
To: Tapani Tarvainen <tapani@tapanitarvainen.fi>
Cc: xfs@oss.sgi.com
Subject: Re: "This is a bug."
Date: Thu, 10 Sep 2015 13:55:58 -0400 [thread overview]
Message-ID: <20150910175557.GE27863@bfoster.bfoster> (raw)
In-Reply-To: <20150910173138.GB18940@tarvainen.info>
On Thu, Sep 10, 2015 at 08:31:38PM +0300, Tapani Tarvainen wrote:
> On Thu, Sep 10, 2015 at 10:51:54AM -0400, Brian Foster (bfoster@redhat.com) wrote:
>
> > First off, I see ~60MB of corruption output before I even get to the
> > reported repair failure, so this appears to be an extremely severe
> > corruption and I wouldn't be surprised if ultimately beyond repair
>
> I assumed as much already.
>
> > I suspect what's more interesting at this point is what happened to
> > cause this level of corruption? What kind of event lead to this? Was it
> > a pure filesystem crash or some kind of hardware/raid failure?
>
> Hardware failure. Details are still a bit unclear but apparently raid
> controller went haywire, offlining the array in the middle of
> heavy filesystem use.
>
> > Also, do you happen to know the geometry (xfs_info) of the original fs?
>
> No (and xfs_info doesn't work on the copy made after crash as it
> can't be mounted).
>
> > Repair was showing agno's up in the 20k's and now that I've mounted the
> > repaired image, xfs_info shows the following:
> [...]
> > So that's a 6TB fs with over 24000 allocation groups of size 256MB, as
> > opposed to the mkfs default of 6 allocation groups of 1TB each. Is that
> > intentional?
>
> Not to my knowledge. Unless I'm mistaken, the filesystem was created
> while the machine was running Debian Squeeze, using whatever defaults
> were back then.
>
Strange... was the filesystem created small and then grown to a much
larger size via xfs_growfs? I just formatted a 1GB fs that started with
4 allocation groups and ends with 24576 (same as above) AGs when grown
to 6TB.
Brian
> --
> Tapani Tarvainen
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
next prev parent reply other threads:[~2015-09-10 17:56 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-09-10 9:18 "This is a bug." Tapani Tarvainen
2015-09-10 10:31 ` Tapani Tarvainen
2015-09-10 11:53 ` Emmanuel Florac
2015-09-10 12:05 ` Tapani Tarvainen
2015-09-10 11:48 ` Emmanuel Florac
2015-09-10 11:55 ` Tapani Tarvainen
2015-09-10 12:30 ` Tapani Tarvainen
2015-09-10 12:36 ` Brian Foster
2015-09-10 12:54 ` Tapani Tarvainen
2015-09-10 13:01 ` Brian Foster
2015-09-10 13:05 ` Tapani Tarvainen
2015-09-10 14:51 ` Brian Foster
2015-09-10 15:05 ` Brian Foster
2015-09-10 17:52 ` Tapani Tarvainen
2015-09-10 18:01 ` Tapani Tarvainen
2015-09-10 17:31 ` Tapani Tarvainen
2015-09-10 17:55 ` Brian Foster [this message]
2015-09-10 18:03 ` Tapani Tarvainen
2015-09-10 18:33 ` Brian Foster
2015-09-11 6:19 ` Tapani Tarvainen
2015-09-11 0:12 ` Eric Sandeen
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20150910175557.GE27863@bfoster.bfoster \
--to=bfoster@redhat.com \
--cc=tapani@tapanitarvainen.fi \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox