public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Dave Chinner <david@fromorbit.com>
To: Alexandru Cardaniuc <cardaniuc@gmail.com>
Cc: xfs@oss.sgi.com
Subject: Re: corruption of in-memory data detected
Date: Tue, 1 Jul 2014 17:02:30 +1000	[thread overview]
Message-ID: <20140701070230.GG4453@dastard> (raw)
In-Reply-To: <CALkk6gqiwyC7HD3QvzUgecVgdr55=zW=bEA-0bqHFR4y28koxA@mail.gmail.com>

On Mon, Jun 30, 2014 at 11:44:45PM -0700, Alexandru Cardaniuc wrote:
> Hi All,
> 
> I am having an issue with an XFS filesystem shutting down under high load
> with very many small files.
> Basically, I have around 3.5 - 4 million files on this filesystem. New
> files are being written to the FS all the time, until I get to 9-11 mln
> small files (35k on average).
> 
> at some point I get the following in dmesg:
> 
> [2870477.695512] Filesystem "sda5": XFS internal error xfs_trans_cancel at
> line 1138 of file fs/xfs/xfs_trans.c.  Caller 0xffffffff8826bb7d
> [2870477.695558]
> [2870477.695559] Call Trace:
> [2870477.695611]  [<ffffffff88262c28>] :xfs:xfs_trans_cancel+0x5b/0xfe
> [2870477.695643]  [<ffffffff8826bb7d>] :xfs:xfs_mkdir+0x57c/0x5d7
> [2870477.695673]  [<ffffffff8822f3f8>] :xfs:xfs_attr_get+0xbf/0xd2
> [2870477.695707]  [<ffffffff88273326>] :xfs:xfs_vn_mknod+0x1e1/0x3bb
> [2870477.695726]  [<ffffffff80264929>] _spin_lock_irqsave+0x9/0x14
> [2870477.695736]  [<ffffffff802230e6>] __up_read+0x19/0x7f
> [2870477.695764]  [<ffffffff8824f8f4>] :xfs:xfs_iunlock+0x57/0x79
> [2870477.695776]  [<ffffffff80264929>] _spin_lock_irqsave+0x9/0x14
> [2870477.695784]  [<ffffffff802230e6>] __up_read+0x19/0x7f
> [2870477.695791]  [<ffffffff80209f4c>] __d_lookup+0xb0/0xff
> [2870477.695803]  [<ffffffff8020cd4a>] _atomic_dec_and_lock+0x39/0x57
> [2870477.695814]  [<ffffffff8022d6db>] mntput_no_expire+0x19/0x89
> [2870477.695829]  [<ffffffff80264929>] _spin_lock_irqsave+0x9/0x14
> [2870477.695837]  [<ffffffff802230e6>] __up_read+0x19/0x7f
> [2870477.695861]  [<ffffffff8824f8f4>] :xfs:xfs_iunlock+0x57/0x79
> [2870477.695887]  [<ffffffff882680af>] :xfs:xfs_access+0x3d/0x46
> [2870477.695899]  [<ffffffff80264929>] _spin_lock_irqsave+0x9/0x14
> [2870477.695923]  [<ffffffff802df4a3>] vfs_mkdir+0xe3/0x152
> [2870477.695933]  [<ffffffff802dfa79>] sys_mkdirat+0xa3/0xe4
> [2870477.695953]  [<ffffffff80260295>] tracesys+0x47/0xb6
> [2870477.695963]  [<ffffffff802602f9>] tracesys+0xab/0xb6
> [2870477.695977]
> [2870477.695985] xfs_force_shutdown(sda5,0x8) called from line 1139 of file
> fs/xfs/xfs_trans.c.  Return address = 0xffffffff88262c46
> [2870477.696452] Filesystem "sda5": Corruption of in-memory data detected.
> Shutting down filesystem: sda5
> [2870477.696464] Please umount the filesystem, and rectify the problem(s)

You've probably fragmented free space to the point where inodes
cannot be allocated anymore, and then it's shutdown because it got
enospc with a dirty inode allocation transaction.

xfs_db -c "freespc -s" <dev>

should tell us whether this is the case or not.

> Using CentOS 5.9 with kernel 2.6.18-348.el5xen

The "enospc with dirty transaction" shutdown bugs have been fixed in
more recent kernels than RHEL5.

> The problem is reproducible and I don't think it's hardware related. The
> problem was reproduced on multiple servers of the same type. So, I doubt
> it's a memory issue or something like that.

Nope, it's not hardware, it's buggy software that has been fixed in
the years since 2.6.18....

> Is that a known issue?

Yes.

> If it is then what's the fix?

If you've fragmented free space, then your ony options are:

	- dump/mkfs/restore
	- remove a large number of files from the filesystem so free
	  space defragments.

If you simply want to avoid the shutdown, then upgrade to a more
recent kernel (3.x of some kind) where all the known issues have
been fixed.

> I went through the
> kernel updates for CentOS 5.10 (newer kernel), but didn't see any xfs
> related fixes since CentOS 5.9

That's something you need to talk to your distro maintainers
about....

Cheers,

Dave.
-- 
Dave Chinner
david@fromorbit.com

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

  reply	other threads:[~2014-07-01  7:03 UTC|newest]

Thread overview: 13+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-07-01  6:44 corruption of in-memory data detected Alexandru Cardaniuc
2014-07-01  7:02 ` Dave Chinner [this message]
2014-07-01  8:29   ` Alexandru Cardaniuc
2014-07-01  9:38     ` Dave Chinner
2014-07-01 20:13       ` Alexandru Cardaniuc
2014-07-01 21:43         ` Dave Chinner
  -- strict thread matches above, loose matches on Subject: below --
2009-01-02  2:46 Corruption " Thomas Gutzler
2009-01-02  3:24 ` Eric Sandeen
2009-03-11  2:44   ` Thomas Gutzler
2009-03-11  4:30     ` Eric Sandeen
2009-03-11 10:42       ` Thomas Gutzler
2009-03-12  2:23         ` Eric Sandeen
2009-03-12  5:06           ` Thomas Gutzler

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20140701070230.GG4453@dastard \
    --to=david@fromorbit.com \
    --cc=cardaniuc@gmail.com \
    --cc=xfs@oss.sgi.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox