From: Dave Chinner <david@fromorbit.com>
To: Anders Ossowicki <aowi@novozymes.com>
Cc: "xfs@oss.sgi.com" <xfs@oss.sgi.com>
Subject: Re: "XFS: possible memory allocation deadlock in kmem_alloc" on high memory machine
Date: Wed, 3 Jun 2015 11:52:45 +1000 [thread overview]
Message-ID: <20150603015245.GN24666@dastard> (raw)
In-Reply-To: <20150602120648.GA22111@otto>
On Tue, Jun 02, 2015 at 02:06:48PM +0200, Anders Ossowicki wrote:
> On Mon, Jun 01, 2015 at 11:01:13PM +0200, Dave Chinner wrote:
> > Nothing should go wrong - XFS will essentially block until it gets
> > the memory it requires.
>
> Good to know, thanks!
>
> > > We're running on 3.18.13, built from kernel.org git.
> >
> > Right around the time that I was seeing all sorts of regressions
> > relating to low memory behaviour and the OOM killer....
>
> We fought with some high cpu load issues back in march, related to
> memory management, and we ended up on a recent longterm kernel.
> http://thread.gmane.org/gmane.linux.kernel.mm/129858
>
> > Ouch. 3TB of memory, and no higher order pages left? Do you have
> > memory compaction turned on? That should be reforming large pages in
> > this situation. What type of machine is it?
>
> Memory compaction is turned on. It's an off-the-shelf dell server with 4
> 12c Xeon processors.
>
> > Yes, memory fragmentation tends to be a MM problem; nothing XFS can
> > do about it.
>
> Ya, knowing we're not in immediate danger of a filesystem meltdown, I
> think we'll tackle the fragmentation issue next.
>
> > Especially as it appears that 2.8TB of your memory is in the page
> > cache and should be reclaimable.
>
> Indeed. I haven't been able to catch the issue while it was ongoing,
> since upgrading to 3.13.18, but my guess is that we're not reclaiming
> the cache fast enough for some reason, possibly because it takes too
> long to find the best reclaimable regions with so many fragment to sift
> through.
You can always try to drop the page cache to see if that solves the
problem...
> As for the pertinent system info:
>
> Linux 3.18.13 (we also saw the issue with 3.18.9)
> xfs_repair version 3.1.7
>
> 4x Intel Xeon E7-8857 v2
>
> $ cat /proc/meminfo
> MemTotal: 3170749444 kB
> MemFree: 18947564 kB
> MemAvailable: 2968870324 kB
> Buffers: 270704 kB
> Cached: 3008702200 kB
> SwapCached: 0 kB
> Active: 1617534420 kB
> Inactive: 1415684856 kB
> Active(anon): 156973416 kB
> Inactive(anon): 4856264 kB
> Active(file): 1460561004 kB
> Inactive(file): 1410828592 kB
This. You've got 2.8GB of reclaimable page cache there.
> Unevictable: 0 kB
> Mlocked: 0 kB
> SwapTotal: 25353212 kB
> SwapFree: 25353212 kB
> Dirty: 1228056 kB
> Writeback: 348024 kB
And very little of it is dirty, so it should all be immediately
reclaimable or compactable.
> Slab: 79729144 kB
> SReclaimable: 79040008 kB
80GB of slab caches as well - what is the output of /proc/slabinfo?
> We have three hardware raid'ed disks with XFS on them, one of which receives
> the bulk of the load. This is a raid 50 volume on SSDs with the raid controller
> running in writethrough mode.
It doesn't seem like writeback of dirty pages is the problem; more
the case that the page cache is rediculously huge and not being
reclaimed in a sane manner. Do you really need 2.8TB of cached file
data in memory for performance?
Cheers,
Dave.
--
Dave Chinner
david@fromorbit.com
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
next prev parent reply other threads:[~2015-06-03 1:52 UTC|newest]
Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-06-01 14:57 "XFS: possible memory allocation deadlock in kmem_alloc" on high memory machine Anders Ossowicki
2015-06-01 21:01 ` Dave Chinner
2015-06-02 12:06 ` Anders Ossowicki
2015-06-03 1:52 ` Dave Chinner [this message]
2015-06-03 7:07 ` Anders Ossowicki
2015-06-03 23:11 ` Dave Chinner
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20150603015245.GN24666@dastard \
--to=david@fromorbit.com \
--cc=aowi@novozymes.com \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox