linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Stefan Bader <stefan.bader@canonical.com>
To: David Rientjes <rientjes@google.com>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>,
	linux-kernel@vger.kernel.org,
	Eric Dumazet <eric.dumazet@gmail.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Peter Zijlstra <a.p.zijlstra@chello.nl>
Subject: Re: fs/stat: Reduce memory requirements for stat_open
Date: Wed, 25 Jun 2014 09:04:10 +0200	[thread overview]
Message-ID: <53AA746A.3030303@canonical.com> (raw)
In-Reply-To: <alpine.DEB.2.02.1406241642280.22030@chino.kir.corp.google.com>

[-- Attachment #1: Type: text/plain, Size: 1640 bytes --]

On 25.06.2014 01:44, David Rientjes wrote:
> On Thu, 12 Jun 2014, Stefan Bader wrote:
> 
>> doh, so you guys have been hit by that before. And I have missed the fact that
>> single_open is special. Which makes the change for the upper limit do the wrong
>> thing. While long-term it sounds like changing it to vmalloc or iterative reads
>> sounds better, maybe the change from possible to online cpus might be something
>> that is better acceptable as a quick fix... Not sure. At least this giving back
>> a bit of attention to the matter and it is not only affecting zSeries. x86
>> starts to see hw that requires a similar high possible cpus... :)
>>
> 
> Heiko's patches that should fix this problem are now in -mm, so it would 
> be nice to know if there are any existing issues that haven't been 
> addressed yet with your bug report.  See 
> http://www.ozlabs.org/~akpm/mmotm/
> 

Heiko and I both had the same issue. Since some x86 hardware also reaches a lot
of CPUs (hyperthreads included), we bumped the possible number of CPUs to 256 at
least for the 64bit kernel. And that resulted in failed accesses to /proc/stat
when memory became fragmented.
So the first patch will avoid this on most systems. I have not seen this myself,
but I would expect him to be happy with 1/2 already. For really excessive
hardware 2/2 will close the gap.
Since this is no critical bug, I am fine with 3.17, too. I have not done so,
yet, but I could let our reporter try the patches (again, probably not verifying
the second part). Just waited to do so to see whether the code settles down to
these changes.

-Stefan


[-- Attachment #2: OpenPGP digital signature --]
[-- Type: application/pgp-signature, Size: 901 bytes --]

  reply	other threads:[~2014-06-25  7:04 UTC|newest]

Thread overview: 9+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-06-12 13:00 fs/stat: Reduce memory requirements for stat_open Stefan Bader
2014-06-12 13:41 ` Heiko Carstens
2014-06-12 14:02   ` Stefan Bader
2014-06-24 23:44     ` David Rientjes
2014-06-25  7:04       ` Stefan Bader [this message]
2014-06-25 22:52         ` David Rientjes
2014-06-25 22:54           ` Andrew Morton
2014-07-08 13:09 ` Peter Zijlstra
2014-07-08 13:30   ` Stefan Bader

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=53AA746A.3030303@canonical.com \
    --to=stefan.bader@canonical.com \
    --cc=a.p.zijlstra@chello.nl \
    --cc=akpm@linux-foundation.org \
    --cc=eric.dumazet@gmail.com \
    --cc=heiko.carstens@de.ibm.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=rientjes@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).