public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Dave Chinner <david@fromorbit.com>
To: Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>
Cc: hannes@cmpxchg.org, dchinner@redhat.com, oleg@redhat.com,
	xfs@oss.sgi.com, mhocko@suse.cz, linux-mm@kvack.org,
	mgorman@suse.de, rientjes@google.com, akpm@linux-foundation.org,
	torvalds@linux-foundation.org
Subject: Re: How to handle TIF_MEMDIE stalls?
Date: Sun, 22 Feb 2015 08:48:23 +1100	[thread overview]
Message-ID: <20150221214823.GJ12722@dastard> (raw)
In-Reply-To: <201502212012.BJJ39083.LQFOtJFSHMVOFO@I-love.SAKURA.ne.jp>

On Sat, Feb 21, 2015 at 08:12:08PM +0900, Tetsuo Handa wrote:
> My main issue is
> 
>   c) whether to oom-kill more processes when the OOM victim cannot be
>      terminated presumably due to the OOM killer deadlock.
> 
> Dave Chinner wrote:
> > On Fri, Feb 20, 2015 at 07:36:33PM +0900, Tetsuo Handa wrote:
> > > Dave Chinner wrote:
> > > > I really don't care about the OOM Killer corner cases - it's
> > > > completely the wrong way line of development to be spending time on
> > > > and you aren't going to convince me otherwise. The OOM killer a
> > > > crutch used to justify having a memory allocation subsystem that
> > > > can't provide forward progress guarantee mechanisms to callers that
> > > > need it.
> > > 
> > > I really care about the OOM Killer corner cases, for I'm
> > > 
> > >   (1) seeing trouble cases which occurred in enterprise systems
> > >       under OOM conditions
> > 
> > You reach OOM, then your SLAs are dead and buried. Reboot the
> > box - its a much more reliable way of returning to a working system
> > than playing Russian Roulette with the OOM killer.
> 
> What Service Level Agreements? Such troubles are occurring on RHEL systems
> where users are not sitting in front of the console. Unless somebody is
> sitting in front of the console in order to do SysRq-b when troubles
> occur, the down time of system will become significantly longer.
>
> What mechanisms are available for minimizing the down time of system
> when troubles under OOM condition occur? Software/hardware watchdog?
> Indeed they may help, but they may be triggered prematurely when the
> system has not entered into the OOM condition. Only the OOM killer knows.

# echo 1 > /proc/sys/vm/panic_on_oom

....

> We have memory cgroups to reduce the possibility of triggering the OOM
> killer, though there will be several bugs remaining in RHEL kernels
> which make administrators hesitate to use memory cgroups.

Fix upstream first, then worry about vendor kernels.

....

> Not only we cannot expect that the OOM killer messages being saved to
> /var/log/messages under the OOM killer deadlock condition, but also

CONFIG_PSTORE=y and configure appropriately from there.

> we do not emit the OOM killer messages if we hit

So add a warning.

> If you want to stop people from playing Russian Roulette with the OOM
> killer, please remove the OOM killer code entirely from RHEL kernels so that
> people must use their systems with hardcoded /proc/sys/vm/panic_on_oom == 1
> setting. Can you do it?

No. You need to go through vendor channels to get a vendor kernel
config change made.

Cheers,

Dave.
-- 
Dave Chinner
david@fromorbit.com

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

  reply	other threads:[~2015-02-21 21:48 UTC|newest]

Thread overview: 83+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20141230112158.GA15546@dhcp22.suse.cz>
     [not found] ` <201502092044.JDG39081.LVFOOtFHQFOMSJ@I-love.SAKURA.ne.jp>
     [not found]   ` <201502102258.IFE09888.OVQFJOMSFtOLFH@I-love.SAKURA.ne.jp>
     [not found]     ` <20150210151934.GA11212@phnom.home.cmpxchg.org>
     [not found]       ` <201502111123.ICD65197.FMLOHSQJFVOtFO@I-love.SAKURA.ne.jp>
     [not found]         ` <201502172123.JIE35470.QOLMVOFJSHOFFt@I-love.SAKURA.ne.jp>
     [not found]           ` <20150217125315.GA14287@phnom.home.cmpxchg.org>
2015-02-17 22:54             ` How to handle TIF_MEMDIE stalls? Dave Chinner
2015-02-17 23:32               ` Dave Chinner
2015-02-18  8:25               ` Michal Hocko
2015-02-18 10:48                 ` Dave Chinner
2015-02-18 12:16                   ` Michal Hocko
2015-02-18 21:31                     ` Dave Chinner
2015-02-19  9:40                       ` Michal Hocko
2015-02-19 22:03                         ` Dave Chinner
2015-02-20  9:27                           ` Michal Hocko
2015-02-19 11:01                     ` Johannes Weiner
2015-02-19 12:29                       ` Michal Hocko
2015-02-19 12:58                         ` Michal Hocko
2015-02-19 15:29                           ` Tetsuo Handa
2015-02-19 21:53                             ` Tetsuo Handa
2015-02-20  9:13                             ` Michal Hocko
2015-02-20 13:37                               ` Stefan Ring
2015-02-19 13:29                         ` Tetsuo Handa
2015-02-20  9:10                           ` Michal Hocko
2015-02-20 12:20                             ` Tetsuo Handa
2015-02-20 12:38                               ` Michal Hocko
2015-02-19 21:43                         ` Dave Chinner
2015-02-20 12:48                           ` Michal Hocko
2015-02-20 23:09                             ` Dave Chinner
2015-02-19 10:24               ` Johannes Weiner
2015-02-19 22:52                 ` Dave Chinner
2015-02-20 10:36                   ` Tetsuo Handa
2015-02-20 23:15                     ` Dave Chinner
2015-02-21  3:20                       ` Theodore Ts'o
2015-02-21  9:19                         ` Andrew Morton
2015-02-21 13:48                           ` Tetsuo Handa
2015-02-21 21:38                           ` Dave Chinner
2015-02-22  0:20                           ` Johannes Weiner
2015-02-23 10:48                             ` Michal Hocko
2015-02-23 11:23                               ` Tetsuo Handa
2015-02-23 21:33                             ` David Rientjes
2015-02-21 12:00                         ` Tetsuo Handa
2015-02-23 10:26                         ` Michal Hocko
2015-02-21 11:12                       ` Tetsuo Handa
2015-02-21 21:48                         ` Dave Chinner [this message]
2015-02-21 23:52                   ` Johannes Weiner
2015-02-23  0:45                     ` Dave Chinner
2015-02-23  1:29                       ` Andrew Morton
2015-02-23  7:32                         ` Dave Chinner
2015-02-27 18:24                           ` Vlastimil Babka
2015-02-28  0:03                             ` Dave Chinner
2015-02-28 15:17                               ` Theodore Ts'o
2015-03-02  9:39                           ` Vlastimil Babka
2015-03-02 22:31                             ` Dave Chinner
2015-03-03  9:13                               ` Vlastimil Babka
2015-03-04  1:33                                 ` Dave Chinner
2015-03-04  8:50                                   ` Vlastimil Babka
2015-03-04 11:03                                     ` Dave Chinner
2015-03-07  0:20                               ` Johannes Weiner
2015-03-07  3:43                                 ` Dave Chinner
2015-03-07 15:08                                   ` Johannes Weiner
2015-03-02 20:22                           ` Johannes Weiner
2015-03-02 23:12                             ` Dave Chinner
2015-03-03  2:50                               ` Johannes Weiner
2015-03-04  6:52                                 ` Dave Chinner
2015-03-04 15:04                                   ` Johannes Weiner
2015-03-04 17:38                                     ` Theodore Ts'o
2015-03-04 23:17                                       ` Dave Chinner
2015-02-28 16:29                       ` Johannes Weiner
2015-02-28 16:41                         ` Theodore Ts'o
2015-02-28 22:15                           ` Johannes Weiner
2015-03-01 11:17                             ` Tetsuo Handa
2015-03-06 11:53                               ` Tetsuo Handa
2015-03-01 13:43                             ` Theodore Ts'o
2015-03-01 16:15                               ` Johannes Weiner
2015-03-01 19:36                                 ` Theodore Ts'o
2015-03-01 20:44                                   ` Johannes Weiner
2015-03-01 20:17                               ` Johannes Weiner
2015-03-01 21:48                             ` Dave Chinner
2015-03-02  0:17                               ` Dave Chinner
2015-03-02 12:46                                 ` Brian Foster
2015-02-28 18:36                       ` Vlastimil Babka
2015-03-02 15:18                       ` Michal Hocko
2015-03-02 16:05                         ` Johannes Weiner
2015-03-02 17:10                           ` Michal Hocko
2015-03-02 17:27                             ` Johannes Weiner
2015-03-02 16:39                         ` Theodore Ts'o
2015-03-02 16:58                           ` Michal Hocko
2015-03-04 12:52                             ` Dave Chinner

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20150221214823.GJ12722@dastard \
    --to=david@fromorbit.com \
    --cc=akpm@linux-foundation.org \
    --cc=dchinner@redhat.com \
    --cc=hannes@cmpxchg.org \
    --cc=linux-mm@kvack.org \
    --cc=mgorman@suse.de \
    --cc=mhocko@suse.cz \
    --cc=oleg@redhat.com \
    --cc=penguin-kernel@I-love.SAKURA.ne.jp \
    --cc=rientjes@google.com \
    --cc=torvalds@linux-foundation.org \
    --cc=xfs@oss.sgi.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox