From: Michal Hocko <mhocko@suse.cz>
To: Nikolay Borisov <kernel@kyup.com>
Cc: Linux MM <linux-mm@kvack.org>
Subject: Re: Softlockup during memory allocation
Date: Wed, 23 Nov 2016 08:50:37 +0100 [thread overview]
Message-ID: <20161123075035.GF2864@dhcp22.suse.cz> (raw)
In-Reply-To: <20161123074947.GE2864@dhcp22.suse.cz>
On Wed 23-11-16 08:49:47, Michal Hocko wrote:
> On Wed 23-11-16 09:44:45, Nikolay Borisov wrote:
> >
> >
> > On 11/22/2016 07:02 PM, Michal Hocko wrote:
> > > On Tue 22-11-16 16:35:38, Nikolay Borisov wrote:
> > >>
> > >>
> > >> On 11/22/2016 04:30 PM, Michal Hocko wrote:
> > >>> On Tue 22-11-16 10:56:51, Nikolay Borisov wrote:
> > >>>>
> > >>>>
> > >>>> On 11/21/2016 07:31 AM, Michal Hocko wrote:
> > >>>>> Hi,
> > >>>>> I am sorry for a late response, but I was offline until this weekend. I
> > >>>>> will try to get to this email ASAP but it might take some time.
> > >>>>
> > >>>> No worries. I did some further digging up and here is what I got, which
> > >>>> I believe is rather strange:
> > >>>>
> > >>>> struct scan_control {
> > >>>> nr_to_reclaim = 32,
> > >>>> gfp_mask = 37880010,
> > >>>> order = 0,
> > >>>> nodemask = 0x0,
> > >>>> target_mem_cgroup = 0xffff8823990d1400,
> > >>>> priority = 7,
> > >>>> may_writepage = 1,
> > >>>> may_unmap = 1,
> > >>>> may_swap = 0,
> > >>>> may_thrash = 1,
> > >>>> hibernation_mode = 0,
> > >>>> compaction_ready = 0,
> > >>>> nr_scanned = 0,
> > >>>> nr_reclaimed = 0
> > >>>> }
> > >>>>
> > >>>> Parsing: 37880010
> > >>>> #define ___GFP_HIGHMEM 0x02
> > >>>> #define ___GFP_MOVABLE 0x08
> > >>>> #define ___GFP_IO 0x40
> > >>>> #define ___GFP_FS 0x80
> > >>>> #define ___GFP_HARDWALL 0x20000
> > >>>> #define ___GFP_DIRECT_RECLAIM 0x400000
> > >>>> #define ___GFP_KSWAPD_RECLAIM 0x2000000
> > >>>>
> > >>>> And initial_priority is 12 (DEF_PRIORITY). Given that nr_scanned is 0
> > >>>> and priority is 7 this means we've gone 5 times through the do {} while
> > >>>> in do_try_to_free_pages. Also total_scanned seems to be 0. Here is the
> > >>>> zone which was being reclaimed :
> > >>
> > >> This is also very strange that total_scanned is 0.
> > >>
> > >>
> > >>>>
> > >>>> http://sprunge.us/hQBi
> > >>>
> > >>> LRUs on that zones seem to be empty from a quick glance. kmem -z in the
> > >>> crash can give you per zone counters much more nicely.
> > >>>
> > >>
> > >> So here are the populated zones:
> > > [...]
> > >> NODE: 0 ZONE: 2 ADDR: ffff88207fffcf00 NAME: "Normal"
> > >> SIZE: 33030144 MIN/LOW/HIGH: 22209/27761/33313
> > >> VM_STAT:
> > >> NR_FREE_PAGES: 62436
> > >> NR_ALLOC_BATCH: 2024
> > >> NR_INACTIVE_ANON: 8177867
> > >> NR_ACTIVE_ANON: 5407176
> > >> NR_INACTIVE_FILE: 5804642
> > >> NR_ACTIVE_FILE: 9694170
> > >
> > > So your LRUs are definitely not empty as I have thought. Having
> > > 0 pages scanned is indeed very strange. We do reset sc->nr_scanned
> > > for each priority but my understanding was that you are looking at a
> > > state where we are somwhere in the middle of shrink_zones. Moreover
> > > total_scanned should be cumulative.
> >
> > So the server began acting wonky. People logged on it and saw the
> > softlockup as per my initial email. They then initiated a crashdump via
> > sysrq since most commands weren't going through (e.g. forking) so
> > crashing it was a last resort measure. After that I start looking at the
> > crashdump and observe that prior to the crash machine seems to have
> > locked up judging from the dmesg logs. However, when I manually inspect
> > the *current* (and current being at the time the crash was actually
> > initiated) state of the processes reported as softlock up they seem to
> > have made progress are now in
> > shrink_zone->shrink_lruvec->shrink_inactive_list->_cond_resched->__schedule
>
> OK, I see.
One more thing. You might get a better picture if you configure your
system to panic on soft lockup so that you get the state at the time of
the problem.
--
Michal Hocko
SUSE Labs
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2016-11-23 7:50 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-11-01 8:12 Softlockup during memory allocation Nikolay Borisov
2016-11-01 8:16 ` Nikolay Borisov
2016-11-02 19:00 ` Vlastimil Babka
2016-11-04 3:46 ` Hugh Dickins
2016-11-04 12:18 ` Nikolay Borisov
2016-11-13 22:02 ` Nikolay Borisov
2016-11-21 5:31 ` Michal Hocko
2016-11-22 8:56 ` Nikolay Borisov
2016-11-22 14:30 ` Michal Hocko
2016-11-22 14:32 ` Michal Hocko
2016-11-22 14:46 ` Nikolay Borisov
2016-11-22 14:35 ` Nikolay Borisov
2016-11-22 17:02 ` Michal Hocko
2016-11-23 7:44 ` Nikolay Borisov
2016-11-23 7:49 ` Michal Hocko
2016-11-23 7:50 ` Michal Hocko [this message]
2016-11-24 11:45 ` Nikolay Borisov
2016-11-24 12:12 ` Michal Hocko
2016-11-24 13:09 ` Nikolay Borisov
2016-11-25 9:00 ` Michal Hocko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20161123075035.GF2864@dhcp22.suse.cz \
--to=mhocko@suse.cz \
--cc=kernel@kyup.com \
--cc=linux-mm@kvack.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).