From: Vivek Goyal <vgoyal@redhat.com>
To: Lisa Mitchell <lisa.mitchell@hp.com>
Cc: "kexec@lists.infradead.org" <kexec@lists.infradead.org>,
Atsushi Kumagai <kumagai-atsushi@mxc.nes.nec.co.jp>,
"Hoemann, Jerry" <jerry.hoemann@hp.com>,
Cliff Wickman <cpw@sgi.com>
Subject: Re: [RFC] makedumpfile-1.5.1 RC
Date: Tue, 20 Nov 2012 16:46:06 -0500 [thread overview]
Message-ID: <20121120214606.GB3823@redhat.com> (raw)
In-Reply-To: <1353416600.13097.152.camel@lisamlinux.fc.hp.com>
On Tue, Nov 20, 2012 at 06:03:20AM -0700, Lisa Mitchell wrote:
> On Tue, 2012-11-20 at 16:35 +0000, Vivek Goyal wrote:
> > On Tue, Nov 20, 2012 at 05:14:55AM -0700, Lisa Mitchell wrote:
> >
> > [..]
> > > I tested this makedumpfile v1.5.1-rc on a 4 TB DL980, on 2.6.32 based
> > > kernel, and got good results. With crashkernel=256M, and default
> > > settings (i.e. no cyclic buffer option selected), the dump successfully
> > > completed in about 2 hours, 40 minutes, and then I specified a cyclic
> > > buffer size of 48 M, and the dump completed in the same time, no
> > > measurable differences within the accuracy of our measurements.
> >
> > This sounds little odd to me.
> >
> > - With smaller buffer size of 48M, it should have taken much more time
> > to finish the dump as compared to when no restriction was put on
> > buffer size. I am assuming that out of 256M reserved, say around 128MB
> > was available for makedumpfile to use.
> >
> > - Also 2 hours 40 minutes sounds a lot. Is it practical to wait that
> > long for a machine to dump before it can be brought into service
> > again? Do you have any data w.r.t older makedumpfile (which did not
> > have cyclic buffer logic).
> >
> > I have some data which I collected in 2008. 128GB system took roughly
> > 4 minutes to filter and save dumpfile. So if we scale it linearly
> > then it should take around 32minutes per TB. Hence around 2 hours
> > 8 minutes for a 4TB systems. Your numbers do seems to be in roughly
> > inline.
> >
> > Still 2-2.5 hours seems too long to be able to filter and save core of a
> > 4TB system. We will probably need to figure out what's taking so much of
> > time. May be we need to look into cliff wickman's idea of kernel returning
> > list of pfns to dump and make dump 20 time faster. I will love to have 4TB
> > system dumped in 6 minutes as opposed to 2 hrs. :-)
> >
> > Thanks
> > Vivek
>
> As I stated, I don't really have precise performance data here, but the
> time I got was comparable to the rough 3-4 hours with a larger
> crashkernel size that I got a successful dump on this same system with a
> makedumpfile v1.4. We haven't made a good apples-apples comparison
> between the two at this point, but this is how long this 4 TB system has
> been taking to dump, dump level =31, so we feel we are in the same
> ballpark with makedumpfile v1.5.1.
>
> It does seem that the "Excluding pages" parts take up a lot of the time
> in the dump, as opposed to the copying, but I don't have a good
> breakdown.
>
> I have added the debug mem_level 3 to kdump.conf file, and have seen
> used memory on this machine recorded right before makedumpfile creates
> the bitmap and starts filtering be around 140 MB, and have seen
> makedumpfile fail, with OOM killer active after this point with a
> crashkernel size of 256 MB or 384 MB using makedumpfile v1.4.
That's sounds right. makedumpfile requires roughly 64MB of memory per TB.
So to be able to filter out 4TB, one needs 256MB of free memory. So no
wonder makedumpfile v1.4 will fail with 140MB free.
>
> So makedumpfile v1.5.1 solves the above problem, and allows us to
> successfully dump a 4 TB system with these smaller crashkernel sizes.
Ok, great. Good to know v1.5.1 is atleast allowing dumping higher memory
systems with smaller reserved amount of memory.
>
> We do need much better performance numbers to insure no regression from
> makedumpfile v1.4, but I wanted you to get the feedback at least of what
> testing we had done, and that it appears it is solving the primary
> problem we were interested in, that we could dump many terabytes of
> memory with crashkernel sizes fixed at 384 MB or below.
Fair enough.
Thanks
Vivek
_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec
next prev parent reply other threads:[~2012-11-20 21:46 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-11-16 8:15 [RFC] makedumpfile-1.5.1 RC Atsushi Kumagai
2012-11-20 12:14 ` Lisa Mitchell
2012-11-20 16:35 ` Vivek Goyal
2012-11-20 13:03 ` Lisa Mitchell
2012-11-20 21:46 ` Vivek Goyal [this message]
2012-11-20 19:05 ` Lisa Mitchell
2012-11-21 13:54 ` Vivek Goyal
2012-11-22 0:49 ` Hatayama, Daisuke
2012-11-26 16:02 ` Vivek Goyal
2012-12-04 13:31 ` Lisa Mitchell
2012-12-07 5:26 ` Atsushi Kumagai
2012-12-10 21:06 ` Lisa Mitchell
2012-12-13 5:06 ` Atsushi Kumagai
2012-12-18 17:20 ` Lisa Mitchell
2012-12-21 6:19 ` Atsushi Kumagai
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20121120214606.GB3823@redhat.com \
--to=vgoyal@redhat.com \
--cc=cpw@sgi.com \
--cc=jerry.hoemann@hp.com \
--cc=kexec@lists.infradead.org \
--cc=kumagai-atsushi@mxc.nes.nec.co.jp \
--cc=lisa.mitchell@hp.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox