public inbox for kexec@lists.infradead.org
 help / color / mirror / Atom feed
From: Lisa Mitchell <lisa.mitchell@hp.com>
To: Vivek Goyal <vgoyal@redhat.com>
Cc: "kexec@lists.infradead.org" <kexec@lists.infradead.org>,
	Atsushi Kumagai <kumagai-atsushi@mxc.nes.nec.co.jp>,
	"Hoemann, Jerry" <jerry.hoemann@hp.com>,
	Cliff Wickman <cpw@sgi.com>
Subject: Re: [RFC] makedumpfile-1.5.1 RC
Date: Tue, 20 Nov 2012 06:03:20 -0700	[thread overview]
Message-ID: <1353416600.13097.152.camel@lisamlinux.fc.hp.com> (raw)
In-Reply-To: <20121120163545.GC30248@redhat.com>

On Tue, 2012-11-20 at 16:35 +0000, Vivek Goyal wrote:
> On Tue, Nov 20, 2012 at 05:14:55AM -0700, Lisa Mitchell wrote:
> 
> [..]
> > I tested this makedumpfile v1.5.1-rc on a 4 TB DL980, on 2.6.32 based
> > kernel, and got good results. With crashkernel=256M, and default
> > settings (i.e. no cyclic buffer option selected), the dump successfully
> > completed in about 2 hours, 40 minutes, and then I specified a cyclic
> > buffer size of 48 M, and the dump completed in the same time, no
> > measurable differences within the accuracy of our measurements. 
> 
> This sounds little odd to me.
> 
> - With smaller buffer size of 48M, it should have taken much more time
>   to finish the dump as compared to when no restriction was put on
>   buffer size. I am assuming that out of 256M reserved, say around 128MB
>   was available for makedumpfile to use.
> 
> - Also 2 hours 40 minutes sounds a lot. Is it practical to wait that
>   long for a machine to dump before it can be brought into service
>   again? Do you have any data w.r.t older makedumpfile (which did not
>   have cyclic buffer logic).
> 
> I have some data which I collected in 2008. 128GB system took roughly
> 4 minutes to filter and save dumpfile. So if we scale it linearly
> then it should take around 32minutes per TB. Hence around 2 hours
> 8 minutes for a 4TB systems. Your numbers do seems to be in roughly
> inline.
> 
> Still 2-2.5 hours seems too long to be able to filter and save core of a
> 4TB system. We will probably need to figure out what's taking so much of
> time. May be we need to look into cliff wickman's idea of kernel returning
> list of pfns to dump and make dump 20 time faster. I will love to have 4TB
> system dumped in 6 minutes as opposed to 2 hrs. :-)
> 
> Thanks
> Vivek

As I stated, I don't really have precise performance data here, but the
time I got was comparable to the rough 3-4 hours with a larger
crashkernel size that I got a successful dump on this same system with a
makedumpfile v1.4.  We haven't made a good apples-apples comparison
between the two at this point, but this is how long this 4 TB system has
been taking to dump, dump level =31, so we feel we are in the same
ballpark with makedumpfile v1.5.1.  

It does seem that the "Excluding pages" parts take up a lot of the time
in the dump, as opposed to the copying, but I don't have a good
breakdown.  

I have added the debug mem_level 3 to kdump.conf file, and have seen
used memory on this machine  recorded right before makedumpfile creates
the bitmap and starts filtering be around 140 MB, and have seen
makedumpfile fail, with OOM killer active after this point with a
crashkernel size of 256 MB or 384 MB using makedumpfile v1.4.  

So makedumpfile v1.5.1 solves the above problem, and allows us to
successfully dump a 4 TB system with these smaller crashkernel sizes.

We do need much better performance numbers  to insure no regression from
makedumpfile v1.4, but I wanted you to get the feedback at least of what
testing we had done, and that it appears it is solving the primary
problem we were interested in, that we could dump many terabytes of
memory with crashkernel sizes fixed at 384 MB or below. 



_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec

  reply	other threads:[~2012-11-20 17:06 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2012-11-16  8:15 [RFC] makedumpfile-1.5.1 RC Atsushi Kumagai
2012-11-20 12:14 ` Lisa Mitchell
2012-11-20 16:35   ` Vivek Goyal
2012-11-20 13:03     ` Lisa Mitchell [this message]
2012-11-20 21:46       ` Vivek Goyal
2012-11-20 19:05         ` Lisa Mitchell
2012-11-21 13:54           ` Vivek Goyal
2012-11-22  0:49             ` Hatayama, Daisuke
2012-11-26 16:02               ` Vivek Goyal
2012-12-04 13:31   ` Lisa Mitchell
2012-12-07  5:26     ` Atsushi Kumagai
2012-12-10 21:06       ` Lisa Mitchell
2012-12-13  5:06         ` Atsushi Kumagai
2012-12-18 17:20           ` Lisa Mitchell
2012-12-21  6:19             ` Atsushi Kumagai

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1353416600.13097.152.camel@lisamlinux.fc.hp.com \
    --to=lisa.mitchell@hp.com \
    --cc=cpw@sgi.com \
    --cc=jerry.hoemann@hp.com \
    --cc=kexec@lists.infradead.org \
    --cc=kumagai-atsushi@mxc.nes.nec.co.jp \
    --cc=vgoyal@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox