public inbox for kexec@lists.infradead.org
 help / color / mirror / Atom feed
From: Vivek Goyal <vgoyal@redhat.com>
To: Lisa Mitchell <lisa.mitchell@hp.com>
Cc: "kexec@lists.infradead.org" <kexec@lists.infradead.org>,
	Atsushi Kumagai <kumagai-atsushi@mxc.nes.nec.co.jp>,
	jerry.hoemann@hp.com, Cliff Wickman <cpw@sgi.com>
Subject: Re: [RFC] makedumpfile-1.5.1 RC
Date: Tue, 20 Nov 2012 11:35:45 -0500	[thread overview]
Message-ID: <20121120163545.GC30248@redhat.com> (raw)
In-Reply-To: <1353413695.13097.131.camel@lisamlinux.fc.hp.com>

On Tue, Nov 20, 2012 at 05:14:55AM -0700, Lisa Mitchell wrote:

[..]
> I tested this makedumpfile v1.5.1-rc on a 4 TB DL980, on 2.6.32 based
> kernel, and got good results. With crashkernel=256M, and default
> settings (i.e. no cyclic buffer option selected), the dump successfully
> completed in about 2 hours, 40 minutes, and then I specified a cyclic
> buffer size of 48 M, and the dump completed in the same time, no
> measurable differences within the accuracy of our measurements. 

This sounds little odd to me.

- With smaller buffer size of 48M, it should have taken much more time
  to finish the dump as compared to when no restriction was put on
  buffer size. I am assuming that out of 256M reserved, say around 128MB
  was available for makedumpfile to use.

- Also 2 hours 40 minutes sounds a lot. Is it practical to wait that
  long for a machine to dump before it can be brought into service
  again? Do you have any data w.r.t older makedumpfile (which did not
  have cyclic buffer logic).

I have some data which I collected in 2008. 128GB system took roughly
4 minutes to filter and save dumpfile. So if we scale it linearly
then it should take around 32minutes per TB. Hence around 2 hours
8 minutes for a 4TB systems. Your numbers do seems to be in roughly
inline.

Still 2-2.5 hours seems too long to be able to filter and save core of a
4TB system. We will probably need to figure out what's taking so much of
time. May be we need to look into cliff wickman's idea of kernel returning
list of pfns to dump and make dump 20 time faster. I will love to have 4TB
system dumped in 6 minutes as opposed to 2 hrs. :-)

Thanks
Vivek

_______________________________________________
kexec mailing list
kexec@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/kexec

  reply	other threads:[~2012-11-20 16:35 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2012-11-16  8:15 [RFC] makedumpfile-1.5.1 RC Atsushi Kumagai
2012-11-20 12:14 ` Lisa Mitchell
2012-11-20 16:35   ` Vivek Goyal [this message]
2012-11-20 13:03     ` Lisa Mitchell
2012-11-20 21:46       ` Vivek Goyal
2012-11-20 19:05         ` Lisa Mitchell
2012-11-21 13:54           ` Vivek Goyal
2012-11-22  0:49             ` Hatayama, Daisuke
2012-11-26 16:02               ` Vivek Goyal
2012-12-04 13:31   ` Lisa Mitchell
2012-12-07  5:26     ` Atsushi Kumagai
2012-12-10 21:06       ` Lisa Mitchell
2012-12-13  5:06         ` Atsushi Kumagai
2012-12-18 17:20           ` Lisa Mitchell
2012-12-21  6:19             ` Atsushi Kumagai

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20121120163545.GC30248@redhat.com \
    --to=vgoyal@redhat.com \
    --cc=cpw@sgi.com \
    --cc=jerry.hoemann@hp.com \
    --cc=kexec@lists.infradead.org \
    --cc=kumagai-atsushi@mxc.nes.nec.co.jp \
    --cc=lisa.mitchell@hp.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox