From: Juan Quintela <quintela@redhat.com>
To: Jitendra Kolhe <jitendra.kolhe@hpe.com>
Cc: qemu-devel@nongnu.org, kwolf@redhat.com,
peter.maydell@linaro.org, armbru@redhat.com,
renganathan.meenakshisundaram@hpe.com,
mohan_parthasarathy@hpe.com, pbonzini@redhat.com
Subject: Re: [Qemu-devel] [PATCH RFC] mem-prealloc: Reduce large guest start-up and migration time.
Date: Fri, 27 Jan 2017 13:53:34 +0100 [thread overview]
Message-ID: <87mvec3bqp.fsf@emacs.mitica> (raw)
In-Reply-To: <1483601042-6435-1-git-send-email-jitendra.kolhe@hpe.com> (Jitendra Kolhe's message of "Thu, 5 Jan 2017 12:54:02 +0530")
Jitendra Kolhe <jitendra.kolhe@hpe.com> wrote:
> Using "-mem-prealloc" option for a very large guest leads to huge guest
> start-up and migration time. This is because with "-mem-prealloc" option
> qemu tries to map every guest page (create address translations), and
> make sure the pages are available during runtime. virsh/libvirt by
> default, seems to use "-mem-prealloc" option in case the guest is
> configured to use huge pages. The patch tries to map all guest pages
> simultaneously by spawning multiple threads. Given the problem is more
> prominent for large guests, the patch limits the changes to the guests
> of at-least 64GB of memory size. Currently limiting the change to QEMU
> library functions on POSIX compliant host only, as we are not sure if
> the problem exists on win32. Below are some stats with "-mem-prealloc"
> option for guest configured to use huge pages.
>
> ------------------------------------------------------------------------
> Idle Guest | Start-up time | Migration time
> ------------------------------------------------------------------------
> Guest stats with 2M HugePage usage - single threaded (existing code)
> ------------------------------------------------------------------------
> 64 Core - 4TB | 54m11.796s | 75m43.843s
^^^^^^^^^^
> 64 Core - 1TB | 8m56.576s | 14m29.049s
> 64 Core - 256GB | 2m11.245s | 3m26.598s
> ------------------------------------------------------------------------
> Guest stats with 2M HugePage usage - map guest pages using 8 threads
> ------------------------------------------------------------------------
> 64 Core - 4TB | 5m1.027s | 34m10.565s
> 64 Core - 1TB | 1m10.366s | 8m28.188s
> 64 Core - 256GB | 0m19.040s | 2m10.148s
> -----------------------------------------------------------------------
> Guest stats with 2M HugePage usage - map guest pages using 16 threads
> -----------------------------------------------------------------------
> 64 Core - 4TB | 1m58.970s | 31m43.400s
^^^^^^^^^
Impressive, not everyday one get an speedup of 20 O:-)
> +static void *do_touch_pages(void *arg)
> +{
> + PageRange *range = (PageRange *)arg;
> + char *start_addr = range->addr;
> + uint64_t numpages = range->numpages;
> + uint64_t hpagesize = range->hpagesize;
> + uint64_t i = 0;
> +
> + for (i = 0; i < numpages; i++) {
> + memset(start_addr + (hpagesize * i), 0, 1);
I would use the range->addr and similar here directly, but it is just a
question of taste.
> - /* MAP_POPULATE silently ignores failures */
> - for (i = 0; i < numpages; i++) {
> - memset(area + (hpagesize * i), 0, 1);
> + /* touch pages simultaneously for memory >= 64G */
> + if (memory < (1ULL << 36)) {
64GB guest already took quite a bit of time, I think I would put it
always as min(num_vcpus, 16). So, we always execute the multiple theard
codepath?
But very nice, thanks.
Later, Juan.
next prev parent reply other threads:[~2017-01-27 12:53 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-01-05 7:24 [Qemu-devel] [PATCH RFC] mem-prealloc: Reduce large guest start-up and migration time Jitendra Kolhe
2017-01-27 12:53 ` Juan Quintela [this message]
2017-01-27 13:06 ` Paolo Bonzini
2017-01-30 8:19 ` Jitendra Kolhe
2017-01-27 13:03 ` Dr. David Alan Gilbert
2017-01-30 8:32 ` Jitendra Kolhe
2017-02-07 7:44 ` Jitendra Kolhe
2017-01-27 13:26 ` Daniel P. Berrange
2017-02-02 9:35 ` Jitendra Kolhe
2017-02-03 18:59 ` Paolo Bonzini
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=87mvec3bqp.fsf@emacs.mitica \
--to=quintela@redhat.com \
--cc=armbru@redhat.com \
--cc=jitendra.kolhe@hpe.com \
--cc=kwolf@redhat.com \
--cc=mohan_parthasarathy@hpe.com \
--cc=pbonzini@redhat.com \
--cc=peter.maydell@linaro.org \
--cc=qemu-devel@nongnu.org \
--cc=renganathan.meenakshisundaram@hpe.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).