qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Peter Xu <peterx@redhat.com>
To: Hao Xiang <hao.xiang@bytedance.com>
Cc: qemu-devel@nongnu.org, farosas@suse.de
Subject: Re: [PATCH 0/6] Introduce multifd zero page checking.
Date: Wed, 7 Feb 2024 11:39:20 +0800	[thread overview]
Message-ID: <ZcL7aIwo_Z8gn6h3@x1n> (raw)
In-Reply-To: <20240206231908.1792529-1-hao.xiang@bytedance.com>

On Tue, Feb 06, 2024 at 11:19:02PM +0000, Hao Xiang wrote:
> This patchset is based on Juan Quintela's old series here
> https://lore.kernel.org/all/20220802063907.18882-1-quintela@redhat.com/
> 
> In the multifd live migration model, there is a single migration main
> thread scanning the page map, queuing the pages to multiple multifd
> sender threads. The migration main thread runs zero page checking on
> every page before queuing the page to the sender threads. Zero page
> checking is a CPU intensive task and hence having a single thread doing
> all that doesn't scale well. This change introduces a new function
> to run the zero page checking on the multifd sender threads. This
> patchset also lays the ground work for future changes to offload zero
> page checking task to accelerator hardwares.
> 
> Use two Intel 4th generation Xeon servers for testing.
> 
> Architecture:        x86_64
> CPU(s):              192
> Thread(s) per core:  2
> Core(s) per socket:  48
> Socket(s):           2
> NUMA node(s):        2
> Vendor ID:           GenuineIntel
> CPU family:          6
> Model:               143
> Model name:          Intel(R) Xeon(R) Platinum 8457C
> Stepping:            8
> CPU MHz:             2538.624
> CPU max MHz:         3800.0000
> CPU min MHz:         800.0000
> 
> Perform multifd live migration with below setup:
> 1. VM has 100GB memory. All pages in the VM are zero pages.
> 2. Use tcp socket for live migratio.
> 3. Use 4 multifd channels and zero page checking on migration main thread.
> 4. Use 1/2/4 multifd channels and zero page checking on multifd sender
> threads.
> 5. Record migration total time from sender QEMU console's "info migrate"
> command.
> 6. Calculate throughput with "100GB / total time".
> 
> +------------------------------------------------------+
> |zero-page-checking | total-time(ms) | throughput(GB/s)|
> +------------------------------------------------------+
> |main-thread        | 9629           | 10.38GB/s       |
> +------------------------------------------------------+
> |multifd-1-threads  | 6182           | 16.17GB/s       |
> +------------------------------------------------------+
> |multifd-2-threads  | 4643           | 21.53GB/s       |
> +------------------------------------------------------+
> |multifd-4-threads  | 4143           | 24.13GB/s       |
> +------------------------------------------------------+

This "throughput" is slightly confusing; I was initially surprised to see a
large throughput for idle guests.  IMHO the "total-time" would explain.
Feel free to drop that column if there's a repost.

Did you check why 4 channels mostly already reached the top line?  Is it
because main thread is already spinning 100%?

Thanks,

-- 
Peter Xu



  parent reply	other threads:[~2024-02-07  3:40 UTC|newest]

Thread overview: 20+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-02-06 23:19 [PATCH 0/6] Introduce multifd zero page checking Hao Xiang
2024-02-06 23:19 ` [PATCH 1/6] migration/multifd: Add new migration option multifd-zero-page Hao Xiang
2024-02-07  3:44   ` Peter Xu
2024-02-08  0:49     ` [External] " Hao Xiang
2024-02-06 23:19 ` [PATCH 2/6] migration/multifd: Add zero pages and zero bytes counter to migration status interface Hao Xiang
2024-02-07  4:13   ` Peter Xu
2024-02-07  4:37     ` Peter Xu
2024-02-07  8:41       ` Jiri Denemark
2024-02-07 23:44         ` [External] " Hao Xiang
2024-02-08  2:51           ` Peter Xu
2024-02-06 23:19 ` [PATCH 3/6] migration/multifd: Support for zero pages transmission in multifd format Hao Xiang
2024-02-07  4:25   ` Peter Xu
2024-02-08 19:03     ` [External] " Hao Xiang
2024-02-06 23:19 ` [PATCH 4/6] migration/multifd: Zero page transmission on the multifd thread Hao Xiang
2024-02-07  4:45   ` Peter Xu
2024-02-06 23:19 ` [PATCH 5/6] migration/multifd: Enable zero page checking from multifd threads Hao Xiang
2024-02-06 23:19 ` [PATCH 6/6] migration/multifd: Add a new migration test case for legacy zero page checking Hao Xiang
2024-02-07  3:39 ` Peter Xu [this message]
2024-02-08  0:47   ` [External] Re: [PATCH 0/6] Introduce multifd " Hao Xiang
2024-02-08  2:36     ` Peter Xu

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=ZcL7aIwo_Z8gn6h3@x1n \
    --to=peterx@redhat.com \
    --cc=farosas@suse.de \
    --cc=hao.xiang@bytedance.com \
    --cc=qemu-devel@nongnu.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).