From: Peter Xu <peterx@redhat.com>
To: "Liu, Yuan1" <yuan1.liu@intel.com>
Cc: "farosas@suse.de" <farosas@suse.de>,
"qemu-devel@nongnu.org" <qemu-devel@nongnu.org>,
"hao.xiang@bytedance.com" <hao.xiang@bytedance.com>,
"bryan.zhang@bytedance.com" <bryan.zhang@bytedance.com>,
"Zou, Nanhai" <nanhai.zou@intel.com>
Subject: Re: [PATCH v5 0/7] Live Migration With IAA
Date: Thu, 28 Mar 2024 11:22:08 -0400 [thread overview]
Message-ID: <ZgWLIJ0U1c0WySio@x1n> (raw)
In-Reply-To: <PH7PR11MB59411F5377A4E087D5FEA719A33B2@PH7PR11MB5941.namprd11.prod.outlook.com>
On Thu, Mar 28, 2024 at 03:02:30AM +0000, Liu, Yuan1 wrote:
> Yes, I will support software fallback to ensure CI testing and users can
> still use qpl compression without IAA hardware.
>
> Although the qpl software solution will have better performance than zlib,
> I still don't think it has a greater advantage than zstd. I don't think there
> is a need to add a migration option to configure the qpl software or hardware path.
> So I will still only use QPL as an independent compression in the next version, and
> no other migration options are needed.
That should be fine.
>
> I will also add a guide to qpl-compression.rst about IAA permission issues and how to
> determine whether the hardware path is available.
OK.
[...]
> > > Yes, I use iperf3 to check the bandwidth for one core, the bandwith is
> > 60Gbps.
> > > [ ID] Interval Transfer Bitrate Retr Cwnd
> > > [ 5] 0.00-1.00 sec 7.00 GBytes 60.1 Gbits/sec 0 2.87 MBytes
> > > [ 5] 1.00-2.00 sec 7.05 GBytes 60.6 Gbits/sec 0 2.87 Mbytes
> > >
> > > And in the live migration test, a multifd thread's CPU utilization is
> > almost 100%
> >
> > This 60Gpbs per-channel is definitely impressive..
> >
> > Have you tried migration without multifd on your system? Would that also
> > perform similarly v.s. 2 channels multifd?
>
> Simple Test result below:
> VM Type: 16vCPU, 64G memory
> Workload in VM: fill 56G memory with Silesia data and vCPUs are idle
> Migration Configurations:
> 1. migrate_set_parameter max-bandwidth 100G
> 2. migrate_set_parameter downtime-limit 300
> 3. migrate_set_capability multifd on (multiFD test case)
> 4. migrate_set_parameter multifd-channels 2 (multiFD test case)
>
> Totaltime (ms) Downtime (ms) Throughput (mbps) Pages-per-second
> without Multifd 23580 307 21221 689588
> Multifd 2 7657 198 65410 2221176
Thanks for the test results.
So I am guessing the migration overheads besides pushing the socket is high
enough to make it drop drastically, even if in this case zero detection
shouldn't play a major role considering most of guest mem is pre-filled.
--
Peter Xu
next prev parent reply other threads:[~2024-03-28 15:22 UTC|newest]
Thread overview: 45+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-03-19 16:45 [PATCH v5 0/7] Live Migration With IAA Yuan Liu
2024-03-19 16:45 ` [PATCH v5 1/7] docs/migration: add qpl compression feature Yuan Liu
2024-03-26 17:58 ` Peter Xu
2024-03-27 2:14 ` Liu, Yuan1
2024-03-19 16:45 ` [PATCH v5 2/7] migration/multifd: put IOV initialization into compression method Yuan Liu
2024-03-20 15:18 ` Fabiano Rosas
2024-03-20 15:32 ` Liu, Yuan1
2024-03-19 16:45 ` [PATCH v5 3/7] configure: add --enable-qpl build option Yuan Liu
2024-03-20 8:55 ` Thomas Huth
2024-03-20 8:56 ` Thomas Huth
2024-03-20 14:34 ` Liu, Yuan1
2024-03-20 10:31 ` Daniel P. Berrangé
2024-03-20 14:42 ` Liu, Yuan1
2024-03-19 16:45 ` [PATCH v5 4/7] migration/multifd: add qpl compression method Yuan Liu
2024-03-27 19:49 ` Peter Xu
2024-03-28 3:03 ` Liu, Yuan1
2024-03-19 16:45 ` [PATCH v5 5/7] migration/multifd: implement initialization of qpl compression Yuan Liu
2024-03-20 10:42 ` Daniel P. Berrangé
2024-03-20 15:02 ` Liu, Yuan1
2024-03-20 15:20 ` Daniel P. Berrangé
2024-03-20 16:04 ` Liu, Yuan1
2024-03-20 15:34 ` Peter Xu
2024-03-20 16:23 ` Liu, Yuan1
2024-03-20 20:31 ` Peter Xu
2024-03-21 1:37 ` Liu, Yuan1
2024-03-21 15:28 ` Peter Xu
2024-03-22 2:06 ` Liu, Yuan1
2024-03-22 14:47 ` Liu, Yuan1
2024-03-22 16:40 ` Peter Xu
2024-03-27 19:25 ` Peter Xu
2024-03-28 2:32 ` Liu, Yuan1
2024-03-28 15:16 ` Peter Xu
2024-03-29 2:04 ` Liu, Yuan1
2024-03-19 16:45 ` [PATCH v5 6/7] migration/multifd: implement qpl compression and decompression Yuan Liu
2024-03-19 16:45 ` [PATCH v5 7/7] tests/migration-test: add qpl compression test Yuan Liu
2024-03-20 10:45 ` Daniel P. Berrangé
2024-03-20 15:30 ` Liu, Yuan1
2024-03-20 15:39 ` Daniel P. Berrangé
2024-03-20 16:26 ` Liu, Yuan1
2024-03-26 20:30 ` [PATCH v5 0/7] Live Migration With IAA Peter Xu
2024-03-27 3:20 ` Liu, Yuan1
2024-03-27 19:46 ` Peter Xu
2024-03-28 3:02 ` Liu, Yuan1
2024-03-28 15:22 ` Peter Xu [this message]
2024-03-29 3:33 ` Liu, Yuan1
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=ZgWLIJ0U1c0WySio@x1n \
--to=peterx@redhat.com \
--cc=bryan.zhang@bytedance.com \
--cc=farosas@suse.de \
--cc=hao.xiang@bytedance.com \
--cc=nanhai.zou@intel.com \
--cc=qemu-devel@nongnu.org \
--cc=yuan1.liu@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).