From: "Michael R. Hines" <mrhines@linux.vnet.ibm.com>
To: Paolo Bonzini <pbonzini@redhat.com>
Cc: aliguori@us.ibm.com, mst@redhat.com, qemu-devel@nongnu.org,
owasserm@redhat.com, abali@us.ibm.com, mrhines@us.ibm.com,
gokul@us.ibm.com
Subject: Re: [Qemu-devel] [RFC PATCH RDMA support v6: 6/7] send pc.ram over RDMA
Date: Wed, 10 Apr 2013 08:38:55 -0400 [thread overview]
Message-ID: <51655D5F.1080901@linux.vnet.ibm.com> (raw)
In-Reply-To: <51651B80.2000900@redhat.com>
Acknowledged.
On 04/10/2013 03:57 AM, Paolo Bonzini wrote:
> Il 10/04/2013 06:29, mrhines@linux.vnet.ibm.com ha scritto:
>> From: "Michael R. Hines" <mrhines@us.ibm.com>
>>
>> All that is left for this part of the patch is:
>>
>> 1. use the new (optionally defined) save_ram_page function pointer
>> to decide what to do with the page if RDMA is enable or not
>> and return ENOTSUP as agreed.
>> 2. invoke hooks from QEMURamControlOps function pointers to hook
>> into the RDMA protocol at the right points in order to perform
>> dynamic page registration.
>> Signed-off-by: Michael R. Hines <mrhines@us.ibm.com>
>> ---
>> arch_init.c | 45 +++++++++++++++++++++++++++++++++++++++++++--
>> 1 file changed, 43 insertions(+), 2 deletions(-)
>>
>> diff --git a/arch_init.c b/arch_init.c
>> index 769ce77..a7d5b16 100644
>> --- a/arch_init.c
>> +++ b/arch_init.c
>> @@ -115,6 +115,7 @@ const uint32_t arch_type = QEMU_ARCH;
>> #define RAM_SAVE_FLAG_EOS 0x10
>> #define RAM_SAVE_FLAG_CONTINUE 0x20
>> #define RAM_SAVE_FLAG_XBZRLE 0x40
>> +#define RAM_SAVE_FLAG_REGISTER 0x80 /* perform hook during iteration */
> Please rename this to RAM_SAVE_FLAG_HOOK.
>
>>
>>
>> static struct defconfig_file {
>> @@ -170,6 +171,13 @@ static struct {
>> .cache = NULL,
>> };
>>
>> +#ifdef CONFIG_RDMA
>> +void qemu_ram_registration_start(QEMUFile *f, void *opaque, int section)
>> +{
>> + DPRINTF("start section: %d\n", section);
>> + qemu_put_be64(f, RAM_SAVE_FLAG_REGISTER);
>> +}
>> +#endif
> Please put this in migration-rdma.c together with the other QEMUFileOps.
>
>> int64_t xbzrle_cache_resize(int64_t new_size)
>> {
>> @@ -447,15 +455,22 @@ static int ram_save_block(QEMUFile *f, bool last_stage)
>> ram_bulk_stage = false;
>> }
>> } else {
>> + bool zero;
>> uint8_t *p;
>> int cont = (block == last_sent_block) ?
>> RAM_SAVE_FLAG_CONTINUE : 0;
>>
>> p = memory_region_get_ram_ptr(mr) + offset;
>>
>> + /* use capability now, defaults to true */
>> + zero = migrate_check_for_zero() ? is_zero_page(p) : false;
>> +
>> /* In doubt sent page as normal */
>> bytes_sent = -1;
>> - if (is_zero_page(p)) {
>> + if ((bytes_sent = ram_control_save_page(f, block->offset,
>> + offset, cont, TARGET_PAGE_SIZE, zero)) >= 0) {
>> + acct_info.norm_pages++;
>> + } else if (zero) {
>> acct_info.dup_pages++;
>> if (!ram_bulk_stage) {
>> bytes_sent = save_block_hdr(f, block, offset, cont,
>> @@ -476,7 +491,7 @@ static int ram_save_block(QEMUFile *f, bool last_stage)
>> }
>>
>> /* XBZRLE overflow or normal page */
>> - if (bytes_sent == -1) {
>> + if (bytes_sent == -1 || bytes_sent == -ENOTSUP) {
>> bytes_sent = save_block_hdr(f, block, offset, cont, RAM_SAVE_FLAG_PAGE);
>> qemu_put_buffer_async(f, p, TARGET_PAGE_SIZE);
>> bytes_sent += TARGET_PAGE_SIZE;
>> @@ -598,6 +613,18 @@ static int ram_save_setup(QEMUFile *f, void *opaque)
>> }
>>
>> qemu_mutex_unlock_ramlist();
>> +
>> + /*
>> + * These following calls generate reserved messages for future expansion of the RDMA
>> + * protocol. If the ops are not defined, nothing will happen.
>> + *
>> + * Please leave in place. They are intended to be used to pre-register
>> + * memory in the future to mitigate the extremely high cost of dynamic page
>> + * registration.
>> + */
>> + ram_control_before_iterate(f, RAM_CONTROL_SETUP);
>> + ram_control_after_iterate(f, RAM_CONTROL_SETUP);
>> +
>> qemu_put_be64(f, RAM_SAVE_FLAG_EOS);
>>
>> return 0;
>> @@ -616,6 +643,8 @@ static int ram_save_iterate(QEMUFile *f, void *opaque)
>> reset_ram_globals();
>> }
>>
>> + ram_control_before_iterate(f, RAM_CONTROL_ROUND);
>> +
>> t0 = qemu_get_clock_ns(rt_clock);
>> i = 0;
>> while ((ret = qemu_file_rate_limit(f)) == 0) {
>> @@ -646,6 +675,12 @@ static int ram_save_iterate(QEMUFile *f, void *opaque)
>>
>> qemu_mutex_unlock_ramlist();
>>
>> + /*
>> + * must occur before EOS (or any QEMUFile operation)
>> + * because of RDMA protocol
>> + */
>> + ram_control_after_iterate(f, RAM_CONTROL_ROUND);
>> +
>> if (ret < 0) {
>> bytes_transferred += total_sent;
>> return ret;
>> @@ -663,6 +698,8 @@ static int ram_save_complete(QEMUFile *f, void *opaque)
>> qemu_mutex_lock_ramlist();
>> migration_bitmap_sync();
>>
>> + ram_control_before_iterate(f, RAM_CONTROL_FINISH);
>> +
>> /* try transferring iterative blocks of memory */
>>
>> /* flush all remaining blocks regardless of rate limiting */
>> @@ -676,6 +713,8 @@ static int ram_save_complete(QEMUFile *f, void *opaque)
>> }
>> bytes_transferred += bytes_sent;
>> }
>> +
>> + ram_control_after_iterate(f, RAM_CONTROL_FINISH);
>> migration_end();
>>
>> qemu_mutex_unlock_ramlist();
>> @@ -864,6 +903,8 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id)
>> ret = -EINVAL;
>> goto done;
>> }
>> + } else if (flags & RAM_SAVE_FLAG_REGISTER) {
>> + ram_control_register_iterate(f, RAM_CONTROL_REGISTER);
> Please rename this function to ram_control_load_hook(f, flags).
>
> Paolo
>
>> }
>> error = qemu_file_get_error(f);
>> if (error) {
>>
next prev parent reply other threads:[~2013-04-10 12:39 UTC|newest]
Thread overview: 30+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-04-10 4:29 [Qemu-devel] [RFC PATCH RDMA support v6: 0/7] additional cleanup and consolidation mrhines
2013-04-10 4:29 ` [Qemu-devel] [RFC PATCH RDMA support v6: 1/7] ./configure and Makefile mrhines
2013-04-10 7:52 ` Paolo Bonzini
2013-04-10 12:37 ` Michael R. Hines
2013-04-10 12:42 ` Paolo Bonzini
2013-04-10 13:11 ` Michael R. Hines
2013-04-10 4:29 ` [Qemu-devel] [RFC PATCH RDMA support v6: 2/7] documentation (docs/rdma.txt) mrhines
2013-04-10 5:35 ` Michael S. Tsirkin
2013-04-10 12:19 ` Michael R. Hines
2013-04-10 4:29 ` [Qemu-devel] [RFC PATCH RDMA support v6: 3/7] Introduce QEMURamControlOps mrhines
2013-04-10 7:52 ` Paolo Bonzini
2013-04-10 12:38 ` Michael R. Hines
2013-04-10 4:29 ` [Qemu-devel] [RFC PATCH RDMA support v6: 4/7] Introduce two new capabilities mrhines
2013-04-10 7:50 ` Paolo Bonzini
2013-04-10 12:34 ` Michael R. Hines
2013-04-10 12:40 ` Paolo Bonzini
2013-04-10 13:10 ` Michael R. Hines
2013-04-10 13:13 ` Paolo Bonzini
2013-04-10 14:03 ` Michael R. Hines
2013-04-10 12:47 ` Orit Wasserman
2013-04-10 13:13 ` Michael R. Hines
2013-04-10 13:20 ` Orit Wasserman
2013-04-10 15:52 ` Michael R. Hines
2013-04-10 4:29 ` [Qemu-devel] [RFC PATCH RDMA support v6: 5/7] core RDMA logic mrhines
2013-04-10 4:29 ` [Qemu-devel] [RFC PATCH RDMA support v6: 6/7] send pc.ram over RDMA mrhines
2013-04-10 7:57 ` Paolo Bonzini
2013-04-10 12:38 ` Michael R. Hines [this message]
2013-04-10 4:29 ` [Qemu-devel] [RFC PATCH RDMA support v6: 7/7] introduce qemu_ram_foreach_block() mrhines
2013-04-10 7:47 ` Paolo Bonzini
2013-04-10 12:36 ` Michael R. Hines
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=51655D5F.1080901@linux.vnet.ibm.com \
--to=mrhines@linux.vnet.ibm.com \
--cc=abali@us.ibm.com \
--cc=aliguori@us.ibm.com \
--cc=gokul@us.ibm.com \
--cc=mrhines@us.ibm.com \
--cc=mst@redhat.com \
--cc=owasserm@redhat.com \
--cc=pbonzini@redhat.com \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).