From: Juan Quintela <quintela@redhat.com>
To: qemu-devel@nongnu.org
Subject: [Qemu-devel] [PATCH 08/22] savevm: split save_live into stage2 and stage3
Date: Fri, 13 Jul 2012 09:23:30 +0200 [thread overview]
Message-ID: <1342164224-32709-9-git-send-email-quintela@redhat.com> (raw)
In-Reply-To: <1342164224-32709-1-git-send-email-quintela@redhat.com>
We split it into 2 functions, foo_live_iterate, and foo_live_complete.
At this point, we only remove the bits that are for the other stage,
functionally this is equivalent to previous code.
Signed-off-by: Juan Quintela <quintela@redhat.com>
---
arch_init.c | 72 +++++++++++++++++++++++++++++---------
block-migration.c | 99 ++++++++++++++++++++++++++++++++---------------------
savevm.c | 10 +++---
vmstate.h | 3 +-
4 files changed, 122 insertions(+), 62 deletions(-)
diff --git a/arch_init.c b/arch_init.c
index 2a4903c..7e04ea5 100644
--- a/arch_init.c
+++ b/arch_init.c
@@ -374,12 +374,13 @@ static int ram_save_setup(QEMUFile *f, void *opaque)
return 0;
}
-static int ram_save_live(QEMUFile *f, int stage, void *opaque)
+static int ram_save_iterate(QEMUFile *f, void *opaque)
{
uint64_t bytes_transferred_last;
double bwidth = 0;
int ret;
int i;
+ uint64_t expected_time;
memory_global_sync_dirty_bitmap(get_system_memory());
@@ -424,28 +425,64 @@ static int ram_save_live(QEMUFile *f, int stage, void *opaque)
bwidth = 0.000001;
}
- /* try transferring iterative blocks of memory */
- if (stage == 3) {
- int bytes_sent;
+ qemu_put_be64(f, RAM_SAVE_FLAG_EOS);
+
+ expected_time = ram_save_remaining() * TARGET_PAGE_SIZE / bwidth;
+
+ DPRINTF("ram_save_live: expected(" PRIu64 ") <= max(" PRIu64 ")?\n",
+ expected_time, migrate_max_downtime());
+
+ return expected_time <= migrate_max_downtime();
+}
+
+static int ram_save_complete(QEMUFile *f, void *opaque)
+{
+ double bwidth = 0;
+ int ret;
+ int i;
+ int bytes_sent;
- /* flush all remaining blocks regardless of rate limiting */
- while ((bytes_sent = ram_save_block(f)) != 0) {
- bytes_transferred += bytes_sent;
+ memory_global_sync_dirty_bitmap(get_system_memory());
+
+ bwidth = qemu_get_clock_ns(rt_clock);
+
+ i = 0;
+ while ((ret = qemu_file_rate_limit(f)) == 0) {
+ bytes_sent = ram_save_block(f);
+ bytes_transferred += bytes_sent;
+ if (bytes_sent == 0) { /* no more blocks */
+ break;
}
- memory_global_dirty_log_stop();
+ /* we want to check in the 1st loop, just in case it was the 1st time
+ and we had to sync the dirty bitmap.
+ qemu_get_clock_ns() is a bit expensive, so we only check each some
+ iterations
+ */
+ if ((i & 63) == 0) {
+ uint64_t t1 = (qemu_get_clock_ns(rt_clock) - bwidth) / 1000000;
+ if (t1 > MAX_WAIT) {
+ DPRINTF("big wait: " PRIu64 " milliseconds, %d iterations\n",
+ t1, i);
+ break;
+ }
+ }
+ i++;
}
- qemu_put_be64(f, RAM_SAVE_FLAG_EOS);
-
- if (stage == 2) {
- uint64_t expected_time;
- expected_time = ram_save_remaining() * TARGET_PAGE_SIZE / bwidth;
+ if (ret < 0) {
+ return ret;
+ }
- DPRINTF("ram_save_live: expected(" PRIu64 ") <= max(" PRIu64 ")?\n",
- expected_time, migrate_max_downtime());
+ /* try transferring iterative blocks of memory */
- return expected_time <= migrate_max_downtime();
+ /* flush all remaining blocks regardless of rate limiting */
+ while ((bytes_sent = ram_save_block(f)) != 0) {
+ bytes_transferred += bytes_sent;
}
+ memory_global_dirty_log_stop();
+
+ qemu_put_be64(f, RAM_SAVE_FLAG_EOS);
+
return 0;
}
@@ -578,7 +615,8 @@ done:
SaveVMHandlers savevm_ram_handlers = {
.save_live_setup = ram_save_setup,
- .save_live_state = ram_save_live,
+ .save_live_iterate = ram_save_iterate,
+ .save_live_complete = ram_save_complete,
.load_state = ram_load,
.cancel = ram_migration_cancel,
};
diff --git a/block-migration.c b/block-migration.c
index fc3d1f4..7def8ab 100644
--- a/block-migration.c
+++ b/block-migration.c
@@ -568,12 +568,12 @@ static int block_save_setup(QEMUFile *f, void *opaque)
return 0;
}
-static int block_save_live(QEMUFile *f, int stage, void *opaque)
+static int block_save_iterate(QEMUFile *f, void *opaque)
{
int ret;
- DPRINTF("Enter save live stage %d submitted %d transferred %d\n",
- stage, block_mig_state.submitted, block_mig_state.transferred);
+ DPRINTF("Enter save live iterate submitted %d transferred %d\n",
+ block_mig_state.submitted, block_mig_state.transferred);
flush_blks(f);
@@ -585,56 +585,76 @@ static int block_save_live(QEMUFile *f, int stage, void *opaque)
blk_mig_reset_dirty_cursor();
- if (stage == 2) {
- /* control the rate of transfer */
- while ((block_mig_state.submitted +
- block_mig_state.read_done) * BLOCK_SIZE <
- qemu_file_get_rate_limit(f)) {
- if (block_mig_state.bulk_completed == 0) {
- /* first finish the bulk phase */
- if (blk_mig_save_bulked_block(f) == 0) {
- /* finished saving bulk on all devices */
- block_mig_state.bulk_completed = 1;
- }
- } else {
- if (blk_mig_save_dirty_block(f, 1) == 0) {
- /* no more dirty blocks */
- break;
- }
+ /* control the rate of transfer */
+ while ((block_mig_state.submitted +
+ block_mig_state.read_done) * BLOCK_SIZE <
+ qemu_file_get_rate_limit(f)) {
+ if (block_mig_state.bulk_completed == 0) {
+ /* first finish the bulk phase */
+ if (blk_mig_save_bulked_block(f) == 0) {
+ /* finished saving bulk on all devices */
+ block_mig_state.bulk_completed = 1;
+ }
+ } else {
+ if (blk_mig_save_dirty_block(f, 1) == 0) {
+ /* no more dirty blocks */
+ break;
}
}
+ }
- flush_blks(f);
+ flush_blks(f);
- ret = qemu_file_get_error(f);
- if (ret) {
- blk_mig_cleanup();
- return ret;
- }
+ ret = qemu_file_get_error(f);
+ if (ret) {
+ blk_mig_cleanup();
+ return ret;
}
- if (stage == 3) {
- /* we know for sure that save bulk is completed and
- all async read completed */
- assert(block_mig_state.submitted == 0);
+ qemu_put_be64(f, BLK_MIG_FLAG_EOS);
+
+ return is_stage2_completed();
+}
+
+static int block_save_complete(QEMUFile *f, void *opaque)
+{
+ int ret;
+
+ DPRINTF("Enter save live complete submitted %d transferred %d\n",
+ block_mig_state.submitted, block_mig_state.transferred);
+
+ flush_blks(f);
- while (blk_mig_save_dirty_block(f, 0) != 0);
+ ret = qemu_file_get_error(f);
+ if (ret) {
blk_mig_cleanup();
+ return ret;
+ }
- /* report completion */
- qemu_put_be64(f, (100 << BDRV_SECTOR_BITS) | BLK_MIG_FLAG_PROGRESS);
+ blk_mig_reset_dirty_cursor();
- ret = qemu_file_get_error(f);
- if (ret) {
- return ret;
- }
+ /* we know for sure that save bulk is completed and
+ all async read completed */
+ assert(block_mig_state.submitted == 0);
+
+ while (blk_mig_save_dirty_block(f, 0) != 0) {
+ /* Do nothing */
+ }
+ blk_mig_cleanup();
+
+ /* report completion */
+ qemu_put_be64(f, (100 << BDRV_SECTOR_BITS) | BLK_MIG_FLAG_PROGRESS);
- DPRINTF("Block migration completed\n");
+ ret = qemu_file_get_error(f);
+ if (ret) {
+ return ret;
}
+ DPRINTF("Block migration completed\n");
+
qemu_put_be64(f, BLK_MIG_FLAG_EOS);
- return ((stage == 2) && is_stage2_completed());
+ return 0;
}
static int block_load(QEMUFile *f, void *opaque, int version_id)
@@ -731,7 +751,8 @@ static bool block_is_active(void *opaque)
SaveVMHandlers savevm_block_handlers = {
.set_params = block_set_params,
.save_live_setup = block_save_setup,
- .save_live_state = block_save_live,
+ .save_live_iterate = block_save_iterate,
+ .save_live_complete = block_save_complete,
.load_state = block_load,
.cancel = block_migration_cancel,
.is_active = block_is_active,
diff --git a/savevm.c b/savevm.c
index 0b80a94..6e82b2d 100644
--- a/savevm.c
+++ b/savevm.c
@@ -1236,7 +1236,7 @@ int register_savevm_live(DeviceState *dev,
se->vmsd = NULL;
se->no_migrate = 0;
/* if this is a live_savem then set is_ram */
- if (ops->save_live_state != NULL) {
+ if (ops->save_live_setup != NULL) {
se->is_ram = 1;
}
@@ -1620,7 +1620,7 @@ int qemu_savevm_state_iterate(QEMUFile *f)
int ret = 1;
QTAILQ_FOREACH(se, &savevm_handlers, entry) {
- if (!se->ops || !se->ops->save_live_state) {
+ if (!se->ops || !se->ops->save_live_iterate) {
continue;
}
if (se->ops && se->ops->is_active) {
@@ -1636,7 +1636,7 @@ int qemu_savevm_state_iterate(QEMUFile *f)
qemu_put_byte(f, QEMU_VM_SECTION_PART);
qemu_put_be32(f, se->section_id);
- ret = se->ops->save_live_state(f, QEMU_VM_SECTION_PART, se->opaque);
+ ret = se->ops->save_live_iterate(f, se->opaque);
trace_savevm_section_end(se->section_id);
if (ret <= 0) {
@@ -1665,7 +1665,7 @@ int qemu_savevm_state_complete(QEMUFile *f)
cpu_synchronize_all_states();
QTAILQ_FOREACH(se, &savevm_handlers, entry) {
- if (!se->ops || !se->ops->save_live_state) {
+ if (!se->ops || !se->ops->save_live_complete) {
continue;
}
if (se->ops && se->ops->is_active) {
@@ -1678,7 +1678,7 @@ int qemu_savevm_state_complete(QEMUFile *f)
qemu_put_byte(f, QEMU_VM_SECTION_END);
qemu_put_be32(f, se->section_id);
- ret = se->ops->save_live_state(f, QEMU_VM_SECTION_END, se->opaque);
+ ret = se->ops->save_live_complete(f, se->opaque);
trace_savevm_section_end(se->section_id);
if (ret < 0) {
return ret;
diff --git a/vmstate.h b/vmstate.h
index 049f2b7..5bd2b76 100644
--- a/vmstate.h
+++ b/vmstate.h
@@ -33,7 +33,8 @@ typedef struct SaveVMHandlers {
void (*set_params)(const MigrationParams *params, void * opaque);
SaveStateHandler *save_state;
int (*save_live_setup)(QEMUFile *f, void *opaque);
- int (*save_live_state)(QEMUFile *f, int stage, void *opaque);
+ int (*save_live_iterate)(QEMUFile *f, void *opaque);
+ int (*save_live_complete)(QEMUFile *f, void *opaque);
void (*cancel)(void *opaque);
LoadStateHandler *load_state;
bool (*is_active)(void *opaque);
--
1.7.10.4
next prev parent reply other threads:[~2012-07-13 7:24 UTC|newest]
Thread overview: 35+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-07-13 7:23 [Qemu-devel] [PATCH 00/22] Migration next Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 01/22] savevm: Use a struct to pass all handlers Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 02/22] savevm: Live migration handlers register the struct directly Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 03/22] savevm: remove SaveSetParamsHandler Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 04/22] savevm: remove SaveLiveStateHandler Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 05/22] savevm: Refactor cancel operation in its own operation Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 06/22] savevm: introduce is_active method Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 07/22] savevm: split save_live_setup from save_live_state Juan Quintela
2012-07-13 7:23 ` Juan Quintela [this message]
2012-07-13 7:23 ` [Qemu-devel] [PATCH 09/22] ram: save_live_setup() don't need to sent pages Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 10/22] ram: save_live_complete() only do one loop Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 11/22] ram: iterate phase Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 12/22] ram: save_live_setup() we don't need to synchronize the dirty bitmap Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 13/22] Add migration capabilities Juan Quintela
2012-07-23 18:23 ` Luiz Capitulino
2012-07-23 19:30 ` Eric Blake
2012-07-24 6:25 ` Orit Wasserman
2012-07-24 12:50 ` Luiz Capitulino
2012-07-24 17:06 ` Orit Wasserman
2012-07-24 18:17 ` Luiz Capitulino
2012-07-25 13:05 ` Orit Wasserman
2012-07-25 13:11 ` Luiz Capitulino
2012-07-25 13:33 ` Orit Wasserman
2012-07-13 7:23 ` [Qemu-devel] [PATCH 14/22] Add XBZRLE documentation Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 15/22] Add cache handling functions Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 16/22] Add uleb encoding/decoding functions Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 17/22] Change ram_save_block to return -1 if there are no more changes Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 18/22] Add xbzrle_encode_buffer and xbzrle_decode_buffer functions Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 19/22] Add XBZRLE to ram_save_block and ram_save_live Juan Quintela
2012-07-13 7:23 ` [Qemu-devel] [PATCH 20/22] Add migrate_set_cachesize command Juan Quintela
2012-07-23 19:21 ` Luiz Capitulino
2012-07-13 7:23 ` [Qemu-devel] [PATCH 21/22] Add XBZRLE statistics Juan Quintela
2012-07-23 19:33 ` Luiz Capitulino
2012-07-24 6:32 ` Orit Wasserman
2012-07-13 7:23 ` [Qemu-devel] [PATCH 22/22] ram: save live optimization Juan Quintela
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1342164224-32709-9-git-send-email-quintela@redhat.com \
--to=quintela@redhat.com \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).