From: "Dr. David Alan Gilbert" <dgilbert@redhat.com>
To: zhanghailiang <zhang.zhanghailiang@huawei.com>
Cc: hangaohuai@huawei.com, Lai Jiangshan <laijs@cn.fujitsu.com>,
Li Zhijian <lizhijian@cn.fujitsu.com>,
yunhong.jiang@intel.com, eddie.dong@intel.com,
peter.huangpeng@huawei.com, qemu-devel@nongnu.org,
Gonglei <arei.gonglei@huawei.com>,
stefanha@redhat.com, pbonzini@redhat.com,
Yang Hongyang <yanghy@cn.fujitsu.com>
Subject: Re: [Qemu-devel] [PATCH RFC v3 13/27] COLO RAM: Flush cached RAM into SVM's memory
Date: Thu, 12 Mar 2015 09:51:17 +0000 [thread overview]
Message-ID: <20150312095116.GB2330@work-vm> (raw)
In-Reply-To: <5500F994.3000300@huawei.com>
* zhanghailiang (zhang.zhanghailiang@huawei.com) wrote:
> On 2015/3/12 4:07, Dr. David Alan Gilbert wrote:
> >* zhanghailiang (zhang.zhanghailiang@huawei.com) wrote:
> >>We only need to flush RAM that is both dirty on PVM and SVM since
> >>last checkpoint. Besides, we must ensure flush RAM cache before load
> >>device state.
> >
> >Actually with a follow up to my previous question, can you explain the 'both'
> >in that description.
> >
>
> The description is wrong,
> It should be 'any page that dirtied by PVM or SVM'. Sorry for my poor english.
That's fine; thank you for the clarification.
Dave
>
> >If a page was dirty on just the PVM, but not the SVM, you would have to copy
> >the new PVM page into the SVM ram before executing with the newly received device
> >state, otherwise the device state would be inconsistent with the RAM state.
> >
> >Dave
> >
> >>Signed-off-by: zhanghailiang <zhang.zhanghailiang@huawei.com>a
> >>Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com>
> >>Signed-off-by: Li Zhijian <lizhijian@cn.fujitsu.com>
> >>Signed-off-by: Yang Hongyang <yanghy@cn.fujitsu.com>
> >>Signed-off-by: Gonglei <arei.gonglei@huawei.com>
> >>---
> >> arch_init.c | 91 +++++++++++++++++++++++++++++++++++++-
> >> include/migration/migration-colo.h | 1 +
> >> migration/colo.c | 1 -
> >> 3 files changed, 91 insertions(+), 2 deletions(-)
> >>
> >>diff --git a/arch_init.c b/arch_init.c
> >>index 4a1d825..f70de23 100644
> >>--- a/arch_init.c
> >>+++ b/arch_init.c
> >>@@ -1100,6 +1100,7 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id)
> >> {
> >> int flags = 0, ret = 0;
> >> static uint64_t seq_iter;
> >>+ bool need_flush = false;
> >>
> >> seq_iter++;
> >>
> >>@@ -1163,6 +1164,7 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id)
> >> break;
> >> }
> >>
> >>+ need_flush = true;
> >> ch = qemu_get_byte(f);
> >> ram_handle_compressed(host, ch, TARGET_PAGE_SIZE);
> >> break;
> >>@@ -1174,6 +1176,7 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id)
> >> break;
> >> }
> >>
> >>+ need_flush = true;
> >> qemu_get_buffer(f, host, TARGET_PAGE_SIZE);
> >> break;
> >> case RAM_SAVE_FLAG_XBZRLE:
> >>@@ -1190,6 +1193,7 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id)
> >> ret = -EINVAL;
> >> break;
> >> }
> >>+ need_flush = true;
> >> break;
> >> case RAM_SAVE_FLAG_EOS:
> >> /* normal exit */
> >>@@ -1207,7 +1211,10 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id)
> >> ret = qemu_file_get_error(f);
> >> }
> >> }
> >>-
> >>+ if (!ret && ram_cache_enable && need_flush) {
> >>+ DPRINTF("Flush ram_cache\n");
> >>+ colo_flush_ram_cache();
> >>+ }
> >> DPRINTF("Completed load of VM with exit code %d seq iteration "
> >> "%" PRIu64 "\n", ret, seq_iter);
> >> return ret;
> >>@@ -1220,6 +1227,7 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id)
> >> void create_and_init_ram_cache(void)
> >> {
> >> RAMBlock *block;
> >>+ int64_t ram_cache_pages = last_ram_offset() >> TARGET_PAGE_BITS;
> >>
> >> QTAILQ_FOREACH(block, &ram_list.blocks, next) {
> >> block->host_cache = g_malloc(block->used_length);
> >>@@ -1227,6 +1235,14 @@ void create_and_init_ram_cache(void)
> >> }
> >>
> >> ram_cache_enable = true;
> >>+ /*
> >>+ * Start dirty log for slave VM, we will use this dirty bitmap together with
> >>+ * VM's cache RAM dirty bitmap to decide which page in cache should be
> >>+ * flushed into VM's RAM.
> >>+ */
> >>+ migration_bitmap = bitmap_new(ram_cache_pages);
> >>+ migration_dirty_pages = 0;
> >>+ memory_global_dirty_log_start();
> >> }
> >>
> >> void release_ram_cache(void)
> >>@@ -1261,6 +1277,79 @@ static void *memory_region_get_ram_cache_ptr(MemoryRegion *mr, RAMBlock *block)
> >> return block->host_cache + (addr - block->offset);
> >> }
> >>
> >>+static inline
> >>+ram_addr_t host_bitmap_find_and_reset_dirty(MemoryRegion *mr,
> >>+ ram_addr_t start)
> >>+{
> >>+ unsigned long base = mr->ram_addr >> TARGET_PAGE_BITS;
> >>+ unsigned long nr = base + (start >> TARGET_PAGE_BITS);
> >>+ unsigned long size = base + (int128_get64(mr->size) >> TARGET_PAGE_BITS);
> >>+
> >>+ unsigned long next;
> >>+
> >>+ next = find_next_bit(ram_list.dirty_memory[DIRTY_MEMORY_MIGRATION],
> >>+ size, nr);
> >>+ if (next < size) {
> >>+ clear_bit(next, ram_list.dirty_memory[DIRTY_MEMORY_MIGRATION]);
> >>+ }
> >>+ return (next - base) << TARGET_PAGE_BITS;
> >>+}
> >>+
> >>+void colo_flush_ram_cache(void)
> >>+{
> >>+ RAMBlock *block = NULL;
> >>+ void *dst_host;
> >>+ void *src_host;
> >>+ ram_addr_t ca = 0, ha = 0;
> >>+ bool got_ca = 0, got_ha = 0;
> >>+ int64_t host_dirty = 0, both_dirty = 0;
> >>+
> >>+ address_space_sync_dirty_bitmap(&address_space_memory);
> >>+
> >>+ block = QTAILQ_FIRST(&ram_list.blocks);
> >>+ while (true) {
> >>+ if (ca < block->used_length && ca <= ha) {
> >>+ ca = migration_bitmap_find_and_reset_dirty(block->mr, ca);
> >>+ if (ca < block->used_length) {
> >>+ got_ca = 1;
> >>+ }
> >>+ }
> >>+ if (ha < block->used_length && ha <= ca) {
> >>+ ha = host_bitmap_find_and_reset_dirty(block->mr, ha);
> >>+ if (ha < block->used_length && ha != ca) {
> >>+ got_ha = 1;
> >>+ }
> >>+ host_dirty += (ha < block->used_length ? 1 : 0);
> >>+ both_dirty += (ha < block->used_length && ha == ca ? 1 : 0);
> >>+ }
> >>+ if (ca >= block->used_length && ha >= block->used_length) {
> >>+ ca = 0;
> >>+ ha = 0;
> >>+ block = QTAILQ_NEXT(block, next);
> >>+ if (!block) {
> >>+ break;
> >>+ }
> >>+ } else {
> >>+ if (got_ha) {
> >>+ got_ha = 0;
> >>+ dst_host = memory_region_get_ram_ptr(block->mr) + ha;
> >>+ src_host = memory_region_get_ram_cache_ptr(block->mr, block)
> >>+ + ha;
> >>+ memcpy(dst_host, src_host, TARGET_PAGE_SIZE);
> >>+ }
> >>+ if (got_ca) {
> >>+ got_ca = 0;
> >>+ dst_host = memory_region_get_ram_ptr(block->mr) + ca;
> >>+ src_host = memory_region_get_ram_cache_ptr(block->mr, block)
> >>+ + ca;
> >>+ memcpy(dst_host, src_host, TARGET_PAGE_SIZE);
> >>+ }
> >>+ }
> >>+ }
> >>+
> >>+ assert(migration_dirty_pages == 0);
> >>+}
> >>+
> >> static SaveVMHandlers savevm_ram_handlers = {
> >> .save_live_setup = ram_save_setup,
> >> .save_live_iterate = ram_save_iterate,
> >>diff --git a/include/migration/migration-colo.h b/include/migration/migration-colo.h
> >>index 7d43aed..2084fe2 100644
> >>--- a/include/migration/migration-colo.h
> >>+++ b/include/migration/migration-colo.h
> >>@@ -36,5 +36,6 @@ void *colo_process_incoming_checkpoints(void *opaque);
> >> bool loadvm_in_colo_state(void);
> >> /* ram cache */
> >> void create_and_init_ram_cache(void);
> >>+void colo_flush_ram_cache(void);
> >> void release_ram_cache(void);
> >> #endif
> >>diff --git a/migration/colo.c b/migration/colo.c
> >>index a0e1b7a..5ff2ee8 100644
> >>--- a/migration/colo.c
> >>+++ b/migration/colo.c
> >>@@ -397,7 +397,6 @@ void *colo_process_incoming_checkpoints(void *opaque)
> >> }
> >> DPRINTF("Finish load all vm state to cache\n");
> >> qemu_mutex_unlock_iothread();
> >>- /* TODO: flush vm state */
> >>
> >> ret = colo_ctl_put(ctl, COLO_CHECKPOINT_LOADED);
> >> if (ret < 0) {
> >>--
> >>1.7.12.4
> >>
> >>
> >--
> >Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK
> >
> >.
> >
>
>
--
Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK
next prev parent reply other threads:[~2015-03-12 9:51 UTC|newest]
Thread overview: 65+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-02-12 3:16 [Qemu-devel] [PATCH RFC v3 00/27] COarse-grain LOck-stepping(COLO) Virtual Machines for Non-stop Service zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 01/27] configure: Add parameter for configure to enable/disable COLO support zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 02/27] migration: Introduce capability 'colo' to migration zhanghailiang
2015-02-16 21:57 ` Eric Blake
2015-02-25 9:19 ` zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 03/27] COLO: migrate colo related info to slave zhanghailiang
2015-02-16 23:20 ` Eric Blake
2015-02-25 6:21 ` zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 04/27] migration: Integrate COLO checkpoint process into migration zhanghailiang
2015-02-16 23:27 ` Eric Blake
2015-02-25 6:43 ` zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 05/27] migration: Integrate COLO checkpoint process into loadvm zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 06/27] migration: Don't send vm description in COLO mode zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 07/27] COLO: Implement colo checkpoint protocol zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 08/27] COLO: Add a new RunState RUN_STATE_COLO zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 09/27] QEMUSizedBuffer: Introduce two help functions for qsb zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 10/27] COLO: Save VM state to slave when do checkpoint zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 11/27] COLO RAM: Load PVM's dirty page into SVM's RAM cache temporarily zhanghailiang
2015-02-12 3:16 ` [Qemu-devel] [PATCH RFC v3 12/27] COLO VMstate: Load VM state into qsb before restore it zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 13/27] COLO RAM: Flush cached RAM into SVM's memory zhanghailiang
2015-03-11 19:08 ` Dr. David Alan Gilbert
2015-03-12 2:02 ` zhanghailiang
2015-03-12 11:49 ` Dr. David Alan Gilbert
2015-03-11 20:07 ` Dr. David Alan Gilbert
2015-03-12 2:27 ` zhanghailiang
2015-03-12 9:51 ` Dr. David Alan Gilbert [this message]
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 14/27] COLO failover: Introduce a new command to trigger a failover zhanghailiang
2015-02-16 23:47 ` Eric Blake
2015-02-25 7:04 ` zhanghailiang
2015-02-25 7:16 ` Hongyang Yang
2015-02-25 7:40 ` Wen Congyang
2015-03-06 16:10 ` Eric Blake
2015-03-09 1:15 ` zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 15/27] COLO failover: Implement COLO master/slave failover work zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 16/27] COLO failover: Don't do failover during loading VM's state zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 17/27] COLO: Add new command parameter 'colo_nicname' 'colo_script' for net zhanghailiang
2015-02-16 23:50 ` Eric Blake
2015-02-24 9:50 ` Wen Congyang
2015-02-24 16:30 ` Eric Blake
2015-02-24 17:24 ` Daniel P. Berrange
2015-02-25 8:21 ` zhanghailiang
2015-02-25 10:09 ` Daniel P. Berrange
2015-02-25 7:50 ` zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 18/27] COLO NIC: Init/remove colo nic devices when add/cleanup tap devices zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 19/27] COLO NIC: Implement colo nic device interface configure() zhanghailiang
2015-02-16 12:03 ` Dr. David Alan Gilbert
2015-02-25 3:44 ` zhanghailiang
2015-02-25 9:08 ` Dr. David Alan Gilbert
2015-02-25 9:38 ` zhanghailiang
2015-02-25 9:40 ` Dr. David Alan Gilbert
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 20/27] COLO NIC : Implement colo nic init/destroy function zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 21/27] COLO NIC: Some init work related with proxy module zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 22/27] COLO: Do checkpoint according to the result of net packets comparing zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 23/27] COLO: Improve checkpoint efficiency by do additional periodic checkpoint zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 24/27] COLO NIC: Implement NIC checkpoint and failover zhanghailiang
2015-03-05 17:12 ` Dr. David Alan Gilbert
2015-03-06 2:35 ` zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 25/27] COLO: Disable qdev hotplug when VM is in COLO mode zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 26/27] COLO: Implement shutdown checkpoint zhanghailiang
2015-02-12 3:17 ` [Qemu-devel] [PATCH RFC v3 27/27] COLO: Add block replication into colo process zhanghailiang
2015-02-16 13:11 ` [Qemu-devel] [PATCH RFC v3 00/27] COarse-grain LOck-stepping(COLO) Virtual Machines for Non-stop Service Dr. David Alan Gilbert
2015-02-25 5:17 ` Gao feng
2015-02-24 11:08 ` Dr. David Alan Gilbert
2015-02-24 20:13 ` Dr. David Alan Gilbert
2015-02-25 3:20 ` Gao feng
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20150312095116.GB2330@work-vm \
--to=dgilbert@redhat.com \
--cc=arei.gonglei@huawei.com \
--cc=eddie.dong@intel.com \
--cc=hangaohuai@huawei.com \
--cc=laijs@cn.fujitsu.com \
--cc=lizhijian@cn.fujitsu.com \
--cc=pbonzini@redhat.com \
--cc=peter.huangpeng@huawei.com \
--cc=qemu-devel@nongnu.org \
--cc=stefanha@redhat.com \
--cc=yanghy@cn.fujitsu.com \
--cc=yunhong.jiang@intel.com \
--cc=zhang.zhanghailiang@huawei.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).