From: "Dr. David Alan Gilbert (git)" <dgilbert@redhat.com>
To: qemu-devel@nongnu.org
Cc: aarcange@redhat.com, yamahata@private.email.ne.jp,
quintela@redhat.com, amit.shah@redhat.com, pbonzini@redhat.com,
david@gibson.dropbear.id.au, yayanghy@cn.fujitsu.com
Subject: [Qemu-devel] [PATCH v6 41/47] Host page!=target page: Cleanup bitmaps
Date: Tue, 14 Apr 2015 18:04:07 +0100 [thread overview]
Message-ID: <1429031053-4454-42-git-send-email-dgilbert@redhat.com> (raw)
In-Reply-To: <1429031053-4454-1-git-send-email-dgilbert@redhat.com>
From: "Dr. David Alan Gilbert" <dgilbert@redhat.com>
Prior to the start of postcopy, ensure that everything that will
be transferred later is a whole host-page in size.
This is accomplished by discarding partially transferred host pages
and marking any that are partially dirty as fully dirty.
Signed-off-by: Dr. David Alan Gilbert <dgilbert@redhat.com>
---
arch_init.c | 271 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++-
1 file changed, 269 insertions(+), 2 deletions(-)
diff --git a/arch_init.c b/arch_init.c
index dc672bf..18253af 100644
--- a/arch_init.c
+++ b/arch_init.c
@@ -850,7 +850,6 @@ static int ram_find_and_save_block(QEMUFile *f, bool last_stage,
int pages = 0;
ram_addr_t dirty_ram_abs; /* Address of the start of the dirty page in
ram_addr_t space */
- unsigned long hps = sysconf(_SC_PAGESIZE);
if (!block) {
block = QLIST_FIRST_RCU(&ram_list.blocks);
@@ -867,7 +866,8 @@ static int ram_find_and_save_block(QEMUFile *f, bool last_stage,
* b) The last sent item was the last target-page in a host page
*/
if (last_was_from_queue || !last_sent_block ||
- ((last_offset & (hps - 1)) == (hps - TARGET_PAGE_SIZE))) {
+ ((last_offset & ~qemu_host_page_mask) ==
+ (qemu_host_page_size - TARGET_PAGE_SIZE))) {
tmpblock = ram_save_unqueue_page(ms, &tmpoffset, &dirty_ram_abs);
}
@@ -1152,6 +1152,265 @@ static int postcopy_each_ram_send_discard(MigrationState *ms)
}
/*
+ * Helper for postcopy_chunk_hostpages where HPS/TPS >= bits-in-long
+ *
+ * !! Untested !!
+ */
+static int hostpage_big_chunk_helper(const char *block_name, void *host_addr,
+ ram_addr_t offset, ram_addr_t length,
+ void *opaque)
+{
+ MigrationState *ms = opaque;
+ unsigned long long_bits = sizeof(long) * 8;
+ unsigned int host_len = (qemu_host_page_size / TARGET_PAGE_SIZE) /
+ long_bits;
+ unsigned long first_long, last_long, cur_long, current_hp;
+ unsigned long first = offset >> TARGET_PAGE_BITS;
+ unsigned long last = (offset + (length - 1)) >> TARGET_PAGE_BITS;
+
+ PostcopyDiscardState *pds = postcopy_discard_send_init(ms,
+ first,
+ block_name);
+ first_long = first / long_bits;
+ last_long = last / long_bits;
+
+ /*
+ * I'm assuming RAMBlocks must start at the start of host pages,
+ * but I guess they might not use the whole of the host page
+ */
+
+ /* Work along one host page at a time */
+ for (current_hp = first_long; current_hp <= last_long;
+ current_hp += host_len) {
+ bool discard = 0;
+ bool redirty = 0;
+ bool has_some_dirty = false;
+ bool has_some_undirty = false;
+ bool has_some_sent = false;
+ bool has_some_unsent = false;
+
+ /*
+ * Check each long of mask for this hp, and see if anything
+ * needs updating.
+ */
+ for (cur_long = current_hp; cur_long < (current_hp + host_len);
+ cur_long++) {
+ /* a chunk of sent pages */
+ unsigned long sdata = ms->sentmap[cur_long];
+ /* a chunk of dirty pages */
+ unsigned long ddata = migration_bitmap[cur_long];
+
+ if (sdata) {
+ has_some_sent = true;
+ }
+ if (sdata != ~0ul) {
+ has_some_unsent = true;
+ }
+ if (ddata) {
+ has_some_dirty = true;
+ }
+ if (ddata != ~0ul) {
+ has_some_undirty = true;
+ }
+
+ }
+
+ if (has_some_sent && has_some_unsent) {
+ /* Partially sent host page */
+ discard = true;
+ redirty = true;
+ }
+
+ if (has_some_dirty && has_some_undirty) {
+ /* Partially dirty host page */
+ redirty = true;
+ }
+
+ if (!discard && !redirty) {
+ /* All consistent - next host page */
+ continue;
+ }
+
+
+ /* Now walk the chunks again, sending discards etc */
+ for (cur_long = current_hp; cur_long < (current_hp + host_len);
+ cur_long++) {
+ unsigned long cur_bits = cur_long * long_bits;
+
+ /* a chunk of sent pages */
+ unsigned long sdata = ms->sentmap[cur_long];
+ /* a chunk of dirty pages */
+ unsigned long ddata = migration_bitmap[cur_long];
+
+ if (discard && sdata) {
+ /* Tell the destination to discard these pages */
+ postcopy_discard_send_range(ms, pds, cur_bits,
+ cur_bits + long_bits - 1);
+ /* And clear them in the sent data structure */
+ ms->sentmap[cur_long] = 0;
+ }
+
+ if (redirty) {
+ migration_bitmap[cur_long] = ~0ul;
+ /* Inc the count of dirty pages */
+ migration_dirty_pages += ctpopl(~ddata);
+ }
+ }
+ }
+
+ postcopy_discard_send_finish(ms, pds);
+
+ return 0;
+}
+
+/*
+ * When working on long chunks of a bitmap where the only valid section
+ * is between start..end (inclusive), generate a mask with only those
+ * valid bits set for the current long word within that bitmask.
+ */
+static unsigned long make_long_mask(unsigned long start, unsigned long end,
+ unsigned long cur_long)
+{
+ unsigned long long_bits = sizeof(long) * 8;
+ unsigned long long_bits_mask = long_bits - 1;
+ unsigned long first_long, last_long;
+ unsigned long mask = ~(unsigned long)0;
+ first_long = start / long_bits ;
+ last_long = end / long_bits;
+
+ if ((cur_long == first_long) && (start & long_bits_mask)) {
+ /* e.g. (start & 31) = 3
+ * 1 << . -> 2^3
+ * . - 1 -> 2^3 - 1 i.e. mask 2..0
+ * ~. -> mask 31..3
+ */
+ mask &= ~((((unsigned long)1) << (start & long_bits_mask)) - 1);
+ }
+
+ if ((cur_long == last_long) && ((end & long_bits_mask) != long_bits_mask)) {
+ /* e.g. (end & 31) = 3
+ * . +1 -> 4
+ * 1 << . -> 2^4
+ * . -1 -> 2^4 - 1
+ * = mask set 3..0
+ */
+ mask &= (((unsigned long)1) << ((end & long_bits_mask) + 1)) - 1;
+ }
+
+ return mask;
+}
+
+/*
+ * Utility for the outgoing postcopy code.
+ *
+ * Discard any partially sent host-page size chunks, mark any partially
+ * dirty host-page size chunks as all dirty.
+ *
+ * Returns: 0 on success
+ */
+static int postcopy_chunk_hostpages(MigrationState *ms)
+{
+ struct RAMBlock *block;
+ unsigned int host_bits = qemu_host_page_size / TARGET_PAGE_SIZE;
+ unsigned long long_bits = sizeof(long) * 8;
+ unsigned long host_mask;
+
+ assert(is_power_of_2(host_bits));
+
+ if (qemu_host_page_size == TARGET_PAGE_SIZE) {
+ /* Easy case - TPS==HPS - nothing to be done */
+ return 0;
+ }
+
+ /* Easiest way to make sure we don't resume in the middle of a host-page */
+ last_seen_block = NULL;
+ last_sent_block = NULL;
+
+ /*
+ * The currently worst known ratio is ARM that has 1kB target pages, and
+ * can have 64kB host pages, which is thus inconveniently larger than a long
+ * on ARM (32bits), and a long is the underlying element of the migration
+ * bitmaps.
+ */
+ if (host_bits >= long_bits) {
+ /* Deal with the odd case separately */
+ return qemu_ram_foreach_block(hostpage_big_chunk_helper, ms);
+ } else {
+ host_mask = (1ul << host_bits) - 1;
+ }
+
+ rcu_read_lock();
+ QLIST_FOREACH_RCU(block, &ram_list.blocks, next) {
+ unsigned long first_long, last_long, cur_long;
+ unsigned long first = block->offset >> TARGET_PAGE_BITS;
+ unsigned long last = (block->offset + (block->used_length - 1))
+ >> TARGET_PAGE_BITS;
+ PostcopyDiscardState *pds = postcopy_discard_send_init(ms,
+ first,
+ block->idstr);
+
+ first_long = first / long_bits;
+ last_long = last / long_bits;
+ for (cur_long = first_long; cur_long <= last_long; cur_long++) {
+ unsigned long current_hp;
+ /* Deal with start/end not on alignment */
+ unsigned long mask = make_long_mask(first, last, cur_long);
+
+ /* a chunk of sent pages */
+ unsigned long sdata = ms->sentmap[cur_long];
+ /* a chunk of dirty pages */
+ unsigned long ddata = migration_bitmap[cur_long];
+ unsigned long discard = 0;
+ unsigned long redirty = 0;
+ sdata &= mask;
+ ddata &= mask;
+
+ for (current_hp = 0; current_hp < long_bits;
+ current_hp += host_bits) {
+ unsigned long host_sent = (sdata >> current_hp) & host_mask;
+ unsigned long host_dirty = (ddata >> current_hp) & host_mask;
+
+ if (host_sent && (host_sent != host_mask)) {
+ /* Partially sent host page */
+ redirty |= host_mask << current_hp;
+ discard |= host_mask << current_hp;
+
+ /* Tell the destination to discard this page */
+ postcopy_discard_send_range(ms, pds,
+ cur_long * long_bits + current_hp /* start */,
+ cur_long * long_bits + current_hp +
+ host_bits - 1 /* end */);
+ } else if (host_dirty && (host_dirty != host_mask)) {
+ /* Partially dirty host page */
+ redirty |= host_mask << current_hp;
+ }
+ }
+ if (discard) {
+ /* clear the page in the sentmap */
+ ms->sentmap[cur_long] &= ~discard;
+ }
+ if (redirty) {
+ /*
+ * Reread original dirty bits and OR in ones we clear; we
+ * must reread since we might be at the start or end of
+ * a RAMBlock that the original 'mask' discarded some
+ * bits from
+ */
+ ddata = migration_bitmap[cur_long];
+ migration_bitmap[cur_long] = ddata | redirty;
+ /* Inc the count of dirty pages */
+ migration_dirty_pages += ctpopl(redirty - (ddata & redirty));
+ }
+ }
+
+ postcopy_discard_send_finish(ms, pds);
+ }
+
+ rcu_read_unlock();
+ return 0;
+}
+
+/*
* Transmit the set of pages to be discarded after precopy to the target
* these are pages that have been sent previously but have been dirtied
* Hopefully this is pretty sparse
@@ -1161,9 +1420,17 @@ int ram_postcopy_send_discard_bitmap(MigrationState *ms)
int ret;
rcu_read_lock();
+
/* This should be our last sync, the src is now paused */
migration_bitmap_sync();
+ /* Deal with TPS != HPS */
+ ret = postcopy_chunk_hostpages(ms);
+ if (ret) {
+ rcu_read_unlock();
+ return ret;
+ }
+
/*
* Update the sentmap to be sentmap&=dirty
*/
--
2.1.0
next prev parent reply other threads:[~2015-04-14 17:05 UTC|newest]
Thread overview: 74+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-04-14 17:03 [Qemu-devel] [PATCH v6 00/47] Postcopy implementation Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 01/47] Start documenting how postcopy works Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 02/47] Split header writing out of qemu_savevm_state_begin Dr. David Alan Gilbert (git)
2015-05-11 11:16 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 03/47] qemu_ram_foreach_block: pass up error value, and down the ramblock name Dr. David Alan Gilbert (git)
2015-05-15 10:38 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 04/47] Add qemu_get_counted_string to read a string prefixed by a count byte Dr. David Alan Gilbert (git)
2015-05-15 13:50 ` Amit Shah
2015-05-15 14:06 ` Dr. David Alan Gilbert
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 05/47] Create MigrationIncomingState Dr. David Alan Gilbert (git)
2015-05-18 6:58 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 06/47] Provide runtime Target page information Dr. David Alan Gilbert (git)
2015-05-18 7:06 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 07/47] Move copy out of qemu_peek_buffer Dr. David Alan Gilbert (git)
2015-05-21 6:47 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 08/47] Add qemu_get_buffer_less_copy to avoid copies some of the time Dr. David Alan Gilbert (git)
2015-05-21 7:09 ` Amit Shah
2015-05-21 8:45 ` Dr. David Alan Gilbert
2015-05-21 8:58 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 09/47] Add wrapper for setting blocking status on a QEMUFile Dr. David Alan Gilbert (git)
2015-05-18 7:35 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 10/47] Rename save_live_complete to save_live_complete_precopy Dr. David Alan Gilbert (git)
2015-05-18 7:35 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 11/47] Return path: Open a return path on QEMUFile for sockets Dr. David Alan Gilbert (git)
2015-06-10 9:00 ` Amit Shah
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 12/47] Return path: socket_writev_buffer: Block even on non-blocking fd's Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 13/47] Migration commands Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 14/47] Return path: Control commands Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 15/47] Return path: Send responses from destination to source Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 16/47] Return path: Source handling of return path Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 17/47] ram_debug_dump_bitmap: Dump a migration bitmap as text Dr. David Alan Gilbert (git)
2015-05-21 9:21 ` Amit Shah
2015-05-21 10:10 ` Dr. David Alan Gilbert
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 18/47] Move loadvm_handlers into MigrationIncomingState Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 19/47] Rework loadvm path for subloops Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 20/47] Add migration-capability boolean for postcopy-ram Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 21/47] Add wrappers and handlers for sending/receiving the postcopy-ram migration messages Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 22/47] MIG_CMD_PACKAGED: Send a packaged chunk of migration stream Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 23/47] migrate_init: Call from savevm Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 24/47] Modify save_live_pending for postcopy Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 25/47] postcopy: OS support test Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 26/47] migrate_start_postcopy: Command to trigger transition to postcopy Dr. David Alan Gilbert (git)
2015-04-14 17:38 ` Eric Blake
2015-04-14 17:40 ` Dr. David Alan Gilbert
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 27/47] MIGRATION_STATUS_POSTCOPY_ACTIVE: Add new migration state Dr. David Alan Gilbert (git)
2015-04-14 17:40 ` Eric Blake
2015-04-14 18:00 ` Dr. David Alan Gilbert
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 28/47] Add qemu_savevm_state_complete_postcopy Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 29/47] Postcopy: Maintain sentmap and calculate discard Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 30/47] postcopy: Incoming initialisation Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 31/47] postcopy: ram_enable_notify to switch on userfault Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 32/47] Postcopy: Postcopy startup in migration thread Dr. David Alan Gilbert (git)
2015-04-14 17:03 ` [Qemu-devel] [PATCH v6 33/47] Postcopy end in migration_thread Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 34/47] Page request: Add MIG_RP_MSG_REQ_PAGES reverse command Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 35/47] Page request: Process incoming page request Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 36/47] Page request: Consume pages off the post-copy queue Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 37/47] postcopy_ram.c: place_page and helpers Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 38/47] Postcopy: Use helpers to map pages during migration Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 39/47] qemu_ram_block_from_host Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 40/47] Don't sync dirty bitmaps in postcopy Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` Dr. David Alan Gilbert (git) [this message]
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 42/47] Postcopy; Handle userfault requests Dr. David Alan Gilbert (git)
2015-05-25 9:18 ` zhanghailiang
2015-05-26 9:50 ` Dr. David Alan Gilbert
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 43/47] Start up a postcopy/listener thread ready for incoming page data Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 44/47] postcopy: Wire up loadvm_postcopy_handle_ commands Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 45/47] End of migration for postcopy Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 46/47] Disable mlock around incoming postcopy Dr. David Alan Gilbert (git)
2015-04-14 17:04 ` [Qemu-devel] [PATCH v6 47/47] Inhibit ballooning during postcopy Dr. David Alan Gilbert (git)
2015-04-27 8:04 ` [Qemu-devel] [PATCH v6 00/47] Postcopy implementation Li, Liang Z
2015-04-29 17:23 ` Dr. David Alan Gilbert
2015-04-30 1:09 ` Li, Liang Z
[not found] ` <20150505150112.GM2126@work-vm>
[not found] ` <F2CBF3009FA73547804AE4C663CAB28E50F0E1@shsmsx102.ccr.corp.intel.com>
[not found] ` <20150506083056.GB2204@work-vm>
2015-05-07 1:21 ` Li, Liang Z
2015-05-07 8:01 ` Dr. David Alan Gilbert
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1429031053-4454-42-git-send-email-dgilbert@redhat.com \
--to=dgilbert@redhat.com \
--cc=aarcange@redhat.com \
--cc=amit.shah@redhat.com \
--cc=david@gibson.dropbear.id.au \
--cc=pbonzini@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=quintela@redhat.com \
--cc=yamahata@private.email.ne.jp \
--cc=yayanghy@cn.fujitsu.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).