qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Juan Quintela <quintela@redhat.com>
To: qemu-devel@nongnu.org
Cc: Paolo Bonzini <pbonzini@redhat.com>,
	Umesh Deshpande <udeshpan@redhat.com>
Subject: [Qemu-devel] [PATCH 03/30] protect the ramlist with a separate mutex
Date: Thu, 18 Oct 2012 09:29:59 +0200	[thread overview]
Message-ID: <1350545426-23172-4-git-send-email-quintela@redhat.com> (raw)
In-Reply-To: <1350545426-23172-1-git-send-email-quintela@redhat.com>

From: Umesh Deshpande <udeshpan@redhat.com>

Add the new mutex that protects shared state between ram_save_live
and the iothread.  If the iothread mutex has to be taken together
with the ramlist mutex, the iothread shall always be _outside_.

Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Umesh Deshpande <udeshpan@redhat.com>
Signed-off-by: Juan Quintela <quintela@redhat.com>
---
 arch_init.c |  9 ++++++++-
 cpu-all.h   |  8 ++++++++
 exec.c      | 23 +++++++++++++++++++++--
 3 files changed, 37 insertions(+), 3 deletions(-)

diff --git a/arch_init.c b/arch_init.c
index b47313d..2d29828 100644
--- a/arch_init.c
+++ b/arch_init.c
@@ -553,7 +553,6 @@ static void ram_migration_cancel(void *opaque)
     migration_end();
 }

-
 static void reset_ram_globals(void)
 {
     last_block = NULL;
@@ -573,6 +572,7 @@ static int ram_save_setup(QEMUFile *f, void *opaque)
     bitmap_set(migration_bitmap, 1, ram_pages);
     migration_dirty_pages = ram_pages;

+    qemu_mutex_lock_ramlist();
     bytes_transferred = 0;
     reset_ram_globals();

@@ -600,6 +600,7 @@ static int ram_save_setup(QEMUFile *f, void *opaque)
         qemu_put_be64(f, block->length);
     }

+    qemu_mutex_unlock_ramlist();
     qemu_put_be64(f, RAM_SAVE_FLAG_EOS);

     return 0;
@@ -614,6 +615,8 @@ static int ram_save_iterate(QEMUFile *f, void *opaque)
     uint64_t expected_downtime;
     MigrationState *s = migrate_get_current();

+    qemu_mutex_lock_ramlist();
+
     if (ram_list.version != last_version) {
         reset_ram_globals();
     }
@@ -662,6 +665,7 @@ static int ram_save_iterate(QEMUFile *f, void *opaque)
         bwidth = 0.000001;
     }

+    qemu_mutex_unlock_ramlist();
     qemu_put_be64(f, RAM_SAVE_FLAG_EOS);

     expected_downtime = ram_save_remaining() * TARGET_PAGE_SIZE / bwidth;
@@ -682,6 +686,8 @@ static int ram_save_complete(QEMUFile *f, void *opaque)
 {
     migration_bitmap_sync();

+    qemu_mutex_lock_ramlist();
+
     /* try transferring iterative blocks of memory */

     /* flush all remaining blocks regardless of rate limiting */
@@ -697,6 +703,7 @@ static int ram_save_complete(QEMUFile *f, void *opaque)
     }
     memory_global_dirty_log_stop();

+    qemu_mutex_unlock_ramlist();
     qemu_put_be64(f, RAM_SAVE_FLAG_EOS);

     g_free(migration_bitmap);
diff --git a/cpu-all.h b/cpu-all.h
index e07c91c..2bafcdd 100644
--- a/cpu-all.h
+++ b/cpu-all.h
@@ -22,6 +22,7 @@
 #include "qemu-common.h"
 #include "qemu-tls.h"
 #include "cpu-common.h"
+#include "qemu-thread.h"

 /* some important defines:
  *
@@ -490,7 +491,9 @@ typedef struct RAMBlock {
     ram_addr_t offset;
     ram_addr_t length;
     uint32_t flags;
+    /* Protected by the iothread lock.  */
     QLIST_ENTRY(RAMBlock) next_mru;
+    /* Protected by the ramlist lock.  */
     QLIST_ENTRY(RAMBlock) next;
     char idstr[256];
 #if defined(__linux__) && !defined(TARGET_S390X)
@@ -499,9 +502,12 @@ typedef struct RAMBlock {
 } RAMBlock;

 typedef struct RAMList {
+    QemuMutex mutex;
+    /* Protected by the iothread lock.  */
     uint8_t *phys_dirty;
     uint32_t version;
     QLIST_HEAD(, RAMBlock) blocks_mru;
+    /* Protected by the ramlist lock.  */
     QLIST_HEAD(, RAMBlock) blocks;
 } RAMList;
 extern RAMList ram_list;
@@ -521,6 +527,8 @@ extern int mem_prealloc;

 void dump_exec_info(FILE *f, fprintf_function cpu_fprintf);
 ram_addr_t last_ram_offset(void);
+void qemu_mutex_lock_ramlist(void);
+void qemu_mutex_unlock_ramlist(void);
 #endif /* !CONFIG_USER_ONLY */

 int cpu_memory_rw_debug(CPUArchState *env, target_ulong addr,
diff --git a/exec.c b/exec.c
index 1e04711..cf9de92 100644
--- a/exec.c
+++ b/exec.c
@@ -637,6 +637,7 @@ bool tcg_enabled(void)
 void cpu_exec_init_all(void)
 {
 #if !defined(CONFIG_USER_ONLY)
+    qemu_mutex_init(&ram_list.mutex);
     memory_map_init();
     io_mem_init();
 #endif
@@ -2329,6 +2330,16 @@ void qemu_flush_coalesced_mmio_buffer(void)
         kvm_flush_coalesced_mmio_buffer();
 }

+void qemu_mutex_lock_ramlist(void)
+{
+    qemu_mutex_lock(&ram_list.mutex);
+}
+
+void qemu_mutex_unlock_ramlist(void)
+{
+    qemu_mutex_unlock(&ram_list.mutex);
+}
+
 #if defined(__linux__) && !defined(TARGET_S390X)

 #include <sys/vfs.h>
@@ -2510,6 +2521,7 @@ void qemu_ram_set_idstr(ram_addr_t addr, const char *name, DeviceState *dev)
     }
     pstrcat(new_block->idstr, sizeof(new_block->idstr), name);

+    qemu_mutex_lock_ramlist();
     QLIST_FOREACH(block, &ram_list.blocks, next) {
         if (block != new_block && !strcmp(block->idstr, new_block->idstr)) {
             fprintf(stderr, "RAMBlock \"%s\" already registered, abort!\n",
@@ -2517,6 +2529,7 @@ void qemu_ram_set_idstr(ram_addr_t addr, const char *name, DeviceState *dev)
             abort();
         }
     }
+    qemu_mutex_unlock_ramlist();
 }

 static int memory_try_enable_merging(void *addr, size_t len)
@@ -2540,6 +2553,7 @@ ram_addr_t qemu_ram_alloc_from_ptr(ram_addr_t size, void *host,
     size = TARGET_PAGE_ALIGN(size);
     new_block = g_malloc0(sizeof(*new_block));

+    qemu_mutex_lock_ramlist();
     new_block->mr = mr;
     new_block->offset = find_ram_offset(size);
     if (host) {
@@ -2575,6 +2589,7 @@ ram_addr_t qemu_ram_alloc_from_ptr(ram_addr_t size, void *host,
     QLIST_INSERT_HEAD(&ram_list.blocks_mru, new_block, next_mru);

     ram_list.version++;
+    qemu_mutex_unlock_ramlist();

     ram_list.phys_dirty = g_realloc(ram_list.phys_dirty,
                                        last_ram_offset() >> TARGET_PAGE_BITS);
@@ -2599,21 +2614,24 @@ void qemu_ram_free_from_ptr(ram_addr_t addr)
 {
     RAMBlock *block;

+    qemu_mutex_lock_ramlist();
     QLIST_FOREACH(block, &ram_list.blocks, next) {
         if (addr == block->offset) {
             QLIST_REMOVE(block, next);
             QLIST_REMOVE(block, next_mru);
             ram_list.version++;
             g_free(block);
-            return;
+            break;
         }
     }
+    qemu_mutex_unlock_ramlist();
 }

 void qemu_ram_free(ram_addr_t addr)
 {
     RAMBlock *block;

+    qemu_mutex_lock_ramlist();
     QLIST_FOREACH(block, &ram_list.blocks, next) {
         if (addr == block->offset) {
             QLIST_REMOVE(block, next);
@@ -2644,9 +2662,10 @@ void qemu_ram_free(ram_addr_t addr)
 #endif
             }
             g_free(block);
-            return;
+            break;
         }
     }
+    qemu_mutex_unlock_ramlist();

 }

-- 
1.7.11.7

  parent reply	other threads:[~2012-10-18  7:31 UTC|newest]

Thread overview: 58+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2012-10-18  7:29 [Qemu-devel] [PATCH 00/30] Migration thread 20121017 edition Juan Quintela
2012-10-18  7:29 ` [Qemu-devel] [PATCH 01/30] split MRU ram list Juan Quintela
2012-10-21 11:58   ` Orit Wasserman
2012-10-21 17:02   ` Peter Maydell
2012-10-18  7:29 ` [Qemu-devel] [PATCH 02/30] add a version number to ram_list Juan Quintela
2012-10-21 12:00   ` Orit Wasserman
2012-10-18  7:29 ` Juan Quintela [this message]
2012-10-21 12:05   ` [Qemu-devel] [PATCH 03/30] protect the ramlist with a separate mutex Orit Wasserman
2012-10-18  7:30 ` [Qemu-devel] [PATCH 04/30] buffered_file: Move from using a timer to use a thread Juan Quintela
2012-10-18  8:56   ` Paolo Bonzini
2012-10-21 12:09   ` Orit Wasserman
2012-11-12 11:42   ` Paolo Bonzini
2012-10-18  7:30 ` [Qemu-devel] [PATCH 05/30] migration: make qemu_fopen_ops_buffered() return void Juan Quintela
2012-10-21 12:10   ` Orit Wasserman
2012-10-18  7:30 ` [Qemu-devel] [PATCH 06/30] migration: stop all cpus correctly Juan Quintela
2012-11-12 11:44   ` Paolo Bonzini
2012-11-14 15:21     ` Paolo Bonzini
2012-12-14 12:36       ` Juan Quintela
2012-12-14 13:53         ` Paolo Bonzini
2012-10-18  7:30 ` [Qemu-devel] [PATCH 07/30] migration: make writes blocking Juan Quintela
2012-11-12 11:52   ` Paolo Bonzini
2012-10-18  7:30 ` [Qemu-devel] [PATCH 08/30] migration: remove unfreeze logic Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 09/30] migration: take finer locking Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 10/30] buffered_file: Unfold the trick to restart generating migration data Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 11/30] buffered_file: don't flush on put buffer Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 12/30] buffered_file: unfold buffered_append in buffered_put_buffer Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 13/30] savevm: New save live migration method: pending Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 14/30] migration: include qemu-file.h Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 15/30] migration-fd: remove duplicate include Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 16/30] migration: move buffered_file.c code into migration.c Juan Quintela
2012-10-18  8:57   ` Paolo Bonzini
2012-10-18  7:30 ` [Qemu-devel] [PATCH 17/30] migration: move migration_fd_put_ready() Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 18/30] migration: Inline qemu_fopen_ops_buffered into migrate_fd_connect Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 19/30] migration: move migration notifier Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 20/30] migration: move begining stage to the migration thread Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 21/30] migration: move exit condition to " Juan Quintela
2012-10-18  8:34   ` Paolo Bonzini
2012-10-26 11:43     ` Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 22/30] migration: unfold rest of migrate_fd_put_ready() into thread Juan Quintela
2012-10-18  8:39   ` Paolo Bonzini
2012-10-18  8:55     ` Paolo Bonzini
2012-10-18  7:30 ` [Qemu-devel] [PATCH 23/30] migration: print times for end phase Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 24/30] ram: rename last_block to last_seen_block Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 25/30] ram: Add last_sent_block Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 26/30] memory: introduce memory_region_test_and_clear_dirty Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 27/30] ram: Use memory_region_test_and_clear_dirty Juan Quintela
2012-10-18 12:52   ` Eric Blake
2012-10-18  7:30 ` [Qemu-devel] [PATCH 28/30] fix memory.c Juan Quintela
2012-10-18  8:43   ` Paolo Bonzini
2012-10-18  7:30 ` [Qemu-devel] [PATCH 29/30] migration: Only go to the iterate stage if there is anything to send Juan Quintela
2012-10-18  7:30 ` [Qemu-devel] [PATCH 30/30] ram: optimize migration bitmap walking Juan Quintela
2012-10-21 13:01   ` Orit Wasserman
2012-10-26 11:39     ` Juan Quintela
2012-10-28  8:35       ` Orit Wasserman
2012-10-30 10:15         ` Orit Wasserman
2012-10-30 15:33           ` Juan Quintela
2012-10-18  9:00 ` [Qemu-devel] [PATCH 00/30] Migration thread 20121017 edition Paolo Bonzini
2012-10-26 13:04 ` Paolo Bonzini

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1350545426-23172-4-git-send-email-quintela@redhat.com \
    --to=quintela@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=qemu-devel@nongnu.org \
    --cc=udeshpan@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).