qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Pavel Dovgalyuk <Pavel.Dovgaluk@ispras.ru>
To: qemu-devel@nongnu.org
Cc: kwolf@redhat.com, peter.maydell@linaro.org,
	boost.lists@gmail.com, quintela@redhat.com, jasowang@redhat.com,
	mst@redhat.com, zuban32s@gmail.com,
	maria.klimushenkova@ispras.ru, dovgaluk@ispras.ru,
	kraxel@redhat.com, pavel.dovgaluk@ispras.ru, pbonzini@redhat.com,
	alex.bennee@linaro.org
Subject: [Qemu-devel] [RFC PATCH v3 16/30] cpus: only take BQL for sleeping threads
Date: Thu, 11 Jan 2018 11:26:26 +0300	[thread overview]
Message-ID: <20180111082626.27295.52919.stgit@pasha-VirtualBox> (raw)
In-Reply-To: <20180111082452.27295.85707.stgit@pasha-VirtualBox>

From: Alex Bennée <alex.bennee@linaro.org>

Now the only real need to hold the BQL is for when we sleep on the
cpu->halt conditional. The lock is actually dropped while the thread
sleeps so the actual window for contention is pretty small. This also
means we can remove the special case hack for exclusive work and
simply declare that work no longer has an implicit BQL held. This
isn't a major problem async work is generally only changing things in
the context of its own vCPU. If it needs to work across vCPUs it
should be using the exclusive mechanism or possibly taking the lock
itself.

Signed-off-by: Alex Bennée <alex.bennee@linaro.org>
Tested-by: Pavel Dovgalyuk <pavel.dovgaluk@ispras.ru>
---
 cpus-common.c |   13 +++++--------
 cpus.c        |   10 ++++------
 2 files changed, 9 insertions(+), 14 deletions(-)

diff --git a/cpus-common.c b/cpus-common.c
index 59f751e..64661c3 100644
--- a/cpus-common.c
+++ b/cpus-common.c
@@ -310,6 +310,11 @@ void async_safe_run_on_cpu(CPUState *cpu, run_on_cpu_func func,
     queue_work_on_cpu(cpu, wi);
 }
 
+/* Work items run outside of the BQL. This is essential for avoiding a
+ * deadlock for exclusive work but also applies to non-exclusive work.
+ * If the work requires cross-vCPU changes then it should use the
+ * exclusive mechanism.
+ */
 void process_queued_cpu_work(CPUState *cpu)
 {
     struct qemu_work_item *wi;
@@ -327,17 +332,9 @@ void process_queued_cpu_work(CPUState *cpu)
         }
         qemu_mutex_unlock(&cpu->work_mutex);
         if (wi->exclusive) {
-            /* Running work items outside the BQL avoids the following deadlock:
-             * 1) start_exclusive() is called with the BQL taken while another
-             * CPU is running; 2) cpu_exec in the other CPU tries to takes the
-             * BQL, so it goes to sleep; start_exclusive() is sleeping too, so
-             * neither CPU can proceed.
-             */
-            qemu_mutex_unlock_iothread();
             start_exclusive();
             wi->func(cpu, wi->data);
             end_exclusive();
-            qemu_mutex_lock_iothread();
         } else {
             wi->func(cpu, wi->data);
         }
diff --git a/cpus.c b/cpus.c
index 82dcbf8..79dda49 100644
--- a/cpus.c
+++ b/cpus.c
@@ -1149,31 +1149,29 @@ static bool qemu_tcg_should_sleep(CPUState *cpu)
 
 static void qemu_tcg_wait_io_event(CPUState *cpu)
 {
-    qemu_mutex_lock_iothread();
 
     while (qemu_tcg_should_sleep(cpu)) {
+        qemu_mutex_lock_iothread();
         stop_tcg_kick_timer();
         qemu_cond_wait(cpu->halt_cond, &qemu_global_mutex);
+        qemu_mutex_unlock_iothread();
     }
 
     start_tcg_kick_timer();
 
     qemu_wait_io_event_common(cpu);
-
-    qemu_mutex_unlock_iothread();
 }
 
 static void qemu_kvm_wait_io_event(CPUState *cpu)
 {
-    qemu_mutex_lock_iothread();
 
     while (cpu_thread_is_idle(cpu)) {
+        qemu_mutex_lock_iothread();
         qemu_cond_wait(cpu->halt_cond, &qemu_global_mutex);
+        qemu_mutex_unlock_iothread();
     }
 
     qemu_wait_io_event_common(cpu);
-
-    qemu_mutex_unlock_iothread();
 }
 
 static void qemu_hvf_wait_io_event(CPUState *cpu)

  parent reply	other threads:[~2018-01-11  8:26 UTC|newest]

Thread overview: 51+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-01-11  8:24 [Qemu-devel] [RFC PATCH v3 00/30] replay additions Pavel Dovgalyuk
2018-01-11  8:24 ` [Qemu-devel] [RFC PATCH v3 01/30] hpet: recover timer offset correctly Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 02/30] cpu: flush TB cache when loading VMState Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 03/30] This patch adds a condition before overwriting exception_index fields Pavel Dovgalyuk
2018-01-11 12:29   ` Paolo Bonzini
2018-01-12  6:12     ` Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 04/30] block: implement bdrv_snapshot_goto for blkreplay Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 05/30] blkreplay: create temporary overlay for underlaying devices Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 06/30] replay: disable default snapshot for record/replay Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 07/30] replay: fix processing async events Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 08/30] replay: fixed replay_enable_events Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 09/30] replay: fix save/load vm for non-empty queue Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 10/30] replay: added replay log format description Pavel Dovgalyuk
2018-01-11  8:25 ` [Qemu-devel] [RFC PATCH v3 11/30] replay: make safe vmstop at record/replay Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 12/30] replay: save prior value of the host clock Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 13/30] icount: fixed saving/restoring of icount warp timers Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 14/30] target/arm/arm-powertctl: drop BQL assertions Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 15/30] cpus: push BQL lock to qemu_*_wait_io_event Pavel Dovgalyuk
2018-01-11 12:50   ` Paolo Bonzini
2018-01-11  8:26 ` Pavel Dovgalyuk [this message]
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 17/30] replay/replay.c: bump REPLAY_VERSION again Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 18/30] replay/replay-internal.c: track holding of replay_lock Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 19/30] replay: make locking visible outside replay code Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 20/30] replay: push replay_mutex_lock up the call tree Pavel Dovgalyuk
2018-01-11  8:26 ` [Qemu-devel] [RFC PATCH v3 21/30] replay: don't destroy mutex at exit Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 22/30] replay: check return values of fwrite Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 23/30] replay: avoid recursive call of checkpoints Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 24/30] scripts/qemu-gdb: add simple tcg lock status helper Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 25/30] util/qemu-thread-*: add qemu_lock, locked and unlock trace events Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 26/30] scripts/analyse-locks-simpletrace.py: script to analyse lock times Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 27/30] scripts/replay-dump.py: replay log dumper Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 28/30] scripts/qemu-gdb/timers.py: new helper to dump timer state Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 29/30] replay: improve replay performance Pavel Dovgalyuk
2018-01-11 12:55   ` Paolo Bonzini
2018-01-11 13:12     ` Pavel Dovgalyuk
2018-01-11 13:22       ` Paolo Bonzini
2018-01-12  6:13         ` Pavel Dovgalyuk
2018-01-12 12:10           ` Paolo Bonzini
2018-01-12 12:41             ` Pavel Dovgalyuk
2018-01-11  8:27 ` [Qemu-devel] [RFC PATCH v3 30/30] replay: don't process async events when warping the clock Pavel Dovgalyuk
2018-01-11 12:37   ` Paolo Bonzini
2018-01-12  7:20     ` Pavel Dovgalyuk
2018-01-11  8:54 ` [Qemu-devel] [RFC PATCH v3 00/30] replay additions no-reply
2018-01-11  8:55 ` no-reply
2018-01-11 12:56 ` Paolo Bonzini
2018-01-12  9:32   ` Paolo Bonzini
2018-01-12 12:07     ` Pavel Dovgalyuk
2018-01-12 12:11       ` Paolo Bonzini
2018-01-19  6:35         ` Pavel Dovgalyuk
2018-01-19  7:41           ` Paolo Bonzini
2018-01-19  7:45             ` Pavel Dovgalyuk

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20180111082626.27295.52919.stgit@pasha-VirtualBox \
    --to=pavel.dovgaluk@ispras.ru \
    --cc=alex.bennee@linaro.org \
    --cc=boost.lists@gmail.com \
    --cc=dovgaluk@ispras.ru \
    --cc=jasowang@redhat.com \
    --cc=kraxel@redhat.com \
    --cc=kwolf@redhat.com \
    --cc=maria.klimushenkova@ispras.ru \
    --cc=mst@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=peter.maydell@linaro.org \
    --cc=qemu-devel@nongnu.org \
    --cc=quintela@redhat.com \
    --cc=zuban32s@gmail.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).