From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from mailman by lists.gnu.org with tmda-scanned (Exim 4.43) id 1Njfop-0006uN-Qv for qemu-devel@nongnu.org; Mon, 22 Feb 2010 16:27:55 -0500 Received: from [199.232.76.173] (port=56403 helo=monty-python.gnu.org) by lists.gnu.org with esmtp (Exim 4.43) id 1Njfop-0006ty-AV for qemu-devel@nongnu.org; Mon, 22 Feb 2010 16:27:55 -0500 Received: from Debian-exim by monty-python.gnu.org with spam-scanned (Exim 4.60) (envelope-from ) id 1Njfoo-000377-B0 for qemu-devel@nongnu.org; Mon, 22 Feb 2010 16:27:55 -0500 Received: from mx1.redhat.com ([209.132.183.28]:29717) by monty-python.gnu.org with esmtp (Exim 4.60) (envelope-from ) id 1Njfon-00036a-71 for qemu-devel@nongnu.org; Mon, 22 Feb 2010 16:27:53 -0500 From: Marcelo Tosatti Date: Mon, 22 Feb 2010 18:26:47 -0300 Message-Id: In-Reply-To: References: Subject: [Qemu-devel] [PATCH 5/8] kvm specific wait_io_event List-Id: qemu-devel.nongnu.org List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: Anthony Liguori Cc: Marcelo Tosatti , qemu-devel@nongnu.org, kvm@vger.kernel.org, Avi Kivity In KVM mode the global mutex is released when vcpus are executing, which means acquiring the fairness mutex is not required. Also for KVM there is one thread per vcpu, so tcg_has_work is meaningless. Add a new qemu_wait_io_event_common function to hold common code between TCG/KVM. Signed-off-by: Marcelo Tosatti Signed-off-by: Avi Kivity --- vl.c | 26 ++++++++++++++++++++------ 1 files changed, 20 insertions(+), 6 deletions(-) diff --git a/vl.c b/vl.c index 3bc618d..af198c1 100644 --- a/vl.c +++ b/vl.c @@ -3383,6 +3383,7 @@ static QemuCond qemu_pause_cond; static void block_io_signals(void); static void unblock_io_signals(void); static int tcg_has_work(void); +static int cpu_has_work(CPUState *env); static int qemu_init_main_loop(void) { @@ -3403,6 +3404,15 @@ static int qemu_init_main_loop(void) return 0; } +static void qemu_wait_io_event_common(CPUState *env) +{ + if (env->stop) { + env->stop = 0; + env->stopped = 1; + qemu_cond_signal(&qemu_pause_cond); + } +} + static void qemu_wait_io_event(CPUState *env) { while (!tcg_has_work()) @@ -3419,11 +3429,15 @@ static void qemu_wait_io_event(CPUState *env) qemu_mutex_unlock(&qemu_fair_mutex); qemu_mutex_lock(&qemu_global_mutex); - if (env->stop) { - env->stop = 0; - env->stopped = 1; - qemu_cond_signal(&qemu_pause_cond); - } + qemu_wait_io_event_common(env); +} + +static void qemu_kvm_wait_io_event(CPUState *env) +{ + while (!cpu_has_work(env)) + qemu_cond_timedwait(env->halt_cond, &qemu_global_mutex, 1000); + + qemu_wait_io_event_common(env); } static int qemu_cpu_exec(CPUState *env); @@ -3449,7 +3463,7 @@ static void *kvm_cpu_thread_fn(void *arg) while (1) { if (cpu_can_run(env)) qemu_cpu_exec(env); - qemu_wait_io_event(env); + qemu_kvm_wait_io_event(env); } return NULL; -- 1.6.6