From: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
To: Rik van Riel <riel@redhat.com>
Cc: "Jeremy Fitzhardinge" <jeremy@goop.org>,
Xen-devel <xen-devel@lists.xensource.com>,
"Mathieu Desnoyers" <mathieu.desnoyers@polymtl.ca>,
"Nick Piggin" <npiggin@kernel.dk>,
kvm@vger.kernel.org, "Peter Zijlstra" <peterz@infradead.org>,
"Linux Kernel Mailing List" <linux-kernel@vger.kernel.org>,
"Jan Beulich" <JBeulich@novell.com>,
"Eric Dumazet" <dada1@cosmosbay.com>,
"Jeremy Fitzhardinge" <jeremy.fitzhardinge@citrix.com>,
suzuki@in.ibm.com, "Avi Kivity" <avi@redhat.com>,
"H. Peter Anvin" <hpa@zytor.com>,
"Américo Wang" <xiyou.wangcong@gmail.com>,
"Linux Virtualization"
<virtualization@lists.linux-foundation.org>
Subject: Re: [PATCH 2/3] kvm hypervisor : Add hypercalls to support pv-ticketlock
Date: Sat, 22 Jan 2011 11:44:17 +0530 [thread overview]
Message-ID: <20110122061417.GA7258@linux.vnet.ibm.com> (raw)
In-Reply-To: <4D399CBD.10506@redhat.com>
On Fri, Jan 21, 2011 at 09:48:29AM -0500, Rik van Riel wrote:
> >>Why? If a VCPU can't make progress because its waiting for some
> >>resource, then why not schedule something else instead?
> >
> >In the process, "something else" can get more share of cpu resource than its
> >entitled to and that's where I was bit concerned. I guess one could
> >employ hard-limits to cap "something else's" bandwidth where it is of real
> >concern (like clouds).
>
> I'd like to think I fixed those things in my yield_task_fair +
> yield_to + kvm_vcpu_on_spin patch series from yesterday.
Speaking of the spinlock-in-virtualized-environment problem as whole, IMHO
I don't think that kvm_vcpu_on_spin + yield changes will provide the best
results, especially where ticketlocks are involved and they are paravirtualized
in a manner being discussed in this thread. An important focus of pv-ticketlocks
is to reduce the lock _acquisition_ time by ensuring that the next-in-line
vcpu gets to run asap when a ticket lock is released. With the way
kvm_vcpu_on_spin+yield_to is implemented, I don't see how we can provide the
best lock acquisition times for threads. It would be nice though to compare
the two approaches (kvm_vcpu_on_spin optimization and the pv-ticketlock scheme)
to get some real-world numbers. I unfortunately don't have access to a PLE
capable hardware which is required to test your kvm_vcpu_on_spin changes?
Also it may be possible for the pv-ticketlocks to track owning vcpu and make use
of a yield-to interface as further optimization to avoid the
"others-get-more-time" problem, but Peterz rightly pointed that PI would be a
better solution there than yield-to. So overall IMO kvm_vcpu_on_spin+yield_to
could be the best solution for unmodified guests, while paravirtualized
ticketlocks + some sort of PI would be a better solution where we have the
luxury of modifying guest sources!
- vatsa
next prev parent reply other threads:[~2011-01-22 6:14 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <cover.1289940821.git.jeremy.fitzhardinge@citrix.com>
2011-01-19 16:44 ` [PATCH 00/14] PV ticket locks without expanding spinlock Srivatsa Vaddagiri
2011-01-19 17:07 ` [PATCH 1/3] debugfs: Add support to print u32 array Srivatsa Vaddagiri
2011-01-19 17:12 ` [PATCH 2/3] kvm hypervisor : Add hypercalls to support pv-ticketlock Srivatsa Vaddagiri
2011-01-19 17:21 ` Peter Zijlstra
2011-01-19 18:29 ` Srivatsa Vaddagiri
2011-01-19 18:53 ` Jeremy Fitzhardinge
2011-01-20 11:42 ` Srivatsa Vaddagiri
2011-01-20 17:49 ` Jeremy Fitzhardinge
2011-01-20 11:59 ` Srivatsa Vaddagiri
2011-01-20 13:41 ` Peter Zijlstra
2011-01-20 14:34 ` Srivatsa Vaddagiri
2011-01-20 17:56 ` Jeremy Fitzhardinge
2011-01-21 14:02 ` Srivatsa Vaddagiri
2011-01-21 14:48 ` Rik van Riel
2011-01-22 6:14 ` Srivatsa Vaddagiri [this message]
2011-01-22 14:53 ` Rik van Riel
2011-01-24 17:49 ` Jeremy Fitzhardinge
2011-01-19 17:23 ` Srivatsa Vaddagiri
2011-01-19 17:50 ` Peter Zijlstra
2011-01-19 17:17 ` [PATCH 3/3] kvm guest : Add support for pv-ticketlocks Srivatsa Vaddagiri
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20110122061417.GA7258@linux.vnet.ibm.com \
--to=vatsa@linux.vnet.ibm.com \
--cc=JBeulich@novell.com \
--cc=avi@redhat.com \
--cc=dada1@cosmosbay.com \
--cc=hpa@zytor.com \
--cc=jeremy.fitzhardinge@citrix.com \
--cc=jeremy@goop.org \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mathieu.desnoyers@polymtl.ca \
--cc=npiggin@kernel.dk \
--cc=peterz@infradead.org \
--cc=riel@redhat.com \
--cc=suzuki@in.ibm.com \
--cc=virtualization@lists.linux-foundation.org \
--cc=xen-devel@lists.xensource.com \
--cc=xiyou.wangcong@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox