From: Stefan Hajnoczi <stefanha@linux.vnet.ibm.com>
To: "Michael S. Tsirkin" <mst@redhat.com>
Cc: Rusty Russell <rusty@rustcorp.com.au>,
virtualization@lists.linux-foundation.org,
Stefan Hajnoczi <stefanha@gmail.com>,
Anthony Liguori <anthony@codemonkey.ws>,
kvm@vger.kernel.org
Subject: Re: [RFC] virtio: Support releasing lock during kick
Date: Fri, 25 Jun 2010 17:05:51 +0100 [thread overview]
Message-ID: <20100625160550.GA30610@stefan-thinkpad.transitives.com> (raw)
In-Reply-To: <20100625153220.GB17911@redhat.com>
On Fri, Jun 25, 2010 at 06:32:20PM +0300, Michael S. Tsirkin wrote:
> On Fri, Jun 25, 2010 at 04:31:44PM +0100, Stefan Hajnoczi wrote:
> > On Fri, Jun 25, 2010 at 01:43:17PM +0300, Michael S. Tsirkin wrote:
> > > On Fri, Jun 25, 2010 at 12:39:21PM +0930, Rusty Russell wrote:
> > > > On Thu, 24 Jun 2010 03:00:30 pm Stefan Hajnoczi wrote:
> > > > > On Wed, Jun 23, 2010 at 11:12 PM, Anthony Liguori <anthony@codemonkey.ws> wrote:
> > > > > > Shouldn't it be possible to just drop the lock before invoking
> > > > > > virtqueue_kick() and reacquire it afterwards? There's nothing in that
> > > > > > virtqueue_kick() path that the lock is protecting AFAICT.
> > > > >
> > > > > No, that would lead to a race condition because vq->num_added is
> > > > > modified by both virtqueue_add_buf_gfp() and virtqueue_kick().
> > > > > Without a lock held during virtqueue_kick() another vcpu could add
> > > > > bufs while vq->num_added is used and cleared by virtqueue_kick():
> > > >
> > > > Right, this dovetails with another proposed change (was it Michael?)
> > > > where we would update the avail idx inside add_buf, rather than waiting
> > > > until kick. This means a barrier inside add_buf, but that's probably
> > > > fine.
> > > >
> > > > If we do that, then we don't need a lock on virtqueue_kick.
> > > >
> > > > Michael, thoughts?
> > >
> > > Maybe not even that: I think we could just do virtio_wmb()
> > > in add, and keep the mb() in kick.
> > >
> > > What I'm a bit worried about is contention on the cacheline
> > > including index and flags: the more we write to that line,
> > > the worse it gets.
> > >
> > > So need to test performance impact of this change:
> > > I didn't find time to do this yet, as I am trying
> > > to finalize the used index publishing patches.
> > > Any takers?
> > >
> > > Do we see performance improvement after making kick lockless?
> >
> > There was no guest CPU reduction or I/O throughput increase with my
> > patch when running 4 dd iflag=direct bs=4k if=/dev/vdb of=/dev/null
> > processes. However, the lock_stat numbers above show clear improvement
> > of the lock hold/wait times.
> >
> > I was hoping to see guest CPU utilization go down and I/O throughput go
> > up, so there is still investigation to do with my patch in isolation.
> > Although I'd like to try it later, putting my patch on top of your avail
> > idx work is too early because it will be harder to reason about the
> > performance with both patches present at the same time.
> >
> > Stefan
>
> What about host CPU utilization?
There is data available for host CPU utilization, I need to dig it up.
> Also, are you using PARAVIRT_SPINLOCKS?
No. I haven't found much documentation on paravirt spinlocks other than
the commit that introduced them:
commit 8efcbab674de2bee45a2e4cdf97de16b8e609ac8
Author: Jeremy Fitzhardinge <jeremy@goop.org>
Date: Mon Jul 7 12:07:51 2008 -0700
paravirt: introduce a "lock-byte" spinlock implementation
PARAVIRT_SPINLOCKS is not set in the config I use, probably because of
the associated performance issue that causes distros to build without
them:
commit b4ecc126991b30fe5f9a59dfacda046aeac124b2
Author: Jeremy Fitzhardinge <jeremy@goop.org>
Date: Wed May 13 17:16:55 2009 -0700
x86: Fix performance regression caused by paravirt_ops on native
kernels
I would expect performance results to be smoother with
PARAVIRT_SPINLOCKS for the guest kernel. I will add it for future runs,
thanks for pointing it out.
Stefan
next prev parent reply other threads:[~2010-06-25 16:05 UTC|newest]
Thread overview: 19+ messages / expand[flat|nested] mbox.gz Atom feed top
2010-06-23 21:24 [RFC] virtio: Support releasing lock during kick Stefan Hajnoczi
2010-06-23 22:12 ` Anthony Liguori
2010-06-24 5:30 ` Stefan Hajnoczi
2010-06-25 3:09 ` Rusty Russell
2010-06-25 6:17 ` Stefan Hajnoczi
2010-06-25 10:43 ` Michael S. Tsirkin
2010-06-25 15:31 ` Stefan Hajnoczi
2010-06-25 15:32 ` Michael S. Tsirkin
2010-06-25 16:05 ` Stefan Hajnoczi [this message]
2010-06-28 15:55 ` Marcelo Tosatti
2010-06-29 7:08 ` Stefan Hajnoczi
2010-06-29 7:12 ` Avi Kivity
2011-06-19 7:14 ` Michael S. Tsirkin
2011-06-20 15:27 ` Stefan Hajnoczi
2011-06-24 9:16 ` Stefan Hajnoczi
2011-08-10 13:18 ` Christoph Hellwig
2011-08-10 14:39 ` Stefan Hajnoczi
2011-06-19 7:48 ` Michael S. Tsirkin
2011-06-19 13:55 ` Christoph Hellwig
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20100625160550.GA30610@stefan-thinkpad.transitives.com \
--to=stefanha@linux.vnet.ibm.com \
--cc=anthony@codemonkey.ws \
--cc=kvm@vger.kernel.org \
--cc=mst@redhat.com \
--cc=rusty@rustcorp.com.au \
--cc=stefanha@gmail.com \
--cc=virtualization@lists.linux-foundation.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox