virtualization.lists.linux-foundation.org archive mirror
 help / color / mirror / Atom feed
* [RFC PATCH TRIVIAL] Reading the virtio code...
@ 2011-04-23 23:13 Rob Landley
  2011-04-27  5:29 ` Rusty Russell
  0 siblings, 1 reply; 3+ messages in thread
From: Rob Landley @ 2011-04-23 23:13 UTC (permalink / raw)
  To: rusty, virtualization

From: Rob Landley <rlandley@parallels.com>

Going indirect for only two buffers isn't likely to be a performance win
because the kmalloc/kfree overhead for the indirect block can't be cheaper
than one extra linked list traversal.

Properly "tuning" the threshold would probably be workload-specific.
(One big downside of not going indirect is extra pressure on the table
entries, and table size varies.)  But I think that in the general case,
2 is a defensible minimum?

Signed-off-by: Rob Landley <rlandley@parallels.com>
---

 drivers/virtio/virtio_ring.c |    2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/drivers/virtio/virtio_ring.c b/drivers/virtio/virtio_ring.c
index b0043fb..2b69441 100644
--- a/drivers/virtio/virtio_ring.c
+++ b/drivers/virtio/virtio_ring.c
@@ -173,7 +173,7 @@ int virtqueue_add_buf_gfp(struct virtqueue *_vq,
 
 	/* If the host supports indirect descriptor tables, and we have multiple
 	 * buffers, then go indirect. FIXME: tune this threshold */
-	if (vq->indirect && (out + in) > 1 && vq->num_free) {
+	if (vq->indirect && (out + in) > 2 && vq->num_free) {
 		head = vring_add_indirect(vq, sg, out, in, gfp);
 		if (likely(head >= 0))
 			goto add_head;

^ permalink raw reply related	[flat|nested] 3+ messages in thread

* Re: [RFC PATCH TRIVIAL] Reading the virtio code...
  2011-04-23 23:13 [RFC PATCH TRIVIAL] Reading the virtio code Rob Landley
@ 2011-04-27  5:29 ` Rusty Russell
  2011-04-27  6:26   ` Michael S. Tsirkin
  0 siblings, 1 reply; 3+ messages in thread
From: Rusty Russell @ 2011-04-27  5:29 UTC (permalink / raw)
  To: Rob Landley, virtualization; +Cc: Michael S. Tsirkin

On Sat, 23 Apr 2011 18:13:34 -0500, Rob Landley <rlandley@parallels.com> wrote:
> From: Rob Landley <rlandley@parallels.com>
> 
> Going indirect for only two buffers isn't likely to be a performance win
> because the kmalloc/kfree overhead for the indirect block can't be cheaper
> than one extra linked list traversal.

Unfortunately it's not completely clear.  QEMU sets fairly small rings,
and the virtio-net driver uses 2 descriptors minimum.  The effect can be
a real bottleneck for small packets.

Now, virtio-net could often stuff the virtio_net_hdr in the space before
the packet data (saving a descriptor) but I think that will need a
feature bit since qemu (incorrectly) used to insist on a separate
descriptor for that header.

> Properly "tuning" the threshold would probably be workload-specific.
> (One big downside of not going indirect is extra pressure on the table
> entries, and table size varies.)  But I think that in the general case,
> 2 is a defensible minimum?

I'd be tempted to say that once we fill the ring, we should drop the
threshold.

Michael?

Thanks,
Rusty.

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [RFC PATCH TRIVIAL] Reading the virtio code...
  2011-04-27  5:29 ` Rusty Russell
@ 2011-04-27  6:26   ` Michael S. Tsirkin
  0 siblings, 0 replies; 3+ messages in thread
From: Michael S. Tsirkin @ 2011-04-27  6:26 UTC (permalink / raw)
  To: Rusty Russell; +Cc: Rob Landley, virtualization

On Wed, Apr 27, 2011 at 02:59:27PM +0930, Rusty Russell wrote:
> On Sat, 23 Apr 2011 18:13:34 -0500, Rob Landley <rlandley@parallels.com> wrote:
> > From: Rob Landley <rlandley@parallels.com>
> > 
> > Going indirect for only two buffers isn't likely to be a performance win
> > because the kmalloc/kfree overhead for the indirect block can't be cheaper
> > than one extra linked list traversal.
> 
> Unfortunately it's not completely clear.  QEMU sets fairly small rings,
> and the virtio-net driver uses 2 descriptors minimum.  The effect can be
> a real bottleneck for small packets.
> 
> Now, virtio-net could often stuff the virtio_net_hdr in the space before
> the packet data (saving a descriptor) but I think that will need a
> feature bit since qemu (incorrectly) used to insist on a separate
> descriptor for that header.
> 
> > Properly "tuning" the threshold would probably be workload-specific.
> > (One big downside of not going indirect is extra pressure on the table
> > entries, and table size varies.)  But I think that in the general case,
> > 2 is a defensible minimum?
> 
> I'd be tempted to say that once we fill the ring, we should drop the
> threshold.
> 
> Michael?
> 
> Thanks,
> Rusty.

Yes, one idea is to use part of a ring (e.g. 1/4 of a ring) for direct
entries, and the rest for indirect. So we end up with a threshold like

	max(1, vq->num_free - in - out - vq->num * N)

(above is pseudo-code, must take care of unsigned vs signed etc)
and I think I'd try with N = 3/4 or maybe N = 1/2


-- 
MST

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2011-04-27  6:26 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2011-04-23 23:13 [RFC PATCH TRIVIAL] Reading the virtio code Rob Landley
2011-04-27  5:29 ` Rusty Russell
2011-04-27  6:26   ` Michael S. Tsirkin

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).