From: Shirley Ma <mashirle@us.ibm.com>
To: "Michael S. Tsirkin" <mst@redhat.com>
Cc: Anthony Liguori <aliguori@us.ibm.com>,
Tom Lendacky <toml@us.ibm.com>,
netdev@vger.kernel.org, Cristian Viana <vianac@br.ibm.com>
Subject: Re: [PATCH 1/2] vhost: allow multiple workers threads
Date: Mon, 20 Feb 2012 20:03:18 -0800 [thread overview]
Message-ID: <1329796998.13141.26.camel@localhost.localdomain> (raw)
In-Reply-To: <20120221032147.GB2502@redhat.com>
On Tue, 2012-02-21 at 05:21 +0200, Michael S. Tsirkin wrote:
> On Mon, Feb 20, 2012 at 05:04:10PM -0800, Shirley Ma wrote:
> > On Mon, 2012-02-20 at 23:00 +0200, Michael S. Tsirkin wrote:
> > >
> > > The point was really to avoid scheduler overhead
> > > as with tcp, tx and rx tend to run on the same cpu.
> >
> > We have tried different approaches in the past, like splitting vhost
> > thread to separate TX, RX threads; create per cpu vhost thread
> instead
> > of creating per VM per virtio_net vhost thread...
> >
> > We think per cpu vhost thread is a better approach based on the data
> we
> > have collected. It will reduce both vhost resource and scheduler
> > overhead. It will not depend on host scheduler, has less various.
> The
> > patch is under testing, we hope we can post it soon.
> >
> > Thanks
> > Shirley
>
> Yes, great, this is definitely interesting. I actually started with
> a per-cpu one - it did not perform well but I did not
> figure out why, switching to a single thread fixed it
> and I did not dig into it.
The patch includes per cpu vhost thread & vhost NUMA aware scheduling
It is very interesting. We are collecting performance data with
different workloads (streams, request/response) related to which VCPU
runs on which CPU, which vhost cpu thread is being scheduled, and which
NIC TX/RX queues is being used. The performance were different when
using different vhost scheduling approach for both TX/RX worker. The
results seems pretty good: like 60 UDP_RRs, the results event more than
doubled in our lab. However the TCP_RRs results couldn't catch up
UDP_RRs.
Thanks
Shirley
next prev parent reply other threads:[~2012-02-21 5:42 UTC|newest]
Thread overview: 23+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-02-17 23:02 [PATCH 0/2][RFC] vhost: improve transmit rate with virtqueue polling Anthony Liguori
2012-02-17 23:02 ` [PATCH 1/2] vhost: allow multiple workers threads Anthony Liguori
2012-02-19 14:41 ` Michael S. Tsirkin
2012-02-20 15:50 ` Tom Lendacky
2012-02-20 19:27 ` Michael S. Tsirkin
2012-02-20 19:46 ` Anthony Liguori
2012-02-20 21:00 ` Michael S. Tsirkin
2012-02-21 1:04 ` Shirley Ma
2012-02-21 3:21 ` Michael S. Tsirkin
2012-02-21 4:03 ` Shirley Ma [this message]
2012-03-05 13:21 ` Anthony Liguori
2012-03-05 20:43 ` Shirley Ma
2012-02-21 4:32 ` Jason Wang
2012-02-21 4:51 ` Jason Wang
2012-02-17 23:02 ` [PATCH 2/2] vhost-net: add a spin_threshold parameter Anthony Liguori
2012-02-19 14:51 ` Michael S. Tsirkin
2012-02-21 1:35 ` Shirley Ma
2012-02-21 5:34 ` Jason Wang
2012-02-21 6:28 ` Shirley Ma
2012-02-21 6:38 ` Jason Wang
2012-02-21 11:09 ` Shirley Ma
2012-02-21 16:08 ` Sridhar Samudrala
2012-03-12 8:12 ` Dor Laor
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1329796998.13141.26.camel@localhost.localdomain \
--to=mashirle@us.ibm.com \
--cc=aliguori@us.ibm.com \
--cc=mst@redhat.com \
--cc=netdev@vger.kernel.org \
--cc=toml@us.ibm.com \
--cc=vianac@br.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).