From: Simon Horman <horms@verge.net.au>
To: Rusty Russell <rusty@rustcorp.com.au>,
Chris Wright <chrisw@sous-sol.org>
Cc: Arnd Bergmann <arnd@arndb.de>,
Herbert Xu <herbert@gondor.apana.org.au>,
kvm@vger.kernel.org
Subject: Re: copyless virtio net thoughts?
Date: Fri, 20 Feb 2009 10:09:33 +1100 [thread overview]
Message-ID: <20090219230933.GE15300@verge.net.au> (raw)
In-Reply-To: <20090219113752.GB25891@sequoia.sous-sol.org> <200902192206.17557.rusty@rustcorp.com.au>
On Thu, Feb 19, 2009 at 10:06:17PM +1030, Rusty Russell wrote:
> On Thursday 19 February 2009 10:01:42 Simon Horman wrote:
> > On Wed, Feb 18, 2009 at 10:08:00PM +1030, Rusty Russell wrote:
> > >
> > > 2) Direct NIC attachment This is particularly interesting with SR-IOV or
> > > other multiqueue nics, but for boutique cases or benchmarks, could be for
> > > normal NICs. So far I have some very sketched-out patches: for the
> > > attached nic dev_alloc_skb() gets an skb from the guest (which supplies
> > > them via some kind of AIO interface), and a branch in netif_receive_skb()
> > > which returned it to the guest. This bypasses all firewalling in the
> > > host though; we're basically having the guest process drive the NIC
> > > directly.
> >
> > Hi Rusty,
> >
> > Can I clarify that the idea with utilising SR-IOV would be to assign
> > virtual functions to guests? That is, something conceptually similar to
> > PCI pass-through in Xen (although I'm not sure that anyone has virtual
> > function pass-through working yet).
>
> Not quite: I think PCI passthrough IMHO is the *wrong* way to do it: it
> makes migrate complicated (if not impossible), and requires emulation or
> the same NIC on the destination host.
>
> This would be the *host* seeing the virtual functions as multiple NICs,
> then the ability to attach a given NIC directly to a process.
>
> This isn't guest-visible: the kvm process is configured to connect
> directly to a NIC, rather than (say) bridging through the host.
Hi Rusty, Hi Chris,
Thanks for the clarification.
I think that the approach that Xen recommends for migration is to
use a bonding device that accesses the pass-through device if present
and a virtual nic.
The idea that you outline above does sound somewhat cleaner :-)
> > If so, wouldn't this also be useful on machines that have multiple
> > NICs?
>
> Yes, but mainly as a benchmark hack AFAICT :)
Ok, I was under the impression that at least in the Xen world it
was something people actually used. But I could easily be mistaken.
> Hope that clarifies, Rusty.
On Thu, Feb 19, 2009 at 03:37:52AM -0800, Chris Wright wrote:
> * Simon Horman (horms@verge.net.au) wrote:
> > On Wed, Feb 18, 2009 at 10:08:00PM +1030, Rusty Russell wrote:
> > > 2) Direct NIC attachment This is particularly interesting with SR-IOV or
> > > other multiqueue nics, but for boutique cases or benchmarks, could be for
> > > normal NICs. So far I have some very sketched-out patches: for the
> > > attached nic dev_alloc_skb() gets an skb from the guest (which supplies
> > > them via some kind of AIO interface), and a branch in netif_receive_skb()
> > > which returned it to the guest. This bypasses all firewalling in the
> > > host though; we're basically having the guest process drive the NIC
> > > directly.
> >
> > Can I clarify that the idea with utilising SR-IOV would be to assign
> > virtual functions to guests? That is, something conceptually similar to
> > PCI pass-through in Xen (although I'm not sure that anyone has virtual
> > function pass-through working yet). If so, wouldn't this also be useful
> > on machines that have multiple NICs?
>
> This would be the typical usecase for sr-iov. But I think Rusty is
> referring to giving a nic "directly" to a guest but the guest is still
> seeing a virtio nic (not pass-through/device-assignment). So there's
> no bridge, and zero copy so the dma buffers are supplied by guest,
> but host has the driver for the physical nic or the VF.
--
Simon Horman
VA Linux Systems Japan K.K., Sydney, Australia Satellite Office
H: www.vergenet.net/~horms/ W: www.valinux.co.jp/en
next prev parent reply other threads:[~2009-02-19 23:09 UTC|newest]
Thread overview: 19+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-02-05 2:07 copyless virtio net thoughts? Chris Wright
2009-02-05 12:37 ` Avi Kivity
2009-02-05 14:25 ` Anthony Liguori
2009-02-06 5:40 ` Herbert Xu
2009-02-06 8:46 ` Avi Kivity
2009-02-06 9:19 ` Herbert Xu
2009-02-06 14:55 ` Avi Kivity
2009-02-07 11:56 ` Arnd Bergmann
2009-02-08 3:01 ` David Miller
2009-02-18 11:38 ` Rusty Russell
2009-02-18 12:17 ` Herbert Xu
2009-02-18 16:24 ` Arnd Bergmann
2009-02-19 10:56 ` Rusty Russell
2009-02-18 23:31 ` Simon Horman
2009-02-19 1:03 ` Dong, Eddie
2009-02-19 11:36 ` Rusty Russell
2009-02-19 14:51 ` Arnd Bergmann
2009-02-19 23:09 ` Simon Horman [this message]
2009-02-19 11:37 ` Chris Wright
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20090219230933.GE15300@verge.net.au \
--to=horms@verge.net.au \
--cc=arnd@arndb.de \
--cc=chrisw@sous-sol.org \
--cc=herbert@gondor.apana.org.au \
--cc=kvm@vger.kernel.org \
--cc=rusty@rustcorp.com.au \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox