* [PATCH 0/3] virtio-net: inline header support @ 2012-09-28 9:26 Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 1/3] virtio: add API to query ring capacity Michael S. Tsirkin ` (4 more replies) 0 siblings, 5 replies; 18+ messages in thread From: Michael S. Tsirkin @ 2012-09-28 9:26 UTC (permalink / raw) To: Thomas Lendacky Cc: kvm, netdev, linux-kernel, virtualization, avi, Sasha Levin Thinking about Sasha's patches, we can reduce ring usage for virtio net small packets dramatically if we put virtio net header inline with the data. This can be done for free in case guest net stack allocated extra head room for the packet, and I don't see why would this have any downsides. Even though with my recent patches qemu no longer requires header to be the first s/g element, we need a new feature bit to detect this. A trivial qemu patch will be sent separately. We could get rid of an extra s/g for big packets too, but since in practice everyone enables mergeable buffers, I don't see much of a point. Rusty, if you decide to pick this up I'll send a (rather trivial) spec patch shortly afterwards, but holidays are beginning here. Considering how simple the guest patch is, I hope it can make it in 3.7? Also note that patch 1 and 2 are IMO a good idea without patch 3. If you decide to defer patch 3 pls consider 1/2 separately. Before: [root@virtlab203 qemu]# ssh robin ./netperf/bin/netperf -t TCP_RR -H 11.0.0.4 TCP REQUEST/RESPONSE TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 11.0.0.4 (11.0.0.4) port 0 AF_INET : demo Local /Remote Socket Size Request Resp. Elapsed Trans. Send Recv Size Size Time Rate bytes Bytes bytes bytes secs. per sec 16384 87380 1 1 10.00 2992.88 16384 87380 After: [root@virtlab203 qemu]# ssh robin ./netperf/bin/netperf -t TCP_RR -H 11.0.0.4 TCP REQUEST/RESPONSE TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 11.0.0.4 (11.0.0.4) port 0 AF_INET : demo Local /Remote Socket Size Request Resp. Elapsed Trans. Send Recv Size Size Time Rate bytes Bytes bytes bytes secs. per sec 16384 87380 1 1 10.00 3195.57 16384 87380 Michael S. Tsirkin (3): virtio: add API to query ring capacity virtio-net: correct capacity math on ring full virtio-net: put virtio net header inline with data drivers/net/virtio_net.c | 57 +++++++++++++++++++++++++++++++------------- drivers/virtio/virtio_ring.c | 19 +++++++++++++++ include/linux/virtio.h | 2 ++ include/linux/virtio_net.h | 5 +++- 4 files changed, 66 insertions(+), 17 deletions(-) -- MST ^ permalink raw reply [flat|nested] 18+ messages in thread
* [PATCH 1/3] virtio: add API to query ring capacity 2012-09-28 9:26 [PATCH 0/3] virtio-net: inline header support Michael S. Tsirkin @ 2012-09-28 9:26 ` Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 2/3] virtio-net: correct capacity math on ring full Michael S. Tsirkin ` (3 subsequent siblings) 4 siblings, 0 replies; 18+ messages in thread From: Michael S. Tsirkin @ 2012-09-28 9:26 UTC (permalink / raw) To: Thomas Lendacky Cc: kvm, netdev, linux-kernel, virtualization, avi, Sasha Levin It's sometimes necessary to query ring capacity after dequeueing a buffer. Add an API for this. Signed-off-by: Michael S. Tsirkin <mst@redhat.com> --- drivers/virtio/virtio_ring.c | 19 +++++++++++++++++++ include/linux/virtio.h | 2 ++ 2 files changed, 21 insertions(+) diff --git a/drivers/virtio/virtio_ring.c b/drivers/virtio/virtio_ring.c index 5aa43c3..ee3d80b 100644 --- a/drivers/virtio/virtio_ring.c +++ b/drivers/virtio/virtio_ring.c @@ -715,4 +715,23 @@ unsigned int virtqueue_get_vring_size(struct virtqueue *_vq) } EXPORT_SYMBOL_GPL(virtqueue_get_vring_size); +/** + * virtqueue_get_capacity - query available ring capacity + * @vq: the struct virtqueue we're talking about. + * + * Caller must ensure we don't call this with other virtqueue operations + * at the same time (except where noted), otherwise result is unreliable. + * + * Returns remaining capacity of queue. + * Note that it only really makes sense to treat all + * return values as "available": indirect buffers mean that + * we can put an entire sg[] array inside a single queue entry. + */ +unsigned int virtqueue_get_capacity(struct virtqueue *_vq) +{ + struct vring_virtqueue *vq = to_vvq(_vq); + return vq->num_free; +} +EXPORT_SYMBOL_GPL(virtqueue_get_capacity); + MODULE_LICENSE("GPL"); diff --git a/include/linux/virtio.h b/include/linux/virtio.h index a1ba8bb..fab61e8 100644 --- a/include/linux/virtio.h +++ b/include/linux/virtio.h @@ -50,6 +50,8 @@ void *virtqueue_detach_unused_buf(struct virtqueue *vq); unsigned int virtqueue_get_vring_size(struct virtqueue *vq); +unsigned int virtqueue_get_capacity(struct virtqueue *vq); + /** * virtio_device - representation of a device using virtio * @index: unique position on the virtio bus -- MST ^ permalink raw reply related [flat|nested] 18+ messages in thread
* [PATCH 2/3] virtio-net: correct capacity math on ring full 2012-09-28 9:26 [PATCH 0/3] virtio-net: inline header support Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 1/3] virtio: add API to query ring capacity Michael S. Tsirkin @ 2012-09-28 9:26 ` Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 3/3] virtio-net: put virtio net header inline with data Michael S. Tsirkin ` (2 subsequent siblings) 4 siblings, 0 replies; 18+ messages in thread From: Michael S. Tsirkin @ 2012-09-28 9:26 UTC (permalink / raw) To: Thomas Lendacky Cc: kvm, netdev, linux-kernel, virtualization, avi, Sasha Levin Capacity math on ring full is wrong: we are looking at num_sg but that might be optimistic because of indirect buffer use. The implementation also penalizes fast path with extra memory accesses for the benefit of ring full condition handling which is slow path. It's easy to query ring capacity so let's do just that. This change also makes it easier to move vnet header for tx around as follow-up patch does. Signed-off-by: Michael S. Tsirkin <mst@redhat.com> --- drivers/net/virtio_net.c | 15 +++++++-------- 1 file changed, 7 insertions(+), 8 deletions(-) diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c index 83d2b0c..316f1be 100644 --- a/drivers/net/virtio_net.c +++ b/drivers/net/virtio_net.c @@ -95,7 +95,6 @@ struct skb_vnet_hdr { struct virtio_net_hdr hdr; struct virtio_net_hdr_mrg_rxbuf mhdr; }; - unsigned int num_sg; }; struct padded_vnet_hdr { @@ -557,10 +556,10 @@ again: return received; } -static unsigned int free_old_xmit_skbs(struct virtnet_info *vi) +static void free_old_xmit_skbs(struct virtnet_info *vi) { struct sk_buff *skb; - unsigned int len, tot_sgs = 0; + unsigned int len; struct virtnet_stats *stats = this_cpu_ptr(vi->stats); while ((skb = virtqueue_get_buf(vi->svq, &len)) != NULL) { @@ -571,16 +570,15 @@ static unsigned int free_old_xmit_skbs(struct virtnet_info *vi) stats->tx_packets++; u64_stats_update_end(&stats->tx_syncp); - tot_sgs += skb_vnet_hdr(skb)->num_sg; dev_kfree_skb_any(skb); } - return tot_sgs; } static int xmit_skb(struct virtnet_info *vi, struct sk_buff *skb) { struct skb_vnet_hdr *hdr = skb_vnet_hdr(skb); const unsigned char *dest = ((struct ethhdr *)skb->data)->h_dest; + unsigned num_sg; pr_debug("%s: xmit %p %pM\n", vi->dev->name, skb, dest); @@ -619,8 +617,8 @@ static int xmit_skb(struct virtnet_info *vi, struct sk_buff *skb) else sg_set_buf(vi->tx_sg, &hdr->hdr, sizeof hdr->hdr); - hdr->num_sg = skb_to_sgvec(skb, vi->tx_sg + 1, 0, skb->len) + 1; - return virtqueue_add_buf(vi->svq, vi->tx_sg, hdr->num_sg, + num_sg = skb_to_sgvec(skb, vi->tx_sg + 1, 0, skb->len) + 1; + return virtqueue_add_buf(vi->svq, vi->tx_sg, num_sg, 0, skb, GFP_ATOMIC); } @@ -664,7 +662,8 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev) netif_stop_queue(dev); if (unlikely(!virtqueue_enable_cb_delayed(vi->svq))) { /* More just got used, free them then recheck. */ - capacity += free_old_xmit_skbs(vi); + free_old_xmit_skbs(vi); + capacity = virtqueue_get_capacity(vi->svq); if (capacity >= 2+MAX_SKB_FRAGS) { netif_start_queue(dev); virtqueue_disable_cb(vi->svq); -- MST ^ permalink raw reply related [flat|nested] 18+ messages in thread
* [PATCH 3/3] virtio-net: put virtio net header inline with data 2012-09-28 9:26 [PATCH 0/3] virtio-net: inline header support Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 1/3] virtio: add API to query ring capacity Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 2/3] virtio-net: correct capacity math on ring full Michael S. Tsirkin @ 2012-09-28 9:26 ` Michael S. Tsirkin 2012-10-03 6:44 ` [PATCH 0/3] virtio-net: inline header support Rusty Russell [not found] ` <87vces2gxq.fsf@rustcorp.com.au> 4 siblings, 0 replies; 18+ messages in thread From: Michael S. Tsirkin @ 2012-09-28 9:26 UTC (permalink / raw) To: Thomas Lendacky Cc: kvm, netdev, linux-kernel, virtualization, avi, Sasha Levin For small packets we can simplify xmit processing by linearizing buffers with the header: most packets seem to have enough head room we can use for this purpose. Since existing hypervisors require that header is the first s/g element, we need a feature bit for this. Signed-off-by: Michael S. Tsirkin <mst@redhat.com> --- drivers/net/virtio_net.c | 44 +++++++++++++++++++++++++++++++++++--------- include/linux/virtio_net.h | 5 ++++- 2 files changed, 39 insertions(+), 10 deletions(-) diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c index 316f1be..6e6e53e 100644 --- a/drivers/net/virtio_net.c +++ b/drivers/net/virtio_net.c @@ -67,6 +67,9 @@ struct virtnet_info { /* Host will merge rx buffers for big packets (shake it! shake it!) */ bool mergeable_rx_bufs; + /* Host can handle any s/g split between our header and packet data */ + bool any_header_sg; + /* enable config space updates */ bool config_enable; @@ -576,11 +579,28 @@ static void free_old_xmit_skbs(struct virtnet_info *vi) static int xmit_skb(struct virtnet_info *vi, struct sk_buff *skb) { - struct skb_vnet_hdr *hdr = skb_vnet_hdr(skb); + struct skb_vnet_hdr *hdr; const unsigned char *dest = ((struct ethhdr *)skb->data)->h_dest; unsigned num_sg; + unsigned hdr_len; + bool can_push; + pr_debug("%s: xmit %p %pM\n", vi->dev->name, skb, dest); + if (vi->mergeable_rx_bufs) + hdr_len = sizeof hdr->mhdr; + else + hdr_len = sizeof hdr->hdr; + + can_push = vi->any_header_sg && + !((unsigned long)skb->data & (__alignof__(*hdr) - 1)) && + !skb_header_cloned(skb) && skb_headroom(skb) >= hdr_len; + /* Even if we can, don't push here yet as this would skew + * csum_start offset below. */ + if (can_push) + hdr = (struct skb_vnet_hdr *)(skb->data - hdr_len); + else + hdr = skb_vnet_hdr(skb); if (skb->ip_summed == CHECKSUM_PARTIAL) { hdr->hdr.flags = VIRTIO_NET_HDR_F_NEEDS_CSUM; @@ -609,15 +629,18 @@ static int xmit_skb(struct virtnet_info *vi, struct sk_buff *skb) hdr->hdr.gso_size = hdr->hdr.hdr_len = 0; } - hdr->mhdr.num_buffers = 0; - - /* Encode metadata header at front. */ if (vi->mergeable_rx_bufs) - sg_set_buf(vi->tx_sg, &hdr->mhdr, sizeof hdr->mhdr); - else - sg_set_buf(vi->tx_sg, &hdr->hdr, sizeof hdr->hdr); + hdr->mhdr.num_buffers = 0; - num_sg = skb_to_sgvec(skb, vi->tx_sg + 1, 0, skb->len) + 1; + if (can_push) { + __skb_push(skb, hdr_len); + num_sg = skb_to_sgvec(skb, vi->tx_sg, 0, skb->len); + /* Pull header back to avoid skew in tx bytes calculations. */ + __skb_pull(skb, hdr_len); + } else { + sg_set_buf(vi->tx_sg, hdr, hdr_len); + num_sg = skb_to_sgvec(skb, vi->tx_sg + 1, 0, skb->len) + 1; + } return virtqueue_add_buf(vi->svq, vi->tx_sg, num_sg, 0, skb, GFP_ATOMIC); } @@ -1128,6 +1151,9 @@ static int virtnet_probe(struct virtio_device *vdev) if (virtio_has_feature(vdev, VIRTIO_NET_F_MRG_RXBUF)) vi->mergeable_rx_bufs = true; + if (virtio_has_feature(vdev, VIRTIO_NET_F_ANY_HEADER_SG)) + vi->any_header_sg = true; + err = init_vqs(vi); if (err) goto free_stats; @@ -1286,7 +1312,7 @@ static unsigned int features[] = { VIRTIO_NET_F_GUEST_ECN, VIRTIO_NET_F_GUEST_UFO, VIRTIO_NET_F_MRG_RXBUF, VIRTIO_NET_F_STATUS, VIRTIO_NET_F_CTRL_VQ, VIRTIO_NET_F_CTRL_RX, VIRTIO_NET_F_CTRL_VLAN, - VIRTIO_NET_F_GUEST_ANNOUNCE, + VIRTIO_NET_F_GUEST_ANNOUNCE, VIRTIO_NET_F_ANY_HEADER_SG }; static struct virtio_driver virtio_net_driver = { diff --git a/include/linux/virtio_net.h b/include/linux/virtio_net.h index 2470f54..16a577b 100644 --- a/include/linux/virtio_net.h +++ b/include/linux/virtio_net.h @@ -51,6 +51,7 @@ #define VIRTIO_NET_F_CTRL_RX_EXTRA 20 /* Extra RX mode control support */ #define VIRTIO_NET_F_GUEST_ANNOUNCE 21 /* Guest can announce device on the * network */ +#define VIRTIO_NET_F_ANY_HEADER_SG 22 /* Host can handle any header s/g */ #define VIRTIO_NET_S_LINK_UP 1 /* Link is up */ #define VIRTIO_NET_S_ANNOUNCE 2 /* Announcement is needed */ @@ -62,7 +63,9 @@ struct virtio_net_config { __u16 status; } __attribute__((packed)); -/* This is the first element of the scatter-gather list. If you don't +/* This header comes first in the scatter-gather list. + * If VIRTIO_NET_F_ANY_HEADER_SG is not negotiated, it must + * be the first element of the scatter-gather list. If you don't * specify GSO or CSUM features, you can simply ignore the header. */ struct virtio_net_hdr { #define VIRTIO_NET_HDR_F_NEEDS_CSUM 1 // Use csum_start, csum_offset -- MST ^ permalink raw reply related [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-09-28 9:26 [PATCH 0/3] virtio-net: inline header support Michael S. Tsirkin ` (2 preceding siblings ...) 2012-09-28 9:26 ` [PATCH 3/3] virtio-net: put virtio net header inline with data Michael S. Tsirkin @ 2012-10-03 6:44 ` Rusty Russell 2012-10-03 10:53 ` Paolo Bonzini [not found] ` <87vces2gxq.fsf@rustcorp.com.au> 4 siblings, 1 reply; 18+ messages in thread From: Rusty Russell @ 2012-10-03 6:44 UTC (permalink / raw) To: Michael S. Tsirkin, Thomas Lendacky Cc: kvm, netdev, linux-kernel, virtualization, avi, Sasha Levin "Michael S. Tsirkin" <mst@redhat.com> writes: > Thinking about Sasha's patches, we can reduce ring usage > for virtio net small packets dramatically if we put > virtio net header inline with the data. > This can be done for free in case guest net stack allocated > extra head room for the packet, and I don't see > why would this have any downsides. I've been wanting to do this for the longest time... but... > Even though with my recent patches qemu > no longer requires header to be the first s/g element, > we need a new feature bit to detect this. > A trivial qemu patch will be sent separately. There's a reason I haven't done this. I really, really dislike "my implemention isn't broken" feature bits. We could have an infinite number of them, for each bug in each device. So my plan was to tie this assumption to the new PCI layout. And have a stress-testing patch like the one below in the kernel (see my virtio-wip branch for stuff like this). Turn it on at boot with "virtio_ring.torture" on the kernel commandline. BTW, I've fixed lguest, but my kvm here (Ubuntu precise, kvm-qemu 1.0) is too old. Building the latest git now... Cheers, Rusty. Subject: virtio: CONFIG_VIRTIO_DEVICE_TORTURE Virtio devices are not supposed to depend on the framing of the scatter-gather lists, but various implementations did. Safeguard this in future by adding an option to deliberately create perverse descriptors. Signed-off-by: Rusty Russell <rusty@rustcorp.com.au> diff --git a/drivers/virtio/Kconfig b/drivers/virtio/Kconfig index 8d5bddb..930a4ea 100644 --- a/drivers/virtio/Kconfig +++ b/drivers/virtio/Kconfig @@ -5,6 +5,15 @@ config VIRTIO bus, such as CONFIG_VIRTIO_PCI, CONFIG_VIRTIO_MMIO, CONFIG_LGUEST, CONFIG_RPMSG or CONFIG_S390_GUEST. +config VIRTIO_DEVICE_TORTURE + bool "Virtio device torture tests" + depends on VIRTIO && DEBUG_KERNEL + help + This makes the virtio_ring implementation creatively change + the format of requests to make sure that devices are + properly implemented. This will make your virtual machine + slow *and* unreliable! Say N. + menu "Virtio drivers" config VIRTIO_PCI diff --git a/drivers/virtio/virtio_ring.c b/drivers/virtio/virtio_ring.c index e639584..8893753 100644 --- a/drivers/virtio/virtio_ring.c +++ b/drivers/virtio/virtio_ring.c @@ -124,6 +124,149 @@ struct vring_virtqueue #define to_vvq(_vq) container_of(_vq, struct vring_virtqueue, vq) +#ifdef CONFIG_VIRTIO_DEVICE_TORTURE +static bool torture; +module_param(torture, bool, 0644); + +struct torture { + unsigned int orig_out, orig_in; + void *orig_data; + struct scatterlist sg[4]; + struct scatterlist orig_sg[]; +}; + +static size_t tot_len(struct scatterlist sg[], unsigned num) +{ + size_t len, i; + + for (len = 0, i = 0; i < num; i++) + len += sg[i].length; + + return len; +} + +static void copy_sg_data(const struct scatterlist *dst, unsigned dnum, + const struct scatterlist *src, unsigned snum) +{ + unsigned len; + struct scatterlist s, d; + + s = *src; + d = *dst; + + while (snum && dnum) { + len = min(s.length, d.length); + memcpy(sg_virt(&d), sg_virt(&s), len); + d.offset += len; + d.length -= len; + s.offset += len; + s.length -= len; + if (!s.length) { + BUG_ON(snum == 0); + src++; + snum--; + s = *src; + } + if (!d.length) { + BUG_ON(dnum == 0); + dst++; + dnum--; + d = *dst; + } + } +} + +static bool torture_replace(struct scatterlist **sg, + unsigned int *out, + unsigned int *in, + void **data, + gfp_t gfp) +{ + static size_t seed; + struct torture *t; + size_t outlen, inlen, ourseed, len1; + void *buf; + + if (!torture) + return true; + + outlen = tot_len(*sg, *out); + inlen = tot_len(*sg + *out, *in); + + /* This will break horribly on large block requests. */ + t = kmalloc(sizeof(*t) + (*out + *in) * sizeof(t->orig_sg[1]) + + outlen + 1 + inlen + 1, gfp); + if (!t) + return false; + + sg_init_table(t->sg, 4); + buf = &t->orig_sg[*out + *in]; + + memcpy(t->orig_sg, *sg, sizeof(**sg) * (*out + *in)); + t->orig_out = *out; + t->orig_in = *in; + t->orig_data = *data; + *data = t; + + ourseed = ACCESS_ONCE(seed); + seed++; + + *sg = t->sg; + if (outlen) { + /* Split outbuf into two parts, one byte apart. */ + *out = 2; + len1 = ourseed % (outlen + 1); + sg_set_buf(&t->sg[0], buf, len1); + buf += len1 + 1; + sg_set_buf(&t->sg[1], buf, outlen - len1); + buf += outlen - len1; + copy_sg_data(t->sg, *out, t->orig_sg, t->orig_out); + } + + if (inlen) { + /* Split inbuf into two parts, one byte apart. */ + *in = 2; + len1 = ourseed % (inlen + 1); + sg_set_buf(&t->sg[*out], buf, len1); + buf += len1 + 1; + sg_set_buf(&t->sg[*out + 1], buf, inlen - len1); + buf += inlen - len1; + } + return true; +} + +static void *torture_done(struct torture *t) +{ + void *data; + + if (!torture) + return t; + + if (t->orig_in) + copy_sg_data(t->orig_sg + t->orig_out, t->orig_in, + t->sg + (t->orig_out ? 2 : 0), 2); + + data = t->orig_data; + kfree(t); + return data; +} + +#else +static bool torture_replace(struct scatterlist **sg, + unsigned int *out, + unsigned int *in, + void **data, + gfp_t gfp) +{ + return true; +} + +static void *torture_done(void *data) +{ + return data; +} +#endif /* CONFIG_VIRTIO_DEVICE_TORTURE */ + /* Set up an indirect table of descriptors and add it to the queue. */ static int vring_add_indirect(struct vring_virtqueue *vq, struct scatterlist sg[], @@ -213,6 +356,9 @@ int virtqueue_add_buf(struct virtqueue *_vq, BUG_ON(data == NULL); + if (!torture_replace(&sg, &out, &in, &data, gfp)) + return -ENOMEM; + #ifdef DEBUG { ktime_t now = ktime_get(); @@ -246,6 +392,7 @@ int virtqueue_add_buf(struct virtqueue *_vq, if (out) vq->notify(&vq->vq); END_USE(vq); + torture_done(data); return -ENOSPC; } @@ -476,7 +623,7 @@ void *virtqueue_get_buf(struct virtqueue *_vq, unsigned int *len) #endif END_USE(vq); - return ret; + return torture_done(ret); } EXPORT_SYMBOL_GPL(virtqueue_get_buf); ^ permalink raw reply related [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-03 6:44 ` [PATCH 0/3] virtio-net: inline header support Rusty Russell @ 2012-10-03 10:53 ` Paolo Bonzini [not found] ` <87bogj2j1b.fsf@rustcorp.com.au> 0 siblings, 1 reply; 18+ messages in thread From: Paolo Bonzini @ 2012-10-03 10:53 UTC (permalink / raw) To: Rusty Russell Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, Sasha Levin, virtualization, Thomas Lendacky, avi Il 03/10/2012 08:44, Rusty Russell ha scritto: > There's a reason I haven't done this. I really, really dislike "my > implemention isn't broken" feature bits. We could have an infinite > number of them, for each bug in each device. However, this bug affects (almost) all implementations and (almost) all devices. It even makes sense to reserve a transport feature bit for it instead of a device feature bit. Paolo ^ permalink raw reply [flat|nested] 18+ messages in thread
[parent not found: <87bogj2j1b.fsf@rustcorp.com.au>]
[parent not found: <506D3610.7000103@redhat.com>]
[parent not found: <87ipaq1jtt.fsf@rustcorp.com.au>]
[parent not found: <506D8DD5.20904@redhat.com>]
* Re: [PATCH 0/3] virtio-net: inline header support [not found] ` <506D8DD5.20904@redhat.com> @ 2012-10-05 5:43 ` Rusty Russell 2012-10-06 12:54 ` Paolo Bonzini 0 siblings, 1 reply; 18+ messages in thread From: Rusty Russell @ 2012-10-05 5:43 UTC (permalink / raw) To: Paolo Bonzini Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, Sasha Levin, virtualization, Thomas Lendacky, avi Paolo Bonzini <pbonzini@redhat.com> writes: > Il 04/10/2012 14:51, Rusty Russell ha scritto: >> Paolo Bonzini <pbonzini@redhat.com> writes: >> >>> Il 04/10/2012 02:11, Rusty Russell ha scritto: >>>>>> There's a reason I haven't done this. I really, really dislike "my >>>>>> implemention isn't broken" feature bits. We could have an infinite >>>>>> number of them, for each bug in each device. >>>>> >>>>> However, this bug affects (almost) all implementations and (almost) all >>>>> devices. It even makes sense to reserve a transport feature bit for it >>>>> instead of a device feature bit. >>>> >>>> Perhaps, but we have to fix the bugs first! >>> >>> Yes. :) Isn't that what mst's patch does? >>> >>>> As I said, my torture patch broke qemu immediately. Since noone has >>>> leapt onto fixing that, I'll take a look now... >>> >>> I can look at virtio-scsi. >> >> Actually, you can't, see my reply to Anthony... >> >> Message-ID: <87lifm1y1n.fsf@rustcorp.com.au> > > struct virtio_scsi_req_cmd { > // Read-only > u8 lun[8]; > u64 id; > u8 task_attr; > u8 prio; > u8 crn; > char cdb[cdb_size]; > char dataout[]; > // Write-only part > u32 sense_len; > u32 residual; > u16 status_qualifier; > u8 status; > u8 response; > u8 sense[sense_size]; > char datain[]; > }; > > where cdb_size and sense_size come from configuration space. The device > right now expects everything before dataout/datain to be in a single > descriptor, but that's in no way part of the spec. Am I missing > something egregious? Since you wrote it, I hope not :) That's good. But virtio_blk's scsi command is insoluble AFAICT. As I said to Anthony, the best rules are "always" and "never", so I'd really rather not have to grandfather that in. Cheers, Rusty. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-05 5:43 ` Rusty Russell @ 2012-10-06 12:54 ` Paolo Bonzini 2012-10-09 4:59 ` Rusty Russell 0 siblings, 1 reply; 18+ messages in thread From: Paolo Bonzini @ 2012-10-06 12:54 UTC (permalink / raw) To: Rusty Russell Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, virtualization, Sasha Levin, avi, Thomas Lendacky Il 05/10/2012 07:43, Rusty Russell ha scritto: >> > struct virtio_scsi_req_cmd { >> > // Read-only >> > u8 lun[8]; >> > u64 id; >> > u8 task_attr; >> > u8 prio; >> > u8 crn; >> > char cdb[cdb_size]; >> > char dataout[]; >> > // Write-only part >> > u32 sense_len; >> > u32 residual; >> > u16 status_qualifier; >> > u8 status; >> > u8 response; >> > u8 sense[sense_size]; >> > char datain[]; >> > }; >> > >> > where cdb_size and sense_size come from configuration space. The device >> > right now expects everything before dataout/datain to be in a single >> > descriptor, but that's in no way part of the spec. Am I missing >> > something egregious? > Since you wrote it, I hope not :) Yeah, I guess the confusion came from cdb_size and sense_size being in configuration space. > That's good. But virtio_blk's scsi command is insoluble AFAICT. As I > said to Anthony, the best rules are "always" and "never", so I'd really > rather not have to grandfather that in. It is, but we can add a rule that if the (transport) flag VIRTIO_RING_F_ANY_HEADER_SG is set, the cdb field is always 32 bytes in virtio-blk. Paolo ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-06 12:54 ` Paolo Bonzini @ 2012-10-09 4:59 ` Rusty Russell 2012-10-09 7:27 ` Paolo Bonzini 0 siblings, 1 reply; 18+ messages in thread From: Rusty Russell @ 2012-10-09 4:59 UTC (permalink / raw) To: Paolo Bonzini Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, virtualization, Sasha Levin, avi, Thomas Lendacky Paolo Bonzini <pbonzini@redhat.com> writes: > Il 05/10/2012 07:43, Rusty Russell ha scritto: >> That's good. But virtio_blk's scsi command is insoluble AFAICT. As I >> said to Anthony, the best rules are "always" and "never", so I'd really >> rather not have to grandfather that in. > > It is, but we can add a rule that if the (transport) flag > VIRTIO_RING_F_ANY_HEADER_SG is set, the cdb field is always 32 bytes in > virtio-blk. Could we do that? It's the cmd length I'm concerned about; is it always 32 in practice for some reason? Currently qemu does: struct sg_io_hdr hdr; memset(&hdr, 0, sizeof(struct sg_io_hdr)); hdr.interface_id = 'S'; hdr.cmd_len = req->elem.out_sg[1].iov_len; hdr.cmdp = req->elem.out_sg[1].iov_base; hdr.dxfer_len = 0; If it's a command which expects more output data, there's no way to guess where the boundary is between that command and the data. Cheers, Rusty. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-09 4:59 ` Rusty Russell @ 2012-10-09 7:27 ` Paolo Bonzini 2012-10-11 0:03 ` Rusty Russell 0 siblings, 1 reply; 18+ messages in thread From: Paolo Bonzini @ 2012-10-09 7:27 UTC (permalink / raw) To: Rusty Russell Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, virtualization, Sasha Levin, avi, Thomas Lendacky Il 09/10/2012 06:59, Rusty Russell ha scritto: > Paolo Bonzini <pbonzini@redhat.com> writes: >> Il 05/10/2012 07:43, Rusty Russell ha scritto: >>> That's good. But virtio_blk's scsi command is insoluble AFAICT. As I >>> said to Anthony, the best rules are "always" and "never", so I'd really >>> rather not have to grandfather that in. >> >> It is, but we can add a rule that if the (transport) flag >> VIRTIO_RING_F_ANY_HEADER_SG is set, the cdb field is always 32 bytes in >> virtio-blk. > > Could we do that? It's the cmd length I'm concerned about; is it always > 32 in practice for some reason? It is always 32 or less except in very obscure cases that are pretty much confined to iSCSI. We don't care about the obscure cases, and the extra bytes don't hurt. BTW, 32 is the default cdb_size used by virtio-scsi. > Currently qemu does: > > struct sg_io_hdr hdr; > memset(&hdr, 0, sizeof(struct sg_io_hdr)); > hdr.interface_id = 'S'; > hdr.cmd_len = req->elem.out_sg[1].iov_len; > hdr.cmdp = req->elem.out_sg[1].iov_base; > hdr.dxfer_len = 0; > > If it's a command which expects more output data, there's no way to > guess where the boundary is between that command and the data. Yep, so I understood the problem right. Paolo ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-09 7:27 ` Paolo Bonzini @ 2012-10-11 0:03 ` Rusty Russell 2012-10-11 11:04 ` Michael S. Tsirkin 0 siblings, 1 reply; 18+ messages in thread From: Rusty Russell @ 2012-10-11 0:03 UTC (permalink / raw) To: Paolo Bonzini Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, virtualization, Sasha Levin, avi, Thomas Lendacky Paolo Bonzini <pbonzini@redhat.com> writes: > Il 09/10/2012 06:59, Rusty Russell ha scritto: >> Paolo Bonzini <pbonzini@redhat.com> writes: >>> Il 05/10/2012 07:43, Rusty Russell ha scritto: >>>> That's good. But virtio_blk's scsi command is insoluble AFAICT. As I >>>> said to Anthony, the best rules are "always" and "never", so I'd really >>>> rather not have to grandfather that in. >>> >>> It is, but we can add a rule that if the (transport) flag >>> VIRTIO_RING_F_ANY_HEADER_SG is set, the cdb field is always 32 bytes in >>> virtio-blk. >> >> Could we do that? It's the cmd length I'm concerned about; is it always >> 32 in practice for some reason? > > It is always 32 or less except in very obscure cases that are pretty > much confined to iSCSI. We don't care about the obscure cases, and the > extra bytes don't hurt. > > BTW, 32 is the default cdb_size used by virtio-scsi. > >> Currently qemu does: >> >> struct sg_io_hdr hdr; >> memset(&hdr, 0, sizeof(struct sg_io_hdr)); >> hdr.interface_id = 'S'; >> hdr.cmd_len = req->elem.out_sg[1].iov_len; >> hdr.cmdp = req->elem.out_sg[1].iov_base; >> hdr.dxfer_len = 0; >> >> If it's a command which expects more output data, there's no way to >> guess where the boundary is between that command and the data. > > Yep, so I understood the problem right. OK. Well, Anthony wants qemu to be robust in this regard, so I am tempted to rework all the qemu drivers to handle arbitrary layouts. They could use a good audit anyway. This would become a glaring exception, but I'm tempted to fix it to 32 bytes at the same time as we get the new pci layout (ie. for the virtio 1.0 spec). The Linux driver would carefully be backwards compatible, of course, and the spec would document why. Cheers, Rusty. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-11 0:03 ` Rusty Russell @ 2012-10-11 11:04 ` Michael S. Tsirkin 2012-10-11 22:37 ` Rusty Russell 0 siblings, 1 reply; 18+ messages in thread From: Michael S. Tsirkin @ 2012-10-11 11:04 UTC (permalink / raw) To: Rusty Russell Cc: kvm, netdev, linux-kernel, virtualization, Sasha Levin, Paolo Bonzini, avi, Thomas Lendacky On Thu, Oct 11, 2012 at 10:33:31AM +1030, Rusty Russell wrote: > Paolo Bonzini <pbonzini@redhat.com> writes: > > Il 09/10/2012 06:59, Rusty Russell ha scritto: > >> Paolo Bonzini <pbonzini@redhat.com> writes: > >>> Il 05/10/2012 07:43, Rusty Russell ha scritto: > >>>> That's good. But virtio_blk's scsi command is insoluble AFAICT. As I > >>>> said to Anthony, the best rules are "always" and "never", so I'd really > >>>> rather not have to grandfather that in. > >>> > >>> It is, but we can add a rule that if the (transport) flag > >>> VIRTIO_RING_F_ANY_HEADER_SG is set, the cdb field is always 32 bytes in > >>> virtio-blk. > >> > >> Could we do that? It's the cmd length I'm concerned about; is it always > >> 32 in practice for some reason? > > > > It is always 32 or less except in very obscure cases that are pretty > > much confined to iSCSI. We don't care about the obscure cases, and the > > extra bytes don't hurt. > > > > BTW, 32 is the default cdb_size used by virtio-scsi. > > > >> Currently qemu does: > >> > >> struct sg_io_hdr hdr; > >> memset(&hdr, 0, sizeof(struct sg_io_hdr)); > >> hdr.interface_id = 'S'; > >> hdr.cmd_len = req->elem.out_sg[1].iov_len; > >> hdr.cmdp = req->elem.out_sg[1].iov_base; > >> hdr.dxfer_len = 0; > >> > >> If it's a command which expects more output data, there's no way to > >> guess where the boundary is between that command and the data. > > > > Yep, so I understood the problem right. > > OK. Well, Anthony wants qemu to be robust in this regard, so I am > tempted to rework all the qemu drivers to handle arbitrary layouts. > They could use a good audit anyway. I agree here. Still trying to understand whether we can agree to use a feature bit for this, or not. > This would become a glaring exception, but I'm tempted to fix it to 32 > bytes at the same time as we get the new pci layout (ie. for the virtio > 1.0 spec). But this isn't a virtio-pci only issue, is it? qemu has s390 bus with same limmitation. How can we tie it to pci layout? Maybe what you mean is to use a transport feature for this and tie *that* to new layout in case of pci? > The Linux driver would carefully be backwards compatible, of > course, and the spec would document why. > > Cheers, > Rusty. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-11 11:04 ` Michael S. Tsirkin @ 2012-10-11 22:37 ` Rusty Russell 2012-10-12 7:38 ` Paolo Bonzini 2012-10-12 11:52 ` Cornelia Huck 0 siblings, 2 replies; 18+ messages in thread From: Rusty Russell @ 2012-10-11 22:37 UTC (permalink / raw) To: Michael S. Tsirkin Cc: kvm, netdev, linux-kernel, virtualization, Sasha Levin, Paolo Bonzini, avi, Thomas Lendacky "Michael S. Tsirkin" <mst@redhat.com> writes: > On Thu, Oct 11, 2012 at 10:33:31AM +1030, Rusty Russell wrote: >> OK. Well, Anthony wants qemu to be robust in this regard, so I am >> tempted to rework all the qemu drivers to handle arbitrary layouts. >> They could use a good audit anyway. > > I agree here. Still trying to understand whether we can agree to use > a feature bit for this, or not. I'd *like* to imply it by the new PCI layout, but if it doesn't work we'll add a new feature bit. I'm resisting a feature bit, since it constrains future implementations which could otherwise assume it. >> This would become a glaring exception, but I'm tempted to fix it to 32 >> bytes at the same time as we get the new pci layout (ie. for the virtio >> 1.0 spec). > > But this isn't a virtio-pci only issue, is it? > qemu has s390 bus with same limmitation. > How can we tie it to pci layout? They can use a transport feature if they need to, of course. But perhaps the timing with ccw will coincide with the fix, in which they don't need to, but it might be a bit late. Cornelia? Cheers, Rusty. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-11 22:37 ` Rusty Russell @ 2012-10-12 7:38 ` Paolo Bonzini 2012-10-12 11:52 ` Cornelia Huck 1 sibling, 0 replies; 18+ messages in thread From: Paolo Bonzini @ 2012-10-12 7:38 UTC (permalink / raw) To: Rusty Russell Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, virtualization, Sasha Levin, avi, Thomas Lendacky Il 12/10/2012 00:37, Rusty Russell ha scritto: > "Michael S. Tsirkin" <mst@redhat.com> writes: >> On Thu, Oct 11, 2012 at 10:33:31AM +1030, Rusty Russell wrote: >>> OK. Well, Anthony wants qemu to be robust in this regard, so I am >>> tempted to rework all the qemu drivers to handle arbitrary layouts. >>> They could use a good audit anyway. >> >> I agree here. Still trying to understand whether we can agree to use >> a feature bit for this, or not. > > I'd *like* to imply it by the new PCI layout, but if it doesn't work > we'll add a new feature bit. > > I'm resisting a feature bit, since it constrains future implementations > which could otherwise assume it. Future implementations may certainly refuse to start if the feature is not there. Whether it's a good idea or not, well, that depends on how much future they are. Paolo >>> This would become a glaring exception, but I'm tempted to fix it to 32 >>> bytes at the same time as we get the new pci layout (ie. for the virtio >>> 1.0 spec). >> >> But this isn't a virtio-pci only issue, is it? >> qemu has s390 bus with same limmitation. >> How can we tie it to pci layout? > > They can use a transport feature if they need to, of course. But > perhaps the timing with ccw will coincide with the fix, in which they > don't need to, but it might be a bit late. > > Cornelia? > > Cheers, > Rusty. > ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support 2012-10-11 22:37 ` Rusty Russell 2012-10-12 7:38 ` Paolo Bonzini @ 2012-10-12 11:52 ` Cornelia Huck 1 sibling, 0 replies; 18+ messages in thread From: Cornelia Huck @ 2012-10-12 11:52 UTC (permalink / raw) To: Rusty Russell Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, virtualization, Sasha Levin, Paolo Bonzini, avi, Thomas Lendacky On Fri, 12 Oct 2012 09:07:46 +1030 Rusty Russell <rusty@rustcorp.com.au> wrote: > "Michael S. Tsirkin" <mst@redhat.com> writes: > > On Thu, Oct 11, 2012 at 10:33:31AM +1030, Rusty Russell wrote: > >> OK. Well, Anthony wants qemu to be robust in this regard, so I am > >> tempted to rework all the qemu drivers to handle arbitrary layouts. > >> They could use a good audit anyway. > > > > I agree here. Still trying to understand whether we can agree to use > > a feature bit for this, or not. > > I'd *like* to imply it by the new PCI layout, but if it doesn't work > we'll add a new feature bit. > > I'm resisting a feature bit, since it constrains future implementations > which could otherwise assume it. > > >> This would become a glaring exception, but I'm tempted to fix it to 32 > >> bytes at the same time as we get the new pci layout (ie. for the virtio > >> 1.0 spec). > > > > But this isn't a virtio-pci only issue, is it? > > qemu has s390 bus with same limmitation. > > How can we tie it to pci layout? > > They can use a transport feature if they need to, of course. But > perhaps the timing with ccw will coincide with the fix, in which they > don't need to, but it might be a bit late. > > Cornelia? My virtio-ccw host code is still going through a bit of rework, so it might well go in after the fix. There's also the existing (non-spec'ed) s390-virtio transport. While it will likely be deprecated sometime in the future, it should probably get a feature bit for consistency's sake. > > Cheers, > Rusty. > ^ permalink raw reply [flat|nested] 18+ messages in thread
[parent not found: <87vces2gxq.fsf@rustcorp.com.au>]
* Re: [PATCH 0/3] virtio-net: inline header support [not found] ` <87vces2gxq.fsf@rustcorp.com.au> @ 2012-10-08 20:41 ` Michael S. Tsirkin [not found] ` <87mx033u74.fsf@codemonkey.ws> 1 sibling, 0 replies; 18+ messages in thread From: Michael S. Tsirkin @ 2012-10-08 20:41 UTC (permalink / raw) To: Rusty Russell Cc: kvm, netdev, linux-kernel, Sasha Levin, virtualization, avi, Thomas Lendacky On Wed, Oct 03, 2012 at 04:14:17PM +0930, Rusty Russell wrote: > "Michael S. Tsirkin" <mst@redhat.com> writes: > > > Thinking about Sasha's patches, we can reduce ring usage > > for virtio net small packets dramatically if we put > > virtio net header inline with the data. > > This can be done for free in case guest net stack allocated > > extra head room for the packet, and I don't see > > why would this have any downsides. > > I've been wanting to do this for the longest time... but... > > > Even though with my recent patches qemu > > no longer requires header to be the first s/g element, > > we need a new feature bit to detect this. > > A trivial qemu patch will be sent separately. > > There's a reason I haven't done this. I really, really dislike "my > implemention isn't broken" feature bits. We could have an infinite > number of them, for each bug in each device. > > So my plan was to tie this assumption to the new PCI layout. I don't object but old qemu has this limitation for s390 as well, and that's not using PCI, right? So how do we detect new hypervisor there? -- MST ^ permalink raw reply [flat|nested] 18+ messages in thread
[parent not found: <87mx033u74.fsf@codemonkey.ws>]
[parent not found: <87391u3o67.fsf@rustcorp.com.au>]
[parent not found: <87wqz6kgg4.fsf@codemonkey.ws>]
[parent not found: <87lifm1y1n.fsf@rustcorp.com.au>]
* Re: [PATCH 0/3] virtio-net: inline header support [not found] ` <87lifm1y1n.fsf@rustcorp.com.au> @ 2012-10-05 7:47 ` Paolo Bonzini 0 siblings, 0 replies; 18+ messages in thread From: Paolo Bonzini @ 2012-10-05 7:47 UTC (permalink / raw) To: Rusty Russell Cc: kvm, Michael S. Tsirkin, netdev, linux-kernel, virtualization, Sasha Levin, Anthony Liguori, avi, Thomas Lendacky Il 04/10/2012 09:44, Rusty Russell ha scritto: > -In particular, no implementation should use the descriptor > -boundaries to determine the size of any header in a request.[footnote: > -The current qemu device implementations mistakenly insist that > -the first descriptor cover the header in these cases exactly, so > -a cautious driver should arrange it so. > +[footnote: > +It was previously asserted that framing should be independent of > +message contents, yet invariably drivers layed out messages in > +reliable ways and devices assumed it. In addition, the > +specifications for virtio_blk and virtio_scsi require intuiting > +field lengths from frame boundaries. > ] Not true for virtio_scsi... Paolo ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [PATCH 0/3] virtio-net: inline header support [not found] ` <87391u3o67.fsf@rustcorp.com.au> [not found] ` <87wqz6kgg4.fsf@codemonkey.ws> @ 2012-10-08 21:31 ` Michael S. Tsirkin 1 sibling, 0 replies; 18+ messages in thread From: Michael S. Tsirkin @ 2012-10-08 21:31 UTC (permalink / raw) To: Rusty Russell Cc: kvm, netdev, linux-kernel, Sasha Levin, virtualization, avi, Anthony Liguori, Thomas Lendacky On Thu, Oct 04, 2012 at 01:04:56PM +0930, Rusty Russell wrote: > Anthony Liguori <anthony@codemonkey.ws> writes: > > Rusty Russell <rusty@rustcorp.com.au> writes: > > > >> "Michael S. Tsirkin" <mst@redhat.com> writes: > >> > >>> Thinking about Sasha's patches, we can reduce ring usage > >>> for virtio net small packets dramatically if we put > >>> virtio net header inline with the data. > >>> This can be done for free in case guest net stack allocated > >>> extra head room for the packet, and I don't see > >>> why would this have any downsides. > >> > >> I've been wanting to do this for the longest time... but... > >> > >>> Even though with my recent patches qemu > >>> no longer requires header to be the first s/g element, > >>> we need a new feature bit to detect this. > >>> A trivial qemu patch will be sent separately. > >> > >> There's a reason I haven't done this. I really, really dislike "my > >> implemention isn't broken" feature bits. We could have an infinite > >> number of them, for each bug in each device. > > > > This is a bug in the specification. > > > > The QEMU implementation pre-dates the specification. All of the actual > > implementations of virtio relied on the semantics of s/g elements and > > still do. > > lguest fix is pending in my queue. lkvm and qemu are broken; lkvm isn't > ever going to be merged, so I'm not sure what its status is? But I'm > determined to fix qemu, and hence my torture patch to make sure this > doesn't creep in again. If you look at my patch you'll notice there's also a comment in virtio_net.h that seems to be broken in this respect: /* This is the first element of the scatter-gather list. If you don't * specify GSO or CSUM features, you can simply ignore the header. */ There is a similar comment in virtio-blk. ^ permalink raw reply [flat|nested] 18+ messages in thread
end of thread, other threads:[~2012-10-12 11:52 UTC | newest] Thread overview: 18+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2012-09-28 9:26 [PATCH 0/3] virtio-net: inline header support Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 1/3] virtio: add API to query ring capacity Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 2/3] virtio-net: correct capacity math on ring full Michael S. Tsirkin 2012-09-28 9:26 ` [PATCH 3/3] virtio-net: put virtio net header inline with data Michael S. Tsirkin 2012-10-03 6:44 ` [PATCH 0/3] virtio-net: inline header support Rusty Russell 2012-10-03 10:53 ` Paolo Bonzini [not found] ` <87bogj2j1b.fsf@rustcorp.com.au> [not found] ` <506D3610.7000103@redhat.com> [not found] ` <87ipaq1jtt.fsf@rustcorp.com.au> [not found] ` <506D8DD5.20904@redhat.com> 2012-10-05 5:43 ` Rusty Russell 2012-10-06 12:54 ` Paolo Bonzini 2012-10-09 4:59 ` Rusty Russell 2012-10-09 7:27 ` Paolo Bonzini 2012-10-11 0:03 ` Rusty Russell 2012-10-11 11:04 ` Michael S. Tsirkin 2012-10-11 22:37 ` Rusty Russell 2012-10-12 7:38 ` Paolo Bonzini 2012-10-12 11:52 ` Cornelia Huck [not found] ` <87vces2gxq.fsf@rustcorp.com.au> 2012-10-08 20:41 ` Michael S. Tsirkin [not found] ` <87mx033u74.fsf@codemonkey.ws> [not found] ` <87391u3o67.fsf@rustcorp.com.au> [not found] ` <87wqz6kgg4.fsf@codemonkey.ws> [not found] ` <87lifm1y1n.fsf@rustcorp.com.au> 2012-10-05 7:47 ` Paolo Bonzini 2012-10-08 21:31 ` Michael S. Tsirkin
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).