qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Wei Xu <wexu@redhat.com>
To: Jason Wang <jasowang@redhat.com>
Cc: qemu-devel@nongnu.org, maxime.coquelin@redhat.com,
	tiwei.bie@intel.com, jfreiman@redhat.com, mst@redhat.com
Subject: Re: [Qemu-devel] [PATCH v4 08/11] virtio: event suppression support for packed ring
Date: Tue, 19 Feb 2019 18:40:44 +0800	[thread overview]
Message-ID: <20190219104044.GC15343@wei-ubt> (raw)
In-Reply-To: <91bf74af-5e22-c6fc-eae5-16b03b1493a1@redhat.com>

On Tue, Feb 19, 2019 at 03:19:58PM +0800, Jason Wang wrote:
> 
> On 2019/2/14 下午12:26, wexu@redhat.com wrote:
> >From: Wei Xu <wexu@redhat.com>
> >
> >Difference between 'avail_wrap_counter' and 'last_avail_wrap_counter':
> >For Tx(guest transmitting), they are the same after each pop of a desc.
> >
> >For Rx(guest receiving), they are also the same when there are enough
> >descriptors to carry the payload for a packet(e.g. usually 16 descs are
> >needed for a 64k packet in typical iperf tcp connection with tso enabled),
> >however, when the ring is running out of descriptors while there are
> >still a few free ones, e.g. 6 descriptors are available which is not
> >enough to carry an entire packet which needs 16 descriptors, in this
> >case the 'avail_wrap_counter' should be set as the first one pending
> >being handled by guest driver in order to get a notification, and the
> >'last_avail_wrap_counter' should stay unchanged to the head of available
> >descriptors, like below:
> >
> >Mark meaning:
> >     | | -- available
> >     |*| -- used
> >
> >A Snapshot of the queue:
> >                                       last_avail_idx = 253
> >                                       last_avail_wrap_counter = 1
> >                                              |
> >     +---------------------------------------------+
> >  0  | | | |*|*|*|*|*|*|*|*|*|*|*|*|*|*|*|*|*| | | | 255
> >     +---------------------------------------------+
> >            |
> >   shadow_avail_idx = 3
> >   avail_wrap_counter = 0
> 
> 
> Well this might not be the good place to describe the difference between
> shadow_avail_idx and last_avail_idx. And the comments above their definition
> looks good enough?

Sorry, I do not get it, can you elaborate more? 

This is one of the buggy parts of packed ring, it is good to make it clear here.

> 
>     /* Next head to pop */
>     uint16_t last_avail_idx;
> 
>     /* Last avail_idx read from VQ. */
>     uint16_t shadow_avail_idx;
> 

What is the meaning of these comment? Do you mean I should replace put them 
to the comments also? thanks.

> Instead, how or why need event suppress is not mentioned ...

Yes, but presumably this knowledge has been acquired from reading through the
spec, so I skipped this part.

Wei

> 
> 
> 
> >
> >Signed-off-by: Wei Xu <wexu@redhat.com>
> >---
> >  hw/virtio/virtio.c | 137 +++++++++++++++++++++++++++++++++++++++++++++++++----
> >  1 file changed, 128 insertions(+), 9 deletions(-)
> >
> >diff --git a/hw/virtio/virtio.c b/hw/virtio/virtio.c
> >index 7e276b4..8cfc7b6 100644
> >--- a/hw/virtio/virtio.c
> >+++ b/hw/virtio/virtio.c
> >@@ -234,6 +234,34 @@ static void vring_desc_read(VirtIODevice *vdev, VRingDesc *desc,
> >      virtio_tswap16s(vdev, &desc->next);
> >  }
> >+static void vring_packed_event_read(VirtIODevice *vdev,
> >+                            MemoryRegionCache *cache, VRingPackedDescEvent *e)
> >+{
> >+    address_space_read_cached(cache, 0, e, sizeof(*e));
> >+    virtio_tswap16s(vdev, &e->off_wrap);
> >+    virtio_tswap16s(vdev, &e->flags);
> >+}
> >+
> >+static void vring_packed_off_wrap_write(VirtIODevice *vdev,
> >+                            MemoryRegionCache *cache, uint16_t off_wrap)
> >+{
> >+    virtio_tswap16s(vdev, &off_wrap);
> >+    address_space_write_cached(cache, offsetof(VRingPackedDescEvent, off_wrap),
> >+                            &off_wrap, sizeof(off_wrap));
> >+    address_space_cache_invalidate(cache,
> >+                offsetof(VRingPackedDescEvent, off_wrap), sizeof(off_wrap));
> >+}
> >+
> >+static void vring_packed_flags_write(VirtIODevice *vdev,
> >+                            MemoryRegionCache *cache, uint16_t flags)
> >+{
> >+    virtio_tswap16s(vdev, &flags);
> >+    address_space_write_cached(cache, offsetof(VRingPackedDescEvent, flags),
> >+                            &flags, sizeof(flags));
> >+    address_space_cache_invalidate(cache,
> >+                        offsetof(VRingPackedDescEvent, flags), sizeof(flags));
> >+}
> >+
> >  static VRingMemoryRegionCaches *vring_get_region_caches(struct VirtQueue *vq)
> >  {
> >      VRingMemoryRegionCaches *caches = atomic_rcu_read(&vq->vring.caches);
> >@@ -340,14 +368,8 @@ static inline void vring_set_avail_event(VirtQueue *vq, uint16_t val)
> >      address_space_cache_invalidate(&caches->used, pa, sizeof(val));
> >  }
> >-void virtio_queue_set_notification(VirtQueue *vq, int enable)
> >+static void virtio_queue_set_notification_split(VirtQueue *vq, int enable)
> >  {
> >-    vq->notification = enable;
> >-
> >-    if (!vq->vring.desc) {
> >-        return;
> >-    }
> >-
> >      rcu_read_lock();
> >      if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> >          vring_set_avail_event(vq, vring_avail_idx(vq));
> >@@ -363,6 +385,57 @@ void virtio_queue_set_notification(VirtQueue *vq, int enable)
> >      rcu_read_unlock();
> >  }
> >+static void virtio_queue_set_notification_packed(VirtQueue *vq, int enable)
> >+{
> >+    VRingPackedDescEvent e;
> >+    VRingMemoryRegionCaches *caches;
> >+
> >+    rcu_read_lock();
> >+    caches  = vring_get_region_caches(vq);
> >+    vring_packed_event_read(vq->vdev, &caches->used, &e);
> >+
> >+    if (!enable) {
> >+        if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> >+            /* no need to write device area since this is outdated. */
> 
> 
> We should advertise off and wrap in this case as well, otherwise we may get
> notifications earlier than expected.

Is it necessary? Supposing offset & wrap_counter are always set before update
notification flags, it is reliable to skip this for disabling, isn't it?

While the logic here is unclear, I did a concise one like below
which doesn't use the 'vq->notification' as in your comment for v2,
I think this should work for packed ring as well, anything I missed?

    if (!enable) {
        e.flags = VRING_PACKED_EVENT_FLAG_DISABLE;
    } else if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
        uint16_t off_wrap;

        off_wrap = vq->shadow_avail_idx | vq->shadow_avail_wrap_counter << 15;
        vring_packed_off_wrap_write(vq->vdev, &caches->used, off_wrap);

        /* Make sure off_wrap is wrote before flags */
        smp_wmb();
        e.flags = VRING_PACKED_EVENT_FLAG_DESC;
    } else {
        e.flags = VRING_PACKED_EVENT_FLAG_ENABLE;
    }

    vring_packed_flags_write(vq->vdev, &caches->used, e.flags);

> 
> 
> >+            goto out;
> >+        }
> >+
> >+        e.flags = VRING_PACKED_EVENT_FLAG_DISABLE;
> >+        goto update;
> >+    }
> >+
> >+    e.flags = VRING_PACKED_EVENT_FLAG_ENABLE;
> 
> 
> Here and the above goto could be eliminated by:
> 
> if (even idx) {
> 
> ...
> 
> } else if (enable) {
> 
> ...
> 
> } else {
> 
> ...
> 
> }
> 

thanks, I have removed it in above snippet.

Wei
> 
> Thanks
> 
> 
> >+    if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> >+        uint16_t off_wrap = vq->shadow_avail_idx | vq->avail_wrap_counter << 15;
> >+
> >+        vring_packed_off_wrap_write(vq->vdev, &caches->used, off_wrap);
> >+        /* Make sure off_wrap is wrote before flags */
> >+        smp_wmb();
> >+
> >+        e.flags = VRING_PACKED_EVENT_FLAG_DESC;
> >+    }
> >+
> >+update:
> >+    vring_packed_flags_write(vq->vdev, &caches->used, e.flags);
> >+out:
> >+    rcu_read_unlock();
> >+}
> >+
> >+void virtio_queue_set_notification(VirtQueue *vq, int enable)
> >+{
> >+    vq->notification = enable;
> >+
> >+    if (!vq->vring.desc) {
> >+        return;
> >+    }
> >+
> >+    if (virtio_vdev_has_feature(vq->vdev, VIRTIO_F_RING_PACKED)) {
> >+        virtio_queue_set_notification_packed(vq, enable);
> >+    } else {
> >+        virtio_queue_set_notification_split(vq, enable);
> >+    }
> >+}
> >+
> >  int virtio_queue_ready(VirtQueue *vq)
> >  {
> >      return vq->vring.avail != 0;
> >@@ -2117,8 +2190,7 @@ static void virtio_set_isr(VirtIODevice *vdev, int value)
> >      }
> >  }
> >-/* Called within rcu_read_lock().  */
> >-static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >+static bool virtio_split_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >  {
> >      uint16_t old, new;
> >      bool v;
> >@@ -2141,6 +2213,53 @@ static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >      return !v || vring_need_event(vring_get_used_event(vq), new, old);
> >  }
> >+static bool vring_packed_need_event(VirtQueue *vq, bool wrap,
> >+                            uint16_t off_wrap, uint16_t new, uint16_t old)
> >+{
> >+    int off = off_wrap & ~(1 << 15);
> >+
> >+    if (wrap != off_wrap >> 15) {
> >+        off -= vq->vring.num;
> >+    }
> >+
> >+    return vring_need_event(off, new, old);
> >+}
> >+
> >+static bool virtio_packed_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >+{
> >+    VRingPackedDescEvent e;
> >+    uint16_t old, new;
> >+    bool v;
> >+    VRingMemoryRegionCaches *caches;
> >+
> >+    caches  = vring_get_region_caches(vq);
> >+    vring_packed_event_read(vdev, &caches->avail, &e);
> >+
> >+    old = vq->signalled_used;
> >+    new = vq->signalled_used = vq->used_idx;
> >+    v = vq->signalled_used_valid;
> >+    vq->signalled_used_valid = true;
> >+
> >+    if (e.flags == VRING_PACKED_EVENT_FLAG_DISABLE) {
> >+        return false;
> >+    } else if (e.flags == VRING_PACKED_EVENT_FLAG_ENABLE) {
> >+        return true;
> >+    }
> >+
> >+    return !v || vring_packed_need_event(vq,
> >+        vq->used_wrap_counter, e.off_wrap, new, old);
> >+}
> >+
> >+/* Called within rcu_read_lock().  */
> >+static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >+{
> >+    if (virtio_vdev_has_feature(vdev, VIRTIO_F_RING_PACKED)) {
> >+        return virtio_packed_should_notify(vdev, vq);
> >+    } else {
> >+        return virtio_split_should_notify(vdev, vq);
> >+    }
> >+}
> >+
> >  void virtio_notify_irqfd(VirtIODevice *vdev, VirtQueue *vq)
> >  {
> >      bool should_notify;
> 

  reply	other threads:[~2019-02-19 10:44 UTC|newest]

Thread overview: 41+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-02-14  4:26 [Qemu-devel] [PATCH v4 00/11] packed ring virtio-net backends support wexu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 01/11] virtio: rename structure for packed ring wexu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 02/11] virtio: device/driver area size calculation helper for split ring wexu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 03/11] virtio: initialize packed ring region wexu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 04/11] virtio: initialize wrap counter for packed ring wexu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 05/11] virtio: queue/descriptor check helpers " wexu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 06/11] virtio: get avail bytes check " wexu
2019-02-18  7:27   ` Jason Wang
2019-02-18 17:07     ` Wei Xu
2019-02-19  6:24       ` Jason Wang
2019-02-19  8:24         ` Wei Xu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 07/11] virtio: fill/flush/pop " wexu
2019-02-18  7:51   ` Jason Wang
2019-02-18 14:46     ` Wei Xu
2019-02-19  6:49       ` Jason Wang
2019-02-19  8:21         ` Wei Xu
2019-02-19  9:33           ` Jason Wang
2019-02-19 11:34             ` Wei Xu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 08/11] virtio: event suppression support " wexu
2019-02-19  7:19   ` Jason Wang
2019-02-19 10:40     ` Wei Xu [this message]
2019-02-19 13:06       ` Jason Wang
2019-02-20  2:17         ` Wei Xu
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 09/11] virtio-net: update the head descriptor in a chain lastly wexu
2019-02-19  7:23   ` Jason Wang
2019-02-19 10:51     ` Wei Xu
2019-02-19 13:09       ` Jason Wang
2019-02-20  1:54         ` Wei Xu
2019-02-20  2:34           ` Jason Wang
2019-02-20  4:01             ` Wei Xu
2019-02-20  7:53               ` Jason Wang
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 10/11] virtio: migration support for packed ring wexu
2019-02-19  7:30   ` Jason Wang
2019-02-19 11:00     ` Wei Xu
2019-02-19 13:12       ` Jason Wang
2019-02-14  4:26 ` [Qemu-devel] [PATCH v4 11/11] virtio: CLI and provide packed ring feature bit by default wexu
2019-02-19  7:32   ` Jason Wang
2019-02-19 11:23     ` Wei Xu
2019-02-19 13:33       ` Jason Wang
2019-02-20  0:46         ` Wei Xu
2019-02-19  7:35 ` [Qemu-devel] [PATCH v4 00/11] packed ring virtio-net backends support Jason Wang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20190219104044.GC15343@wei-ubt \
    --to=wexu@redhat.com \
    --cc=jasowang@redhat.com \
    --cc=jfreiman@redhat.com \
    --cc=maxime.coquelin@redhat.com \
    --cc=mst@redhat.com \
    --cc=qemu-devel@nongnu.org \
    --cc=tiwei.bie@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).