From: Wei Xu <wexu@redhat.com>
To: Jason Wang <jasowang@redhat.com>
Cc: qemu-devel@nongnu.org, maxime.coquelin@redhat.com,
tiwei.bie@intel.com, jfreiman@redhat.com, mst@redhat.com
Subject: Re: [Qemu-devel] [PATCH v4 08/11] virtio: event suppression support for packed ring
Date: Tue, 19 Feb 2019 18:40:44 +0800 [thread overview]
Message-ID: <20190219104044.GC15343@wei-ubt> (raw)
In-Reply-To: <91bf74af-5e22-c6fc-eae5-16b03b1493a1@redhat.com>
On Tue, Feb 19, 2019 at 03:19:58PM +0800, Jason Wang wrote:
>
> On 2019/2/14 下午12:26, wexu@redhat.com wrote:
> >From: Wei Xu <wexu@redhat.com>
> >
> >Difference between 'avail_wrap_counter' and 'last_avail_wrap_counter':
> >For Tx(guest transmitting), they are the same after each pop of a desc.
> >
> >For Rx(guest receiving), they are also the same when there are enough
> >descriptors to carry the payload for a packet(e.g. usually 16 descs are
> >needed for a 64k packet in typical iperf tcp connection with tso enabled),
> >however, when the ring is running out of descriptors while there are
> >still a few free ones, e.g. 6 descriptors are available which is not
> >enough to carry an entire packet which needs 16 descriptors, in this
> >case the 'avail_wrap_counter' should be set as the first one pending
> >being handled by guest driver in order to get a notification, and the
> >'last_avail_wrap_counter' should stay unchanged to the head of available
> >descriptors, like below:
> >
> >Mark meaning:
> > | | -- available
> > |*| -- used
> >
> >A Snapshot of the queue:
> > last_avail_idx = 253
> > last_avail_wrap_counter = 1
> > |
> > +---------------------------------------------+
> > 0 | | | |*|*|*|*|*|*|*|*|*|*|*|*|*|*|*|*|*| | | | 255
> > +---------------------------------------------+
> > |
> > shadow_avail_idx = 3
> > avail_wrap_counter = 0
>
>
> Well this might not be the good place to describe the difference between
> shadow_avail_idx and last_avail_idx. And the comments above their definition
> looks good enough?
Sorry, I do not get it, can you elaborate more?
This is one of the buggy parts of packed ring, it is good to make it clear here.
>
> /* Next head to pop */
> uint16_t last_avail_idx;
>
> /* Last avail_idx read from VQ. */
> uint16_t shadow_avail_idx;
>
What is the meaning of these comment? Do you mean I should replace put them
to the comments also? thanks.
> Instead, how or why need event suppress is not mentioned ...
Yes, but presumably this knowledge has been acquired from reading through the
spec, so I skipped this part.
Wei
>
>
>
> >
> >Signed-off-by: Wei Xu <wexu@redhat.com>
> >---
> > hw/virtio/virtio.c | 137 +++++++++++++++++++++++++++++++++++++++++++++++++----
> > 1 file changed, 128 insertions(+), 9 deletions(-)
> >
> >diff --git a/hw/virtio/virtio.c b/hw/virtio/virtio.c
> >index 7e276b4..8cfc7b6 100644
> >--- a/hw/virtio/virtio.c
> >+++ b/hw/virtio/virtio.c
> >@@ -234,6 +234,34 @@ static void vring_desc_read(VirtIODevice *vdev, VRingDesc *desc,
> > virtio_tswap16s(vdev, &desc->next);
> > }
> >+static void vring_packed_event_read(VirtIODevice *vdev,
> >+ MemoryRegionCache *cache, VRingPackedDescEvent *e)
> >+{
> >+ address_space_read_cached(cache, 0, e, sizeof(*e));
> >+ virtio_tswap16s(vdev, &e->off_wrap);
> >+ virtio_tswap16s(vdev, &e->flags);
> >+}
> >+
> >+static void vring_packed_off_wrap_write(VirtIODevice *vdev,
> >+ MemoryRegionCache *cache, uint16_t off_wrap)
> >+{
> >+ virtio_tswap16s(vdev, &off_wrap);
> >+ address_space_write_cached(cache, offsetof(VRingPackedDescEvent, off_wrap),
> >+ &off_wrap, sizeof(off_wrap));
> >+ address_space_cache_invalidate(cache,
> >+ offsetof(VRingPackedDescEvent, off_wrap), sizeof(off_wrap));
> >+}
> >+
> >+static void vring_packed_flags_write(VirtIODevice *vdev,
> >+ MemoryRegionCache *cache, uint16_t flags)
> >+{
> >+ virtio_tswap16s(vdev, &flags);
> >+ address_space_write_cached(cache, offsetof(VRingPackedDescEvent, flags),
> >+ &flags, sizeof(flags));
> >+ address_space_cache_invalidate(cache,
> >+ offsetof(VRingPackedDescEvent, flags), sizeof(flags));
> >+}
> >+
> > static VRingMemoryRegionCaches *vring_get_region_caches(struct VirtQueue *vq)
> > {
> > VRingMemoryRegionCaches *caches = atomic_rcu_read(&vq->vring.caches);
> >@@ -340,14 +368,8 @@ static inline void vring_set_avail_event(VirtQueue *vq, uint16_t val)
> > address_space_cache_invalidate(&caches->used, pa, sizeof(val));
> > }
> >-void virtio_queue_set_notification(VirtQueue *vq, int enable)
> >+static void virtio_queue_set_notification_split(VirtQueue *vq, int enable)
> > {
> >- vq->notification = enable;
> >-
> >- if (!vq->vring.desc) {
> >- return;
> >- }
> >-
> > rcu_read_lock();
> > if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> > vring_set_avail_event(vq, vring_avail_idx(vq));
> >@@ -363,6 +385,57 @@ void virtio_queue_set_notification(VirtQueue *vq, int enable)
> > rcu_read_unlock();
> > }
> >+static void virtio_queue_set_notification_packed(VirtQueue *vq, int enable)
> >+{
> >+ VRingPackedDescEvent e;
> >+ VRingMemoryRegionCaches *caches;
> >+
> >+ rcu_read_lock();
> >+ caches = vring_get_region_caches(vq);
> >+ vring_packed_event_read(vq->vdev, &caches->used, &e);
> >+
> >+ if (!enable) {
> >+ if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> >+ /* no need to write device area since this is outdated. */
>
>
> We should advertise off and wrap in this case as well, otherwise we may get
> notifications earlier than expected.
Is it necessary? Supposing offset & wrap_counter are always set before update
notification flags, it is reliable to skip this for disabling, isn't it?
While the logic here is unclear, I did a concise one like below
which doesn't use the 'vq->notification' as in your comment for v2,
I think this should work for packed ring as well, anything I missed?
if (!enable) {
e.flags = VRING_PACKED_EVENT_FLAG_DISABLE;
} else if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
uint16_t off_wrap;
off_wrap = vq->shadow_avail_idx | vq->shadow_avail_wrap_counter << 15;
vring_packed_off_wrap_write(vq->vdev, &caches->used, off_wrap);
/* Make sure off_wrap is wrote before flags */
smp_wmb();
e.flags = VRING_PACKED_EVENT_FLAG_DESC;
} else {
e.flags = VRING_PACKED_EVENT_FLAG_ENABLE;
}
vring_packed_flags_write(vq->vdev, &caches->used, e.flags);
>
>
> >+ goto out;
> >+ }
> >+
> >+ e.flags = VRING_PACKED_EVENT_FLAG_DISABLE;
> >+ goto update;
> >+ }
> >+
> >+ e.flags = VRING_PACKED_EVENT_FLAG_ENABLE;
>
>
> Here and the above goto could be eliminated by:
>
> if (even idx) {
>
> ...
>
> } else if (enable) {
>
> ...
>
> } else {
>
> ...
>
> }
>
thanks, I have removed it in above snippet.
Wei
>
> Thanks
>
>
> >+ if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> >+ uint16_t off_wrap = vq->shadow_avail_idx | vq->avail_wrap_counter << 15;
> >+
> >+ vring_packed_off_wrap_write(vq->vdev, &caches->used, off_wrap);
> >+ /* Make sure off_wrap is wrote before flags */
> >+ smp_wmb();
> >+
> >+ e.flags = VRING_PACKED_EVENT_FLAG_DESC;
> >+ }
> >+
> >+update:
> >+ vring_packed_flags_write(vq->vdev, &caches->used, e.flags);
> >+out:
> >+ rcu_read_unlock();
> >+}
> >+
> >+void virtio_queue_set_notification(VirtQueue *vq, int enable)
> >+{
> >+ vq->notification = enable;
> >+
> >+ if (!vq->vring.desc) {
> >+ return;
> >+ }
> >+
> >+ if (virtio_vdev_has_feature(vq->vdev, VIRTIO_F_RING_PACKED)) {
> >+ virtio_queue_set_notification_packed(vq, enable);
> >+ } else {
> >+ virtio_queue_set_notification_split(vq, enable);
> >+ }
> >+}
> >+
> > int virtio_queue_ready(VirtQueue *vq)
> > {
> > return vq->vring.avail != 0;
> >@@ -2117,8 +2190,7 @@ static void virtio_set_isr(VirtIODevice *vdev, int value)
> > }
> > }
> >-/* Called within rcu_read_lock(). */
> >-static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >+static bool virtio_split_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> > {
> > uint16_t old, new;
> > bool v;
> >@@ -2141,6 +2213,53 @@ static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> > return !v || vring_need_event(vring_get_used_event(vq), new, old);
> > }
> >+static bool vring_packed_need_event(VirtQueue *vq, bool wrap,
> >+ uint16_t off_wrap, uint16_t new, uint16_t old)
> >+{
> >+ int off = off_wrap & ~(1 << 15);
> >+
> >+ if (wrap != off_wrap >> 15) {
> >+ off -= vq->vring.num;
> >+ }
> >+
> >+ return vring_need_event(off, new, old);
> >+}
> >+
> >+static bool virtio_packed_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >+{
> >+ VRingPackedDescEvent e;
> >+ uint16_t old, new;
> >+ bool v;
> >+ VRingMemoryRegionCaches *caches;
> >+
> >+ caches = vring_get_region_caches(vq);
> >+ vring_packed_event_read(vdev, &caches->avail, &e);
> >+
> >+ old = vq->signalled_used;
> >+ new = vq->signalled_used = vq->used_idx;
> >+ v = vq->signalled_used_valid;
> >+ vq->signalled_used_valid = true;
> >+
> >+ if (e.flags == VRING_PACKED_EVENT_FLAG_DISABLE) {
> >+ return false;
> >+ } else if (e.flags == VRING_PACKED_EVENT_FLAG_ENABLE) {
> >+ return true;
> >+ }
> >+
> >+ return !v || vring_packed_need_event(vq,
> >+ vq->used_wrap_counter, e.off_wrap, new, old);
> >+}
> >+
> >+/* Called within rcu_read_lock(). */
> >+static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> >+{
> >+ if (virtio_vdev_has_feature(vdev, VIRTIO_F_RING_PACKED)) {
> >+ return virtio_packed_should_notify(vdev, vq);
> >+ } else {
> >+ return virtio_split_should_notify(vdev, vq);
> >+ }
> >+}
> >+
> > void virtio_notify_irqfd(VirtIODevice *vdev, VirtQueue *vq)
> > {
> > bool should_notify;
>
next prev parent reply other threads:[~2019-02-19 10:44 UTC|newest]
Thread overview: 41+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-02-14 4:26 [Qemu-devel] [PATCH v4 00/11] packed ring virtio-net backends support wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 01/11] virtio: rename structure for packed ring wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 02/11] virtio: device/driver area size calculation helper for split ring wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 03/11] virtio: initialize packed ring region wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 04/11] virtio: initialize wrap counter for packed ring wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 05/11] virtio: queue/descriptor check helpers " wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 06/11] virtio: get avail bytes check " wexu
2019-02-18 7:27 ` Jason Wang
2019-02-18 17:07 ` Wei Xu
2019-02-19 6:24 ` Jason Wang
2019-02-19 8:24 ` Wei Xu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 07/11] virtio: fill/flush/pop " wexu
2019-02-18 7:51 ` Jason Wang
2019-02-18 14:46 ` Wei Xu
2019-02-19 6:49 ` Jason Wang
2019-02-19 8:21 ` Wei Xu
2019-02-19 9:33 ` Jason Wang
2019-02-19 11:34 ` Wei Xu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 08/11] virtio: event suppression support " wexu
2019-02-19 7:19 ` Jason Wang
2019-02-19 10:40 ` Wei Xu [this message]
2019-02-19 13:06 ` Jason Wang
2019-02-20 2:17 ` Wei Xu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 09/11] virtio-net: update the head descriptor in a chain lastly wexu
2019-02-19 7:23 ` Jason Wang
2019-02-19 10:51 ` Wei Xu
2019-02-19 13:09 ` Jason Wang
2019-02-20 1:54 ` Wei Xu
2019-02-20 2:34 ` Jason Wang
2019-02-20 4:01 ` Wei Xu
2019-02-20 7:53 ` Jason Wang
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 10/11] virtio: migration support for packed ring wexu
2019-02-19 7:30 ` Jason Wang
2019-02-19 11:00 ` Wei Xu
2019-02-19 13:12 ` Jason Wang
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 11/11] virtio: CLI and provide packed ring feature bit by default wexu
2019-02-19 7:32 ` Jason Wang
2019-02-19 11:23 ` Wei Xu
2019-02-19 13:33 ` Jason Wang
2019-02-20 0:46 ` Wei Xu
2019-02-19 7:35 ` [Qemu-devel] [PATCH v4 00/11] packed ring virtio-net backends support Jason Wang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190219104044.GC15343@wei-ubt \
--to=wexu@redhat.com \
--cc=jasowang@redhat.com \
--cc=jfreiman@redhat.com \
--cc=maxime.coquelin@redhat.com \
--cc=mst@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=tiwei.bie@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).