From: Jason Wang <jasowang@redhat.com>
To: wexu@redhat.com, qemu-devel@nongnu.org
Cc: mst@redhat.com, jfreiman@redhat.com, maxime.coquelin@redhat.com,
tiwei.bie@intel.com
Subject: Re: [Qemu-devel] [PATCH v4 08/11] virtio: event suppression support for packed ring
Date: Tue, 19 Feb 2019 15:19:58 +0800 [thread overview]
Message-ID: <91bf74af-5e22-c6fc-eae5-16b03b1493a1@redhat.com> (raw)
In-Reply-To: <1550118402-4057-9-git-send-email-wexu@redhat.com>
On 2019/2/14 下午12:26, wexu@redhat.com wrote:
> From: Wei Xu <wexu@redhat.com>
>
> Difference between 'avail_wrap_counter' and 'last_avail_wrap_counter':
> For Tx(guest transmitting), they are the same after each pop of a desc.
>
> For Rx(guest receiving), they are also the same when there are enough
> descriptors to carry the payload for a packet(e.g. usually 16 descs are
> needed for a 64k packet in typical iperf tcp connection with tso enabled),
> however, when the ring is running out of descriptors while there are
> still a few free ones, e.g. 6 descriptors are available which is not
> enough to carry an entire packet which needs 16 descriptors, in this
> case the 'avail_wrap_counter' should be set as the first one pending
> being handled by guest driver in order to get a notification, and the
> 'last_avail_wrap_counter' should stay unchanged to the head of available
> descriptors, like below:
>
> Mark meaning:
> | | -- available
> |*| -- used
>
> A Snapshot of the queue:
> last_avail_idx = 253
> last_avail_wrap_counter = 1
> |
> +---------------------------------------------+
> 0 | | | |*|*|*|*|*|*|*|*|*|*|*|*|*|*|*|*|*| | | | 255
> +---------------------------------------------+
> |
> shadow_avail_idx = 3
> avail_wrap_counter = 0
Well this might not be the good place to describe the difference between
shadow_avail_idx and last_avail_idx. And the comments above their
definition looks good enough?
/* Next head to pop */
uint16_t last_avail_idx;
/* Last avail_idx read from VQ. */
uint16_t shadow_avail_idx;
Instead, how or why need event suppress is not mentioned ...
>
> Signed-off-by: Wei Xu <wexu@redhat.com>
> ---
> hw/virtio/virtio.c | 137 +++++++++++++++++++++++++++++++++++++++++++++++++----
> 1 file changed, 128 insertions(+), 9 deletions(-)
>
> diff --git a/hw/virtio/virtio.c b/hw/virtio/virtio.c
> index 7e276b4..8cfc7b6 100644
> --- a/hw/virtio/virtio.c
> +++ b/hw/virtio/virtio.c
> @@ -234,6 +234,34 @@ static void vring_desc_read(VirtIODevice *vdev, VRingDesc *desc,
> virtio_tswap16s(vdev, &desc->next);
> }
>
> +static void vring_packed_event_read(VirtIODevice *vdev,
> + MemoryRegionCache *cache, VRingPackedDescEvent *e)
> +{
> + address_space_read_cached(cache, 0, e, sizeof(*e));
> + virtio_tswap16s(vdev, &e->off_wrap);
> + virtio_tswap16s(vdev, &e->flags);
> +}
> +
> +static void vring_packed_off_wrap_write(VirtIODevice *vdev,
> + MemoryRegionCache *cache, uint16_t off_wrap)
> +{
> + virtio_tswap16s(vdev, &off_wrap);
> + address_space_write_cached(cache, offsetof(VRingPackedDescEvent, off_wrap),
> + &off_wrap, sizeof(off_wrap));
> + address_space_cache_invalidate(cache,
> + offsetof(VRingPackedDescEvent, off_wrap), sizeof(off_wrap));
> +}
> +
> +static void vring_packed_flags_write(VirtIODevice *vdev,
> + MemoryRegionCache *cache, uint16_t flags)
> +{
> + virtio_tswap16s(vdev, &flags);
> + address_space_write_cached(cache, offsetof(VRingPackedDescEvent, flags),
> + &flags, sizeof(flags));
> + address_space_cache_invalidate(cache,
> + offsetof(VRingPackedDescEvent, flags), sizeof(flags));
> +}
> +
> static VRingMemoryRegionCaches *vring_get_region_caches(struct VirtQueue *vq)
> {
> VRingMemoryRegionCaches *caches = atomic_rcu_read(&vq->vring.caches);
> @@ -340,14 +368,8 @@ static inline void vring_set_avail_event(VirtQueue *vq, uint16_t val)
> address_space_cache_invalidate(&caches->used, pa, sizeof(val));
> }
>
> -void virtio_queue_set_notification(VirtQueue *vq, int enable)
> +static void virtio_queue_set_notification_split(VirtQueue *vq, int enable)
> {
> - vq->notification = enable;
> -
> - if (!vq->vring.desc) {
> - return;
> - }
> -
> rcu_read_lock();
> if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> vring_set_avail_event(vq, vring_avail_idx(vq));
> @@ -363,6 +385,57 @@ void virtio_queue_set_notification(VirtQueue *vq, int enable)
> rcu_read_unlock();
> }
>
> +static void virtio_queue_set_notification_packed(VirtQueue *vq, int enable)
> +{
> + VRingPackedDescEvent e;
> + VRingMemoryRegionCaches *caches;
> +
> + rcu_read_lock();
> + caches = vring_get_region_caches(vq);
> + vring_packed_event_read(vq->vdev, &caches->used, &e);
> +
> + if (!enable) {
> + if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> + /* no need to write device area since this is outdated. */
We should advertise off and wrap in this case as well, otherwise we may
get notifications earlier than expected.
> + goto out;
> + }
> +
> + e.flags = VRING_PACKED_EVENT_FLAG_DISABLE;
> + goto update;
> + }
> +
> + e.flags = VRING_PACKED_EVENT_FLAG_ENABLE;
Here and the above goto could be eliminated by:
if (even idx) {
...
} else if (enable) {
...
} else {
...
}
Thanks
> + if (virtio_vdev_has_feature(vq->vdev, VIRTIO_RING_F_EVENT_IDX)) {
> + uint16_t off_wrap = vq->shadow_avail_idx | vq->avail_wrap_counter << 15;
> +
> + vring_packed_off_wrap_write(vq->vdev, &caches->used, off_wrap);
> + /* Make sure off_wrap is wrote before flags */
> + smp_wmb();
> +
> + e.flags = VRING_PACKED_EVENT_FLAG_DESC;
> + }
> +
> +update:
> + vring_packed_flags_write(vq->vdev, &caches->used, e.flags);
> +out:
> + rcu_read_unlock();
> +}
> +
> +void virtio_queue_set_notification(VirtQueue *vq, int enable)
> +{
> + vq->notification = enable;
> +
> + if (!vq->vring.desc) {
> + return;
> + }
> +
> + if (virtio_vdev_has_feature(vq->vdev, VIRTIO_F_RING_PACKED)) {
> + virtio_queue_set_notification_packed(vq, enable);
> + } else {
> + virtio_queue_set_notification_split(vq, enable);
> + }
> +}
> +
> int virtio_queue_ready(VirtQueue *vq)
> {
> return vq->vring.avail != 0;
> @@ -2117,8 +2190,7 @@ static void virtio_set_isr(VirtIODevice *vdev, int value)
> }
> }
>
> -/* Called within rcu_read_lock(). */
> -static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> +static bool virtio_split_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> {
> uint16_t old, new;
> bool v;
> @@ -2141,6 +2213,53 @@ static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> return !v || vring_need_event(vring_get_used_event(vq), new, old);
> }
>
> +static bool vring_packed_need_event(VirtQueue *vq, bool wrap,
> + uint16_t off_wrap, uint16_t new, uint16_t old)
> +{
> + int off = off_wrap & ~(1 << 15);
> +
> + if (wrap != off_wrap >> 15) {
> + off -= vq->vring.num;
> + }
> +
> + return vring_need_event(off, new, old);
> +}
> +
> +static bool virtio_packed_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> +{
> + VRingPackedDescEvent e;
> + uint16_t old, new;
> + bool v;
> + VRingMemoryRegionCaches *caches;
> +
> + caches = vring_get_region_caches(vq);
> + vring_packed_event_read(vdev, &caches->avail, &e);
> +
> + old = vq->signalled_used;
> + new = vq->signalled_used = vq->used_idx;
> + v = vq->signalled_used_valid;
> + vq->signalled_used_valid = true;
> +
> + if (e.flags == VRING_PACKED_EVENT_FLAG_DISABLE) {
> + return false;
> + } else if (e.flags == VRING_PACKED_EVENT_FLAG_ENABLE) {
> + return true;
> + }
> +
> + return !v || vring_packed_need_event(vq,
> + vq->used_wrap_counter, e.off_wrap, new, old);
> +}
> +
> +/* Called within rcu_read_lock(). */
> +static bool virtio_should_notify(VirtIODevice *vdev, VirtQueue *vq)
> +{
> + if (virtio_vdev_has_feature(vdev, VIRTIO_F_RING_PACKED)) {
> + return virtio_packed_should_notify(vdev, vq);
> + } else {
> + return virtio_split_should_notify(vdev, vq);
> + }
> +}
> +
> void virtio_notify_irqfd(VirtIODevice *vdev, VirtQueue *vq)
> {
> bool should_notify;
next prev parent reply other threads:[~2019-02-19 7:20 UTC|newest]
Thread overview: 41+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-02-14 4:26 [Qemu-devel] [PATCH v4 00/11] packed ring virtio-net backends support wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 01/11] virtio: rename structure for packed ring wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 02/11] virtio: device/driver area size calculation helper for split ring wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 03/11] virtio: initialize packed ring region wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 04/11] virtio: initialize wrap counter for packed ring wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 05/11] virtio: queue/descriptor check helpers " wexu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 06/11] virtio: get avail bytes check " wexu
2019-02-18 7:27 ` Jason Wang
2019-02-18 17:07 ` Wei Xu
2019-02-19 6:24 ` Jason Wang
2019-02-19 8:24 ` Wei Xu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 07/11] virtio: fill/flush/pop " wexu
2019-02-18 7:51 ` Jason Wang
2019-02-18 14:46 ` Wei Xu
2019-02-19 6:49 ` Jason Wang
2019-02-19 8:21 ` Wei Xu
2019-02-19 9:33 ` Jason Wang
2019-02-19 11:34 ` Wei Xu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 08/11] virtio: event suppression support " wexu
2019-02-19 7:19 ` Jason Wang [this message]
2019-02-19 10:40 ` Wei Xu
2019-02-19 13:06 ` Jason Wang
2019-02-20 2:17 ` Wei Xu
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 09/11] virtio-net: update the head descriptor in a chain lastly wexu
2019-02-19 7:23 ` Jason Wang
2019-02-19 10:51 ` Wei Xu
2019-02-19 13:09 ` Jason Wang
2019-02-20 1:54 ` Wei Xu
2019-02-20 2:34 ` Jason Wang
2019-02-20 4:01 ` Wei Xu
2019-02-20 7:53 ` Jason Wang
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 10/11] virtio: migration support for packed ring wexu
2019-02-19 7:30 ` Jason Wang
2019-02-19 11:00 ` Wei Xu
2019-02-19 13:12 ` Jason Wang
2019-02-14 4:26 ` [Qemu-devel] [PATCH v4 11/11] virtio: CLI and provide packed ring feature bit by default wexu
2019-02-19 7:32 ` Jason Wang
2019-02-19 11:23 ` Wei Xu
2019-02-19 13:33 ` Jason Wang
2019-02-20 0:46 ` Wei Xu
2019-02-19 7:35 ` [Qemu-devel] [PATCH v4 00/11] packed ring virtio-net backends support Jason Wang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=91bf74af-5e22-c6fc-eae5-16b03b1493a1@redhat.com \
--to=jasowang@redhat.com \
--cc=jfreiman@redhat.com \
--cc=maxime.coquelin@redhat.com \
--cc=mst@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=tiwei.bie@intel.com \
--cc=wexu@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).