From: Jason Wang <jasowang@redhat.com>
To: Eugenio Perez Martin <eperezma@redhat.com>,
"Michael S. Tsirkin" <mst@redhat.com>
Cc: qemu-level <qemu-devel@nongnu.org>,
Gautam Dawar <gdawar@xilinx.com>,
Martin Petrus Hubertus Habets <martinh@xilinx.com>,
Harpreet Singh Anand <hanand@xilinx.com>,
tanujk@xilinx.com, pabloc@xilinx.com,
Gautam Dawar <gautam.dawar@xilinx.com>
Subject: Re: [PATCH 2/2] Allow VIRTIO_F_IN_ORDER to be negotiated for vdpa devices
Date: Mon, 21 Feb 2022 12:31:37 +0800 [thread overview]
Message-ID: <681533f1-e2e6-fac2-c9a2-dadd07d1f7c4@redhat.com> (raw)
In-Reply-To: <CAJaqyWcdv3jTP7FzZ9NP29K1YYFp8GYOQEt3s-ySz3bHQDfucQ@mail.gmail.com>
在 2022/2/18 下午6:22, Eugenio Perez Martin 写道:
> On Thu, Feb 17, 2022 at 8:32 AM Michael S. Tsirkin <mst@redhat.com> wrote:
>> On Tue, Feb 15, 2022 at 12:52:31PM +0530, Gautam Dawar wrote:
>>> This patch adds the ability to negotiate VIRTIO_F_IN_ORDER bit
>>> for vhost-vdpa backend when the underlying device supports this
>>> feature.
>>> This would aid in reaping performance benefits with HW devices
>>> that implement this feature. At the same time, it shouldn't have
>>> any negative impact as vhost-vdpa backend doesn't involve any
>>> userspace virtqueue operations.
>>>
>>> Signed-off-by: Gautam Dawar <gdawar@xilinx.com>
>> Having features that hardware implements but qemu does not
>> means we can't migrate between them.
>> So I'd rather see a userspace implementation.
>>
> While I totally agree the userspace implementation is a better option,
> would it be a problem if we implement it as a cmdline option as Jason
> proposed?
>
> I see other backends have similar issues with migration. For example
> it's possible to run qemu with
> -device=virtio-net-pci,...,indirect_desc=on and use a vhost-kernel
> backend without indirect support in their features. I also understand
> qemu emulated backend as "the base" somehow, but it should work
> similarly to my example if cmdline parameter is off by default.
We had a lot of issues with the command line compatibility like this
(e.g qemu may silently clear a host feature) which is probably worth to
fix in the future.
>
> On the other hand, It may be worth thinking if it's worth waiting for
> GSoC though, so we avoid this problem entirely at the moment. But I
> feel that is going to come back with a different feature set with the
> advent of more out of qemu devices and the fast adding of features of
> VirtIO.
>
> Thoughts?
Not sure we had sufficient resources on the Qemu/kernel implementation
of in-order. But if we decide not to wait we can change the GSoC to
implement other stuffs like e.g NOTIFICATION_DATA.
Thanks
>
> Thanks!
>
>>> ---
>>> hw/net/virtio-net.c | 10 ++++++++++
>>> net/vhost-vdpa.c | 1 +
>>> 2 files changed, 11 insertions(+)
>>>
>>> diff --git a/hw/net/virtio-net.c b/hw/net/virtio-net.c
>>> index cf8ab0f8af..a1089d06f6 100644
>>> --- a/hw/net/virtio-net.c
>>> +++ b/hw/net/virtio-net.c
>>> @@ -3507,11 +3507,21 @@ static void virtio_net_device_realize(DeviceState *dev, Error **errp)
>>> nc->rxfilter_notify_enabled = 1;
>>>
>>> if (nc->peer && nc->peer->info->type == NET_CLIENT_DRIVER_VHOST_VDPA) {
>>> + uint64_t features = BIT_ULL(VIRTIO_F_IN_ORDER);
>>> struct virtio_net_config netcfg = {};
>>> +
>>> memcpy(&netcfg.mac, &n->nic_conf.macaddr, ETH_ALEN);
>>> vhost_net_set_config(get_vhost_net(nc->peer),
>>> (uint8_t *)&netcfg, 0, ETH_ALEN, VHOST_SET_CONFIG_TYPE_MASTER);
>>> +
>>> + /*
>>> + * For vhost-vdpa, if underlying device supports IN_ORDER feature,
>>> + * make it available for negotiation.
>>> + */
>>> + features = vhost_net_get_features(get_vhost_net(nc->peer), features);
>>> + n->host_features |= features;
>>> }
>>> +
>>> QTAILQ_INIT(&n->rsc_chains);
>>> n->qdev = dev;
>>>
>>> diff --git a/net/vhost-vdpa.c b/net/vhost-vdpa.c
>>> index 25dd6dd975..2886cba5ec 100644
>>> --- a/net/vhost-vdpa.c
>>> +++ b/net/vhost-vdpa.c
>>> @@ -62,6 +62,7 @@ const int vdpa_feature_bits[] = {
>>> VIRTIO_NET_F_CTRL_VQ,
>>> VIRTIO_F_IOMMU_PLATFORM,
>>> VIRTIO_F_RING_PACKED,
>>> + VIRTIO_F_IN_ORDER,
>>> VIRTIO_NET_F_RSS,
>>> VIRTIO_NET_F_HASH_REPORT,
>>> VIRTIO_NET_F_GUEST_ANNOUNCE,
>>> --
>>> 2.30.1
next prev parent reply other threads:[~2022-02-21 4:32 UTC|newest]
Thread overview: 17+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-02-15 7:22 [PATCH 0/2] Allow VIRTIO_F_IN_ORDER negotiation with vhost-vdpa Gautam Dawar
2022-02-15 7:22 ` [PATCH 1/2] linux headers: update against Linux 5.17-rc4 Gautam Dawar
2022-02-17 7:17 ` Jason Wang
2022-02-17 9:06 ` Gautam Dawar
2022-02-15 7:22 ` [PATCH 2/2] Allow VIRTIO_F_IN_ORDER to be negotiated for vdpa devices Gautam Dawar
2022-02-15 15:22 ` Eugenio Perez Martin
2022-02-17 7:16 ` Jason Wang
2022-02-17 8:27 ` Gautam Dawar
2022-02-17 8:32 ` Michael S. Tsirkin
2022-02-17 8:54 ` Eugenio Perez Martin
2022-02-17 7:32 ` Michael S. Tsirkin
2022-02-17 14:29 ` Stefano Garzarella
2022-02-18 10:24 ` Eugenio Perez Martin
2022-02-18 16:57 ` Stefano Garzarella
2022-02-18 10:22 ` Eugenio Perez Martin
2022-02-21 4:31 ` Jason Wang [this message]
2022-02-18 11:09 ` Michael S. Tsirkin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=681533f1-e2e6-fac2-c9a2-dadd07d1f7c4@redhat.com \
--to=jasowang@redhat.com \
--cc=eperezma@redhat.com \
--cc=gautam.dawar@xilinx.com \
--cc=gdawar@xilinx.com \
--cc=hanand@xilinx.com \
--cc=martinh@xilinx.com \
--cc=mst@redhat.com \
--cc=pabloc@xilinx.com \
--cc=qemu-devel@nongnu.org \
--cc=tanujk@xilinx.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).