From: "Michael S. Tsirkin" <mst@redhat.com>
To: Demi Marie Obenour <demiobenour@gmail.com>
Cc: "virtio-comment@lists.linux.dev" <virtio-comment@lists.linux.dev>
Subject: Re: Should there be a mode in which the virtqueue -> MSI mapping is fixed?
Date: Sun, 5 Apr 2026 16:15:17 -0400 [thread overview]
Message-ID: <20260405155501-mutt-send-email-mst@kernel.org> (raw)
In-Reply-To: <b01e0d9b-086b-4a60-8ada-1ec895d46139@gmail.com>
On Sun, Apr 05, 2026 at 01:50:25PM -0400, Demi Marie Obenour wrote:
> On 4/4/26 20:56, Michael S. Tsirkin wrote:
> > On Sat, Apr 04, 2026 at 05:19:41PM -0400, Demi Marie Obenour wrote:
> >> Cloud Hypervisor's vhost-user frontend does not implement MSI-X
> >> properly [1]. Specifically:
> >>
> >> 1. Reads from the Pending Bit Array (PBA) always return 0.
> >> 2. Changes to the MSI associated with a virtqueue after the device
> >> is activated are ignored.
> >>
> >> Amazingly, there have not been any reports of this causing breakage.
> >> I have a fix for the first [2], which actually decreases the amount
> >> of code. However, the second is trickier and I'm tempted to not
> >> bother unless it causes real-world problems.
> >>
> >> Are there real-world drivers that will run into either of the above
> >> bugs? Linux seems to only choose anything else as a fallback, which
> >> presumably is not triggered.
> >
> > It will sometimes trigger.
>
> Would it be possible to provide an example? A reproducible test
> case would be ideal, but conditions under which this will trigger
> are also sufficient.
I am not sure what does "is activated" mean.
For example, on latest Linux:
vq = vp_find_one_vq_msix(vdev, avq->vq_index, vp_modern_avq_done,
avq->name, false, true, &allocated_vectors,
vector_policy, &vp_dev->admin_vq.info);
if (IS_ERR(vq)) {
err = PTR_ERR(vq);
goto error_find;
}
return 0;
error_find:
vp_del_vqs(vdev);
return err;
}
And
static void del_vq(struct virtio_pci_vq_info *info)
{
struct virtqueue *vq = info->vq;
struct virtio_pci_device *vp_dev = to_vp_device(vq->vdev);
struct virtio_pci_modern_device *mdev = &vp_dev->mdev;
if (vp_dev->msix_enabled)
vp_modern_queue_vector(mdev, vq->index,
VIRTIO_MSI_NO_VECTOR);
if (!mdev->notify_base)
pci_iounmap(mdev->pci_dev, (void __force __iomem *)vq->priv);
vring_del_virtqueue(vq);
}
and this happens after feature negotiation.
It's before device_ready, however.
> >> One reason I am asking is that I am working on an updated
> >> virtio-vhost-user spec, which I've renamed vhost-guest. A vhost-guest
> >> device implements a vhost-user server, and requires one MSI for each
> >> virtqueue of _the device being implemented_. The existing spec
> >> allows the guest to select which MSIs are used, but that seems to
> >> be pointless additional complexity. A simpler option would be to
> >> hard-code the MSI assignments:
> >>
> >> - 0: Configuration change interrupt.
> >>
> >> - 1..N (inclusive): Queue interrupts for the N virtqueues provided
> >> by the vhost-guest device.
> >>
> >> - N+1..N+M (inclusive): Buffer availability interrupts for each of
> >> the M virtqueues that the driver is implementing.
> >
> >
> > you can do this, and imply ask drivers to share msi vector values.
>
> Are there drivers that cannot do this? Is this the reason that the
> virtio spec allows using the same MSI for multiple virtqueues?
>
No, it's because of the devices:
1. it's easier for device to detect sharing when it's explicit,
rather than matching msi vectors which can change at any time at all.
2. we thought it might be more portable to non pci transports.
> > but sharing has to work because # of vectors in the system is limited.
>
> Sharing vectors will indeed work. It's just sharing MSIs that won't.
>
> >> I expect that all real-world drivers for vhost-guest will select
> >> MSIs in this manner. Hard-coding it makes the device implementation
> >> (and specification!) simpler. I'd also rather not ship the first
> >> implementation with a known bug!
> >>
> >> [1]: https://github.com/cloud-hypervisor/cloud-hypervisor/issues/7813
> >> [2]: https://github.com/cloud-hypervisor/cloud-hypervisor/pull/7963
> --
> Sincerely,
> Demi Marie Obenour (she/her/hers)
next prev parent reply other threads:[~2026-04-05 20:15 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-04 21:19 Should there be a mode in which the virtqueue -> MSI mapping is fixed? Demi Marie Obenour
2026-04-05 0:56 ` Michael S. Tsirkin
2026-04-05 17:50 ` Demi Marie Obenour
2026-04-05 20:15 ` Michael S. Tsirkin [this message]
2026-04-05 20:58 ` Demi Marie Obenour
2026-04-05 21:09 ` Michael S. Tsirkin
2026-04-05 21:47 ` Demi Marie Obenour
2026-04-05 21:50 ` Michael S. Tsirkin
2026-04-05 22:28 ` Demi Marie Obenour
2026-04-06 9:25 ` Michael S. Tsirkin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260405155501-mutt-send-email-mst@kernel.org \
--to=mst@redhat.com \
--cc=demiobenour@gmail.com \
--cc=virtio-comment@lists.linux.dev \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox