qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v1] vhost: Fix used memslot tracking when destroying a vhost device
@ 2025-06-03 11:13 David Hildenbrand
  2025-06-11  8:21 ` Igor Mammedov
  2025-07-16 13:31 ` Michael Tokarev
  0 siblings, 2 replies; 4+ messages in thread
From: David Hildenbrand @ 2025-06-03 11:13 UTC (permalink / raw)
  To: qemu-devel
  Cc: David Hildenbrand, yuanminghao, Igor Mammedov, Michael S. Tsirkin,
	Stefano Garzarella

When we unplug a vhost device, we end up calling vhost_dev_cleanup()
where we do a memory_listener_unregister().

This memory_listener_unregister() call will end up disconnecting the
listener from the address space through listener_del_address_space().

In that process, we effectively communicate the removal of all memory
regions from that listener, resulting in region_del() + commit()
callbacks getting triggered.

So in case of vhost, we end up calling vhost_commit() with no remaining
memory slots (0).

In vhost_commit() we end up overwriting the global variables
used_memslots / used_shared_memslots, used for detecting the number
of free memslots. With used_memslots / used_shared_memslots set to 0
by vhost_commit() during device removal, we'll later assume that the
other vhost devices still have plenty of memslots left when calling
vhost_get_free_memslots().

Let's fix it by simply removing the global variables and depending
only on the actual per-device count.

Easy to reproduce by adding two vhost-user devices to a VM and then
hot-unplugging one of them.

While at it, detect unexpected underflows in vhost_get_free_memslots()
and issue a warning.

Reported-by: yuanminghao <yuanmh12@chinatelecom.cn>
Link: https://lore.kernel.org/qemu-devel/20241121060755.164310-1-yuanmh12@chinatelecom.cn/
Fixes: 2ce68e4cf5be ("vhost: add vhost_has_free_slot() interface")
Cc: Igor Mammedov <imammedo@redhat.com>
Cc: Michael S. Tsirkin <mst@redhat.com>
Cc: Stefano Garzarella <sgarzare@redhat.com>
Signed-off-by: David Hildenbrand <david@redhat.com>
---

I assume the problem existed in some form when used_memslots was
introduced. However, I did not check the old behavior of memory listener
unregistration etc.

---
 hw/virtio/vhost.c | 37 ++++++++++---------------------------
 1 file changed, 10 insertions(+), 27 deletions(-)

diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c
index fc43853704..c87861b31f 100644
--- a/hw/virtio/vhost.c
+++ b/hw/virtio/vhost.c
@@ -47,12 +47,6 @@ static struct vhost_log *vhost_log[VHOST_BACKEND_TYPE_MAX];
 static struct vhost_log *vhost_log_shm[VHOST_BACKEND_TYPE_MAX];
 static QLIST_HEAD(, vhost_dev) vhost_log_devs[VHOST_BACKEND_TYPE_MAX];
 
-/* Memslots used by backends that support private memslots (without an fd). */
-static unsigned int used_memslots;
-
-/* Memslots used by backends that only support shared memslots (with an fd). */
-static unsigned int used_shared_memslots;
-
 static QLIST_HEAD(, vhost_dev) vhost_devices =
     QLIST_HEAD_INITIALIZER(vhost_devices);
 
@@ -74,15 +68,15 @@ unsigned int vhost_get_free_memslots(void)
 
     QLIST_FOREACH(hdev, &vhost_devices, entry) {
         unsigned int r = hdev->vhost_ops->vhost_backend_memslots_limit(hdev);
-        unsigned int cur_free;
+        unsigned int cur_free = r - hdev->mem->nregions;
 
-        if (hdev->vhost_ops->vhost_backend_no_private_memslots &&
-            hdev->vhost_ops->vhost_backend_no_private_memslots(hdev)) {
-            cur_free = r - used_shared_memslots;
+        if (unlikely(r < hdev->mem->nregions)) {
+            warn_report_once("used (%u) vhost backend memory slots exceed"
+                             " the device limit (%u).", hdev->mem->nregions, r);
+            free = 0;
         } else {
-            cur_free = r - used_memslots;
+            free = MIN(free, cur_free);
         }
-        free = MIN(free, cur_free);
     }
     return free;
 }
@@ -666,13 +660,6 @@ static void vhost_commit(MemoryListener *listener)
     dev->mem = g_realloc(dev->mem, regions_size);
     dev->mem->nregions = dev->n_mem_sections;
 
-    if (dev->vhost_ops->vhost_backend_no_private_memslots &&
-        dev->vhost_ops->vhost_backend_no_private_memslots(dev)) {
-        used_shared_memslots = dev->mem->nregions;
-    } else {
-        used_memslots = dev->mem->nregions;
-    }
-
     for (i = 0; i < dev->n_mem_sections; i++) {
         struct vhost_memory_region *cur_vmr = dev->mem->regions + i;
         struct MemoryRegionSection *mrs = dev->mem_sections + i;
@@ -1619,15 +1606,11 @@ int vhost_dev_init(struct vhost_dev *hdev, void *opaque,
     QLIST_INSERT_HEAD(&vhost_devices, hdev, entry);
 
     /*
-     * The listener we registered properly updated the corresponding counter.
-     * So we can trust that these values are accurate.
+     * The listener we registered properly setup the number of required
+     * memslots in vhost_commit().
      */
-    if (hdev->vhost_ops->vhost_backend_no_private_memslots &&
-        hdev->vhost_ops->vhost_backend_no_private_memslots(hdev)) {
-        used = used_shared_memslots;
-    } else {
-        used = used_memslots;
-    }
+    used = hdev->mem->nregions;
+
     /*
      * We assume that all reserved memslots actually require a real memslot
      * in our vhost backend. This might not be true, for example, if the
-- 
2.49.0



^ permalink raw reply related	[flat|nested] 4+ messages in thread

* Re: [PATCH v1] vhost: Fix used memslot tracking when destroying a vhost device
  2025-06-03 11:13 [PATCH v1] vhost: Fix used memslot tracking when destroying a vhost device David Hildenbrand
@ 2025-06-11  8:21 ` Igor Mammedov
  2025-07-16 13:31 ` Michael Tokarev
  1 sibling, 0 replies; 4+ messages in thread
From: Igor Mammedov @ 2025-06-11  8:21 UTC (permalink / raw)
  To: David Hildenbrand
  Cc: qemu-devel, yuanminghao, Michael S. Tsirkin, Stefano Garzarella

On Tue,  3 Jun 2025 13:13:36 +0200
David Hildenbrand <david@redhat.com> wrote:

> When we unplug a vhost device, we end up calling vhost_dev_cleanup()
> where we do a memory_listener_unregister().
> 
> This memory_listener_unregister() call will end up disconnecting the
> listener from the address space through listener_del_address_space().
> 
> In that process, we effectively communicate the removal of all memory
> regions from that listener, resulting in region_del() + commit()
> callbacks getting triggered.
> 
> So in case of vhost, we end up calling vhost_commit() with no remaining
> memory slots (0).
> 
> In vhost_commit() we end up overwriting the global variables
> used_memslots / used_shared_memslots, used for detecting the number
> of free memslots. With used_memslots / used_shared_memslots set to 0
> by vhost_commit() during device removal, we'll later assume that the
> other vhost devices still have plenty of memslots left when calling
> vhost_get_free_memslots().
> 
> Let's fix it by simply removing the global variables and depending
> only on the actual per-device count.
> 
> Easy to reproduce by adding two vhost-user devices to a VM and then
> hot-unplugging one of them.
> 
> While at it, detect unexpected underflows in vhost_get_free_memslots()
> and issue a warning.
> 
> Reported-by: yuanminghao <yuanmh12@chinatelecom.cn>
> Link: https://lore.kernel.org/qemu-devel/20241121060755.164310-1-yuanmh12@chinatelecom.cn/
> Fixes: 2ce68e4cf5be ("vhost: add vhost_has_free_slot() interface")
> Cc: Igor Mammedov <imammedo@redhat.com>
> Cc: Michael S. Tsirkin <mst@redhat.com>
> Cc: Stefano Garzarella <sgarzare@redhat.com>
> Signed-off-by: David Hildenbrand <david@redhat.com>

Reviewed-by: Igor Mammedov <imammedo@redhat.com>

> ---
> 
> I assume the problem existed in some form when used_memslots was
> introduced. However, I did not check the old behavior of memory listener
> unregistration etc.
> 
> ---
>  hw/virtio/vhost.c | 37 ++++++++++---------------------------
>  1 file changed, 10 insertions(+), 27 deletions(-)
> 
> diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c
> index fc43853704..c87861b31f 100644
> --- a/hw/virtio/vhost.c
> +++ b/hw/virtio/vhost.c
> @@ -47,12 +47,6 @@ static struct vhost_log *vhost_log[VHOST_BACKEND_TYPE_MAX];
>  static struct vhost_log *vhost_log_shm[VHOST_BACKEND_TYPE_MAX];
>  static QLIST_HEAD(, vhost_dev) vhost_log_devs[VHOST_BACKEND_TYPE_MAX];
>  
> -/* Memslots used by backends that support private memslots (without an fd). */
> -static unsigned int used_memslots;
> -
> -/* Memslots used by backends that only support shared memslots (with an fd). */
> -static unsigned int used_shared_memslots;
> -
>  static QLIST_HEAD(, vhost_dev) vhost_devices =
>      QLIST_HEAD_INITIALIZER(vhost_devices);
>  
> @@ -74,15 +68,15 @@ unsigned int vhost_get_free_memslots(void)
>  
>      QLIST_FOREACH(hdev, &vhost_devices, entry) {
>          unsigned int r = hdev->vhost_ops->vhost_backend_memslots_limit(hdev);
> -        unsigned int cur_free;
> +        unsigned int cur_free = r - hdev->mem->nregions;
>  
> -        if (hdev->vhost_ops->vhost_backend_no_private_memslots &&
> -            hdev->vhost_ops->vhost_backend_no_private_memslots(hdev)) {
> -            cur_free = r - used_shared_memslots;
> +        if (unlikely(r < hdev->mem->nregions)) {
> +            warn_report_once("used (%u) vhost backend memory slots exceed"
> +                             " the device limit (%u).", hdev->mem->nregions, r);
> +            free = 0;
>          } else {
> -            cur_free = r - used_memslots;
> +            free = MIN(free, cur_free);
>          }
> -        free = MIN(free, cur_free);
>      }
>      return free;
>  }
> @@ -666,13 +660,6 @@ static void vhost_commit(MemoryListener *listener)
>      dev->mem = g_realloc(dev->mem, regions_size);
>      dev->mem->nregions = dev->n_mem_sections;
>  
> -    if (dev->vhost_ops->vhost_backend_no_private_memslots &&
> -        dev->vhost_ops->vhost_backend_no_private_memslots(dev)) {
> -        used_shared_memslots = dev->mem->nregions;
> -    } else {
> -        used_memslots = dev->mem->nregions;
> -    }
> -
>      for (i = 0; i < dev->n_mem_sections; i++) {
>          struct vhost_memory_region *cur_vmr = dev->mem->regions + i;
>          struct MemoryRegionSection *mrs = dev->mem_sections + i;
> @@ -1619,15 +1606,11 @@ int vhost_dev_init(struct vhost_dev *hdev, void *opaque,
>      QLIST_INSERT_HEAD(&vhost_devices, hdev, entry);
>  
>      /*
> -     * The listener we registered properly updated the corresponding counter.
> -     * So we can trust that these values are accurate.
> +     * The listener we registered properly setup the number of required
> +     * memslots in vhost_commit().
>       */
> -    if (hdev->vhost_ops->vhost_backend_no_private_memslots &&
> -        hdev->vhost_ops->vhost_backend_no_private_memslots(hdev)) {
> -        used = used_shared_memslots;
> -    } else {
> -        used = used_memslots;
> -    }
> +    used = hdev->mem->nregions;
> +
>      /*
>       * We assume that all reserved memslots actually require a real memslot
>       * in our vhost backend. This might not be true, for example, if the



^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH v1] vhost: Fix used memslot tracking when destroying a vhost device
  2025-06-03 11:13 [PATCH v1] vhost: Fix used memslot tracking when destroying a vhost device David Hildenbrand
  2025-06-11  8:21 ` Igor Mammedov
@ 2025-07-16 13:31 ` Michael Tokarev
  2025-07-16 13:41   ` David Hildenbrand
  1 sibling, 1 reply; 4+ messages in thread
From: Michael Tokarev @ 2025-07-16 13:31 UTC (permalink / raw)
  To: David Hildenbrand, qemu-devel
  Cc: yuanminghao, Igor Mammedov, Michael S. Tsirkin,
	Stefano Garzarella, qemu-stable

On 03.06.2025 14:13, David Hildenbrand wrote:
> When we unplug a vhost device, we end up calling vhost_dev_cleanup()
> where we do a memory_listener_unregister().
> 
> This memory_listener_unregister() call will end up disconnecting the
> listener from the address space through listener_del_address_space().
> 
> In that process, we effectively communicate the removal of all memory
> regions from that listener, resulting in region_del() + commit()
> callbacks getting triggered.
> 
> So in case of vhost, we end up calling vhost_commit() with no remaining
> memory slots (0).
> 
> In vhost_commit() we end up overwriting the global variables
> used_memslots / used_shared_memslots, used for detecting the number
> of free memslots. With used_memslots / used_shared_memslots set to 0
> by vhost_commit() during device removal, we'll later assume that the
> other vhost devices still have plenty of memslots left when calling
> vhost_get_free_memslots().
> 
> Let's fix it by simply removing the global variables and depending
> only on the actual per-device count.
> 
> Easy to reproduce by adding two vhost-user devices to a VM and then
> hot-unplugging one of them.
> 
> While at it, detect unexpected underflows in vhost_get_free_memslots()
> and issue a warning.
> 
> Reported-by: yuanminghao <yuanmh12@chinatelecom.cn>
> Link: https://lore.kernel.org/qemu-devel/20241121060755.164310-1-yuanmh12@chinatelecom.cn/
> Fixes: 2ce68e4cf5be ("vhost: add vhost_has_free_slot() interface")
> Cc: Igor Mammedov <imammedo@redhat.com>
> Cc: Michael S. Tsirkin <mst@redhat.com>
> Cc: Stefano Garzarella <sgarzare@redhat.com>
> Signed-off-by: David Hildenbrand <david@redhat.com>

Is it a stable material (10.0.x) too?

Thanks,

/mjt


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH v1] vhost: Fix used memslot tracking when destroying a vhost device
  2025-07-16 13:31 ` Michael Tokarev
@ 2025-07-16 13:41   ` David Hildenbrand
  0 siblings, 0 replies; 4+ messages in thread
From: David Hildenbrand @ 2025-07-16 13:41 UTC (permalink / raw)
  To: Michael Tokarev, qemu-devel
  Cc: yuanminghao, Igor Mammedov, Michael S. Tsirkin,
	Stefano Garzarella, qemu-stable

On 16.07.25 15:31, Michael Tokarev wrote:
> On 03.06.2025 14:13, David Hildenbrand wrote:
>> When we unplug a vhost device, we end up calling vhost_dev_cleanup()
>> where we do a memory_listener_unregister().
>>
>> This memory_listener_unregister() call will end up disconnecting the
>> listener from the address space through listener_del_address_space().
>>
>> In that process, we effectively communicate the removal of all memory
>> regions from that listener, resulting in region_del() + commit()
>> callbacks getting triggered.
>>
>> So in case of vhost, we end up calling vhost_commit() with no remaining
>> memory slots (0).
>>
>> In vhost_commit() we end up overwriting the global variables
>> used_memslots / used_shared_memslots, used for detecting the number
>> of free memslots. With used_memslots / used_shared_memslots set to 0
>> by vhost_commit() during device removal, we'll later assume that the
>> other vhost devices still have plenty of memslots left when calling
>> vhost_get_free_memslots().
>>
>> Let's fix it by simply removing the global variables and depending
>> only on the actual per-device count.
>>
>> Easy to reproduce by adding two vhost-user devices to a VM and then
>> hot-unplugging one of them.
>>
>> While at it, detect unexpected underflows in vhost_get_free_memslots()
>> and issue a warning.
>>
>> Reported-by: yuanminghao <yuanmh12@chinatelecom.cn>
>> Link: https://lore.kernel.org/qemu-devel/20241121060755.164310-1-yuanmh12@chinatelecom.cn/
>> Fixes: 2ce68e4cf5be ("vhost: add vhost_has_free_slot() interface")
>> Cc: Igor Mammedov <imammedo@redhat.com>
>> Cc: Michael S. Tsirkin <mst@redhat.com>
>> Cc: Stefano Garzarella <sgarzare@redhat.com>
>> Signed-off-by: David Hildenbrand <david@redhat.com>
> 
> Is it a stable material (10.0.x) too?

Probably better to have it in stable, yes. Thanks!

-- 
Cheers,

David / dhildenb



^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2025-07-16 13:45 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-06-03 11:13 [PATCH v1] vhost: Fix used memslot tracking when destroying a vhost device David Hildenbrand
2025-06-11  8:21 ` Igor Mammedov
2025-07-16 13:31 ` Michael Tokarev
2025-07-16 13:41   ` David Hildenbrand

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).