qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: "Michael S. Tsirkin" <mst@redhat.com>
To: Stefan Hajnoczi <stefanha@redhat.com>
Cc: Alexander Bulekov <alxndr@bu.edu>,
	qemu-stable@nongnu.org, Cornelia Huck <cohuck@redhat.com>,
	qemu-devel@nongnu.org, Paolo Bonzini <pbonzini@redhat.com>
Subject: Re: [PATCH v2] virtio: gracefully handle invalid region caches
Date: Thu, 27 Feb 2020 03:51:41 -0500	[thread overview]
Message-ID: <20200227034505-mutt-send-email-mst@kernel.org> (raw)
In-Reply-To: <20200207104619.164892-1-stefanha@redhat.com>

On Fri, Feb 07, 2020 at 10:46:19AM +0000, Stefan Hajnoczi wrote:
> The virtqueue code sets up MemoryRegionCaches to access the virtqueue
> guest RAM data structures.  The code currently assumes that
> VRingMemoryRegionCaches is initialized before device emulation code
> accesses the virtqueue.  An assertion will fail in
> vring_get_region_caches() when this is not true.  Device fuzzing found a
> case where this assumption is false (see below).
> 
> Virtqueue guest RAM addresses can also be changed from a vCPU thread
> while an IOThread is accessing the virtqueue.  This breaks the same
> assumption but this time the caches could become invalid partway through
> the virtqueue code.  The code fetches the caches RCU pointer multiple
> times so we will need to validate the pointer every time it is fetched.
> 
> Add checks each time we call vring_get_region_caches() and treat invalid
> caches as a nop: memory stores are ignored and memory reads return 0.
> 
> The fuzz test failure is as follows:
> 
>   $ qemu -M pc -device virtio-blk-pci,id=drv0,drive=drive0,addr=4.0 \
>          -drive if=none,id=drive0,file=null-co://,format=raw,auto-read-only=off \
>          -drive if=none,id=drive1,file=null-co://,file.read-zeroes=on,format=raw \
>          -display none \
>          -qtest stdio
>   endianness
>   outl 0xcf8 0x80002020
>   outl 0xcfc 0xe0000000
>   outl 0xcf8 0x80002004
>   outw 0xcfc 0x7
>   write 0xe0000000 0x24 0x00ffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffab5cffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffabffffffab0000000001
>   inb 0x4
>   writew 0xe000001c 0x1
>   write 0xe0000014 0x1 0x0d
> 
> The following error message is produced:
> 
>   qemu-system-x86_64: /home/stefanha/qemu/hw/virtio/virtio.c:286: vring_get_region_caches: Assertion `caches != NULL' failed.
> 
> The backtrace looks like this:
> 
>   #0  0x00007ffff5520625 in raise () at /lib64/libc.so.6
>   #1  0x00007ffff55098d9 in abort () at /lib64/libc.so.6
>   #2  0x00007ffff55097a9 in _nl_load_domain.cold () at /lib64/libc.so.6
>   #3  0x00007ffff5518a66 in annobin_assert.c_end () at /lib64/libc.so.6
>   #4  0x00005555559073da in vring_get_region_caches (vq=<optimized out>) at qemu/hw/virtio/virtio.c:286
>   #5  vring_get_region_caches (vq=<optimized out>) at qemu/hw/virtio/virtio.c:283
>   #6  0x000055555590818d in vring_used_flags_set_bit (mask=1, vq=0x5555575ceea0) at qemu/hw/virtio/virtio.c:398
>   #7  virtio_queue_split_set_notification (enable=0, vq=0x5555575ceea0) at qemu/hw/virtio/virtio.c:398
>   #8  virtio_queue_set_notification (vq=vq@entry=0x5555575ceea0, enable=enable@entry=0) at qemu/hw/virtio/virtio.c:451
>   #9  0x0000555555908512 in virtio_queue_set_notification (vq=vq@entry=0x5555575ceea0, enable=enable@entry=0) at qemu/hw/virtio/virtio.c:444
>   #10 0x00005555558c697a in virtio_blk_handle_vq (s=0x5555575c57e0, vq=0x5555575ceea0) at qemu/hw/block/virtio-blk.c:775
>   #11 0x0000555555907836 in virtio_queue_notify_aio_vq (vq=0x5555575ceea0) at qemu/hw/virtio/virtio.c:2244
>   #12 0x0000555555cb5dd7 in aio_dispatch_handlers (ctx=ctx@entry=0x55555671a420) at util/aio-posix.c:429
>   #13 0x0000555555cb67a8 in aio_dispatch (ctx=0x55555671a420) at util/aio-posix.c:460
>   #14 0x0000555555cb307e in aio_ctx_dispatch (source=<optimized out>, callback=<optimized out>, user_data=<optimized out>) at util/async.c:260
>   #15 0x00007ffff7bbc510 in g_main_context_dispatch () at /lib64/libglib-2.0.so.0
>   #16 0x0000555555cb5848 in glib_pollfds_poll () at util/main-loop.c:219
>   #17 os_host_main_loop_wait (timeout=<optimized out>) at util/main-loop.c:242
>   #18 main_loop_wait (nonblocking=<optimized out>) at util/main-loop.c:518
>   #19 0x00005555559b20c9 in main_loop () at vl.c:1683
>   #20 0x0000555555838115 in main (argc=<optimized out>, argv=<optimized out>, envp=<optimized out>) at vl.c:4441
> 
> Reported-by: Alexander Bulekov <alxndr@bu.edu>
> Cc: Michael Tsirkin <mst@redhat.com>
> Cc: Cornelia Huck <cohuck@redhat.com>
> Cc: Paolo Bonzini <pbonzini@redhat.com>
> Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>

I sent a pull already, but I just realized it's needed for stable.
In case it gets merged earlier, CC stable now.

> ---
> An alternative solution is to keep the vring.caches pointer non-NULL all
> the time so no checks are necessary.  We would need to initialize it to
> a VRingMemoryRegionCaches object that points to unassigned_mem.  This
> way virtio.c never hits NULL pointers and all memory loads/stores become
> nop when caches are invalid.
> 
> I think this solution is cleaner but couldn't see a reasonable way of
> initializing MemoryRegionCache objects so that they point to a 64-bit
> unassigned_mem MemoryRegion.  Maybe someone who knows the memory API
> better knows whether this is doable?
> 
> Michael: We discussed changing vring.desc checks, but I think that's no
> longer necessary with this patch.  If a guest gets past a vring.desc
> check then it can no longer trigger the assertion failure.
> ---
>  hw/virtio/virtio.c | 99 ++++++++++++++++++++++++++++++++++++++++++----
>  1 file changed, 91 insertions(+), 8 deletions(-)
> 
> diff --git a/hw/virtio/virtio.c b/hw/virtio/virtio.c
> index 2c5410e981..00d444699d 100644
> --- a/hw/virtio/virtio.c
> +++ b/hw/virtio/virtio.c
> @@ -282,15 +282,19 @@ static void vring_packed_flags_write(VirtIODevice *vdev,
>  /* Called within rcu_read_lock().  */
>  static VRingMemoryRegionCaches *vring_get_region_caches(struct VirtQueue *vq)
>  {
> -    VRingMemoryRegionCaches *caches = atomic_rcu_read(&vq->vring.caches);
> -    assert(caches != NULL);
> -    return caches;
> +    return atomic_rcu_read(&vq->vring.caches);
>  }
> +
>  /* Called within rcu_read_lock().  */
>  static inline uint16_t vring_avail_flags(VirtQueue *vq)
>  {
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      hwaddr pa = offsetof(VRingAvail, flags);
> +
> +    if (!caches) {
> +        return 0;
> +    }
> +
>      return virtio_lduw_phys_cached(vq->vdev, &caches->avail, pa);
>  }
>  
> @@ -299,6 +303,11 @@ static inline uint16_t vring_avail_idx(VirtQueue *vq)
>  {
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      hwaddr pa = offsetof(VRingAvail, idx);
> +
> +    if (!caches) {
> +        return 0;
> +    }
> +
>      vq->shadow_avail_idx = virtio_lduw_phys_cached(vq->vdev, &caches->avail, pa);
>      return vq->shadow_avail_idx;
>  }
> @@ -308,6 +317,11 @@ static inline uint16_t vring_avail_ring(VirtQueue *vq, int i)
>  {
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      hwaddr pa = offsetof(VRingAvail, ring[i]);
> +
> +    if (!caches) {
> +        return 0;
> +    }
> +
>      return virtio_lduw_phys_cached(vq->vdev, &caches->avail, pa);
>  }
>  
> @@ -323,6 +337,11 @@ static inline void vring_used_write(VirtQueue *vq, VRingUsedElem *uelem,
>  {
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      hwaddr pa = offsetof(VRingUsed, ring[i]);
> +
> +    if (!caches) {
> +        return;
> +    }
> +
>      virtio_tswap32s(vq->vdev, &uelem->id);
>      virtio_tswap32s(vq->vdev, &uelem->len);
>      address_space_write_cached(&caches->used, pa, uelem, sizeof(VRingUsedElem));
> @@ -334,6 +353,11 @@ static uint16_t vring_used_idx(VirtQueue *vq)
>  {
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      hwaddr pa = offsetof(VRingUsed, idx);
> +
> +    if (!caches) {
> +        return 0;
> +    }
> +
>      return virtio_lduw_phys_cached(vq->vdev, &caches->used, pa);
>  }
>  
> @@ -342,8 +366,12 @@ static inline void vring_used_idx_set(VirtQueue *vq, uint16_t val)
>  {
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      hwaddr pa = offsetof(VRingUsed, idx);
> -    virtio_stw_phys_cached(vq->vdev, &caches->used, pa, val);
> -    address_space_cache_invalidate(&caches->used, pa, sizeof(val));
> +
> +    if (caches) {
> +        virtio_stw_phys_cached(vq->vdev, &caches->used, pa, val);
> +        address_space_cache_invalidate(&caches->used, pa, sizeof(val));
> +    }
> +
>      vq->used_idx = val;
>  }
>  
> @@ -353,8 +381,13 @@ static inline void vring_used_flags_set_bit(VirtQueue *vq, int mask)
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      VirtIODevice *vdev = vq->vdev;
>      hwaddr pa = offsetof(VRingUsed, flags);
> -    uint16_t flags = virtio_lduw_phys_cached(vq->vdev, &caches->used, pa);
> +    uint16_t flags;
>  
> +    if (!caches) {
> +        return;
> +    }
> +
> +    flags = virtio_lduw_phys_cached(vq->vdev, &caches->used, pa);
>      virtio_stw_phys_cached(vdev, &caches->used, pa, flags | mask);
>      address_space_cache_invalidate(&caches->used, pa, sizeof(flags));
>  }
> @@ -365,8 +398,13 @@ static inline void vring_used_flags_unset_bit(VirtQueue *vq, int mask)
>      VRingMemoryRegionCaches *caches = vring_get_region_caches(vq);
>      VirtIODevice *vdev = vq->vdev;
>      hwaddr pa = offsetof(VRingUsed, flags);
> -    uint16_t flags = virtio_lduw_phys_cached(vq->vdev, &caches->used, pa);
> +    uint16_t flags;
>  
> +    if (!caches) {
> +        return;
> +    }
> +
> +    flags = virtio_lduw_phys_cached(vq->vdev, &caches->used, pa);
>      virtio_stw_phys_cached(vdev, &caches->used, pa, flags & ~mask);
>      address_space_cache_invalidate(&caches->used, pa, sizeof(flags));
>  }
> @@ -381,6 +419,10 @@ static inline void vring_set_avail_event(VirtQueue *vq, uint16_t val)
>      }
>  
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        return;
> +    }
> +
>      pa = offsetof(VRingUsed, ring[vq->vring.num]);
>      virtio_stw_phys_cached(vq->vdev, &caches->used, pa, val);
>      address_space_cache_invalidate(&caches->used, pa, sizeof(val));
> @@ -410,7 +452,11 @@ static void virtio_queue_packed_set_notification(VirtQueue *vq, int enable)
>      VRingMemoryRegionCaches *caches;
>  
>      RCU_READ_LOCK_GUARD();
> -    caches  = vring_get_region_caches(vq);
> +    caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        return;
> +    }
> +
>      vring_packed_event_read(vq->vdev, &caches->used, &e);
>  
>      if (!enable) {
> @@ -597,6 +643,10 @@ static int virtio_queue_packed_empty_rcu(VirtQueue *vq)
>      }
>  
>      cache = vring_get_region_caches(vq);
> +    if (!cache) {
> +        return 1;
> +    }
> +
>      vring_packed_desc_read_flags(vq->vdev, &desc.flags, &cache->desc,
>                                   vq->last_avail_idx);
>  
> @@ -777,6 +827,10 @@ static void virtqueue_packed_fill_desc(VirtQueue *vq,
>      }
>  
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        return;
> +    }
> +
>      vring_packed_desc_write(vq->vdev, &desc, &caches->desc, head, strict_order);
>  }
>  
> @@ -949,6 +1003,10 @@ static void virtqueue_split_get_avail_bytes(VirtQueue *vq,
>  
>      max = vq->vring.num;
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        goto err;
> +    }
> +
>      while ((rc = virtqueue_num_heads(vq, idx)) > 0) {
>          MemoryRegionCache *desc_cache = &caches->desc;
>          unsigned int num_bufs;
> @@ -1089,6 +1147,9 @@ static void virtqueue_packed_get_avail_bytes(VirtQueue *vq,
>  
>      max = vq->vring.num;
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        goto err;
> +    }
>  
>      for (;;) {
>          unsigned int num_bufs = total_bufs;
> @@ -1194,6 +1255,10 @@ void virtqueue_get_avail_bytes(VirtQueue *vq, unsigned int *in_bytes,
>      }
>  
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        goto err;
> +    }
> +
>      desc_size = virtio_vdev_has_feature(vq->vdev, VIRTIO_F_RING_PACKED) ?
>                                  sizeof(VRingPackedDesc) : sizeof(VRingDesc);
>      if (caches->desc.len < vq->vring.num * desc_size) {
> @@ -1387,6 +1452,11 @@ static void *virtqueue_split_pop(VirtQueue *vq, size_t sz)
>      i = head;
>  
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        virtio_error(vdev, "Region caches not initialized");
> +        goto done;
> +    }
> +
>      if (caches->desc.len < max * sizeof(VRingDesc)) {
>          virtio_error(vdev, "Cannot map descriptor ring");
>          goto done;
> @@ -1509,6 +1579,11 @@ static void *virtqueue_packed_pop(VirtQueue *vq, size_t sz)
>      i = vq->last_avail_idx;
>  
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        virtio_error(vdev, "Region caches not initialized");
> +        goto done;
> +    }
> +
>      if (caches->desc.len < max * sizeof(VRingDesc)) {
>          virtio_error(vdev, "Cannot map descriptor ring");
>          goto done;
> @@ -1628,6 +1703,10 @@ static unsigned int virtqueue_packed_drop_all(VirtQueue *vq)
>      VRingPackedDesc desc;
>  
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        return 0;
> +    }
> +
>      desc_cache = &caches->desc;
>  
>      virtio_queue_set_notification(vq, 0);
> @@ -2412,6 +2491,10 @@ static bool virtio_packed_should_notify(VirtIODevice *vdev, VirtQueue *vq)
>      VRingMemoryRegionCaches *caches;
>  
>      caches = vring_get_region_caches(vq);
> +    if (!caches) {
> +        return false;
> +    }
> +
>      vring_packed_event_read(vdev, &caches->avail, &e);
>  
>      old = vq->signalled_used;
> -- 
> 2.24.1



      parent reply	other threads:[~2020-02-27  8:52 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-02-07 10:46 [PATCH v2] virtio: gracefully handle invalid region caches Stefan Hajnoczi
2020-02-24 13:35 ` Stefan Hajnoczi
2020-02-24 13:37   ` Michael S. Tsirkin
2020-02-27  8:51 ` Michael S. Tsirkin [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20200227034505-mutt-send-email-mst@kernel.org \
    --to=mst@redhat.com \
    --cc=alxndr@bu.edu \
    --cc=cohuck@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=qemu-devel@nongnu.org \
    --cc=qemu-stable@nongnu.org \
    --cc=stefanha@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).