From: Kevin Wolf <kwolf@redhat.com>
To: qemu-block@nongnu.org
Cc: kwolf@redhat.com, peter.maydell@linaro.org, qemu-devel@nongnu.org
Subject: [PULL 04/43] virtio-blk: On restart, process queued requests in the proper context
Date: Wed, 17 Jun 2020 16:48:30 +0200 [thread overview]
Message-ID: <20200617144909.192176-5-kwolf@redhat.com> (raw)
In-Reply-To: <20200617144909.192176-1-kwolf@redhat.com>
From: Sergio Lopez <slp@redhat.com>
On restart, we were scheduling a BH to process queued requests, which
would run before starting up the data plane, leading to those requests
being assigned and started on coroutines on the main context.
This could cause requests to be wrongly processed in parallel from
different threads (the main thread and the iothread managing the data
plane), potentially leading to multiple issues.
For example, stopping and resuming a VM multiple times while the guest
is generating I/O on a virtio_blk device can trigger a crash with a
stack tracing looking like this one:
<------>
Thread 2 (Thread 0x7ff736765700 (LWP 1062503)):
#0 0x00005567a13b99d6 in iov_memset
(iov=0x6563617073206f4e, iov_cnt=1717922848, offset=516096, fillc=0, bytes=7018105756081554803)
at util/iov.c:69
#1 0x00005567a13bab73 in qemu_iovec_memset
(qiov=0x7ff73ec99748, offset=516096, fillc=0, bytes=7018105756081554803) at util/iov.c:530
#2 0x00005567a12f411c in qemu_laio_process_completion (laiocb=0x7ff6512ee6c0) at block/linux-aio.c:86
#3 0x00005567a12f42ff in qemu_laio_process_completions (s=0x7ff7182e8420) at block/linux-aio.c:217
#4 0x00005567a12f480d in ioq_submit (s=0x7ff7182e8420) at block/linux-aio.c:323
#5 0x00005567a12f43d9 in qemu_laio_process_completions_and_submit (s=0x7ff7182e8420)
at block/linux-aio.c:236
#6 0x00005567a12f44c2 in qemu_laio_poll_cb (opaque=0x7ff7182e8430) at block/linux-aio.c:267
#7 0x00005567a13aed83 in run_poll_handlers_once (ctx=0x5567a2b58c70, timeout=0x7ff7367645f8)
at util/aio-posix.c:520
#8 0x00005567a13aee9f in run_poll_handlers (ctx=0x5567a2b58c70, max_ns=16000, timeout=0x7ff7367645f8)
at util/aio-posix.c:562
#9 0x00005567a13aefde in try_poll_mode (ctx=0x5567a2b58c70, timeout=0x7ff7367645f8)
at util/aio-posix.c:597
#10 0x00005567a13af115 in aio_poll (ctx=0x5567a2b58c70, blocking=true) at util/aio-posix.c:639
#11 0x00005567a109acca in iothread_run (opaque=0x5567a2b29760) at iothread.c:75
#12 0x00005567a13b2790 in qemu_thread_start (args=0x5567a2b694c0) at util/qemu-thread-posix.c:519
#13 0x00007ff73eedf2de in start_thread () at /lib64/libpthread.so.0
#14 0x00007ff73ec10e83 in clone () at /lib64/libc.so.6
Thread 1 (Thread 0x7ff743986f00 (LWP 1062500)):
#0 0x00005567a13b99d6 in iov_memset
(iov=0x6563617073206f4e, iov_cnt=1717922848, offset=516096, fillc=0, bytes=7018105756081554803)
at util/iov.c:69
#1 0x00005567a13bab73 in qemu_iovec_memset
(qiov=0x7ff73ec99748, offset=516096, fillc=0, bytes=7018105756081554803) at util/iov.c:530
#2 0x00005567a12f411c in qemu_laio_process_completion (laiocb=0x7ff6512ee6c0) at block/linux-aio.c:86
#3 0x00005567a12f42ff in qemu_laio_process_completions (s=0x7ff7182e8420) at block/linux-aio.c:217
#4 0x00005567a12f480d in ioq_submit (s=0x7ff7182e8420) at block/linux-aio.c:323
#5 0x00005567a12f4a2f in laio_do_submit (fd=19, laiocb=0x7ff5f4ff9ae0, offset=472363008, type=2)
at block/linux-aio.c:375
#6 0x00005567a12f4af2 in laio_co_submit
(bs=0x5567a2b8c460, s=0x7ff7182e8420, fd=19, offset=472363008, qiov=0x7ff5f4ff9ca0, type=2)
at block/linux-aio.c:394
#7 0x00005567a12f1803 in raw_co_prw
(bs=0x5567a2b8c460, offset=472363008, bytes=20480, qiov=0x7ff5f4ff9ca0, type=2)
at block/file-posix.c:1892
#8 0x00005567a12f1941 in raw_co_pwritev
(bs=0x5567a2b8c460, offset=472363008, bytes=20480, qiov=0x7ff5f4ff9ca0, flags=0)
at block/file-posix.c:1925
#9 0x00005567a12fe3e1 in bdrv_driver_pwritev
(bs=0x5567a2b8c460, offset=472363008, bytes=20480, qiov=0x7ff5f4ff9ca0, qiov_offset=0, flags=0)
at block/io.c:1183
#10 0x00005567a1300340 in bdrv_aligned_pwritev
(child=0x5567a2b5b070, req=0x7ff5f4ff9db0, offset=472363008, bytes=20480, align=512, qiov=0x7ff72c0425b8, qiov_offset=0, flags=0) at block/io.c:1980
#11 0x00005567a1300b29 in bdrv_co_pwritev_part
(child=0x5567a2b5b070, offset=472363008, bytes=20480, qiov=0x7ff72c0425b8, qiov_offset=0, flags=0)
at block/io.c:2137
#12 0x00005567a12baba1 in qcow2_co_pwritev_task
(bs=0x5567a2b92740, file_cluster_offset=472317952, offset=487305216, bytes=20480, qiov=0x7ff72c0425b8, qiov_offset=0, l2meta=0x0) at block/qcow2.c:2444
#13 0x00005567a12bacdb in qcow2_co_pwritev_task_entry (task=0x5567a2b48540) at block/qcow2.c:2475
#14 0x00005567a13167d8 in aio_task_co (opaque=0x5567a2b48540) at block/aio_task.c:45
#15 0x00005567a13cf00c in coroutine_trampoline (i0=738245600, i1=32759) at util/coroutine-ucontext.c:115
#16 0x00007ff73eb622e0 in __start_context () at /lib64/libc.so.6
#17 0x00007ff6626f1350 in ()
#18 0x0000000000000000 in ()
<------>
This is also known to cause crashes with this message (assertion
failed):
aio_co_schedule: Co-routine was already scheduled in 'aio_co_schedule'
RHBZ: https://bugzilla.redhat.com/show_bug.cgi?id=1812765
Signed-off-by: Sergio Lopez <slp@redhat.com>
Message-Id: <20200603093240.40489-3-slp@redhat.com>
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
---
include/hw/virtio/virtio-blk.h | 2 +-
hw/block/dataplane/virtio-blk.c | 8 ++++++++
hw/block/virtio-blk.c | 18 ++++++++++++------
3 files changed, 21 insertions(+), 7 deletions(-)
diff --git a/include/hw/virtio/virtio-blk.h b/include/hw/virtio/virtio-blk.h
index f584ad9b86..b1334c3904 100644
--- a/include/hw/virtio/virtio-blk.h
+++ b/include/hw/virtio/virtio-blk.h
@@ -86,6 +86,6 @@ typedef struct MultiReqBuffer {
} MultiReqBuffer;
bool virtio_blk_handle_vq(VirtIOBlock *s, VirtQueue *vq);
-void virtio_blk_process_queued_requests(VirtIOBlock *s);
+void virtio_blk_process_queued_requests(VirtIOBlock *s, bool is_bh);
#endif
diff --git a/hw/block/dataplane/virtio-blk.c b/hw/block/dataplane/virtio-blk.c
index 1b52e8159c..37499c5564 100644
--- a/hw/block/dataplane/virtio-blk.c
+++ b/hw/block/dataplane/virtio-blk.c
@@ -220,6 +220,9 @@ int virtio_blk_data_plane_start(VirtIODevice *vdev)
goto fail_guest_notifiers;
}
+ /* Process queued requests before the ones in vring */
+ virtio_blk_process_queued_requests(vblk, false);
+
/* Kick right away to begin processing requests already in vring */
for (i = 0; i < nvqs; i++) {
VirtQueue *vq = virtio_get_queue(s->vdev, i);
@@ -239,6 +242,11 @@ int virtio_blk_data_plane_start(VirtIODevice *vdev)
return 0;
fail_guest_notifiers:
+ /*
+ * If we failed to set up the guest notifiers queued requests will be
+ * processed on the main context.
+ */
+ virtio_blk_process_queued_requests(vblk, false);
vblk->dataplane_disabled = true;
s->starting = false;
vblk->dataplane_started = true;
diff --git a/hw/block/virtio-blk.c b/hw/block/virtio-blk.c
index 978574e4da..8882a1d1d4 100644
--- a/hw/block/virtio-blk.c
+++ b/hw/block/virtio-blk.c
@@ -819,7 +819,7 @@ static void virtio_blk_handle_output(VirtIODevice *vdev, VirtQueue *vq)
virtio_blk_handle_output_do(s, vq);
}
-void virtio_blk_process_queued_requests(VirtIOBlock *s)
+void virtio_blk_process_queued_requests(VirtIOBlock *s, bool is_bh)
{
VirtIOBlockReq *req = s->rq;
MultiReqBuffer mrb = {};
@@ -847,7 +847,9 @@ void virtio_blk_process_queued_requests(VirtIOBlock *s)
if (mrb.num_reqs) {
virtio_blk_submit_multireq(s->blk, &mrb);
}
- blk_dec_in_flight(s->conf.conf.blk);
+ if (is_bh) {
+ blk_dec_in_flight(s->conf.conf.blk);
+ }
aio_context_release(blk_get_aio_context(s->conf.conf.blk));
}
@@ -858,21 +860,25 @@ static void virtio_blk_dma_restart_bh(void *opaque)
qemu_bh_delete(s->bh);
s->bh = NULL;
- virtio_blk_process_queued_requests(s);
+ virtio_blk_process_queued_requests(s, true);
}
static void virtio_blk_dma_restart_cb(void *opaque, int running,
RunState state)
{
VirtIOBlock *s = opaque;
+ BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(s)));
+ VirtioBusState *bus = VIRTIO_BUS(qbus);
if (!running) {
return;
}
- if (!s->bh) {
- /* FIXME The data plane is not started yet, so these requests are
- * processed in the main thread. */
+ /*
+ * If ioeventfd is enabled, don't schedule the BH here as queued
+ * requests will be processed while starting the data plane.
+ */
+ if (!s->bh && !virtio_bus_ioeventfd_enabled(bus)) {
s->bh = aio_bh_new(blk_get_aio_context(s->conf.conf.blk),
virtio_blk_dma_restart_bh, s);
blk_inc_in_flight(s->conf.conf.blk);
--
2.25.4
next prev parent reply other threads:[~2020-06-17 14:51 UTC|newest]
Thread overview: 48+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-17 14:48 [PULL 00/43] Block layer patches Kevin Wolf
2020-06-17 14:48 ` [PULL 01/43] hw/ide: Make IDEDMAOps handlers take a const IDEDMA pointer Kevin Wolf
2020-06-17 14:48 ` [PULL 02/43] icount: make dma reads deterministic Kevin Wolf
2020-06-17 14:48 ` [PULL 03/43] virtio-blk: Refactor the code that processes queued requests Kevin Wolf
2020-06-17 14:48 ` Kevin Wolf [this message]
2020-06-17 14:48 ` [PULL 05/43] block: Refactor subdirectory recursion during make Kevin Wolf
2020-06-17 14:48 ` [PULL 06/43] qcow2: Tweak comments on qcow2_get_persistent_dirty_bitmap_size Kevin Wolf
2020-06-17 14:48 ` [PULL 07/43] hw/block/nvme: fix pci doorbell size calculation Kevin Wolf
2020-06-17 14:48 ` [PULL 08/43] hw/block/nvme: rename trace events to pci_nvme Kevin Wolf
2020-06-17 14:48 ` [PULL 09/43] hw/block/nvme: remove superfluous breaks Kevin Wolf
2020-06-17 14:48 ` [PULL 10/43] hw/block/nvme: move device parameters to separate struct Kevin Wolf
2020-06-17 14:48 ` [PULL 11/43] hw/block/nvme: use constants in identify Kevin Wolf
2020-06-17 14:48 ` [PULL 12/43] hw/block/nvme: refactor nvme_addr_read Kevin Wolf
2020-06-17 14:48 ` [PULL 13/43] hw/block/nvme: fix pin-based interrupt behavior Kevin Wolf
2020-06-17 14:48 ` [PULL 14/43] hw/block/nvme: add max_ioqpairs device parameter Kevin Wolf
2020-06-17 14:48 ` [PULL 15/43] hw/block/nvme: remove redundant cmbloc/cmbsz members Kevin Wolf
2020-06-17 14:48 ` [PULL 16/43] hw/block/nvme: factor out property/constraint checks Kevin Wolf
2020-06-17 14:48 ` [PULL 17/43] hw/block/nvme: factor out device state setup Kevin Wolf
2020-06-17 14:48 ` [PULL 18/43] hw/block/nvme: factor out block backend setup Kevin Wolf
2020-06-17 14:48 ` [PULL 19/43] hw/block/nvme: add namespace helpers Kevin Wolf
2020-06-17 14:48 ` [PULL 20/43] hw/block/nvme: factor out namespace setup Kevin Wolf
2020-06-17 14:48 ` [PULL 21/43] hw/block/nvme: factor out pci setup Kevin Wolf
2020-06-17 14:48 ` [PULL 22/43] hw/block/nvme: factor out cmb setup Kevin Wolf
2020-06-17 14:48 ` [PULL 23/43] hw/block/nvme: factor out pmr setup Kevin Wolf
2020-06-17 14:48 ` [PULL 24/43] hw/block/nvme: do cmb/pmr init as part of pci init Kevin Wolf
2020-06-17 14:48 ` [PULL 25/43] hw/block/nvme: factor out controller identify setup Kevin Wolf
2020-06-17 14:48 ` [PULL 26/43] hw/block/nvme: Verify msix_vector_use() returned value Kevin Wolf
2020-06-17 14:48 ` [PULL 27/43] hw/block/nvme: add msix_qsize parameter Kevin Wolf
2020-06-17 14:48 ` [PULL 28/43] hw/block/nvme: verify msix_init_exclusive_bar() return value Kevin Wolf
2020-06-17 14:48 ` [PULL 29/43] .gitignore: Ignore storage-daemon files Kevin Wolf
2020-06-17 14:48 ` [PULL 30/43] virtio-blk: store opt_io_size with correct size Kevin Wolf
2020-06-17 14:48 ` [PULL 31/43] block: consolidate blocksize properties consistency checks Kevin Wolf
2020-06-17 14:48 ` [PULL 32/43] qdev-properties: blocksize: use same limits in code and description Kevin Wolf
2020-06-17 14:48 ` [PULL 33/43] qdev-properties: add size32 property type Kevin Wolf
2020-06-17 14:49 ` [PULL 34/43] qdev-properties: make blocksize accept size suffixes Kevin Wolf
2020-06-17 14:49 ` [PULL 35/43] block: make BlockConf size props 32bit and " Kevin Wolf
2020-06-17 14:49 ` [PULL 36/43] qdev-properties: add getter for size32 and blocksize Kevin Wolf
2020-06-17 14:49 ` [PULL 37/43] block: lift blocksize property limit to 2 MiB Kevin Wolf
2020-06-17 14:49 ` [PULL 38/43] iotests.py: Add skip_for_formats() decorator Kevin Wolf
2020-06-17 14:49 ` [PULL 39/43] iotests/041: Skip test_small_target for qed Kevin Wolf
2020-06-17 14:49 ` [PULL 40/43] iotests/292: data_file is unsupported Kevin Wolf
2020-06-17 14:49 ` [PULL 41/43] iotests/229: " Kevin Wolf
2020-06-17 14:49 ` [PULL 42/43] iotests/{190,291}: compat=0.10 " Kevin Wolf
2020-06-17 14:49 ` [PULL 43/43] iotests: Add copyright line in qcow2.py Kevin Wolf
2020-06-17 15:48 ` [PULL 00/43] Block layer patches no-reply
2020-06-18 14:30 ` Peter Maydell
2020-06-25 8:39 ` Klaus Jensen
2020-06-25 10:39 ` Kevin Wolf
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200617144909.192176-5-kwolf@redhat.com \
--to=kwolf@redhat.com \
--cc=peter.maydell@linaro.org \
--cc=qemu-block@nongnu.org \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).