From: Kevin Wolf <kwolf@redhat.com>
To: qemu-block@nongnu.org
Cc: kwolf@redhat.com, richard.henderson@linaro.org, qemu-devel@nongnu.org
Subject: [PULL v2 01/28] aio-posix: fix race between io_uring CQE and AioHandler deletion
Date: Tue, 11 Nov 2025 22:32:11 +0100 [thread overview]
Message-ID: <20251111213238.181992-2-kwolf@redhat.com> (raw)
In-Reply-To: <20251111213238.181992-1-kwolf@redhat.com>
From: Stefan Hajnoczi <stefanha@redhat.com>
When an AioHandler is enqueued on ctx->submit_list for removal, the
fill_sq_ring() function will submit an io_uring POLL_REMOVE operation to
cancel the in-flight POLL_ADD operation.
There is a race when another thread enqueues an AioHandler for deletion
on ctx->submit_list when the POLL_ADD CQE has already appeared. In that
case POLL_REMOVE is unnecessary. The code already handled this, but
forgot that the AioHandler itself is still on ctx->submit_list when the
POLL_ADD CQE is being processed. It's unsafe to delete the AioHandler at
that point in time (use-after-free).
Solve this problem by keeping the AioHandler alive but setting a flag so
that it will be deleted by fill_sq_ring() when it runs.
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
Reviewed-by: Eric Blake <eblake@redhat.com>
Reviewed-by: Kevin Wolf <kwolf@redhat.com>
Message-ID: <20251104022933.618123-2-stefanha@redhat.com>
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
---
util/fdmon-io_uring.c | 33 ++++++++++++++++++++++++++-------
1 file changed, 26 insertions(+), 7 deletions(-)
diff --git a/util/fdmon-io_uring.c b/util/fdmon-io_uring.c
index b0d68bdc44..ad89160f31 100644
--- a/util/fdmon-io_uring.c
+++ b/util/fdmon-io_uring.c
@@ -52,9 +52,10 @@ enum {
FDMON_IO_URING_ENTRIES = 128, /* sq/cq ring size */
/* AioHandler::flags */
- FDMON_IO_URING_PENDING = (1 << 0),
- FDMON_IO_URING_ADD = (1 << 1),
- FDMON_IO_URING_REMOVE = (1 << 2),
+ FDMON_IO_URING_PENDING = (1 << 0),
+ FDMON_IO_URING_ADD = (1 << 1),
+ FDMON_IO_URING_REMOVE = (1 << 2),
+ FDMON_IO_URING_DELETE_AIO_HANDLER = (1 << 3),
};
static inline int poll_events_from_pfd(int pfd_events)
@@ -218,6 +219,16 @@ static void fill_sq_ring(AioContext *ctx)
if (flags & FDMON_IO_URING_REMOVE) {
add_poll_remove_sqe(ctx, node);
}
+ if (flags & FDMON_IO_URING_DELETE_AIO_HANDLER) {
+ /*
+ * process_cqe() sets this flag after ADD and REMOVE have been
+ * cleared. They cannot be set again, so they must be clear.
+ */
+ assert(!(flags & FDMON_IO_URING_ADD));
+ assert(!(flags & FDMON_IO_URING_REMOVE));
+
+ QLIST_INSERT_HEAD_RCU(&ctx->deleted_aio_handlers, node, node_deleted);
+ }
}
}
@@ -241,7 +252,12 @@ static bool process_cqe(AioContext *ctx,
*/
flags = qatomic_fetch_and(&node->flags, ~FDMON_IO_URING_REMOVE);
if (flags & FDMON_IO_URING_REMOVE) {
- QLIST_INSERT_HEAD_RCU(&ctx->deleted_aio_handlers, node, node_deleted);
+ if (flags & FDMON_IO_URING_PENDING) {
+ /* Still on ctx->submit_list, defer deletion until fill_sq_ring() */
+ qatomic_or(&node->flags, FDMON_IO_URING_DELETE_AIO_HANDLER);
+ } else {
+ QLIST_INSERT_HEAD_RCU(&ctx->deleted_aio_handlers, node, node_deleted);
+ }
return false;
}
@@ -347,10 +363,13 @@ void fdmon_io_uring_destroy(AioContext *ctx)
unsigned flags = qatomic_fetch_and(&node->flags,
~(FDMON_IO_URING_PENDING |
FDMON_IO_URING_ADD |
- FDMON_IO_URING_REMOVE));
+ FDMON_IO_URING_REMOVE |
+ FDMON_IO_URING_DELETE_AIO_HANDLER));
- if (flags & FDMON_IO_URING_REMOVE) {
- QLIST_INSERT_HEAD_RCU(&ctx->deleted_aio_handlers, node, node_deleted);
+ if ((flags & FDMON_IO_URING_REMOVE) ||
+ (flags & FDMON_IO_URING_DELETE_AIO_HANDLER)) {
+ QLIST_INSERT_HEAD_RCU(&ctx->deleted_aio_handlers,
+ node, node_deleted);
}
QSLIST_REMOVE_HEAD_RCU(&ctx->submit_list, node_submitted);
--
2.51.1
next prev parent reply other threads:[~2025-11-11 21:33 UTC|newest]
Thread overview: 32+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-11 21:32 [PULL v2 00/28] Block layer patches Kevin Wolf
2025-11-11 21:32 ` Kevin Wolf [this message]
2025-11-11 21:32 ` [PULL v2 02/28] aio-posix: fix fdmon-io_uring.c timeout stack variable lifetime Kevin Wolf
2025-11-11 21:32 ` [PULL v2 03/28] aio-posix: fix spurious return from ->wait() due to signals Kevin Wolf
2025-11-11 21:32 ` [PULL v2 04/28] aio-posix: keep polling enabled with fdmon-io_uring.c Kevin Wolf
2025-11-11 21:32 ` [PULL v2 05/28] tests/unit: skip test-nested-aio-poll with io_uring Kevin Wolf
2025-11-11 21:32 ` [PULL v2 06/28] aio-posix: integrate fdmon into glib event loop Kevin Wolf
2025-11-11 21:32 ` [PULL v2 07/28] aio: remove aio_context_use_g_source() Kevin Wolf
2025-11-11 21:32 ` [PULL v2 08/28] aio: free AioContext when aio_context_new() fails Kevin Wolf
2025-11-11 21:32 ` [PULL v2 09/28] aio: add errp argument to aio_context_setup() Kevin Wolf
2025-11-11 21:32 ` [PULL v2 10/28] aio-posix: gracefully handle io_uring_queue_init() failure Kevin Wolf
2025-11-11 21:32 ` [PULL v2 11/28] aio-posix: unindent fdmon_io_uring_destroy() Kevin Wolf
2025-11-11 21:32 ` [PULL v2 12/28] aio-posix: add fdmon_ops->dispatch() Kevin Wolf
2025-11-11 21:32 ` [PULL v2 13/28] aio-posix: add aio_add_sqe() API for user-defined io_uring requests Kevin Wolf
2025-11-11 21:32 ` [PULL v2 14/28] block/io_uring: use aio_add_sqe() Kevin Wolf
2025-11-11 21:32 ` [PULL v2 15/28] block/io_uring: use non-vectored read/write when possible Kevin Wolf
2025-11-11 21:32 ` [PULL v2 16/28] block: replace TABs with space Kevin Wolf
2025-11-11 21:32 ` [PULL v2 17/28] block: Drop detach_subchain for bdrv_replace_node Kevin Wolf
2025-11-11 21:32 ` [PULL v2 18/28] iotests: Test resizing file node under raw with size/offset Kevin Wolf
2025-11-11 21:32 ` [PULL v2 19/28] qemu-img: Fix amend option parse error handling Kevin Wolf
2025-11-11 21:32 ` [PULL v2 20/28] iotests: Run iotests with sanitizers Kevin Wolf
2025-11-11 21:32 ` [PULL v2 21/28] qcow2: rename update_refcount_discard to queue_discard Kevin Wolf
2025-11-11 21:32 ` [PULL v2 22/28] qcow2: put discards in discard queue when discard-no-unref is enabled Kevin Wolf
2025-11-11 21:32 ` [PULL v2 23/28] tests/qemu-iotests/184: Fix skip message for qemu-img without throttle Kevin Wolf
2025-11-11 21:32 ` [PULL v2 24/28] tests/qemu-iotests: Improve the dry run list to speed up thorough testing Kevin Wolf
2025-11-11 21:32 ` [PULL v2 25/28] tests/qemu-iotest: Add more image formats to the " Kevin Wolf
2025-11-11 21:32 ` [PULL v2 26/28] block: Allow drivers to control protocol prefix at creation Kevin Wolf
2025-11-11 21:32 ` [PULL v2 27/28] qcow2, vmdk: Restrict creation with secondary file using protocol Kevin Wolf
2025-11-11 21:32 ` [PULL v2 28/28] qemu-img rebase: don't exceed IO_BUF_SIZE in one operation Kevin Wolf
2025-11-12 16:40 ` Thomas Huth
2025-11-12 16:42 ` Thomas Huth
2025-11-12 15:08 ` [PULL v2 00/28] Block layer patches Richard Henderson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20251111213238.181992-2-kwolf@redhat.com \
--to=kwolf@redhat.com \
--cc=qemu-block@nongnu.org \
--cc=qemu-devel@nongnu.org \
--cc=richard.henderson@linaro.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).