From: "Philippe Mathieu-Daudé" <philmd@redhat.com>
To: qemu-devel@nongnu.org, Stefan Hajnoczi <stefanha@redhat.com>
Cc: "Fam Zheng" <fam@euphon.net>, "Kevin Wolf" <kwolf@redhat.com>,
qemu-block@nongnu.org, "Max Reitz" <mreitz@redhat.com>,
"Maxim Levitsky" <mlevitsk@redhat.com>,
"Philippe Mathieu-Daudé" <philmd@redhat.com>
Subject: [RFC PATCH 15/17] block/nvme: Use per-queue AIO context
Date: Thu, 25 Jun 2020 20:48:36 +0200 [thread overview]
Message-ID: <20200625184838.28172-16-philmd@redhat.com> (raw)
In-Reply-To: <20200625184838.28172-1-philmd@redhat.com>
To be able to use multiple queues on the same hardware,
we need to have each queue able to receive IRQ notifications
in the correct AIO context.
The context has to be proper to each queue, not to the block
driver. Move aio_context from BDRVNVMeState to NVMeQueuePair.
Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com>
---
RFC because I'm not familiar with AIO context
block/nvme.c | 35 ++++++++++++++++++-----------------
1 file changed, 18 insertions(+), 17 deletions(-)
diff --git a/block/nvme.c b/block/nvme.c
index ac933cafd0..0f7cc568ef 100644
--- a/block/nvme.c
+++ b/block/nvme.c
@@ -51,6 +51,7 @@ typedef struct {
} NVMeRequest;
typedef struct {
+ AioContext *aio_context;
CoQueue free_req_queue;
QemuMutex lock;
@@ -93,7 +94,6 @@ QEMU_BUILD_BUG_ON(offsetof(NVMeRegs, doorbells) != 0x1000);
#define QUEUE_INDEX_IO(n) (1 + n)
typedef struct {
- AioContext *aio_context;
QEMUVFIOState *vfio;
NVMeRegs *regs;
/* The submission/completion queue pairs.
@@ -190,6 +190,7 @@ static void nvme_free_req_queue_cb(void *opaque)
}
static NVMeQueuePair *nvme_create_queue_pair(BDRVNVMeState *s,
+ AioContext *aio_context,
int idx, int size,
Error **errp)
{
@@ -207,6 +208,7 @@ static NVMeQueuePair *nvme_create_queue_pair(BDRVNVMeState *s,
if (!q->prp_list_pages) {
goto fail;
}
+ q->aio_context = aio_context;
memset(q->prp_list_pages, 0, s->page_size * NVME_QUEUE_SIZE);
qemu_mutex_init(&q->lock);
q->index = idx;
@@ -365,7 +367,7 @@ static bool nvme_process_completion(BDRVNVMeState *s, NVMeQueuePair *q)
smp_mb_release();
*q->cq.doorbell = cpu_to_le32(q->cq.head);
if (!qemu_co_queue_empty(&q->free_req_queue)) {
- replay_bh_schedule_oneshot_event(s->aio_context,
+ replay_bh_schedule_oneshot_event(q->aio_context,
nvme_free_req_queue_cb, q);
}
}
@@ -419,7 +421,6 @@ static void nvme_cmd_sync_cb(void *opaque, int ret)
static int nvme_cmd_sync(BlockDriverState *bs, NVMeQueuePair *q,
NvmeCmd *cmd)
{
- AioContext *aio_context = bdrv_get_aio_context(bs);
NVMeRequest *req;
BDRVNVMeState *s = bs->opaque;
int ret = -EINPROGRESS;
@@ -429,7 +430,7 @@ static int nvme_cmd_sync(BlockDriverState *bs, NVMeQueuePair *q,
}
nvme_submit_command(s, q, req, cmd, nvme_cmd_sync_cb, &ret);
- AIO_WAIT_WHILE(aio_context, ret == -EINPROGRESS);
+ AIO_WAIT_WHILE(q->aio_context, ret == -EINPROGRESS);
return ret;
}
@@ -547,7 +548,8 @@ static void nvme_handle_event(EventNotifier *n)
nvme_poll_queues(s);
}
-static bool nvme_add_io_queue(BlockDriverState *bs, Error **errp)
+static bool nvme_add_io_queue(BlockDriverState *bs,
+ AioContext *aio_context, Error **errp)
{
BDRVNVMeState *s = bs->opaque;
int n = s->nr_queues;
@@ -555,7 +557,7 @@ static bool nvme_add_io_queue(BlockDriverState *bs, Error **errp)
NvmeCmd cmd;
int queue_size = NVME_QUEUE_SIZE;
- q = nvme_create_queue_pair(s, n, queue_size, errp);
+ q = nvme_create_queue_pair(s, aio_context, n, queue_size, errp);
if (!q) {
return false;
}
@@ -600,6 +602,7 @@ static int nvme_init(BlockDriverState *bs, const char *device, int namespace,
Error **errp)
{
BDRVNVMeState *s = bs->opaque;
+ AioContext *aio_context = bdrv_get_aio_context(bs);
int ret;
uint64_t cap;
uint64_t timeout_ms;
@@ -610,7 +613,6 @@ static int nvme_init(BlockDriverState *bs, const char *device, int namespace,
qemu_co_queue_init(&s->dma_flush_queue);
s->device = g_strdup(device);
s->nsid = namespace;
- s->aio_context = bdrv_get_aio_context(bs);
ret = event_notifier_init(&s->irq_notifier, 0);
if (ret) {
error_setg(errp, "Failed to init event notifier");
@@ -660,7 +662,7 @@ static int nvme_init(BlockDriverState *bs, const char *device, int namespace,
/* Set up admin queue. */
s->queues = g_new(NVMeQueuePair *, 1);
- s->queues[QUEUE_INDEX_ADMIN] = nvme_create_queue_pair(s, 0,
+ s->queues[QUEUE_INDEX_ADMIN] = nvme_create_queue_pair(s, aio_context, 0,
NVME_QUEUE_SIZE,
errp);
if (!s->queues[QUEUE_INDEX_ADMIN]) {
@@ -695,7 +697,7 @@ static int nvme_init(BlockDriverState *bs, const char *device, int namespace,
if (ret) {
goto out;
}
- aio_set_event_notifier(bdrv_get_aio_context(bs), &s->irq_notifier,
+ aio_set_event_notifier(aio_context, &s->irq_notifier,
false, nvme_handle_event, nvme_poll_cb);
nvme_identify(bs, namespace, &local_err);
@@ -706,7 +708,7 @@ static int nvme_init(BlockDriverState *bs, const char *device, int namespace,
}
/* Set up command queues. */
- if (!nvme_add_io_queue(bs, errp)) {
+ if (!nvme_add_io_queue(bs, aio_context, errp)) {
ret = -EIO;
}
out:
@@ -775,11 +777,11 @@ static void nvme_close(BlockDriverState *bs)
BDRVNVMeState *s = bs->opaque;
for (i = 0; i < s->nr_queues; ++i) {
+ aio_set_event_notifier(s->queues[i]->aio_context,
+ &s->irq_notifier, false, NULL, NULL);
nvme_free_queue_pair(s->queues[i]);
}
g_free(s->queues);
- aio_set_event_notifier(bdrv_get_aio_context(bs), &s->irq_notifier,
- false, NULL, NULL);
event_notifier_cleanup(&s->irq_notifier);
qemu_vfio_pci_unmap_bar(s->vfio, 0, (void *)s->regs, 0, NVME_BAR_SIZE);
qemu_vfio_close(s->vfio);
@@ -992,7 +994,7 @@ static coroutine_fn int nvme_co_prw_aligned(BlockDriverState *bs,
.cdw12 = cpu_to_le32(cdw12),
};
NVMeCoData data = {
- .ctx = bdrv_get_aio_context(bs),
+ .ctx = ioq->aio_context,
.ret = -EINPROGRESS,
};
@@ -1101,7 +1103,7 @@ static coroutine_fn int nvme_co_flush(BlockDriverState *bs)
.nsid = cpu_to_le32(s->nsid),
};
NVMeCoData data = {
- .ctx = bdrv_get_aio_context(bs),
+ .ctx = ioq->aio_context,
.ret = -EINPROGRESS,
};
@@ -1142,7 +1144,7 @@ static coroutine_fn int nvme_co_pwrite_zeroes(BlockDriverState *bs,
};
NVMeCoData data = {
- .ctx = bdrv_get_aio_context(bs),
+ .ctx = ioq->aio_context,
.ret = -EINPROGRESS,
};
@@ -1192,7 +1194,7 @@ static int coroutine_fn nvme_co_pdiscard(BlockDriverState *bs,
};
NVMeCoData data = {
- .ctx = bdrv_get_aio_context(bs),
+ .ctx = ioq->aio_context,
.ret = -EINPROGRESS,
};
@@ -1289,7 +1291,6 @@ static void nvme_attach_aio_context(BlockDriverState *bs,
{
BDRVNVMeState *s = bs->opaque;
- s->aio_context = new_context;
aio_set_event_notifier(new_context, &s->irq_notifier,
false, nvme_handle_event, nvme_poll_cb);
}
--
2.21.3
next prev parent reply other threads:[~2020-06-25 19:02 UTC|newest]
Thread overview: 43+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-25 18:48 [PATCH 00/17] block/nvme: Various cleanups required to use multiple queues Philippe Mathieu-Daudé
2020-06-25 18:48 ` [PATCH 01/17] block/nvme: Avoid further processing if trace event not enabled Philippe Mathieu-Daudé
2020-06-26 10:36 ` Stefan Hajnoczi
2020-06-26 14:02 ` Philippe Mathieu-Daudé
2020-06-29 13:02 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 02/17] block/nvme: Let nvme_create_queue_pair() fail gracefully Philippe Mathieu-Daudé
2020-06-26 11:11 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 03/17] block/nvme: Define QUEUE_INDEX macros to ease code review Philippe Mathieu-Daudé
2020-06-26 11:12 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 04/17] block/nvme: Be explicit we share NvmeIdCtrl / NvmeIdNs structures Philippe Mathieu-Daudé
2020-06-26 11:19 ` Stefan Hajnoczi
2020-06-26 12:45 ` Philippe Mathieu-Daudé
2020-06-25 18:48 ` [PATCH 05/17] block/nvme: Replace qemu_try_blockalign0 by qemu_try_blockalign/memset Philippe Mathieu-Daudé
2020-06-26 12:20 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 06/17] block/nvme: Replace qemu_try_blockalign(bs) by qemu_try_memalign(pg_sz) Philippe Mathieu-Daudé
2020-06-26 12:24 ` Stefan Hajnoczi
2020-06-26 12:48 ` Philippe Mathieu-Daudé
2020-06-29 13:07 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 07/17] block/nvme: Move code around Philippe Mathieu-Daudé
2020-06-26 12:25 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 08/17] block/nvme: Use correct type void* Philippe Mathieu-Daudé
2020-06-26 12:31 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 09/17] block/nvme: Remove unused argument from nvme_free_queue_pair() Philippe Mathieu-Daudé
2020-06-26 12:31 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 10/17] block/nvme: Simplify nvme_init_queue() arguments Philippe Mathieu-Daudé
2020-06-26 12:31 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 11/17] block/nvme: Simplify nvme_create_queue_pair() arguments Philippe Mathieu-Daudé
2020-06-26 12:31 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 12/17] block/nvme: Simplify nvme_kick trace event Philippe Mathieu-Daudé
2020-06-26 12:33 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 13/17] block/nvme: Simplify completion trace events Philippe Mathieu-Daudé
2020-06-26 12:34 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 14/17] block/nvme: Replace BDRV_POLL_WHILE by AIO_WAIT_WHILE Philippe Mathieu-Daudé
2020-06-26 12:35 ` Stefan Hajnoczi
2020-06-25 18:48 ` Philippe Mathieu-Daudé [this message]
2020-06-26 12:42 ` [RFC PATCH 15/17] block/nvme: Use per-queue AIO context Stefan Hajnoczi
2020-06-26 12:59 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 16/17] block/nvme: Check BDRVNVMeState::plugged out of nvme_kick() Philippe Mathieu-Daudé
2020-06-26 12:43 ` Stefan Hajnoczi
2020-06-25 18:48 ` [PATCH 17/17] block/nvme: Check BDRVNVMeState::plugged out of nvme_process_completion Philippe Mathieu-Daudé
2020-06-26 12:46 ` Stefan Hajnoczi
2020-06-25 19:27 ` [PATCH 00/17] block/nvme: Various cleanups required to use multiple queues no-reply
2020-06-26 9:18 ` Philippe Mathieu-Daudé
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200625184838.28172-16-philmd@redhat.com \
--to=philmd@redhat.com \
--cc=fam@euphon.net \
--cc=kwolf@redhat.com \
--cc=mlevitsk@redhat.com \
--cc=mreitz@redhat.com \
--cc=qemu-block@nongnu.org \
--cc=qemu-devel@nongnu.org \
--cc=stefanha@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).