From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from mailman by lists.gnu.org with tmda-scanned (Exim 4.43) id 1LjCbF-0006BB-Hj for qemu-devel@nongnu.org; Mon, 16 Mar 2009 09:11:25 -0400 Received: from exim by lists.gnu.org with spam-scanned (Exim 4.43) id 1LjCb9-00068u-IG for qemu-devel@nongnu.org; Mon, 16 Mar 2009 09:11:23 -0400 Received: from [199.232.76.173] (port=41550 helo=monty-python.gnu.org) by lists.gnu.org with esmtp (Exim 4.43) id 1LjCb9-00068l-A3 for qemu-devel@nongnu.org; Mon, 16 Mar 2009 09:11:19 -0400 Received: from mx2.redhat.com ([66.187.237.31]:39032) by monty-python.gnu.org with esmtp (Exim 4.60) (envelope-from ) id 1LjCb8-00023v-MC for qemu-devel@nongnu.org; Mon, 16 Mar 2009 09:11:19 -0400 From: Avi Kivity Date: Mon, 16 Mar 2009 15:11:09 +0200 Message-Id: <1237209071-26942-4-git-send-email-avi@redhat.com> In-Reply-To: <1237209071-26942-1-git-send-email-avi@redhat.com> References: <1237209071-26942-1-git-send-email-avi@redhat.com> Subject: [Qemu-devel] [PATCH 3/5] Implement cancellation method for dma async I/O Reply-To: qemu-devel@nongnu.org List-Id: qemu-devel.nongnu.org List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: Anthony Liguori Cc: qemu-devel@nongnu.org Move the dma helpers to a private aio pool, and implement a cancellation method for them. Should prevent issues when cancelling I/O while dma is in progress. Signed-off-by: Avi Kivity --- dma-helpers.c | 15 ++++++++++++++- dma.h | 1 + vl.c | 1 + 3 files changed, 16 insertions(+), 1 deletions(-) diff --git a/dma-helpers.c b/dma-helpers.c index b2ade19..19fa4f0 100644 --- a/dma-helpers.c +++ b/dma-helpers.c @@ -10,6 +10,8 @@ #include "dma.h" #include "block_int.h" +static AIOPool dma_aio_pool; + void qemu_sglist_init(QEMUSGList *qsg, int alloc_hint) { qsg->sg = qemu_malloc(alloc_hint * sizeof(ScatterGatherEntry)); @@ -126,7 +128,7 @@ static BlockDriverAIOCB *dma_bdrv_io( DMABlockState *dbs = qemu_malloc(sizeof(*dbs)); dbs->bs = bs; - dbs->acb = qemu_aio_get(bs, cb, opaque); + dbs->acb = qemu_aio_get_pool(&dma_aio_pool, bs, cb, opaque); dbs->sg = sg; dbs->sector_num = sector_num; dbs->sg_cur_index = 0; @@ -153,3 +155,14 @@ BlockDriverAIOCB *dma_bdrv_write(BlockDriverState *bs, return dma_bdrv_io(bs, sg, sector, cb, opaque, 1); } +static void dma_aio_cancel(BlockDriverAIOCB *acb) +{ + DMABlockState *dbs = (DMABlockState *)acb->opaque; + + bdrv_aio_cancel(dbs->acb); +} + +void dma_helper_init(void) +{ + aio_pool_init(&dma_aio_pool, sizeof(BlockDriverAIOCB), dma_aio_cancel); +} diff --git a/dma.h b/dma.h index d596717..7d2ab45 100644 --- a/dma.h +++ b/dma.h @@ -37,5 +37,6 @@ BlockDriverAIOCB *dma_bdrv_read(BlockDriverState *bs, BlockDriverAIOCB *dma_bdrv_write(BlockDriverState *bs, QEMUSGList *sg, uint64_t sector, BlockDriverCompletionFunc *cb, void *opaque); +void dma_helper_init(void); #endif diff --git a/vl.c b/vl.c index 2adfebf..4016c37 100644 --- a/vl.c +++ b/vl.c @@ -5517,6 +5517,7 @@ int main(int argc, char **argv, char **envp) cpu_exec_init_all(tb_size * 1024 * 1024); bdrv_init(); + dma_helper_init(); /* we always create the cdrom drive, even if no disk is there */ -- 1.6.1.1