qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Stefano Garzarella <sgarzare@redhat.com>
To: Stefan Hajnoczi <stefanha@redhat.com>
Cc: Fam Zheng <fam@euphon.net>, Kevin Wolf <kwolf@redhat.com>,
	Maxim Levitsky <maximlevitsky@gmail.com>,
	qemu-block@nongnu.org, oleksandr@redhat.com,
	Julia Suvorova <jusual@redhat.com>,
	qemu-devel@nongnu.org, Markus Armbruster <armbru@redhat.com>,
	Paolo Bonzini <pbonzini@redhat.com>,
	Max Reitz <mreitz@redhat.com>,
	Aarushi Mehta <mehta.aaru20@gmail.com>
Subject: Re: [PATCH v3 08/15] block/file-posix.c: extend to use io_uring
Date: Mon, 13 Jan 2020 12:49:27 +0100	[thread overview]
Message-ID: <20200113114927.zbse5mlua5mq4hct@steredhat> (raw)
In-Reply-To: <20191218163228.1613099-9-stefanha@redhat.com>

On Wed, Dec 18, 2019 at 04:32:21PM +0000, Stefan Hajnoczi wrote:
> From: Aarushi Mehta <mehta.aaru20@gmail.com>
> 
> Signed-off-by: Aarushi Mehta <mehta.aaru20@gmail.com>
> Reviewed-by: Maxim Levitsky <maximlevitsky@gmail.com>
> Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
> ---
>  block/file-posix.c | 95 ++++++++++++++++++++++++++++++++++++----------
>  1 file changed, 75 insertions(+), 20 deletions(-)
> 
> diff --git a/block/file-posix.c b/block/file-posix.c
> index 1b805bd938..a42a90e59d 100644
> --- a/block/file-posix.c
> +++ b/block/file-posix.c
> @@ -156,6 +156,7 @@ typedef struct BDRVRawState {
>      bool has_write_zeroes:1;
>      bool discard_zeroes:1;
>      bool use_linux_aio:1;
> +    bool use_linux_io_uring:1;
>      bool page_cache_inconsistent:1;
>      bool has_fallocate;
>      bool needs_alignment;
> @@ -444,7 +445,7 @@ static QemuOptsList raw_runtime_opts = {
>          {
>              .name = "aio",
>              .type = QEMU_OPT_STRING,
> -            .help = "host AIO implementation (threads, native)",
> +            .help = "host AIO implementation (threads, native, io_uring)",
>          },
>          {
>              .name = "locking",
> @@ -503,9 +504,11 @@ static int raw_open_common(BlockDriverState *bs, QDict *options,
>          goto fail;
>      }
>  
> -    aio_default = (bdrv_flags & BDRV_O_NATIVE_AIO)
> -                  ? BLOCKDEV_AIO_OPTIONS_NATIVE
> -                  : BLOCKDEV_AIO_OPTIONS_THREADS;
> +    if (bdrv_flags & BDRV_O_NATIVE_AIO) {
> +        aio_default = BLOCKDEV_AIO_OPTIONS_NATIVE;
> +    } else {
> +        aio_default = BLOCKDEV_AIO_OPTIONS_THREADS;
> +    }

This is only a cosmetic change?

>      aio = qapi_enum_parse(&BlockdevAioOptions_lookup,
>                            qemu_opt_get(opts, "aio"),
>                            aio_default, &local_err);
> @@ -514,7 +517,11 @@ static int raw_open_common(BlockDriverState *bs, QDict *options,
>          ret = -EINVAL;
>          goto fail;
>      }
> +
>      s->use_linux_aio = (aio == BLOCKDEV_AIO_OPTIONS_NATIVE);
> +#ifdef CONFIG_LINUX_IO_URING
> +    s->use_linux_io_uring = (aio == BLOCKDEV_AIO_OPTIONS_IO_URING);
> +#endif
>  
>      locking = qapi_enum_parse(&OnOffAuto_lookup,
>                                qemu_opt_get(opts, "locking"),
> @@ -578,7 +585,7 @@ static int raw_open_common(BlockDriverState *bs, QDict *options,
>      s->shared_perm = BLK_PERM_ALL;
>  
>  #ifdef CONFIG_LINUX_AIO
> -     /* Currently Linux does AIO only for files opened with O_DIRECT */
> +    /* Currently Linux does AIO only for files opened with O_DIRECT */

Also this is a not related fix, if you respin maybe we should split in a
new patch or say something in the commit message.

>      if (s->use_linux_aio) {
>          if (!(s->open_flags & O_DIRECT)) {
>              error_setg(errp, "aio=native was specified, but it requires "
> @@ -600,6 +607,22 @@ static int raw_open_common(BlockDriverState *bs, QDict *options,
>      }
>  #endif /* !defined(CONFIG_LINUX_AIO) */
>  
> +#ifdef CONFIG_LINUX_IO_URING
> +    if (s->use_linux_io_uring) {
> +        if (!aio_setup_linux_io_uring(bdrv_get_aio_context(bs), errp)) {
> +            error_prepend(errp, "Unable to use io_uring: ");
> +            goto fail;
> +        }
> +    }
> +#else
> +    if (s->use_linux_io_uring) {
> +        error_setg(errp, "aio=io_uring was specified, but is not supported "
> +                         "in this build.");
> +        ret = -EINVAL;
> +        goto fail;
> +    }
> +#endif /* !defined(CONFIG_LINUX_IO_URING) */
> +
>      s->has_discard = true;
>      s->has_write_zeroes = true;
>      if ((bs->open_flags & BDRV_O_NOCACHE) != 0) {
> @@ -1877,21 +1900,25 @@ static int coroutine_fn raw_co_prw(BlockDriverState *bs, uint64_t offset,
>          return -EIO;
>  
>      /*
> -     * Check if the underlying device requires requests to be aligned,
> -     * and if the request we are trying to submit is aligned or not.
> -     * If this is the case tell the low-level driver that it needs
> -     * to copy the buffer.
> +     * When using O_DIRECT, the request must be aligned to be able to use
> +     * either libaio or io_uring interface. If not fail back to regular thread
> +     * pool read/write code which emulates this for us if we
> +     * set QEMU_AIO_MISALIGNED.
>       */
> -    if (s->needs_alignment) {
> -        if (!bdrv_qiov_is_aligned(bs, qiov)) {
> -            type |= QEMU_AIO_MISALIGNED;
> +    if (s->needs_alignment && !bdrv_qiov_is_aligned(bs, qiov)) {
> +        type |= QEMU_AIO_MISALIGNED;
> +#ifdef CONFIG_LINUX_IO_URING
> +    } else if (s->use_linux_io_uring) {
> +        LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
> +        assert(qiov->size == bytes);
> +        return luring_co_submit(bs, aio, s->fd, offset, qiov, type);
> +#endif
>  #ifdef CONFIG_LINUX_AIO
> -        } else if (s->use_linux_aio) {
> -            LinuxAioState *aio = aio_get_linux_aio(bdrv_get_aio_context(bs));
> -            assert(qiov->size == bytes);
> -            return laio_co_submit(bs, aio, s->fd, offset, qiov, type);
> +    } else if (s->use_linux_aio) {

This code block was executed if "s->needs_alignment" was true, now we don't
check it. Could this be a problem?

> +        LinuxAioState *aio = aio_get_linux_aio(bdrv_get_aio_context(bs));
> +        assert(qiov->size == bytes);
> +        return laio_co_submit(bs, aio, s->fd, offset, qiov, type);
>  #endif
> -        }
>      }
>  
>      acb = (RawPosixAIOData) {
> @@ -1927,24 +1954,36 @@ static int coroutine_fn raw_co_pwritev(BlockDriverState *bs, uint64_t offset,
>  
>  static void raw_aio_plug(BlockDriverState *bs)
>  {
> +    BDRVRawState __attribute__((unused)) *s = bs->opaque;
>  #ifdef CONFIG_LINUX_AIO
> -    BDRVRawState *s = bs->opaque;
>      if (s->use_linux_aio) {
>          LinuxAioState *aio = aio_get_linux_aio(bdrv_get_aio_context(bs));
>          laio_io_plug(bs, aio);
>      }
>  #endif
> +#ifdef CONFIG_LINUX_IO_URING
> +    if (s->use_linux_io_uring) {
> +        LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
> +        luring_io_plug(bs, aio);
> +    }
> +#endif
>  }
>  
>  static void raw_aio_unplug(BlockDriverState *bs)
>  {
> +    BDRVRawState __attribute__((unused)) *s = bs->opaque;
>  #ifdef CONFIG_LINUX_AIO
> -    BDRVRawState *s = bs->opaque;
>      if (s->use_linux_aio) {
>          LinuxAioState *aio = aio_get_linux_aio(bdrv_get_aio_context(bs));
>          laio_io_unplug(bs, aio);
>      }
>  #endif
> +#ifdef CONFIG_LINUX_IO_URING
> +    if (s->use_linux_io_uring) {
> +        LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
> +        luring_io_unplug(bs, aio);
> +    }
> +#endif
>  }
>  
>  static int raw_co_flush_to_disk(BlockDriverState *bs)
> @@ -1964,14 +2003,20 @@ static int raw_co_flush_to_disk(BlockDriverState *bs)
>          .aio_type       = QEMU_AIO_FLUSH,
>      };
>  
> +#ifdef CONFIG_LINUX_IO_URING
> +    if (s->use_linux_io_uring) {
> +        LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
> +        return luring_co_submit(bs, aio, s->fd, 0, NULL, QEMU_AIO_FLUSH);
> +    }
> +#endif
>      return raw_thread_pool_submit(bs, handle_aiocb_flush, &acb);
>  }
>  
>  static void raw_aio_attach_aio_context(BlockDriverState *bs,
>                                         AioContext *new_context)
>  {
> +    BDRVRawState __attribute__((unused)) *s = bs->opaque;
>  #ifdef CONFIG_LINUX_AIO
> -    BDRVRawState *s = bs->opaque;
>      if (s->use_linux_aio) {
>          Error *local_err = NULL;
>          if (!aio_setup_linux_aio(new_context, &local_err)) {
> @@ -1981,6 +2026,16 @@ static void raw_aio_attach_aio_context(BlockDriverState *bs,
>          }
>      }
>  #endif
> +#ifdef CONFIG_LINUX_IO_URING
> +    if (s->use_linux_io_uring) {
> +        Error *local_err;
> +        if (!aio_setup_linux_io_uring(new_context, &local_err)) {
> +            error_reportf_err(local_err, "Unable to use linux io_uring, "
> +                                         "falling back to thread pool: ");
> +            s->use_linux_io_uring = false;
> +        }
> +    }
> +#endif
>  }
>  
>  static void raw_close(BlockDriverState *bs)
> -- 
> 2.23.0
> 
> 



  reply	other threads:[~2020-01-13 11:50 UTC|newest]

Thread overview: 22+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-12-18 16:32 [PATCH v3 00/15] io_uring: add Linux io_uring AIO engine Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 01/15] configure: permit use of io_uring Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 02/15] qapi/block-core: add option for io_uring Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 03/15] block/block: add BDRV flag " Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 04/15] block/io_uring: implements interfaces " Stefan Hajnoczi
2020-01-13 11:24   ` Stefano Garzarella
2020-01-14 10:40     ` Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 05/15] stubs: add stubs for io_uring interface Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 06/15] util/async: add aio interfaces for io_uring Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 07/15] blockdev: adds bdrv_parse_aio to use io_uring Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 08/15] block/file-posix.c: extend " Stefan Hajnoczi
2020-01-13 11:49   ` Stefano Garzarella [this message]
2020-01-14 10:37     ` Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 09/15] block: add trace events for io_uring Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 10/15] block/io_uring: adds userspace completion polling Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 11/15] qemu-io: adds option to use aio engine Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 12/15] qemu-img: adds option to use aio engine for benchmarking Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 13/15] qemu-nbd: adds option for aio engines Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 14/15] tests/qemu-iotests: enable testing with aio options Stefan Hajnoczi
2019-12-18 16:32 ` [PATCH v3 15/15] tests/qemu-iotests: use AIOMODE with various tests Stefan Hajnoczi
2020-01-10  9:55 ` [PATCH v3 00/15] io_uring: add Linux io_uring AIO engine Stefan Hajnoczi
2020-01-13 11:58 ` Stefano Garzarella

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20200113114927.zbse5mlua5mq4hct@steredhat \
    --to=sgarzare@redhat.com \
    --cc=armbru@redhat.com \
    --cc=fam@euphon.net \
    --cc=jusual@redhat.com \
    --cc=kwolf@redhat.com \
    --cc=maximlevitsky@gmail.com \
    --cc=mehta.aaru20@gmail.com \
    --cc=mreitz@redhat.com \
    --cc=oleksandr@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=qemu-block@nongnu.org \
    --cc=qemu-devel@nongnu.org \
    --cc=stefanha@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).