qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: "Benoît Canet" <benoit.canet@irqsave.net>
To: Fam Zheng <famz@redhat.com>
Cc: Kevin Wolf <kwolf@redhat.com>,
	Chrysostomos Nanakos <cnanakos@grnet.gr>,
	Ronnie Sahlberg <ronniesahlberg@gmail.com>,
	Peter Lieven <pl@kamp.de>,
	qemu-devel@nongnu.org, Paolo Bonzini <pbonzini@redhat.com>,
	Stefan Hajnoczi <stefanha@redhat.com>,
	Josh Durgin <josh.durgin@inktank.com>,
	Liu Yuan <namei.unix@gmail.com>,
	MORITA Kazutaka <morita.kazutaka@lab.ntt.co.jp>
Subject: Re: [Qemu-devel] [PATCH v4 05/20] thread-pool: Convert thread_pool_aiocb_info.cancel to cancel_async
Date: Thu, 4 Sep 2014 17:29:13 +0200	[thread overview]
Message-ID: <20140904152913.GD8094@irqsave.net> (raw)
In-Reply-To: <1409743435-21155-6-git-send-email-famz@redhat.com>

The Wednesday 03 Sep 2014 à 19:23:40 (+0800), Fam Zheng wrote :

> The .cancel_async has the same the first half with .cancel: try to steal

"The .cancel_async share the same first half with .cancel" ?

> the request if not submitted yet. In this case set the elem to
> THREAD_DONE status and ret to -ECANCELED, which means
> thread_pool_completion_bh will call the cb with -ECANCELED.
> 
> If the request is already submitted, do nothing, as we know the normal
> completion will happen in the future.
> 
> Testing code update:
> 
> Before, done_cb is only called if the request is already submitted by
> thread pool. Now done_cb is always called, even before it is submitted,
> because we emulate bdrv_aio_cancel with bdrv_aio_cancel_async. So also
> update the test criteria accordingly.
> 
> Signed-off-by: Fam Zheng <famz@redhat.com>
> ---
>  tests/test-thread-pool.c | 34 ++++++++++++++++++++++++++--------
>  thread-pool.c            | 32 ++++++++++++++------------------
>  2 files changed, 40 insertions(+), 26 deletions(-)
> 
> diff --git a/tests/test-thread-pool.c b/tests/test-thread-pool.c
> index f40b7fc..ed2b25b 100644
> --- a/tests/test-thread-pool.c
> +++ b/tests/test-thread-pool.c
> @@ -33,7 +33,7 @@ static int long_cb(void *opaque)
>  static void done_cb(void *opaque, int ret)
>  {
>      WorkerTestData *data = opaque;
> -    g_assert_cmpint(data->ret, ==, -EINPROGRESS);
> +    g_assert(data->ret == -EINPROGRESS || data->ret == -ECANCELED);
>      data->ret = ret;
>      data->aiocb = NULL;
>  
> @@ -132,7 +132,7 @@ static void test_submit_many(void)
>      }
>  }
>  
> -static void test_cancel(void)
> +static void do_test_cancel(bool sync)
>  {
>      WorkerTestData data[100];
>      int num_canceled;
> @@ -170,18 +170,25 @@ static void test_cancel(void)
>      for (i = 0; i < 100; i++) {
>          if (atomic_cmpxchg(&data[i].n, 0, 3) == 0) {
>              data[i].ret = -ECANCELED;
> -            bdrv_aio_cancel(data[i].aiocb);
> -            active--;
> +            if (sync) {
> +                bdrv_aio_cancel(data[i].aiocb);
> +            } else {
> +                bdrv_aio_cancel_async(data[i].aiocb);
> +            }
>              num_canceled++;
>          }
>      }
>      g_assert_cmpint(active, >, 0);
>      g_assert_cmpint(num_canceled, <, 100);
>  
> -    /* Canceling the others will be a blocking operation.  */
>      for (i = 0; i < 100; i++) {
>          if (data[i].aiocb && data[i].n != 3) {
> -            bdrv_aio_cancel(data[i].aiocb);
> +            if (sync) {
> +                /* Canceling the others will be a blocking operation.  */
> +                bdrv_aio_cancel(data[i].aiocb);
> +            } else {
> +                bdrv_aio_cancel_async(data[i].aiocb);
> +            }
>          }
>      }
>  
> @@ -193,15 +200,25 @@ static void test_cancel(void)
>      for (i = 0; i < 100; i++) {
>          if (data[i].n == 3) {
>              g_assert_cmpint(data[i].ret, ==, -ECANCELED);
> -            g_assert(data[i].aiocb != NULL);
> +            g_assert(data[i].aiocb == NULL);
>          } else {
>              g_assert_cmpint(data[i].n, ==, 2);
> -            g_assert_cmpint(data[i].ret, ==, 0);
> +            g_assert(data[i].ret == 0 || data[i].ret == -ECANCELED);
>              g_assert(data[i].aiocb == NULL);
>          }
>      }
>  }
>  
> +static void test_cancel(void)
> +{
> +    do_test_cancel(true);
> +}
> +
> +static void test_cancel_async(void)
> +{
> +    do_test_cancel(false);
> +}
> +
>  int main(int argc, char **argv)
>  {
>      int ret;
> @@ -217,6 +234,7 @@ int main(int argc, char **argv)
>      g_test_add_func("/thread-pool/submit-co", test_submit_co);
>      g_test_add_func("/thread-pool/submit-many", test_submit_many);
>      g_test_add_func("/thread-pool/cancel", test_cancel);
> +    g_test_add_func("/thread-pool/cancel-async", test_cancel_async);
>  
>      ret = g_test_run();
>  
> diff --git a/thread-pool.c b/thread-pool.c
> index 23888dc..6afd343 100644
> --- a/thread-pool.c
> +++ b/thread-pool.c
> @@ -32,7 +32,6 @@ enum ThreadState {
>      THREAD_QUEUED,
>      THREAD_ACTIVE,
>      THREAD_DONE,
> -    THREAD_CANCELED,
>  };
>  
>  struct ThreadPoolElement {
> @@ -59,7 +58,6 @@ struct ThreadPool {
>      AioContext *ctx;
>      QEMUBH *completion_bh;
>      QemuMutex lock;
> -    QemuCond check_cancel;
>      QemuCond worker_stopped;
>      QemuSemaphore sem;
>      int max_threads;
> @@ -74,7 +72,6 @@ struct ThreadPool {
>      int idle_threads;
>      int new_threads;     /* backlog of threads we need to create */
>      int pending_threads; /* threads created but not running yet */
> -    int pending_cancellations; /* whether we need a cond_broadcast */
>      bool stopping;
>  };
>  
> @@ -114,9 +111,6 @@ static void *worker_thread(void *opaque)
>          req->state = THREAD_DONE;
>  
>          qemu_mutex_lock(&pool->lock);
> -        if (pool->pending_cancellations) {
> -            qemu_cond_broadcast(&pool->check_cancel);
> -        }
>  
>          qemu_bh_schedule(pool->completion_bh);
>      }
> @@ -174,7 +168,7 @@ static void thread_pool_completion_bh(void *opaque)
>  
>  restart:
>      QLIST_FOREACH_SAFE(elem, &pool->head, all, next) {
> -        if (elem->state != THREAD_CANCELED && elem->state != THREAD_DONE) {
> +        if (elem->state != THREAD_DONE) {
>              continue;
>          }
>          if (elem->state == THREAD_DONE) {
> @@ -218,22 +212,26 @@ static void thread_pool_cancel(BlockDriverAIOCB *acb)
>           */
>          qemu_sem_timedwait(&pool->sem, 0) == 0) {
>          QTAILQ_REMOVE(&pool->request_list, elem, reqs);
> -        elem->state = THREAD_CANCELED;
>          qemu_bh_schedule(pool->completion_bh);
> -    } else {
> -        pool->pending_cancellations++;
> -        while (elem->state != THREAD_CANCELED && elem->state != THREAD_DONE) {
> -            qemu_cond_wait(&pool->check_cancel, &pool->lock);
> -        }
> -        pool->pending_cancellations--;
> +
> +        elem->state = THREAD_DONE;
> +        elem->ret = -ECANCELED;
>      }
> +
>      qemu_mutex_unlock(&pool->lock);
> -    thread_pool_completion_bh(pool);
> +}
> +
> +static AioContext *thread_pool_get_aio_context(BlockDriverAIOCB *acb)
> +{
> +    ThreadPoolElement *elem = (ThreadPoolElement *)acb;
> +    ThreadPool *pool = elem->pool;
> +    return pool->ctx;
>  }
>  
>  static const AIOCBInfo thread_pool_aiocb_info = {
>      .aiocb_size         = sizeof(ThreadPoolElement),
> -    .cancel             = thread_pool_cancel,
> +    .cancel_async       = thread_pool_cancel,
> +    .get_aio_context    = thread_pool_get_aio_context,
>  };
>  
>  BlockDriverAIOCB *thread_pool_submit_aio(ThreadPool *pool,
> @@ -300,7 +298,6 @@ static void thread_pool_init_one(ThreadPool *pool, AioContext *ctx)
>      pool->ctx = ctx;
>      pool->completion_bh = aio_bh_new(ctx, thread_pool_completion_bh, pool);
>      qemu_mutex_init(&pool->lock);
> -    qemu_cond_init(&pool->check_cancel);
>      qemu_cond_init(&pool->worker_stopped);
>      qemu_sem_init(&pool->sem, 0);
>      pool->max_threads = 64;
> @@ -343,7 +340,6 @@ void thread_pool_free(ThreadPool *pool)
>  
>      qemu_bh_delete(pool->completion_bh);
>      qemu_sem_destroy(&pool->sem);
> -    qemu_cond_destroy(&pool->check_cancel);
>      qemu_cond_destroy(&pool->worker_stopped);
>      qemu_mutex_destroy(&pool->lock);
>      g_free(pool);
> -- 
> 2.1.0.27.g96db324
> 
> 

  reply	other threads:[~2014-09-04 15:30 UTC|newest]

Thread overview: 32+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-09-03 11:23 [Qemu-devel] [PATCH v4 00/20] block: Asynchronous request cancellation Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 01/20] block: Add refcnt in BlockDriverAIOCB Fam Zheng
2014-09-04 15:07   ` Benoît Canet
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 02/20] block: Add bdrv_aio_cancel_async Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 03/20] block: Drop bdrv_em_co_aiocb_info.cancel Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 04/20] block: Convert bdrv_em_aiocb_info.cancel to .cancel_async Fam Zheng
2014-09-03 11:28   ` Paolo Bonzini
2014-09-04  1:20     ` Fam Zheng
2014-09-04  5:45       ` Paolo Bonzini
2014-09-04 15:21   ` Benoît Canet
2014-09-05 10:55     ` Fam Zheng
2014-09-05 11:06       ` Benoît Canet
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 05/20] thread-pool: Convert thread_pool_aiocb_info.cancel to cancel_async Fam Zheng
2014-09-04 15:29   ` Benoît Canet [this message]
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 06/20] linux-aio: Convert laio_aiocb_info.cancel to .cancel_async Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 07/20] dma: Check iov pointer before unmap memory Fam Zheng
2014-09-03 11:38   ` Paolo Bonzini
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 08/20] dma: Convert dma_aiocb_info.cancel to .cancel_async Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 09/20] iscsi: Convert iscsi_aiocb_info.cancel " Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 10/20] archipelago: Drop archipelago_aiocb_info.cancel Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 11/20] blkdebug: Convert blkdebug_aiocb_info.cancel to .cancel_async Fam Zheng
2014-09-03 11:42   ` Paolo Bonzini
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 12/20] blkverify: Drop blkverify_aiocb_info.cancel Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 13/20] curl: Drop curl_aiocb_info.cancel Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 14/20] qed: Drop qed_aiocb_info.cancel Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 15/20] quorum: Convert quorum_aiocb_info.cancel to .cancel_async Fam Zheng
2014-09-04 15:40   ` Benoît Canet
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 16/20] rbd: Drop rbd_aiocb_info.cancel Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 17/20] sheepdog: Convert sd_aiocb_info.cancel to .cancel_async Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 18/20] win32-aio: Drop win32_aiocb_info.cancel Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 19/20] ide: Convert trim_aiocb_info.cancel to .cancel_async Fam Zheng
2014-09-03 11:23 ` [Qemu-devel] [PATCH v4 20/20] block: Drop AIOCBInfo.cancel Fam Zheng

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20140904152913.GD8094@irqsave.net \
    --to=benoit.canet@irqsave.net \
    --cc=cnanakos@grnet.gr \
    --cc=famz@redhat.com \
    --cc=josh.durgin@inktank.com \
    --cc=kwolf@redhat.com \
    --cc=morita.kazutaka@lab.ntt.co.jp \
    --cc=namei.unix@gmail.com \
    --cc=pbonzini@redhat.com \
    --cc=pl@kamp.de \
    --cc=qemu-devel@nongnu.org \
    --cc=ronniesahlberg@gmail.com \
    --cc=stefanha@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).