From: Paolo Bonzini <pbonzini@redhat.com>
To: Ming Lei <tom.leiming@gmail.com>
Cc: Linux SCSI List <linux-scsi@vger.kernel.org>,
Wanlong Gao <gaowanlong@cn.fujitsu.com>,
"James E.J. Bottomley" <JBottomley@parallels.com>,
Rusty Russell <rusty@rustcorp.com.au>
Subject: Re: virtio-scsi: two questions related with picking up queue
Date: Wed, 07 May 2014 18:43:45 +0200 [thread overview]
Message-ID: <536A62C1.6000905@redhat.com> (raw)
In-Reply-To: <20140508002437.0dd549e8@tom-ThinkPad-T410>
Il 07/05/2014 18:24, Ming Lei ha scritto:
> On Tue, 06 May 2014 15:17:15 +0200
> Paolo Bonzini <pbonzini@redhat.com> wrote:
>
>> Il 06/05/2014 11:26, Ming Lei ha scritto:
>>> Hi Paolo and All,
>>>
>>> One question is about ACCESS_ONCE() in virtscsi_pick_vq(),
>>> looks it needn't since both reading and writing tgt->req_vq holds
>>> tgt->tgt_lock.
>>
>> You're right. It should be possible to avoid the lock in
>> virtscsi_pick_vq like this:
>>
>> value = atomic_read(&tgt->reqs);
>> retry:
>> if (value != 0) {
>> old_value = atomic_cmpxchg(&tgt->regs, value, value + 1)
>> if (old_value != value) {
>> value = old_value;
>> goto retry;
>> }
>>
>> smp_mb__after_atomic_cmpxchg(); // you get the idea :)
>> vq = ACCESS_ONCE(tgt->req_vq);
>> } else {
>> spin_lock_irqsave(&tgt->tgt_lock, flags);
>>
>> // tgt->reqs may not be 0 anymore, need to recheck
>> value = atomic_read(&tgt->reqs);
>> if (atomic_read(&tgt->reqs) != 0) {
>> spin_unlock_irqrestore(&tgt->tgt_lock, flags);
>> goto retry;
>> }
>>
>> // tgt->reqs now will remain fixed to 0.
>> ...
>> tgt->req_vq = vq = ...;
>> smp_wmb();
>> atomic_set(&tgt->reqs, 1);
>> spin_unlock_irqrestore(&tgt->tgt_lock, flags);
>> }
>>
>> return vq;
>>
>
> Another approach I thought of is to use percpu spinlock, and
> the idea is simple:
>
> - all perpcu locks are held for writing req_vq, and
> - only percpu lock is needed for reading req_vq.
>
> What do you think about the below patch?
Per-CPU spinlocks have bad scalability problems, especially if you're
overcommitting. Writing req_vq is not at all rare.
Paolo
> diff --git a/drivers/scsi/virtio_scsi.c b/drivers/scsi/virtio_scsi.c
> index 697fa53..00deab4 100644
> --- a/drivers/scsi/virtio_scsi.c
> +++ b/drivers/scsi/virtio_scsi.c
> @@ -82,7 +82,7 @@ struct virtio_scsi_vq {
> */
> struct virtio_scsi_target_state {
> /* This spinlock never held at the same time as vq_lock. */
> - spinlock_t tgt_lock;
> + spinlock_t __percpu *lock;
>
> /* Count of outstanding requests. */
> atomic_t reqs;
> @@ -517,21 +517,46 @@ static struct virtio_scsi_vq *virtscsi_pick_vq(struct virtio_scsi *vscsi,
> {
> struct virtio_scsi_vq *vq;
> unsigned long flags;
> - u32 queue_num;
> + u32 cpu = get_cpu();
> + spinlock_t *lock = per_cpu_ptr(tgt->lock, cpu);
>
> - spin_lock_irqsave(&tgt->tgt_lock, flags);
> + spin_lock_irqsave(lock, flags);
>
> if (atomic_inc_return(&tgt->reqs) > 1)
> vq = tgt->req_vq;
> else {
> - queue_num = smp_processor_id();
> + u32 queue_num = cpu;
> + int i;
> +
> while (unlikely(queue_num >= vscsi->num_queues))
> queue_num -= vscsi->num_queues;
>
> - tgt->req_vq = vq = &vscsi->req_vqs[queue_num];
> + /*
> + * there should be only one writing because of atomic
> + * counter, so we don't worry about deadlock, but
> + * might need to teach lockdep to not complain it
> + */
> + for_each_possible_cpu(i) {
> + spinlock_t *other = per_cpu_ptr(tgt->lock, i);
> + if (i != cpu)
> + spin_lock(other);
> + }
> +
> + /* only update req_vq when reqs is one*/
> + if (atomic_read(&tgt->reqs) == 1)
> + tgt->req_vq = vq = &vscsi->req_vqs[queue_num];
> + else
> + vq = tgt->req_vq;
> +
> + for_each_possible_cpu(i) {
> + spinlock_t *other = per_cpu_ptr(tgt->lock, i);
> + if (i != cpu)
> + spin_unlock(other);
> + }
> }
>
> - spin_unlock_irqrestore(&tgt->tgt_lock, flags);
> + spin_unlock_irqrestore(lock, flags);
> + put_cpu();
> return vq;
> }
>
> @@ -618,10 +643,22 @@ static int virtscsi_target_alloc(struct scsi_target *starget)
> {
> struct virtio_scsi_target_state *tgt =
> kmalloc(sizeof(*tgt), GFP_KERNEL);
> + int i;
> +
> if (!tgt)
> return -ENOMEM;
>
> - spin_lock_init(&tgt->tgt_lock);
> + tgt->lock = alloc_percpu(spinlock_t);
> + if (!tgt->lock) {
> + kfree(tgt);
> + return -ENOMEM;
> + }
> +
> + for_each_possible_cpu(i) {
> + spinlock_t *lock = per_cpu_ptr(tgt->lock, i);
> + spin_lock_init(lock);
> + }
> +
> atomic_set(&tgt->reqs, 0);
> tgt->req_vq = NULL;
>
> @@ -632,6 +669,7 @@ static int virtscsi_target_alloc(struct scsi_target *starget)
> static void virtscsi_target_destroy(struct scsi_target *starget)
> {
> struct virtio_scsi_target_state *tgt = starget->hostdata;
> + free_percpu(tgt->lock);
> kfree(tgt);
> }
>
>
> Thanks,
>
next prev parent reply other threads:[~2014-05-07 16:44 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-05-06 9:26 virtio-scsi: two questions related with picking up queue Ming Lei
2014-05-06 13:17 ` Paolo Bonzini
2014-05-07 1:07 ` Ming Lei
2014-05-07 7:10 ` Paolo Bonzini
2014-05-07 16:24 ` Ming Lei
2014-05-07 16:43 ` Paolo Bonzini [this message]
2014-05-08 10:44 ` Ming Lei
2014-05-08 12:17 ` Paolo Bonzini
2014-05-08 12:55 ` Ming Lei
2014-05-08 13:21 ` Paolo Bonzini
2014-05-08 14:00 ` Ming Lei
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=536A62C1.6000905@redhat.com \
--to=pbonzini@redhat.com \
--cc=JBottomley@parallels.com \
--cc=gaowanlong@cn.fujitsu.com \
--cc=linux-scsi@vger.kernel.org \
--cc=rusty@rustcorp.com.au \
--cc=tom.leiming@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox