* [PATCH] virtio-blk: Disable callback in virtblk_done()
@ 2012-09-25 2:36 Asias He
2012-09-25 14:08 ` Asias He
0 siblings, 1 reply; 9+ messages in thread
From: Asias He @ 2012-09-25 2:36 UTC (permalink / raw)
To: Rusty Russell; +Cc: virtualization, kvm, Michael S. Tsirkin
This reduces unnecessary interrupts that host could send to guest while
guest is in the progress of irq handling.
If one vcpu is handling the irq, while another interrupt comes, in
handle_edge_irq(), the guest will mask the interrupt via mask_msi_irq()
which is a very heavy operation that goes all the way down to host.
Signed-off-by: Asias He <asias@redhat.com>
---
drivers/block/virtio_blk.c | 19 +++++++++++--------
1 file changed, 11 insertions(+), 8 deletions(-)
diff --git a/drivers/block/virtio_blk.c b/drivers/block/virtio_blk.c
index 53b81d5..0bdde8f 100644
--- a/drivers/block/virtio_blk.c
+++ b/drivers/block/virtio_blk.c
@@ -274,15 +274,18 @@ static void virtblk_done(struct virtqueue *vq)
unsigned int len;
spin_lock_irqsave(vblk->disk->queue->queue_lock, flags);
- while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
- if (vbr->bio) {
- virtblk_bio_done(vbr);
- bio_done = true;
- } else {
- virtblk_request_done(vbr);
- req_done = true;
+ do {
+ virtqueue_disable_cb(vq);
+ while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
+ if (vbr->bio) {
+ virtblk_bio_done(vbr);
+ bio_done = true;
+ } else {
+ virtblk_request_done(vbr);
+ req_done = true;
+ }
}
- }
+ } while (!virtqueue_enable_cb(vq));
/* In case queue is stopped waiting for more buffers. */
if (req_done)
blk_start_queue(vblk->disk->queue);
--
1.7.11.4
^ permalink raw reply related [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
2012-09-25 2:36 [PATCH] virtio-blk: Disable callback in virtblk_done() Asias He
@ 2012-09-25 14:08 ` Asias He
2012-09-27 0:10 ` Rusty Russell
0 siblings, 1 reply; 9+ messages in thread
From: Asias He @ 2012-09-25 14:08 UTC (permalink / raw)
To: Rusty Russell; +Cc: virtualization, kvm, Michael S. Tsirkin
On 09/25/2012 10:36 AM, Asias He wrote:
> This reduces unnecessary interrupts that host could send to guest while
> guest is in the progress of irq handling.
>
> If one vcpu is handling the irq, while another interrupt comes, in
> handle_edge_irq(), the guest will mask the interrupt via mask_msi_irq()
> which is a very heavy operation that goes all the way down to host.
>
> Signed-off-by: Asias He <asias@redhat.com>
> ---
Here are some performance numbers on qemu:
Before:
-------------------------------------
seq-read : io=0 B, bw=269730KB/s, iops=67432 , runt= 62200msec
seq-write : io=0 B, bw=339716KB/s, iops=84929 , runt= 49386msec
rand-read : io=0 B, bw=270435KB/s, iops=67608 , runt= 62038msec
rand-write: io=0 B, bw=354436KB/s, iops=88608 , runt= 47335msec
clat (usec): min=101 , max=138052 , avg=14822.09, stdev=11771.01
clat (usec): min=96 , max=81543 , avg=11798.94, stdev=7735.60
clat (usec): min=128 , max=140043 , avg=14835.85, stdev=11765.33
clat (usec): min=109 , max=147207 , avg=11337.09, stdev=5990.35
cpu : usr=15.93%, sys=60.37%, ctx=7764972, majf=0, minf=54
cpu : usr=32.73%, sys=120.49%, ctx=7372945, majf=0, minf=1
cpu : usr=18.84%, sys=58.18%, ctx=7775420, majf=0, minf=1
cpu : usr=24.20%, sys=59.85%, ctx=8307886, majf=0, minf=0
vdb: ios=8389107/8368136, merge=0/0, ticks=19457874/14616506,
in_queue=34206098, util=99.68%
43: interrupt in total: 887320
fio --exec_prerun="echo 3 > /proc/sys/vm/drop_caches" --group_reporting
--ioscheduler=noop --thread --bs=4k --size=512MB --direct=1 --numjobs=16
--ioengine=libaio --iodepth=64 --loops=3 --ramp_time=0
--filename=/dev/vdb --name=seq-read --stonewall --rw=read
--name=seq-write --stonewall --rw=write --name=rnd-read --stonewall
--rw=randread --name=rnd-write --stonewall --rw=randwrite
After:
-------------------------------------
seq-read : io=0 B, bw=309503KB/s, iops=77375 , runt= 54207msec
seq-write : io=0 B, bw=448205KB/s, iops=112051 , runt= 37432msec
rand-read : io=0 B, bw=311254KB/s, iops=77813 , runt= 53902msec
rand-write: io=0 B, bw=377152KB/s, iops=94287 , runt= 44484msec
clat (usec): min=81 , max=90588 , avg=12946.06, stdev=9085.94
clat (usec): min=57 , max=72264 , avg=8967.97, stdev=5951.04
clat (usec): min=29 , max=101046 , avg=12889.95, stdev=9067.91
clat (usec): min=52 , max=106152 , avg=10660.56, stdev=4778.19
cpu : usr=15.05%, sys=57.92%, ctx=7710941, majf=0, minf=54
cpu : usr=26.78%, sys=101.40%, ctx=7387891, majf=0, minf=2
cpu : usr=19.03%, sys=58.17%, ctx=7681976, majf=0, minf=8
cpu : usr=24.65%, sys=58.34%, ctx=8442632, majf=0, minf=4
vdb: ios=8389086/8361888, merge=0/0, ticks=17243780/12742010,
in_queue=30078377, util=99.59%
43: interrupt in total: 1259639
fio --exec_prerun="echo 3 > /proc/sys/vm/drop_caches" --group_reporting
--ioscheduler=noop --thread --bs=4k --size=512MB --direct=1 --numjobs=16
--ioengine=libaio --iodepth=64 --loops=3 --ramp_time=0
--filename=/dev/vdb --name=seq-read --stonewall --rw=read
--name=seq-write --stonewall --rw=write --name=rnd-read --stonewall
--rw=randread --name=rnd-write --stonewall --rw=randwrite
> drivers/block/virtio_blk.c | 19 +++++++++++--------
> 1 file changed, 11 insertions(+), 8 deletions(-)
>
> diff --git a/drivers/block/virtio_blk.c b/drivers/block/virtio_blk.c
> index 53b81d5..0bdde8f 100644
> --- a/drivers/block/virtio_blk.c
> +++ b/drivers/block/virtio_blk.c
> @@ -274,15 +274,18 @@ static void virtblk_done(struct virtqueue *vq)
> unsigned int len;
>
> spin_lock_irqsave(vblk->disk->queue->queue_lock, flags);
> - while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
> - if (vbr->bio) {
> - virtblk_bio_done(vbr);
> - bio_done = true;
> - } else {
> - virtblk_request_done(vbr);
> - req_done = true;
> + do {
> + virtqueue_disable_cb(vq);
> + while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
> + if (vbr->bio) {
> + virtblk_bio_done(vbr);
> + bio_done = true;
> + } else {
> + virtblk_request_done(vbr);
> + req_done = true;
> + }
> }
> - }
> + } while (!virtqueue_enable_cb(vq));
> /* In case queue is stopped waiting for more buffers. */
> if (req_done)
> blk_start_queue(vblk->disk->queue);
>
--
Asias
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
2012-09-25 14:08 ` Asias He
@ 2012-09-27 0:10 ` Rusty Russell
2012-09-27 6:48 ` Asias He
` (3 more replies)
0 siblings, 4 replies; 9+ messages in thread
From: Rusty Russell @ 2012-09-27 0:10 UTC (permalink / raw)
To: Asias He; +Cc: virtualization, kvm, Michael S. Tsirkin
Asias He <asias@redhat.com> writes:
> On 09/25/2012 10:36 AM, Asias He wrote:
>> This reduces unnecessary interrupts that host could send to guest while
>> guest is in the progress of irq handling.
>>
>> If one vcpu is handling the irq, while another interrupt comes, in
>> handle_edge_irq(), the guest will mask the interrupt via mask_msi_irq()
>> which is a very heavy operation that goes all the way down to host.
>>
>> Signed-off-by: Asias He <asias@redhat.com>
>> ---
>
> Here are some performance numbers on qemu:
I assume this is with qemu using kvm, not qemu in soft emulation? :)
> Before:
> -------------------------------------
> seq-read : io=0 B, bw=269730KB/s, iops=67432 , runt= 62200msec
> seq-write : io=0 B, bw=339716KB/s, iops=84929 , runt= 49386msec
> rand-read : io=0 B, bw=270435KB/s, iops=67608 , runt= 62038msec
> rand-write: io=0 B, bw=354436KB/s, iops=88608 , runt= 47335msec
> clat (usec): min=101 , max=138052 , avg=14822.09, stdev=11771.01
> clat (usec): min=96 , max=81543 , avg=11798.94, stdev=7735.60
> clat (usec): min=128 , max=140043 , avg=14835.85, stdev=11765.33
> clat (usec): min=109 , max=147207 , avg=11337.09, stdev=5990.35
> cpu : usr=15.93%, sys=60.37%, ctx=7764972, majf=0, minf=54
> cpu : usr=32.73%, sys=120.49%, ctx=7372945, majf=0, minf=1
> cpu : usr=18.84%, sys=58.18%, ctx=7775420, majf=0, minf=1
> cpu : usr=24.20%, sys=59.85%, ctx=8307886, majf=0, minf=0
> vdb: ios=8389107/8368136, merge=0/0, ticks=19457874/14616506,
> in_queue=34206098, util=99.68%
> 43: interrupt in total: 887320
> fio --exec_prerun="echo 3 > /proc/sys/vm/drop_caches" --group_reporting
> --ioscheduler=noop --thread --bs=4k --size=512MB --direct=1 --numjobs=16
> --ioengine=libaio --iodepth=64 --loops=3 --ramp_time=0
> --filename=/dev/vdb --name=seq-read --stonewall --rw=read
> --name=seq-write --stonewall --rw=write --name=rnd-read --stonewall
> --rw=randread --name=rnd-write --stonewall --rw=randwrite
>
> After:
> -------------------------------------
> seq-read : io=0 B, bw=309503KB/s, iops=77375 , runt= 54207msec
> seq-write : io=0 B, bw=448205KB/s, iops=112051 , runt= 37432msec
> rand-read : io=0 B, bw=311254KB/s, iops=77813 , runt= 53902msec
> rand-write: io=0 B, bw=377152KB/s, iops=94287 , runt= 44484msec
> clat (usec): min=81 , max=90588 , avg=12946.06, stdev=9085.94
> clat (usec): min=57 , max=72264 , avg=8967.97, stdev=5951.04
> clat (usec): min=29 , max=101046 , avg=12889.95, stdev=9067.91
> clat (usec): min=52 , max=106152 , avg=10660.56, stdev=4778.19
> cpu : usr=15.05%, sys=57.92%, ctx=7710941, majf=0, minf=54
> cpu : usr=26.78%, sys=101.40%, ctx=7387891, majf=0, minf=2
> cpu : usr=19.03%, sys=58.17%, ctx=7681976, majf=0, minf=8
> cpu : usr=24.65%, sys=58.34%, ctx=8442632, majf=0, minf=4
> vdb: ios=8389086/8361888, merge=0/0, ticks=17243780/12742010,
> in_queue=30078377, util=99.59%
> 43: interrupt in total: 1259639
> fio --exec_prerun="echo 3 > /proc/sys/vm/drop_caches" --group_reporting
> --ioscheduler=noop --thread --bs=4k --size=512MB --direct=1 --numjobs=16
> --ioengine=libaio --iodepth=64 --loops=3 --ramp_time=0
> --filename=/dev/vdb --name=seq-read --stonewall --rw=read
> --name=seq-write --stonewall --rw=write --name=rnd-read --stonewall
> --rw=randread --name=rnd-write --stonewall --rw=randwrite
>
>> drivers/block/virtio_blk.c | 19 +++++++++++--------
>> 1 file changed, 11 insertions(+), 8 deletions(-)
>>
>> diff --git a/drivers/block/virtio_blk.c b/drivers/block/virtio_blk.c
>> index 53b81d5..0bdde8f 100644
>> --- a/drivers/block/virtio_blk.c
>> +++ b/drivers/block/virtio_blk.c
>> @@ -274,15 +274,18 @@ static void virtblk_done(struct virtqueue *vq)
>> unsigned int len;
>>
>> spin_lock_irqsave(vblk->disk->queue->queue_lock, flags);
>> - while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
>> - if (vbr->bio) {
>> - virtblk_bio_done(vbr);
>> - bio_done = true;
>> - } else {
>> - virtblk_request_done(vbr);
>> - req_done = true;
>> + do {
>> + virtqueue_disable_cb(vq);
>> + while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
>> + if (vbr->bio) {
>> + virtblk_bio_done(vbr);
>> + bio_done = true;
>> + } else {
>> + virtblk_request_done(vbr);
>> + req_done = true;
>> + }
>> }
>> - }
>> + } while (!virtqueue_enable_cb(vq));
>> /* In case queue is stopped waiting for more buffers. */
>> if (req_done)
>> blk_start_queue(vblk->disk->queue);
Fascinating. Please just confirm that VIRTIO_RING_F_EVENT_IDX is
enabled?
I forgot about the cool hack which MST put in to defer event updates
using disable_cb/enable_cb.
Applied!
Rusty.
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
2012-09-27 0:10 ` Rusty Russell
@ 2012-09-27 6:48 ` Asias He
2012-09-28 6:08 ` Rusty Russell
2012-09-27 10:01 ` Paolo Bonzini
` (2 subsequent siblings)
3 siblings, 1 reply; 9+ messages in thread
From: Asias He @ 2012-09-27 6:48 UTC (permalink / raw)
To: Rusty Russell; +Cc: virtualization, kvm, Michael S. Tsirkin
On 09/27/2012 08:10 AM, Rusty Russell wrote:
> Asias He <asias@redhat.com> writes:
>
>> On 09/25/2012 10:36 AM, Asias He wrote:
>>> This reduces unnecessary interrupts that host could send to guest while
>>> guest is in the progress of irq handling.
>>>
>>> If one vcpu is handling the irq, while another interrupt comes, in
>>> handle_edge_irq(), the guest will mask the interrupt via mask_msi_irq()
>>> which is a very heavy operation that goes all the way down to host.
>>>
>>> Signed-off-by: Asias He <asias@redhat.com>
>>> ---
>>
>> Here are some performance numbers on qemu:
>
> I assume this is with qemu using kvm, not qemu in soft emulation? :)
Of course.
>
>> Before:
>> -------------------------------------
>> seq-read : io=0 B, bw=269730KB/s, iops=67432 , runt= 62200msec
>> seq-write : io=0 B, bw=339716KB/s, iops=84929 , runt= 49386msec
>> rand-read : io=0 B, bw=270435KB/s, iops=67608 , runt= 62038msec
>> rand-write: io=0 B, bw=354436KB/s, iops=88608 , runt= 47335msec
>> clat (usec): min=101 , max=138052 , avg=14822.09, stdev=11771.01
>> clat (usec): min=96 , max=81543 , avg=11798.94, stdev=7735.60
>> clat (usec): min=128 , max=140043 , avg=14835.85, stdev=11765.33
>> clat (usec): min=109 , max=147207 , avg=11337.09, stdev=5990.35
>> cpu : usr=15.93%, sys=60.37%, ctx=7764972, majf=0, minf=54
>> cpu : usr=32.73%, sys=120.49%, ctx=7372945, majf=0, minf=1
>> cpu : usr=18.84%, sys=58.18%, ctx=7775420, majf=0, minf=1
>> cpu : usr=24.20%, sys=59.85%, ctx=8307886, majf=0, minf=0
>> vdb: ios=8389107/8368136, merge=0/0, ticks=19457874/14616506,
>> in_queue=34206098, util=99.68%
>> 43: interrupt in total: 887320
>> fio --exec_prerun="echo 3 > /proc/sys/vm/drop_caches" --group_reporting
>> --ioscheduler=noop --thread --bs=4k --size=512MB --direct=1 --numjobs=16
>> --ioengine=libaio --iodepth=64 --loops=3 --ramp_time=0
>> --filename=/dev/vdb --name=seq-read --stonewall --rw=read
>> --name=seq-write --stonewall --rw=write --name=rnd-read --stonewall
>> --rw=randread --name=rnd-write --stonewall --rw=randwrite
>>
>> After:
>> -------------------------------------
>> seq-read : io=0 B, bw=309503KB/s, iops=77375 , runt= 54207msec
>> seq-write : io=0 B, bw=448205KB/s, iops=112051 , runt= 37432msec
>> rand-read : io=0 B, bw=311254KB/s, iops=77813 , runt= 53902msec
>> rand-write: io=0 B, bw=377152KB/s, iops=94287 , runt= 44484msec
>> clat (usec): min=81 , max=90588 , avg=12946.06, stdev=9085.94
>> clat (usec): min=57 , max=72264 , avg=8967.97, stdev=5951.04
>> clat (usec): min=29 , max=101046 , avg=12889.95, stdev=9067.91
>> clat (usec): min=52 , max=106152 , avg=10660.56, stdev=4778.19
>> cpu : usr=15.05%, sys=57.92%, ctx=7710941, majf=0, minf=54
>> cpu : usr=26.78%, sys=101.40%, ctx=7387891, majf=0, minf=2
>> cpu : usr=19.03%, sys=58.17%, ctx=7681976, majf=0, minf=8
>> cpu : usr=24.65%, sys=58.34%, ctx=8442632, majf=0, minf=4
>> vdb: ios=8389086/8361888, merge=0/0, ticks=17243780/12742010,
>> in_queue=30078377, util=99.59%
>> 43: interrupt in total: 1259639
>> fio --exec_prerun="echo 3 > /proc/sys/vm/drop_caches" --group_reporting
>> --ioscheduler=noop --thread --bs=4k --size=512MB --direct=1 --numjobs=16
>> --ioengine=libaio --iodepth=64 --loops=3 --ramp_time=0
>> --filename=/dev/vdb --name=seq-read --stonewall --rw=read
>> --name=seq-write --stonewall --rw=write --name=rnd-read --stonewall
>> --rw=randread --name=rnd-write --stonewall --rw=randwrite
>>
>>> drivers/block/virtio_blk.c | 19 +++++++++++--------
>>> 1 file changed, 11 insertions(+), 8 deletions(-)
>>>
>>> diff --git a/drivers/block/virtio_blk.c b/drivers/block/virtio_blk.c
>>> index 53b81d5..0bdde8f 100644
>>> --- a/drivers/block/virtio_blk.c
>>> +++ b/drivers/block/virtio_blk.c
>>> @@ -274,15 +274,18 @@ static void virtblk_done(struct virtqueue *vq)
>>> unsigned int len;
>>>
>>> spin_lock_irqsave(vblk->disk->queue->queue_lock, flags);
>>> - while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
>>> - if (vbr->bio) {
>>> - virtblk_bio_done(vbr);
>>> - bio_done = true;
>>> - } else {
>>> - virtblk_request_done(vbr);
>>> - req_done = true;
>>> + do {
>>> + virtqueue_disable_cb(vq);
>>> + while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
>>> + if (vbr->bio) {
>>> + virtblk_bio_done(vbr);
>>> + bio_done = true;
>>> + } else {
>>> + virtblk_request_done(vbr);
>>> + req_done = true;
>>> + }
>>> }
>>> - }
>>> + } while (!virtqueue_enable_cb(vq));
>>> /* In case queue is stopped waiting for more buffers. */
>>> if (req_done)
>>> blk_start_queue(vblk->disk->queue);
>
> Fascinating. Please just confirm that VIRTIO_RING_F_EVENT_IDX is
> enabled?
Sure. It is enabled ;-)
>
> I forgot about the cool hack which MST put in to defer event updates
> using disable_cb/enable_cb.
Hmm, are you talking about virtqueue_enable_cb_delayed()?
>
> Applied!
> Rusty.
>
--
Asias
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
2012-09-27 0:10 ` Rusty Russell
2012-09-27 6:48 ` Asias He
@ 2012-09-27 10:01 ` Paolo Bonzini
2012-09-28 8:32 ` Michael S. Tsirkin
[not found] ` <50642400.5030906@redhat.com>
3 siblings, 0 replies; 9+ messages in thread
From: Paolo Bonzini @ 2012-09-27 10:01 UTC (permalink / raw)
To: Rusty Russell; +Cc: Michael S. Tsirkin, kvm, virtualization
Il 27/09/2012 02:10, Rusty Russell ha scritto:
>>> >> + do {
>>> >> + virtqueue_disable_cb(vq);
>>> >> + while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
>>> >> + if (vbr->bio) {
>>> >> + virtblk_bio_done(vbr);
>>> >> + bio_done = true;
>>> >> + } else {
>>> >> + virtblk_request_done(vbr);
>>> >> + req_done = true;
>>> >> + }
>>> >> }
>>> >> - }
>>> >> + } while (!virtqueue_enable_cb(vq));
>>> >> /* In case queue is stopped waiting for more buffers. */
>>> >> if (req_done)
>>> >> blk_start_queue(vblk->disk->queue);
> Fascinating. Please just confirm that VIRTIO_RING_F_EVENT_IDX is
> enabled?
Yeah, it's a nice and cheap trick. Stefan, I see that you had this in
virtio-scsi since even before I picked it up. Do you remember how you
came up with it?
Paolo
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
2012-09-27 6:48 ` Asias He
@ 2012-09-28 6:08 ` Rusty Russell
2012-09-28 7:03 ` Asias He
0 siblings, 1 reply; 9+ messages in thread
From: Rusty Russell @ 2012-09-28 6:08 UTC (permalink / raw)
To: Asias He; +Cc: virtualization, kvm, Michael S. Tsirkin
Asias He <asias@redhat.com> writes:
>> I forgot about the cool hack which MST put in to defer event updates
>> using disable_cb/enable_cb.
>
> Hmm, are you talking about virtqueue_enable_cb_delayed()?
Just the fact that virtqueue_disable_cb() prevents updates of
used_index, and then we do the update in virtqueue_enable_cb().
Cheers,
Rusty.
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
2012-09-28 6:08 ` Rusty Russell
@ 2012-09-28 7:03 ` Asias He
0 siblings, 0 replies; 9+ messages in thread
From: Asias He @ 2012-09-28 7:03 UTC (permalink / raw)
To: Rusty Russell; +Cc: virtualization, kvm, Michael S. Tsirkin
On 09/28/2012 02:08 PM, Rusty Russell wrote:
> Asias He <asias@redhat.com> writes:
>>> I forgot about the cool hack which MST put in to defer event updates
>>> using disable_cb/enable_cb.
>>
>> Hmm, are you talking about virtqueue_enable_cb_delayed()?
>
> Just the fact that virtqueue_disable_cb() prevents updates of
> used_index, and then we do the update in virtqueue_enable_cb().
Okay.
--
Asias
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
2012-09-27 0:10 ` Rusty Russell
2012-09-27 6:48 ` Asias He
2012-09-27 10:01 ` Paolo Bonzini
@ 2012-09-28 8:32 ` Michael S. Tsirkin
[not found] ` <50642400.5030906@redhat.com>
3 siblings, 0 replies; 9+ messages in thread
From: Michael S. Tsirkin @ 2012-09-28 8:32 UTC (permalink / raw)
To: Rusty Russell; +Cc: kvm, virtualization
On Thu, Sep 27, 2012 at 09:40:03AM +0930, Rusty Russell wrote:
> I forgot about the cool hack which MST put in to defer event updates
> using disable_cb/enable_cb.
I considered sticking some invalid value
in event index on disable but in my testing it did not seem to
give any gain, and knowing actual index of the other side
is better for debugging.
--
MST
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH] virtio-blk: Disable callback in virtblk_done()
[not found] ` <50642400.5030906@redhat.com>
@ 2012-10-02 14:42 ` Stefan Hajnoczi
0 siblings, 0 replies; 9+ messages in thread
From: Stefan Hajnoczi @ 2012-10-02 14:42 UTC (permalink / raw)
To: Paolo Bonzini; +Cc: Michael S. Tsirkin, kvm, virtualization
On Thu, Sep 27, 2012 at 12:01 PM, Paolo Bonzini <pbonzini@redhat.com> wrote:
> Il 27/09/2012 02:10, Rusty Russell ha scritto:
>>>> >> + do {
>>>> >> + virtqueue_disable_cb(vq);
>>>> >> + while ((vbr = virtqueue_get_buf(vblk->vq, &len)) != NULL) {
>>>> >> + if (vbr->bio) {
>>>> >> + virtblk_bio_done(vbr);
>>>> >> + bio_done = true;
>>>> >> + } else {
>>>> >> + virtblk_request_done(vbr);
>>>> >> + req_done = true;
>>>> >> + }
>>>> >> }
>>>> >> - }
>>>> >> + } while (!virtqueue_enable_cb(vq));
>>>> >> /* In case queue is stopped waiting for more buffers. */
>>>> >> if (req_done)
>>>> >> blk_start_queue(vblk->disk->queue);
>> Fascinating. Please just confirm that VIRTIO_RING_F_EVENT_IDX is
>> enabled?
>
> Yeah, it's a nice and cheap trick. Stefan, I see that you had this in
> virtio-scsi since even before I picked it up. Do you remember how you
> came up with it?
I've played with disable_cb/enable_cb previously in virtio-blk and
virtio-scsi when aliguori had suggested reducing notifies. I
definitely didn't invent it :).
Stefan
^ permalink raw reply [flat|nested] 9+ messages in thread
end of thread, other threads:[~2012-10-02 14:42 UTC | newest]
Thread overview: 9+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2012-09-25 2:36 [PATCH] virtio-blk: Disable callback in virtblk_done() Asias He
2012-09-25 14:08 ` Asias He
2012-09-27 0:10 ` Rusty Russell
2012-09-27 6:48 ` Asias He
2012-09-28 6:08 ` Rusty Russell
2012-09-28 7:03 ` Asias He
2012-09-27 10:01 ` Paolo Bonzini
2012-09-28 8:32 ` Michael S. Tsirkin
[not found] ` <50642400.5030906@redhat.com>
2012-10-02 14:42 ` Stefan Hajnoczi
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).