* [PATCH] cfq-iosched: Revert the logic of deep queues
@ 2010-05-19 20:33 Vivek Goyal
2010-05-19 23:51 ` Corrado Zoccolo
0 siblings, 1 reply; 8+ messages in thread
From: Vivek Goyal @ 2010-05-19 20:33 UTC (permalink / raw)
To: linux kernel mailing list, Jens Axboe; +Cc: Corrado Zoccolo, Moyer Jeff Moyer
o This patch basically reverts following commit.
76280af cfq-iosched: idling on deep seeky sync queues
o Idling in CFQ is bad on high end storage. This is especially more true of
random reads. Idling works very well for SATA disks with single
spindle but harms a lot on powerful storage boxes.
So even if deep queues can be little unfair to other random workload with
shallow depths, treat deep queues as sync-noidle workload and not sync,
because with sync workload we dispatch IO from only one queue at a time
and idle and we don't drive enough queue depth to keep the array busy.
o I am running aio-stress (random reads) as follows.
aio-stress -s 2g -O -t 4 -r 64k aio5 aio6 aio7 aio8 -o 3
Following are results with various combinations.
deadline: 232.94 MB/s
without patch
-------------
cfq default 75.32 MB/s
cfq, quantum=64 134.58 MB/s
with patch
----------
cfq default 78.37 MB/s
cfq, quantum=64 213.94 MB
Note that with the patch applied, cfq really scales well if "quantum" is
increased and comes close to deadline performance.
o Point being that on powerful arrays one queue is not sufficient to keep
array busy. This is already a bottleneck for sequential workloads. Lets
not aggravate the problem by marking random read queues as sync and
giving them exclusive access and hence effectively serializing the
access to array.
Signed-off-by: Vivek Goyal <vgoyal@redhat.com>
---
block/cfq-iosched.c | 12 +-----------
1 files changed, 1 insertions(+), 11 deletions(-)
diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
index 5f127cf..3336bd7 100644
--- a/block/cfq-iosched.c
+++ b/block/cfq-iosched.c
@@ -313,7 +313,6 @@ enum cfqq_state_flags {
CFQ_CFQQ_FLAG_sync, /* synchronous queue */
CFQ_CFQQ_FLAG_coop, /* cfqq is shared */
CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */
- CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */
CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */
};
@@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new);
CFQ_CFQQ_FNS(sync);
CFQ_CFQQ_FNS(coop);
CFQ_CFQQ_FNS(split_coop);
-CFQ_CFQQ_FNS(deep);
CFQ_CFQQ_FNS(wait_busy);
#undef CFQ_CFQQ_FNS
@@ -3036,11 +3034,8 @@ cfq_update_idle_window(struct cfq_data *cfqd, struct cfq_queue *cfqq,
enable_idle = old_idle = cfq_cfqq_idle_window(cfqq);
- if (cfqq->queued[0] + cfqq->queued[1] >= 4)
- cfq_mark_cfqq_deep(cfqq);
-
if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle ||
- (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq)))
+ CFQQ_SEEKY(cfqq))
enable_idle = 0;
else if (sample_valid(cic->ttime_samples)) {
if (cic->ttime_mean > cfqd->cfq_slice_idle)
@@ -3593,11 +3588,6 @@ static void cfq_idle_slice_timer(unsigned long data)
*/
if (!RB_EMPTY_ROOT(&cfqq->sort_list))
goto out_kick;
-
- /*
- * Queue depth flag is reset only when the idle didn't succeed
- */
- cfq_clear_cfqq_deep(cfqq);
}
expire:
cfq_slice_expired(cfqd, timed_out);
--
1.6.5.2
^ permalink raw reply related [flat|nested] 8+ messages in thread* Re: [PATCH] cfq-iosched: Revert the logic of deep queues 2010-05-19 20:33 [PATCH] cfq-iosched: Revert the logic of deep queues Vivek Goyal @ 2010-05-19 23:51 ` Corrado Zoccolo 2010-05-20 13:18 ` Vivek Goyal 0 siblings, 1 reply; 8+ messages in thread From: Corrado Zoccolo @ 2010-05-19 23:51 UTC (permalink / raw) To: Vivek Goyal; +Cc: linux kernel mailing list, Jens Axboe, Moyer Jeff Moyer On Wed, May 19, 2010 at 10:33 PM, Vivek Goyal <vgoyal@redhat.com> wrote: > o This patch basically reverts following commit. > > 76280af cfq-iosched: idling on deep seeky sync queues > > o Idling in CFQ is bad on high end storage. This is especially more true of > random reads. Idling works very well for SATA disks with single > spindle but harms a lot on powerful storage boxes. > > So even if deep queues can be little unfair to other random workload with > shallow depths, treat deep queues as sync-noidle workload and not sync, > because with sync workload we dispatch IO from only one queue at a time > and idle and we don't drive enough queue depth to keep the array busy. Maybe we should have a tunable for this behavior, like we have group_isolation for groups? It seems the same trade off to me (fairness vs. performance). And we could let it default off (on NCQ disks), since I guess deep seeky queues are more frequent on high end hardware than on our desktops. Note that on non-NCQ disks there is not only a fairness problem, but also a latency problem by having those deep queues as noidle (so I would say that non-NCQ disks should default to on, instead). Thanks Corrado > > o I am running aio-stress (random reads) as follows. > > aio-stress -s 2g -O -t 4 -r 64k aio5 aio6 aio7 aio8 -o 3 > > Following are results with various combinations. > > deadline: 232.94 MB/s > > without patch > ------------- > cfq default 75.32 MB/s > cfq, quantum=64 134.58 MB/s > > with patch > ---------- > cfq default 78.37 MB/s > cfq, quantum=64 213.94 MB > > Note that with the patch applied, cfq really scales well if "quantum" is > increased and comes close to deadline performance. > > o Point being that on powerful arrays one queue is not sufficient to keep > array busy. This is already a bottleneck for sequential workloads. Lets > not aggravate the problem by marking random read queues as sync and > giving them exclusive access and hence effectively serializing the > access to array. > > Signed-off-by: Vivek Goyal <vgoyal@redhat.com> > --- > block/cfq-iosched.c | 12 +----------- > 1 files changed, 1 insertions(+), 11 deletions(-) > > diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c > index 5f127cf..3336bd7 100644 > --- a/block/cfq-iosched.c > +++ b/block/cfq-iosched.c > @@ -313,7 +313,6 @@ enum cfqq_state_flags { > CFQ_CFQQ_FLAG_sync, /* synchronous queue */ > CFQ_CFQQ_FLAG_coop, /* cfqq is shared */ > CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */ > - CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */ > CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */ > }; > > @@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new); > CFQ_CFQQ_FNS(sync); > CFQ_CFQQ_FNS(coop); > CFQ_CFQQ_FNS(split_coop); > -CFQ_CFQQ_FNS(deep); > CFQ_CFQQ_FNS(wait_busy); > #undef CFQ_CFQQ_FNS > > @@ -3036,11 +3034,8 @@ cfq_update_idle_window(struct cfq_data *cfqd, struct cfq_queue *cfqq, > > enable_idle = old_idle = cfq_cfqq_idle_window(cfqq); > > - if (cfqq->queued[0] + cfqq->queued[1] >= 4) > - cfq_mark_cfqq_deep(cfqq); > - > if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle || > - (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq))) > + CFQQ_SEEKY(cfqq)) > enable_idle = 0; > else if (sample_valid(cic->ttime_samples)) { > if (cic->ttime_mean > cfqd->cfq_slice_idle) > @@ -3593,11 +3588,6 @@ static void cfq_idle_slice_timer(unsigned long data) > */ > if (!RB_EMPTY_ROOT(&cfqq->sort_list)) > goto out_kick; > - > - /* > - * Queue depth flag is reset only when the idle didn't succeed > - */ > - cfq_clear_cfqq_deep(cfqq); > } > expire: > cfq_slice_expired(cfqd, timed_out); > -- > 1.6.5.2 > > -- __________________________________________________________________________ dott. Corrado Zoccolo mailto:czoccolo@gmail.com PhD - Department of Computer Science - University of Pisa, Italy -------------------------------------------------------------------------- The self-confidence of a warrior is not the self-confidence of the average man. The average man seeks certainty in the eyes of the onlooker and calls that self-confidence. The warrior seeks impeccability in his own eyes and calls that humbleness. Tales of Power - C. Castaneda ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH] cfq-iosched: Revert the logic of deep queues 2010-05-19 23:51 ` Corrado Zoccolo @ 2010-05-20 13:18 ` Vivek Goyal 2010-05-20 14:01 ` Corrado Zoccolo 0 siblings, 1 reply; 8+ messages in thread From: Vivek Goyal @ 2010-05-20 13:18 UTC (permalink / raw) To: Corrado Zoccolo; +Cc: linux kernel mailing list, Jens Axboe, Moyer Jeff Moyer On Thu, May 20, 2010 at 01:51:49AM +0200, Corrado Zoccolo wrote: > On Wed, May 19, 2010 at 10:33 PM, Vivek Goyal <vgoyal@redhat.com> wrote: > > o This patch basically reverts following commit. > > > > 76280af cfq-iosched: idling on deep seeky sync queues > > > > o Idling in CFQ is bad on high end storage. This is especially more true of > > random reads. Idling works very well for SATA disks with single > > spindle but harms a lot on powerful storage boxes. > > > > So even if deep queues can be little unfair to other random workload with > > shallow depths, treat deep queues as sync-noidle workload and not sync, > > because with sync workload we dispatch IO from only one queue at a time > > and idle and we don't drive enough queue depth to keep the array busy. > > Maybe we should have a tunable for this behavior, like we have > group_isolation for groups? It seems the same trade off to me > (fairness vs. performance). > And we could let it default off (on NCQ disks), since I guess deep > seeky queues are more frequent on high end hardware than on our > desktops. > Note that on non-NCQ disks there is not only a fairness problem, but > also a latency problem by having those deep queues as noidle (so I > would say that non-NCQ disks should default to on, instead). Hi Corrado, Deep queues can happen often on high end storage. One case I can think of is multiple kvm virt machines running and doing IO using AIO. I am not too keen on introducing a tunable at this point of time. Reason being that somebody having a SATA disk and driving deep queue depths is not very practical thing to do. At the same time we have fixed a theoritical problem in the past. If somebody really runs into the issue of deep queue starving other random IO, then we can fix it. Even if we have to fix it, I think instead of a tunable, a better solution would be to expire the deep queue after one round of dispatch (after having dispatched "quantum" number of requests from queue). That way no single sync-noidle queue will starve other queues and they will get to dispatch IO very nicely without intorducing any bottlenecks. Thanks Vivek > > Thanks > Corrado > > > > o I am running aio-stress (random reads) as follows. > > > > aio-stress -s 2g -O -t 4 -r 64k aio5 aio6 aio7 aio8 -o 3 > > > > Following are results with various combinations. > > > > deadline: 232.94 MB/s > > > > without patch > > ------------- > > cfq default 75.32 MB/s > > cfq, quantum=64 134.58 MB/s > > > > with patch > > ---------- > > cfq default 78.37 MB/s > > cfq, quantum=64 213.94 MB > > > > Note that with the patch applied, cfq really scales well if "quantum" is > > increased and comes close to deadline performance. > > > > o Point being that on powerful arrays one queue is not sufficient to keep > > array busy. This is already a bottleneck for sequential workloads. Lets > > not aggravate the problem by marking random read queues as sync and > > giving them exclusive access and hence effectively serializing the > > access to array. > > > > Signed-off-by: Vivek Goyal <vgoyal@redhat.com> > > --- > > block/cfq-iosched.c | 12 +----------- > > 1 files changed, 1 insertions(+), 11 deletions(-) > > > > diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c > > index 5f127cf..3336bd7 100644 > > --- a/block/cfq-iosched.c > > +++ b/block/cfq-iosched.c > > @@ -313,7 +313,6 @@ enum cfqq_state_flags { > > CFQ_CFQQ_FLAG_sync, /* synchronous queue */ > > CFQ_CFQQ_FLAG_coop, /* cfqq is shared */ > > CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */ > > - CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */ > > CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */ > > }; > > > > @@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new); > > CFQ_CFQQ_FNS(sync); > > CFQ_CFQQ_FNS(coop); > > CFQ_CFQQ_FNS(split_coop); > > -CFQ_CFQQ_FNS(deep); > > CFQ_CFQQ_FNS(wait_busy); > > #undef CFQ_CFQQ_FNS > > > > @@ -3036,11 +3034,8 @@ cfq_update_idle_window(struct cfq_data *cfqd, struct cfq_queue *cfqq, > > > > enable_idle = old_idle = cfq_cfqq_idle_window(cfqq); > > > > - if (cfqq->queued[0] + cfqq->queued[1] >= 4) > > - cfq_mark_cfqq_deep(cfqq); > > - > > if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle || > > - (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq))) > > + CFQQ_SEEKY(cfqq)) > > enable_idle = 0; > > else if (sample_valid(cic->ttime_samples)) { > > if (cic->ttime_mean > cfqd->cfq_slice_idle) > > @@ -3593,11 +3588,6 @@ static void cfq_idle_slice_timer(unsigned long data) > > */ > > if (!RB_EMPTY_ROOT(&cfqq->sort_list)) > > goto out_kick; > > - > > - /* > > - * Queue depth flag is reset only when the idle didn't succeed > > - */ > > - cfq_clear_cfqq_deep(cfqq); > > } > > expire: > > cfq_slice_expired(cfqd, timed_out); > > -- > > 1.6.5.2 > > > > > > > > -- > __________________________________________________________________________ > > dott. Corrado Zoccolo mailto:czoccolo@gmail.com > PhD - Department of Computer Science - University of Pisa, Italy > -------------------------------------------------------------------------- > The self-confidence of a warrior is not the self-confidence of the average > man. The average man seeks certainty in the eyes of the onlooker and calls > that self-confidence. The warrior seeks impeccability in his own eyes and > calls that humbleness. > Tales of Power - C. Castaneda ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH] cfq-iosched: Revert the logic of deep queues 2010-05-20 13:18 ` Vivek Goyal @ 2010-05-20 14:01 ` Corrado Zoccolo 2010-05-20 14:50 ` Vivek Goyal 0 siblings, 1 reply; 8+ messages in thread From: Corrado Zoccolo @ 2010-05-20 14:01 UTC (permalink / raw) To: Vivek Goyal; +Cc: linux kernel mailing list, Jens Axboe, Moyer Jeff Moyer On Thu, May 20, 2010 at 3:18 PM, Vivek Goyal <vgoyal@redhat.com> wrote: > On Thu, May 20, 2010 at 01:51:49AM +0200, Corrado Zoccolo wrote: > Hi Corrado, > > Deep queues can happen often on high end storage. One case I can think of is > multiple kvm virt machines running and doing IO using AIO. > > I am not too keen on introducing a tunable at this point of time. Reason > being that somebody having a SATA disk and driving deep queue depths is > not very practical thing to do. At the same time we have fixed a theoritical > problem in the past. If somebody really runs into the issue of deep queue > starving other random IO, then we can fix it. > > Even if we have to fix it, I think instead of a tunable, a better solution > would be to expire the deep queue after one round of dispatch (after > having dispatched "quantum" number of requests from queue). That way no > single sync-noidle queue will starve other queues and they will get to > dispatch IO very nicely without intorducing any bottlenecks. Can you implement this solution in the patch? It seems that this will solve both the performance issue as well as non-reintroducing the theoretical starvation problem. If we don't mind some more tree operations, the queue could be expired at every dispatch (if there are other queues in the service tree), instead of every quantum dispatches, to cycle through all no-idle queues more quickly and more fairly. Thanks, Corrado > > Thanks > Vivek > ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH] cfq-iosched: Revert the logic of deep queues 2010-05-20 14:01 ` Corrado Zoccolo @ 2010-05-20 14:50 ` Vivek Goyal 2010-05-20 18:57 ` Vivek Goyal 0 siblings, 1 reply; 8+ messages in thread From: Vivek Goyal @ 2010-05-20 14:50 UTC (permalink / raw) To: Corrado Zoccolo; +Cc: linux kernel mailing list, Jens Axboe, Moyer Jeff Moyer On Thu, May 20, 2010 at 04:01:55PM +0200, Corrado Zoccolo wrote: > On Thu, May 20, 2010 at 3:18 PM, Vivek Goyal <vgoyal@redhat.com> wrote: > > On Thu, May 20, 2010 at 01:51:49AM +0200, Corrado Zoccolo wrote: > > Hi Corrado, > > > > Deep queues can happen often on high end storage. One case I can think of is > > multiple kvm virt machines running and doing IO using AIO. > > > > I am not too keen on introducing a tunable at this point of time. Reason > > being that somebody having a SATA disk and driving deep queue depths is > > not very practical thing to do. At the same time we have fixed a theoritical > > problem in the past. If somebody really runs into the issue of deep queue > > starving other random IO, then we can fix it. > > > > Even if we have to fix it, I think instead of a tunable, a better solution > > would be to expire the deep queue after one round of dispatch (after > > having dispatched "quantum" number of requests from queue). That way no > > single sync-noidle queue will starve other queues and they will get to > > dispatch IO very nicely without intorducing any bottlenecks. > > Can you implement this solution in the patch? It seems that this will > solve both the performance issue as well as non-reintroducing the > theoretical starvation problem. > If we don't mind some more tree operations, the queue could be expired > at every dispatch (if there are other queues in the service tree), > instead of every quantum dispatches, to cycle through all no-idle > queues more quickly and more fairly. Alright. Following is a copile tested only patch. I have yet to do the testing to make sure it works. But I think it should address your concern of a deep queue starving other shallow sync-noidle queues. Does this one look good? Thanks Vivek Index: linux-2.6/block/cfq-iosched.c =================================================================== --- linux-2.6.orig/block/cfq-iosched.c +++ linux-2.6/block/cfq-iosched.c @@ -313,7 +313,6 @@ enum cfqq_state_flags { CFQ_CFQQ_FLAG_sync, /* synchronous queue */ CFQ_CFQQ_FLAG_coop, /* cfqq is shared */ CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */ - CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */ CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */ }; @@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new); CFQ_CFQQ_FNS(sync); CFQ_CFQQ_FNS(coop); CFQ_CFQQ_FNS(split_coop); -CFQ_CFQQ_FNS(deep); CFQ_CFQQ_FNS(wait_busy); #undef CFQ_CFQQ_FNS @@ -2377,6 +2375,17 @@ static int cfq_dispatch_requests(struct cfq_class_idle(cfqq))) { cfqq->slice_end = jiffies + 1; cfq_slice_expired(cfqd, 0); + } else if (cfqq_type(cfqq) == SYNC_NOIDLE_WORKLOAD + && cfqq->service_tree->count > 1 + && cfqq->slice_dispatch >= cfq_prio_to_maxrq(cfqd, cfqq)/2) { + /* + * Expire a sync-noidle queue immediately if it has already + * dispatched many requests. This will make sure one deep + * sync-noidle queue will not starve other shallow sync-noidle + * queues. + */ + cfqq->slice_end = jiffies + 1; + cfq_slice_expired(cfqd, 0); } cfq_log_cfqq(cfqd, cfqq, "dispatched a request"); @@ -3036,11 +3045,8 @@ cfq_update_idle_window(struct cfq_data * enable_idle = old_idle = cfq_cfqq_idle_window(cfqq); - if (cfqq->queued[0] + cfqq->queued[1] >= 4) - cfq_mark_cfqq_deep(cfqq); - if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle || - (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq))) + CFQQ_SEEKY(cfqq)) enable_idle = 0; else if (sample_valid(cic->ttime_samples)) { if (cic->ttime_mean > cfqd->cfq_slice_idle) @@ -3593,11 +3599,6 @@ static void cfq_idle_slice_timer(unsigne */ if (!RB_EMPTY_ROOT(&cfqq->sort_list)) goto out_kick; - - /* - * Queue depth flag is reset only when the idle didn't succeed - */ - cfq_clear_cfqq_deep(cfqq); } expire: cfq_slice_expired(cfqd, timed_out); ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH] cfq-iosched: Revert the logic of deep queues 2010-05-20 14:50 ` Vivek Goyal @ 2010-05-20 18:57 ` Vivek Goyal 2010-05-20 20:09 ` Nauman Rafique 0 siblings, 1 reply; 8+ messages in thread From: Vivek Goyal @ 2010-05-20 18:57 UTC (permalink / raw) To: Corrado Zoccolo; +Cc: linux kernel mailing list, Jens Axboe, Moyer Jeff Moyer On Thu, May 20, 2010 at 10:50:24AM -0400, Vivek Goyal wrote: > On Thu, May 20, 2010 at 04:01:55PM +0200, Corrado Zoccolo wrote: > > On Thu, May 20, 2010 at 3:18 PM, Vivek Goyal <vgoyal@redhat.com> wrote: > > > On Thu, May 20, 2010 at 01:51:49AM +0200, Corrado Zoccolo wrote: > > > Hi Corrado, > > > > > > Deep queues can happen often on high end storage. One case I can think of is > > > multiple kvm virt machines running and doing IO using AIO. > > > > > > I am not too keen on introducing a tunable at this point of time. Reason > > > being that somebody having a SATA disk and driving deep queue depths is > > > not very practical thing to do. At the same time we have fixed a theoritical > > > problem in the past. If somebody really runs into the issue of deep queue > > > starving other random IO, then we can fix it. > > > > > > Even if we have to fix it, I think instead of a tunable, a better solution > > > would be to expire the deep queue after one round of dispatch (after > > > having dispatched "quantum" number of requests from queue). That way no > > > single sync-noidle queue will starve other queues and they will get to > > > dispatch IO very nicely without intorducing any bottlenecks. > > > > Can you implement this solution in the patch? It seems that this will > > solve both the performance issue as well as non-reintroducing the > > theoretical starvation problem. > > If we don't mind some more tree operations, the queue could be expired > > at every dispatch (if there are other queues in the service tree), > > instead of every quantum dispatches, to cycle through all no-idle > > queues more quickly and more fairly. > > Alright. Following is a copile tested only patch. I have yet to do the > testing to make sure it works. But I think it should address your concern > of a deep queue starving other shallow sync-noidle queues. > > Does this one look good? Well, this patch does not work. The reason being that over a period of time we drive deeper queue depths (due to all the queues doing IO at same time) and that means somebody doing 1 random read is stuck somewhere in that 32 requests dispatched to disk and it does not issue next IO till previous one is completed. So even if you expire the deep queue early, it will be selected again and it will do more dispatches (cfq quantum seems to work only for one round and for next round we don't keep track how many requests have already been disptached from this queue). Even your solution of tunable and turning it off by default on NCQ will not work as most of the SATA disks are now NCQ enabled. Its hard to find non-NCQ hardware. CFQ's tunable and everything is so much geared towards SATA disks that it is hard to use it with servers. Now IO controller feature is primarily useful for servers and it is dependent on CFQ. This seems to be a bad situation. Now CFQ has to scale so that one can comfortably use with high end storage without losing performance. So far folks will easily switch to dealine but that take IO controller out of the picture. Autotuning seems to be only reliable way. Default can be set right only for one type of usage and other type of users will be left with the exercise of setting the tunables right. Sadly, getting autotuning right is hard. Vivek > > Index: linux-2.6/block/cfq-iosched.c > =================================================================== > --- linux-2.6.orig/block/cfq-iosched.c > +++ linux-2.6/block/cfq-iosched.c > @@ -313,7 +313,6 @@ enum cfqq_state_flags { > CFQ_CFQQ_FLAG_sync, /* synchronous queue */ > CFQ_CFQQ_FLAG_coop, /* cfqq is shared */ > CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */ > - CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */ > CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */ > }; > > @@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new); > CFQ_CFQQ_FNS(sync); > CFQ_CFQQ_FNS(coop); > CFQ_CFQQ_FNS(split_coop); > -CFQ_CFQQ_FNS(deep); > CFQ_CFQQ_FNS(wait_busy); > #undef CFQ_CFQQ_FNS > > @@ -2377,6 +2375,17 @@ static int cfq_dispatch_requests(struct > cfq_class_idle(cfqq))) { > cfqq->slice_end = jiffies + 1; > cfq_slice_expired(cfqd, 0); > + } else if (cfqq_type(cfqq) == SYNC_NOIDLE_WORKLOAD > + && cfqq->service_tree->count > 1 > + && cfqq->slice_dispatch >= cfq_prio_to_maxrq(cfqd, cfqq)/2) { > + /* > + * Expire a sync-noidle queue immediately if it has already > + * dispatched many requests. This will make sure one deep > + * sync-noidle queue will not starve other shallow sync-noidle > + * queues. > + */ > + cfqq->slice_end = jiffies + 1; > + cfq_slice_expired(cfqd, 0); > } > > cfq_log_cfqq(cfqd, cfqq, "dispatched a request"); > @@ -3036,11 +3045,8 @@ cfq_update_idle_window(struct cfq_data * > > enable_idle = old_idle = cfq_cfqq_idle_window(cfqq); > > - if (cfqq->queued[0] + cfqq->queued[1] >= 4) > - cfq_mark_cfqq_deep(cfqq); > - > if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle || > - (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq))) > + CFQQ_SEEKY(cfqq)) > enable_idle = 0; > else if (sample_valid(cic->ttime_samples)) { > if (cic->ttime_mean > cfqd->cfq_slice_idle) > @@ -3593,11 +3599,6 @@ static void cfq_idle_slice_timer(unsigne > */ > if (!RB_EMPTY_ROOT(&cfqq->sort_list)) > goto out_kick; > - > - /* > - * Queue depth flag is reset only when the idle didn't succeed > - */ > - cfq_clear_cfqq_deep(cfqq); > } > expire: > cfq_slice_expired(cfqd, timed_out); ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH] cfq-iosched: Revert the logic of deep queues 2010-05-20 18:57 ` Vivek Goyal @ 2010-05-20 20:09 ` Nauman Rafique 2010-05-20 20:29 ` Vivek Goyal 0 siblings, 1 reply; 8+ messages in thread From: Nauman Rafique @ 2010-05-20 20:09 UTC (permalink / raw) To: Vivek Goyal Cc: Corrado Zoccolo, linux kernel mailing list, Jens Axboe, Moyer Jeff Moyer On Thu, May 20, 2010 at 11:57 AM, Vivek Goyal <vgoyal@redhat.com> wrote: > On Thu, May 20, 2010 at 10:50:24AM -0400, Vivek Goyal wrote: >> On Thu, May 20, 2010 at 04:01:55PM +0200, Corrado Zoccolo wrote: >> > On Thu, May 20, 2010 at 3:18 PM, Vivek Goyal <vgoyal@redhat.com> wrote: >> > > On Thu, May 20, 2010 at 01:51:49AM +0200, Corrado Zoccolo wrote: >> > > Hi Corrado, >> > > >> > > Deep queues can happen often on high end storage. One case I can think of is >> > > multiple kvm virt machines running and doing IO using AIO. >> > > >> > > I am not too keen on introducing a tunable at this point of time. Reason >> > > being that somebody having a SATA disk and driving deep queue depths is >> > > not very practical thing to do. At the same time we have fixed a theoritical >> > > problem in the past. If somebody really runs into the issue of deep queue >> > > starving other random IO, then we can fix it. >> > > >> > > Even if we have to fix it, I think instead of a tunable, a better solution >> > > would be to expire the deep queue after one round of dispatch (after >> > > having dispatched "quantum" number of requests from queue). That way no >> > > single sync-noidle queue will starve other queues and they will get to >> > > dispatch IO very nicely without intorducing any bottlenecks. >> > >> > Can you implement this solution in the patch? It seems that this will >> > solve both the performance issue as well as non-reintroducing the >> > theoretical starvation problem. >> > If we don't mind some more tree operations, the queue could be expired >> > at every dispatch (if there are other queues in the service tree), >> > instead of every quantum dispatches, to cycle through all no-idle >> > queues more quickly and more fairly. >> >> Alright. Following is a copile tested only patch. I have yet to do the >> testing to make sure it works. But I think it should address your concern >> of a deep queue starving other shallow sync-noidle queues. >> >> Does this one look good? > > Well, this patch does not work. The reason being that over a period of > time we drive deeper queue depths (due to all the queues doing IO at > same time) and that means somebody doing 1 random read is stuck somewhere > in that 32 requests dispatched to disk and it does not issue next IO > till previous one is completed. > > So even if you expire the deep queue early, it will be selected again and > it will do more dispatches (cfq quantum seems to work only for one round > and for next round we don't keep track how many requests have already been > disptached from this queue). > > Even your solution of tunable and turning it off by default on NCQ will > not work as most of the SATA disks are now NCQ enabled. Its hard to find > non-NCQ hardware. I do not agree with the some of the statements you are making here. Even if drives generally support NCQ, it can still be turned off. And there are advantages associated with turning off NCQ, like more predictable latencies for IOs. > > CFQ's tunable and everything is so much geared towards SATA disks that it > is hard to use it with servers. Now IO controller feature is primarily > useful for servers and it is dependent on CFQ. This seems to be a bad > situation. Now CFQ has to scale so that one can comfortably use with > high end storage without losing performance. So far folks will easily > switch to dealine but that take IO controller out of the picture. I beg to disagree here too. IO controller is useful for plain SATA disks too. Actually I would go one step further and argue that doing resource management is more important for resources that see higher level of contention, due to lesser capacity than the demand. That's why IO controller would be more useful on traditional hardware. > > Autotuning seems to be only reliable way. Default can be set right only > for one type of usage and other type of users will be left with the > exercise of setting the tunables right. Sadly, getting autotuning right > is hard. > > Vivek > >> >> Index: linux-2.6/block/cfq-iosched.c >> =================================================================== >> --- linux-2.6.orig/block/cfq-iosched.c >> +++ linux-2.6/block/cfq-iosched.c >> @@ -313,7 +313,6 @@ enum cfqq_state_flags { >> CFQ_CFQQ_FLAG_sync, /* synchronous queue */ >> CFQ_CFQQ_FLAG_coop, /* cfqq is shared */ >> CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */ >> - CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */ >> CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */ >> }; >> >> @@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new); >> CFQ_CFQQ_FNS(sync); >> CFQ_CFQQ_FNS(coop); >> CFQ_CFQQ_FNS(split_coop); >> -CFQ_CFQQ_FNS(deep); >> CFQ_CFQQ_FNS(wait_busy); >> #undef CFQ_CFQQ_FNS >> >> @@ -2377,6 +2375,17 @@ static int cfq_dispatch_requests(struct >> cfq_class_idle(cfqq))) { >> cfqq->slice_end = jiffies + 1; >> cfq_slice_expired(cfqd, 0); >> + } else if (cfqq_type(cfqq) == SYNC_NOIDLE_WORKLOAD >> + && cfqq->service_tree->count > 1 >> + && cfqq->slice_dispatch >= cfq_prio_to_maxrq(cfqd, cfqq)/2) { >> + /* >> + * Expire a sync-noidle queue immediately if it has already >> + * dispatched many requests. This will make sure one deep >> + * sync-noidle queue will not starve other shallow sync-noidle >> + * queues. >> + */ >> + cfqq->slice_end = jiffies + 1; >> + cfq_slice_expired(cfqd, 0); >> } >> >> cfq_log_cfqq(cfqd, cfqq, "dispatched a request"); >> @@ -3036,11 +3045,8 @@ cfq_update_idle_window(struct cfq_data * >> >> enable_idle = old_idle = cfq_cfqq_idle_window(cfqq); >> >> - if (cfqq->queued[0] + cfqq->queued[1] >= 4) >> - cfq_mark_cfqq_deep(cfqq); >> - >> if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle || >> - (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq))) >> + CFQQ_SEEKY(cfqq)) >> enable_idle = 0; >> else if (sample_valid(cic->ttime_samples)) { >> if (cic->ttime_mean > cfqd->cfq_slice_idle) >> @@ -3593,11 +3599,6 @@ static void cfq_idle_slice_timer(unsigne >> */ >> if (!RB_EMPTY_ROOT(&cfqq->sort_list)) >> goto out_kick; >> - >> - /* >> - * Queue depth flag is reset only when the idle didn't succeed >> - */ >> - cfq_clear_cfqq_deep(cfqq); >> } >> expire: >> cfq_slice_expired(cfqd, timed_out); > -- > To unsubscribe from this list: send the line "unsubscribe linux-kernel" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html > Please read the FAQ at http://www.tux.org/lkml/ > ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH] cfq-iosched: Revert the logic of deep queues 2010-05-20 20:09 ` Nauman Rafique @ 2010-05-20 20:29 ` Vivek Goyal 0 siblings, 0 replies; 8+ messages in thread From: Vivek Goyal @ 2010-05-20 20:29 UTC (permalink / raw) To: Nauman Rafique Cc: Corrado Zoccolo, linux kernel mailing list, Jens Axboe, Moyer Jeff Moyer On Thu, May 20, 2010 at 01:09:22PM -0700, Nauman Rafique wrote: > On Thu, May 20, 2010 at 11:57 AM, Vivek Goyal <vgoyal@redhat.com> wrote: > > On Thu, May 20, 2010 at 10:50:24AM -0400, Vivek Goyal wrote: > >> On Thu, May 20, 2010 at 04:01:55PM +0200, Corrado Zoccolo wrote: > >> > On Thu, May 20, 2010 at 3:18 PM, Vivek Goyal <vgoyal@redhat.com> wrote: > >> > > On Thu, May 20, 2010 at 01:51:49AM +0200, Corrado Zoccolo wrote: > >> > > Hi Corrado, > >> > > > >> > > Deep queues can happen often on high end storage. One case I can think of is > >> > > multiple kvm virt machines running and doing IO using AIO. > >> > > > >> > > I am not too keen on introducing a tunable at this point of time. Reason > >> > > being that somebody having a SATA disk and driving deep queue depths is > >> > > not very practical thing to do. At the same time we have fixed a theoritical > >> > > problem in the past. If somebody really runs into the issue of deep queue > >> > > starving other random IO, then we can fix it. > >> > > > >> > > Even if we have to fix it, I think instead of a tunable, a better solution > >> > > would be to expire the deep queue after one round of dispatch (after > >> > > having dispatched "quantum" number of requests from queue). That way no > >> > > single sync-noidle queue will starve other queues and they will get to > >> > > dispatch IO very nicely without intorducing any bottlenecks. > >> > > >> > Can you implement this solution in the patch? It seems that this will > >> > solve both the performance issue as well as non-reintroducing the > >> > theoretical starvation problem. > >> > If we don't mind some more tree operations, the queue could be expired > >> > at every dispatch (if there are other queues in the service tree), > >> > instead of every quantum dispatches, to cycle through all no-idle > >> > queues more quickly and more fairly. > >> > >> Alright. Following is a copile tested only patch. I have yet to do the > >> testing to make sure it works. But I think it should address your concern > >> of a deep queue starving other shallow sync-noidle queues. > >> > >> Does this one look good? > > > > Well, this patch does not work. The reason being that over a period of > > time we drive deeper queue depths (due to all the queues doing IO at > > same time) and that means somebody doing 1 random read is stuck somewhere > > in that 32 requests dispatched to disk and it does not issue next IO > > till previous one is completed. > > > > So even if you expire the deep queue early, it will be selected again and > > it will do more dispatches (cfq quantum seems to work only for one round > > and for next round we don't keep track how many requests have already been > > disptached from this queue). > > > > Even your solution of tunable and turning it off by default on NCQ will > > not work as most of the SATA disks are now NCQ enabled. Its hard to find > > non-NCQ hardware. > > I do not agree with the some of the statements you are making here. > Even if drives generally support NCQ, it can still be turned off. And > there are advantages associated with turning off NCQ, like more > predictable latencies for IOs. I am not sure what are you disagreeing with. If you turn off NCQ, then my patch should work well. Issue happens only when less capable SATA disks advertise themselves as having queue depth of 32 and IO scheduler pumps in those many requests. So are you agreeing that we can treat deep queues as sync-idle on non NCQ hardware or hardware where NCQ has been turned off. On NCQ hardware we will by default treat it as sync-noidle and that can starve other shallow readers and for better response user should turn off NCQ? > > > > > CFQ's tunable and everything is so much geared towards SATA disks that it > > is hard to use it with servers. Now IO controller feature is primarily > > useful for servers and it is dependent on CFQ. This seems to be a bad > > situation. Now CFQ has to scale so that one can comfortably use with > > high end storage without losing performance. So far folks will easily > > switch to dealine but that take IO controller out of the picture. > > I beg to disagree here too. IO controller is useful for plain SATA > disks too. So are you creating some kind of software RAID of SATA disks and that's why it is useful? In enterprise environment I will assume it is not reasonable to run lots of things on single SATA disk. > Actually I would go one step further and argue that doing > resource management is more important for resources that see higher > level of contention, due to lesser capacity than the demand. That's > why IO controller would be more useful on traditional hardware. Couple of thoughts. - One can put more applicatoions on same piece of hardware once proper IO control mechanism is there and create more demand and again get into a situation where demand is more than capacity. - Secondly, one might still want to priotize the applications so that one application gets more bandwidth. Especially to protect against some really heavy writer and that is independent of SATA vs enterprise array. Vivek > > > > > Autotuning seems to be only reliable way. Default can be set right only > > for one type of usage and other type of users will be left with the > > exercise of setting the tunables right. Sadly, getting autotuning right > > is hard. > > > > Vivek > > > >> > >> Index: linux-2.6/block/cfq-iosched.c > >> =================================================================== > >> --- linux-2.6.orig/block/cfq-iosched.c > >> +++ linux-2.6/block/cfq-iosched.c > >> @@ -313,7 +313,6 @@ enum cfqq_state_flags { > >> CFQ_CFQQ_FLAG_sync, /* synchronous queue */ > >> CFQ_CFQQ_FLAG_coop, /* cfqq is shared */ > >> CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */ > >> - CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */ > >> CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */ > >> }; > >> > >> @@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new); > >> CFQ_CFQQ_FNS(sync); > >> CFQ_CFQQ_FNS(coop); > >> CFQ_CFQQ_FNS(split_coop); > >> -CFQ_CFQQ_FNS(deep); > >> CFQ_CFQQ_FNS(wait_busy); > >> #undef CFQ_CFQQ_FNS > >> > >> @@ -2377,6 +2375,17 @@ static int cfq_dispatch_requests(struct > >> cfq_class_idle(cfqq))) { > >> cfqq->slice_end = jiffies + 1; > >> cfq_slice_expired(cfqd, 0); > >> + } else if (cfqq_type(cfqq) == SYNC_NOIDLE_WORKLOAD > >> + && cfqq->service_tree->count > 1 > >> + && cfqq->slice_dispatch >= cfq_prio_to_maxrq(cfqd, cfqq)/2) { > >> + /* > >> + * Expire a sync-noidle queue immediately if it has already > >> + * dispatched many requests. This will make sure one deep > >> + * sync-noidle queue will not starve other shallow sync-noidle > >> + * queues. > >> + */ > >> + cfqq->slice_end = jiffies + 1; > >> + cfq_slice_expired(cfqd, 0); > >> } > >> > >> cfq_log_cfqq(cfqd, cfqq, "dispatched a request"); > >> @@ -3036,11 +3045,8 @@ cfq_update_idle_window(struct cfq_data * > >> > >> enable_idle = old_idle = cfq_cfqq_idle_window(cfqq); > >> > >> - if (cfqq->queued[0] + cfqq->queued[1] >= 4) > >> - cfq_mark_cfqq_deep(cfqq); > >> - > >> if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle || > >> - (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq))) > >> + CFQQ_SEEKY(cfqq)) > >> enable_idle = 0; > >> else if (sample_valid(cic->ttime_samples)) { > >> if (cic->ttime_mean > cfqd->cfq_slice_idle) > >> @@ -3593,11 +3599,6 @@ static void cfq_idle_slice_timer(unsigne > >> */ > >> if (!RB_EMPTY_ROOT(&cfqq->sort_list)) > >> goto out_kick; > >> - > >> - /* > >> - * Queue depth flag is reset only when the idle didn't succeed > >> - */ > >> - cfq_clear_cfqq_deep(cfqq); > >> } > >> expire: > >> cfq_slice_expired(cfqd, timed_out); > > -- > > To unsubscribe from this list: send the line "unsubscribe linux-kernel" in > > the body of a message to majordomo@vger.kernel.org > > More majordomo info at http://vger.kernel.org/majordomo-info.html > > Please read the FAQ at http://www.tux.org/lkml/ > > ^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2010-05-20 20:35 UTC | newest] Thread overview: 8+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2010-05-19 20:33 [PATCH] cfq-iosched: Revert the logic of deep queues Vivek Goyal 2010-05-19 23:51 ` Corrado Zoccolo 2010-05-20 13:18 ` Vivek Goyal 2010-05-20 14:01 ` Corrado Zoccolo 2010-05-20 14:50 ` Vivek Goyal 2010-05-20 18:57 ` Vivek Goyal 2010-05-20 20:09 ` Nauman Rafique 2010-05-20 20:29 ` Vivek Goyal
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).