* scheduler interactivity - does this patch help?
@ 2003-06-09 5:43 Martin J. Bligh
2003-06-09 19:22 ` Felipe Alfaro Solana
0 siblings, 1 reply; 5+ messages in thread
From: Martin J. Bligh @ 2003-06-09 5:43 UTC (permalink / raw)
To: linux-kernel
I've had this patch (I think from Ingo) kicking around in -mjb
for a while. I'm going to drop it unless someone thinks it's useful
for some testcase you have ... anyone interested?
Thanks,
M.
diff -urpN -X /home/fletch/.diff.exclude 400-reiserfs_dio/kernel/sched.c 420-sched_interactive/kernel/sched.c
--- 400-reiserfs_dio/kernel/sched.c Fri May 30 19:26:34 2003
+++ 420-sched_interactive/kernel/sched.c Fri May 30 19:28:06 2003
@@ -89,6 +89,8 @@ int node_threshold = 125;
#define STARVATION_LIMIT (starvation_limit)
#define NODE_THRESHOLD (node_threshold)
+#define TIMESLICE_GRANULARITY (HZ/20 ?: 1)
+
/*
* If a task is 'interactive' then we reinsert it in the active
* array after it has expired its current timeslice. (it will not
@@ -1365,6 +1367,27 @@ void scheduler_tick(int user_ticks, int
enqueue_task(p, rq->expired);
} else
enqueue_task(p, rq->active);
+ } else {
+ /*
+ * Prevent a too long timeslice allowing a task to monopolize
+ * the CPU. We do this by splitting up the timeslice into
+ * smaller pieces.
+ *
+ * Note: this does not mean the task's timeslices expire or
+ * get lost in any way, they just might be preempted by
+ * another task of equal priority. (one with higher
+ * priority would have preempted this task already.) We
+ * requeue this task to the end of the list on this priority
+ * level, which is in essence a round-robin of tasks with
+ * equal priority.
+ */
+ if (!(p->time_slice % TIMESLICE_GRANULARITY) &&
+ (p->array == rq->active)) {
+ dequeue_task(p, rq->active);
+ set_tsk_need_resched(p);
+ p->prio = effective_prio(p);
+ enqueue_task(p, rq->active);
+ }
}
out_unlock:
spin_unlock(&rq->lock);
^ permalink raw reply [flat|nested] 5+ messages in thread* Re: scheduler interactivity - does this patch help? 2003-06-09 5:43 scheduler interactivity - does this patch help? Martin J. Bligh @ 2003-06-09 19:22 ` Felipe Alfaro Solana 2003-06-10 9:39 ` Con Kolivas 0 siblings, 1 reply; 5+ messages in thread From: Felipe Alfaro Solana @ 2003-06-09 19:22 UTC (permalink / raw) To: Martin J. Bligh; +Cc: linux-kernel On Mon, 2003-06-09 at 07:43, Martin J. Bligh wrote: > I've had this patch (I think from Ingo) kicking around in -mjb > for a while. I'm going to drop it unless someone thinks it's useful > for some testcase you have ... anyone interested? > > Thanks, > > M. > > diff -urpN -X /home/fletch/.diff.exclude 400-reiserfs_dio/kernel/sched.c 420-sched_interactive/kernel/sched.c > --- 400-reiserfs_dio/kernel/sched.c Fri May 30 19:26:34 2003 > +++ 420-sched_interactive/kernel/sched.c Fri May 30 19:28:06 2003 > @@ -89,6 +89,8 @@ int node_threshold = 125; > #define STARVATION_LIMIT (starvation_limit) > #define NODE_THRESHOLD (node_threshold) > > +#define TIMESLICE_GRANULARITY (HZ/20 ?: 1) > + > /* > * If a task is 'interactive' then we reinsert it in the active > * array after it has expired its current timeslice. (it will not > @@ -1365,6 +1367,27 @@ void scheduler_tick(int user_ticks, int > enqueue_task(p, rq->expired); > } else > enqueue_task(p, rq->active); > + } else { > + /* > + * Prevent a too long timeslice allowing a task to monopolize > + * the CPU. We do this by splitting up the timeslice into > + * smaller pieces. > + * > + * Note: this does not mean the task's timeslices expire or > + * get lost in any way, they just might be preempted by > + * another task of equal priority. (one with higher > + * priority would have preempted this task already.) We > + * requeue this task to the end of the list on this priority > + * level, which is in essence a round-robin of tasks with > + * equal priority. > + */ > + if (!(p->time_slice % TIMESLICE_GRANULARITY) && > + (p->array == rq->active)) { > + dequeue_task(p, rq->active); > + set_tsk_need_resched(p); > + p->prio = effective_prio(p); > + enqueue_task(p, rq->active); > + } > } > out_unlock: > spin_unlock(&rq->lock); > I'm currently testing it on a modified 2.5.70-mm6 kernel (with HZ set to 1000) and seems to help a little with XMMS's chunky audio playback when X is reniced to -20. ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: scheduler interactivity - does this patch help? 2003-06-09 19:22 ` Felipe Alfaro Solana @ 2003-06-10 9:39 ` Con Kolivas 2003-06-10 11:37 ` Mike Galbraith 0 siblings, 1 reply; 5+ messages in thread From: Con Kolivas @ 2003-06-10 9:39 UTC (permalink / raw) To: Felipe Alfaro Solana, Martin J. Bligh; +Cc: linux-kernel On Tue, 10 Jun 2003 05:22, Felipe Alfaro Solana wrote: > On Mon, 2003-06-09 at 07:43, Martin J. Bligh wrote: > > I've had this patch (I think from Ingo) kicking around in -mjb > > for a while. I'm going to drop it unless someone thinks it's useful > > for some testcase you have ... anyone interested? > > > > Thanks, > > > > M. > > > > diff -urpN -X /home/fletch/.diff.exclude 400-reiserfs_dio/kernel/sched.c > > 420-sched_interactive/kernel/sched.c --- > > 400-reiserfs_dio/kernel/sched.c Fri May 30 19:26:34 2003 > > +++ 420-sched_interactive/kernel/sched.c Fri May 30 19:28:06 2003 > > @@ -89,6 +89,8 @@ int node_threshold = 125; > > #define STARVATION_LIMIT (starvation_limit) > > #define NODE_THRESHOLD (node_threshold) > > > > +#define TIMESLICE_GRANULARITY (HZ/20 ?: 1) > > + > > /* > > * If a task is 'interactive' then we reinsert it in the active > > * array after it has expired its current timeslice. (it will not > > @@ -1365,6 +1367,27 @@ void scheduler_tick(int user_ticks, int > > enqueue_task(p, rq->expired); > > } else > > enqueue_task(p, rq->active); > > + } else { > > + /* > > + * Prevent a too long timeslice allowing a task to monopolize > > + * the CPU. We do this by splitting up the timeslice into > > + * smaller pieces. > > + * > > + * Note: this does not mean the task's timeslices expire or > > + * get lost in any way, they just might be preempted by > > + * another task of equal priority. (one with higher > > + * priority would have preempted this task already.) We > > + * requeue this task to the end of the list on this priority > > + * level, which is in essence a round-robin of tasks with > > + * equal priority. > > + */ > > + if (!(p->time_slice % TIMESLICE_GRANULARITY) && > > + (p->array == rq->active)) { > > + dequeue_task(p, rq->active); > > + set_tsk_need_resched(p); > > + p->prio = effective_prio(p); > > + enqueue_task(p, rq->active); > > + } > > } > > out_unlock: > > spin_unlock(&rq->lock); > > I'm currently testing it on a modified 2.5.70-mm6 kernel (with HZ set to > 1000) and seems to help a little with XMMS's chunky audio playback when > X is reniced to -20. I tried this patch way back when mingo first posted it and found it helped a little. Have a close look at it, though; all it does is limit max timeslice to 50ms when other tasks are running at the same priority. A better effect can and is obtained by changing max_timeslice to 50ms... Con ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: scheduler interactivity - does this patch help? 2003-06-10 9:39 ` Con Kolivas @ 2003-06-10 11:37 ` Mike Galbraith 2003-06-10 11:39 ` Con Kolivas 0 siblings, 1 reply; 5+ messages in thread From: Mike Galbraith @ 2003-06-10 11:37 UTC (permalink / raw) To: Con Kolivas; +Cc: Felipe Alfaro Solana, Martin J. Bligh, linux-kernel At 07:39 PM 6/10/2003 +1000, Con Kolivas wrote: >On Tue, 10 Jun 2003 05:22, Felipe Alfaro Solana wrote: > > On Mon, 2003-06-09 at 07:43, Martin J. Bligh wrote: > > > I've had this patch (I think from Ingo) kicking around in -mjb > > > for a while. I'm going to drop it unless someone thinks it's useful > > > for some testcase you have ... anyone interested? > > > > > > Thanks, > > > > > > M. > > > > > > diff -urpN -X /home/fletch/.diff.exclude 400-reiserfs_dio/kernel/sched.c > > > 420-sched_interactive/kernel/sched.c --- > > > 400-reiserfs_dio/kernel/sched.c Fri May 30 19:26:34 2003 > > > +++ 420-sched_interactive/kernel/sched.c Fri May 30 19:28:06 2003 > > > @@ -89,6 +89,8 @@ int node_threshold = 125; > > > #define STARVATION_LIMIT (starvation_limit) > > > #define NODE_THRESHOLD (node_threshold) > > > > > > +#define TIMESLICE_GRANULARITY (HZ/20 ?: 1) > > > + > > > /* > > > * If a task is 'interactive' then we reinsert it in the active > > > * array after it has expired its current timeslice. (it will not > > > @@ -1365,6 +1367,27 @@ void scheduler_tick(int user_ticks, int > > > enqueue_task(p, rq->expired); > > > } else > > > enqueue_task(p, rq->active); > > > + } else { > > > + /* > > > + * Prevent a too long timeslice allowing a task to monopolize > > > + * the CPU. We do this by splitting up the timeslice into > > > + * smaller pieces. > > > + * > > > + * Note: this does not mean the task's timeslices expire or > > > + * get lost in any way, they just might be preempted by > > > + * another task of equal priority. (one with higher > > > + * priority would have preempted this task already.) We > > > + * requeue this task to the end of the list on this priority > > > + * level, which is in essence a round-robin of tasks with > > > + * equal priority. > > > + */ > > > + if (!(p->time_slice % TIMESLICE_GRANULARITY) && > > > + (p->array == rq->active)) { > > > + dequeue_task(p, rq->active); > > > + set_tsk_need_resched(p); > > > + p->prio = effective_prio(p); > > > + enqueue_task(p, rq->active); > > > + } > > > } > > > out_unlock: > > > spin_unlock(&rq->lock); > > > > I'm currently testing it on a modified 2.5.70-mm6 kernel (with HZ set to > > 1000) and seems to help a little with XMMS's chunky audio playback when > > X is reniced to -20. > >I tried this patch way back when mingo first posted it and found it helped a >little. Have a close look at it, though; all it does is limit max timeslice >to 50ms when other tasks are running at the same priority. A better effect >can and is obtained by changing max_timeslice to 50ms... It also drops priority somewhat sooner. If you reduce max to 50ms, normal task timeslice becomes tiny, which won't do anything good for throughput. -Mike ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: scheduler interactivity - does this patch help? 2003-06-10 11:37 ` Mike Galbraith @ 2003-06-10 11:39 ` Con Kolivas 0 siblings, 0 replies; 5+ messages in thread From: Con Kolivas @ 2003-06-10 11:39 UTC (permalink / raw) To: Mike Galbraith Cc: Felipe Alfaro Solana, Martin J. Bligh, linux-kernel, Zwane Mwaikambo On Tue, 10 Jun 2003 21:37, Mike Galbraith wrote: > At 07:39 PM 6/10/2003 +1000, Con Kolivas wrote: > >On Tue, 10 Jun 2003 05:22, Felipe Alfaro Solana wrote: > > > On Mon, 2003-06-09 at 07:43, Martin J. Bligh wrote: > > > > I've had this patch (I think from Ingo) kicking around in -mjb > > > > for a while. I'm going to drop it unless someone thinks it's useful > > > > for some testcase you have ... anyone interested? > > > > > > > > Thanks, > > > > > > > > M. > > > > > > > > diff -urpN -X /home/fletch/.diff.exclude > > > > 400-reiserfs_dio/kernel/sched.c 420-sched_interactive/kernel/sched.c > > > > --- > > > > 400-reiserfs_dio/kernel/sched.c Fri May 30 19:26:34 2003 > > > > +++ 420-sched_interactive/kernel/sched.c Fri May 30 19:28:06 2003 > > > > @@ -89,6 +89,8 @@ int node_threshold = 125; > > > > #define STARVATION_LIMIT (starvation_limit) > > > > #define NODE_THRESHOLD (node_threshold) > > > > > > > > +#define TIMESLICE_GRANULARITY (HZ/20 ?: 1) > > > > + > > > > /* > > > > * If a task is 'interactive' then we reinsert it in the active > > > > * array after it has expired its current timeslice. (it will not > > > > @@ -1365,6 +1367,27 @@ void scheduler_tick(int user_ticks, int > > > > enqueue_task(p, rq->expired); > > > > } else > > > > enqueue_task(p, rq->active); > > > > + } else { > > > > + /* > > > > + * Prevent a too long timeslice allowing a task to > > > > monopolize + * the CPU. We do this by splitting up the > > > > timeslice into + * smaller pieces. > > > > + * > > > > + * Note: this does not mean the task's timeslices expire > > > > or + * get lost in any way, they just might be preempted by > > > > + * another task of equal priority. (one with higher + > > > > * priority would have preempted this task already.) We + > > > > * requeue this task to the end of the list on this priority + > > > > * level, which is in essence a round-robin of tasks with + > > > > * equal priority. > > > > + */ > > > > + if (!(p->time_slice % TIMESLICE_GRANULARITY) && > > > > + (p->array == rq->active)) { > > > > + dequeue_task(p, rq->active); > > > > + set_tsk_need_resched(p); > > > > + p->prio = effective_prio(p); > > > > + enqueue_task(p, rq->active); > > > > + } > > > > } > > > > out_unlock: > > > > spin_unlock(&rq->lock); > > > > > > I'm currently testing it on a modified 2.5.70-mm6 kernel (with HZ set > > > to 1000) and seems to help a little with XMMS's chunky audio playback > > > when X is reniced to -20. > > > >I tried this patch way back when mingo first posted it and found it helped > > a little. Have a close look at it, though; all it does is limit max > > timeslice to 50ms when other tasks are running at the same priority. A > > better effect can and is obtained by changing max_timeslice to 50ms... > > It also drops priority somewhat sooner. If you reduce max to 50ms, normal > task timeslice becomes tiny, which won't do anything good for throughput. I wasn't advocating this as a fix, just making an observation. I do think the effect on throughput is overrated though. A few benchmarks I did on cpu intensive cache heavy tasks shows only very slight but measurable improvements as timeslices get beyond 7ms on a P4 2.53. Below this however throughput very rapidly drops off. P3 733 showed comparable effects. I haven't tested lower spec machines than this though. Con ^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2003-06-10 11:31 UTC | newest] Thread overview: 5+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2003-06-09 5:43 scheduler interactivity - does this patch help? Martin J. Bligh 2003-06-09 19:22 ` Felipe Alfaro Solana 2003-06-10 9:39 ` Con Kolivas 2003-06-10 11:37 ` Mike Galbraith 2003-06-10 11:39 ` Con Kolivas
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox