* [patch] CFS scheduler, -v18
@ 2007-06-22 22:02 Ingo Molnar
2007-06-22 22:09 ` S.Çağlar Onur
` (6 more replies)
0 siblings, 7 replies; 43+ messages in thread
From: Ingo Molnar @ 2007-06-22 22:02 UTC (permalink / raw)
To: linux-kernel
Cc: Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven,
Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri
i'm pleased to announce release -v18 of the CFS scheduler patchset.
The rolled-up CFS patch against today's -git kernel, v2.6.22-rc5,
v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be downloaded from the
usual place:
http://people.redhat.com/mingo/cfs-scheduler/
The biggest change in -v18 are various performance related improvements.
Thomas Gleixner has eliminated expensive 64-bit divisions by converting
the arithmetics to scaled math (without impacting the quality of
calculations). Srivatsa Vaddagiri and Dmitry Adamushko have continued
the abstraction and cleanup work. Srivatsa Vaddagiri and Christoph
Lameter fixed the NUMA balancing bug reported by Paul McKenney. There
were also a good number of other refinements to the CFS code. (No
reproducible behavioral regressions were reported against -v17 so far,
so the 'behavioral' bits are mostly unchanged.)
Changes since -v17:
- implement scaled math speedups for CFS. (Thomas Gleixner)
- lots of core code updates, cleanups and streamlining.
(Srivatsa Vaddagiri, Dmitry Adamushko, me.)
- bugfix: fix NUMA balancing. (Srivatsa Vaddagiri, Christoph Lameter,
Paul E. McKenney)
- feature: SCHED_IDLE now also implies block-scheduler (CFQ)
idle-IO-priority. (suggested by Thomas Sattler, picked up from -ck)
- build fix for ppc32. (reported, tested and confirmed fixed by
Art Haas)
- ARM fix. (reported and debugged by Thomas Gleixner)
- cleanup: implemented idle_sched_class in kernel/sched_idletask.c as a
way to separate out rq->idle handling out of the core scheduler. This
made a good deal of idle-task related special-cases go away.
- debug: make the sysctls safer by introducing high and low limits.
- cleanup: move some of the debug counters to under CONFIG_SCHEDSTATS.
- speedup: various micro-optimizations
- various other small updates.
As usual, any sort of feedback, bugreport, fix and suggestion is more
than welcome!
Ingo
^ permalink raw reply [flat|nested] 43+ messages in thread* Re: [patch] CFS scheduler, -v18 2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar @ 2007-06-22 22:09 ` S.Çağlar Onur 2007-06-22 22:16 ` S.Çağlar Onur 2007-06-22 23:08 ` Gene Heskett ` (5 subsequent siblings) 6 siblings, 1 reply; 43+ messages in thread From: S.Çağlar Onur @ 2007-06-22 22:09 UTC (permalink / raw) To: Ingo Molnar Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri [-- Attachment #1: Type: text/plain, Size: 1244 bytes --] Hi Ingo; 23 Haz 2007 Cts tarihinde, Ingo Molnar şunları yazmıştı: > As usual, any sort of feedback, bugreport, fix and suggestion is more > than welcome! caglar@zangetsu linux-2.6 $ LC_ALL=C make CHK include/linux/version.h CHK include/linux/utsrelease.h CALL scripts/checksyscalls.sh CHK include/linux/compile.h CC kernel/sched.o kernel/sched.c:745:28: sched_idletask.c: No such file or directory kernel/sched.c: In function `init_idle_bootup_task': kernel/sched.c:4659: error: `idle_sched_class' undeclared (first use in this function) kernel/sched.c:4659: error: (Each undeclared identifier is reported only once kernel/sched.c:4659: error: for each function it appears in.) kernel/sched.c: In function `init_idle': kernel/sched.c:4698: error: `idle_sched_class' undeclared (first use in this function) kernel/sched.c: In function `sched_init': kernel/sched.c:6196: error: `idle_sched_class' undeclared (first use in this function) make[1]: *** [kernel/sched.o] Error 1 make: *** [kernel] Error 2 Cheers -- S.Çağlar Onur <caglar@pardus.org.tr> http://cekirdek.pardus.org.tr/~caglar/ Linux is like living in a teepee. No Windows, no Gates and an Apache in house! [-- Attachment #2: This is a digitally signed message part. --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:09 ` S.Çağlar Onur @ 2007-06-22 22:16 ` S.Çağlar Onur 2007-06-22 22:20 ` Ingo Molnar 0 siblings, 1 reply; 43+ messages in thread From: S.Çağlar Onur @ 2007-06-22 22:16 UTC (permalink / raw) To: Ingo Molnar Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri [-- Attachment #1: Type: text/plain, Size: 2780 bytes --] 23 Haz 2007 Cts tarihinde, S.Çağlar Onur şunları yazmıştı: > Hi Ingo; > > 23 Haz 2007 Cts tarihinde, Ingo Molnar şunları yazmıştı: > > As usual, any sort of feedback, bugreport, fix and suggestion is more > > than welcome! > > caglar@zangetsu linux-2.6 $ LC_ALL=C make > CHK include/linux/version.h > CHK include/linux/utsrelease.h > CALL scripts/checksyscalls.sh > CHK include/linux/compile.h > CC kernel/sched.o > kernel/sched.c:745:28: sched_idletask.c: No such file or directory Ahh and this happens with [1], grabbing sched_idletask.c from .18 one solves the problem... Index: linux/kernel/sched_idletask.c =================================================================== --- /dev/null +++ linux/kernel/sched_idletask.c @@ -0,0 +1,68 @@ +/* + * idle-task scheduling class. + * + * (NOTE: these are not related to SCHED_IDLE tasks which are + * handled in sched_fair.c) + */ + +/* + * Idle tasks are unconditionally rescheduled: + */ +static void check_preempt_curr_idle(struct rq *rq, struct task_struct *p) +{ + resched_task(rq->idle); +} + +static struct task_struct *pick_next_task_idle(struct rq *rq, u64 now) +{ + schedstat_inc(rq, sched_goidle); + + return rq->idle; +} + +/* + * It is not legal to sleep in the idle task - print a warning + * message if some code attempts to do it: + */ +static void +dequeue_task_idle(struct rq *rq, struct task_struct *p, int sleep, u64 now) +{ + spin_unlock_irq(&rq->lock); + printk(KERN_ERR "bad: scheduling from the idle thread!\n"); + dump_stack(); + spin_lock_irq(&rq->lock); +} + +static void put_prev_task_idle(struct rq *rq, struct task_struct *prev, u64 now) +{ +} + +static struct task_struct *load_balance_start_idle(struct rq *rq) +{ + return NULL; +} + +static void task_tick_idle(struct rq *rq, struct task_struct *curr) +{ +} + +/* + * Simple, special scheduling class for the per-CPU idle tasks: + */ +struct sched_class idle_sched_class __read_mostly = { + /* no enqueue/yield_task for idle tasks */ + + /* dequeue is not valid, we print a debug message there: */ + .dequeue_task = dequeue_task_idle, + + .check_preempt_curr = check_preempt_curr_idle, + + .pick_next_task = pick_next_task_idle, + .put_prev_task = put_prev_task_idle, + + .load_balance_start = load_balance_start_idle, + /* no .load_balance_next for idle tasks */ + + .task_tick = task_tick_idle, + /* no .task_new for idle tasks */ +}; [1] http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-git-v18.patch Cheers -- S.Çağlar Onur <caglar@pardus.org.tr> http://cekirdek.pardus.org.tr/~caglar/ Linux is like living in a teepee. No Windows, no Gates and an Apache in house! [-- Attachment #2: This is a digitally signed message part. --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:16 ` S.Çağlar Onur @ 2007-06-22 22:20 ` Ingo Molnar 2007-06-26 3:02 ` Andrew Morton 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-06-22 22:20 UTC (permalink / raw) To: S.Çağlar Onur Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri * S.Çağlar Onur <caglar@pardus.org.tr> wrote: > > kernel/sched.c:745:28: sched_idletask.c: No such file or directory > > Ahh and this happens with [1], grabbing sched_idletask.c from .18 one solves > the problem... oops, indeed - i've fixed up the -git patch: http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-git-v18.patch Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:20 ` Ingo Molnar @ 2007-06-26 3:02 ` Andrew Morton 2007-06-26 8:38 ` Ingo Molnar 0 siblings, 1 reply; 43+ messages in thread From: Andrew Morton @ 2007-06-26 3:02 UTC (permalink / raw) To: Ingo Molnar Cc: S.Çağlar Onur, linux-kernel, Linus Torvalds, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri On Sat, 23 Jun 2007 00:20:36 +0200 Ingo Molnar <mingo@elte.hu> wrote: > > * S.Çağlar Onur <caglar@pardus.org.tr> wrote: > > > > kernel/sched.c:745:28: sched_idletask.c: No such file or directory > > > > Ahh and this happens with [1], grabbing sched_idletask.c from .18 one solves > > the problem... > > oops, indeed - i've fixed up the -git patch: > > http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-git-v18.patch > So I locally generated the diff to take -mm up to the above version of CFS. - sys_sched_yield_to() went away? I guess I missed that. - Curious. the simplification of task_tick_rt() seems to go only halfway. Could do if (p->policy != SCHED_RR) return; if (--p->time_slice) return; /* stuff goes here */ - dud macro: #define is_rt_policy(p) ((p) == SCHED_FIFO || (p) == SCHED_RR) It evaluates its arg twice and could and should be coded in C. There are a bunch of other don't-need-to-be-implemented-as-a-macro macros around there too. Generally, I suggest you review all the patchset for macros-which-don't-need-to-be-macros. - Extraneous newline: enum cpu_idle_type { - Style thing: struct sched_entity { struct load_weight load; /* for nice- load-balancing purposes */ int on_rq; struct rb_node run_node; unsigned long delta_exec; s64 delta_fair; u64 wait_start_fair; u64 wait_start; u64 exec_start; u64 sleep_start, sleep_start_fair; u64 block_start; u64 sleep_max; u64 block_max; u64 exec_max; u64 wait_max; u64 last_ran; s64 wait_runtime; u64 sum_exec_runtime; s64 fair_key; s64 sum_wait_runtime, sum_sleep_runtime; unsigned long wait_runtime_overruns, wait_runtime_underruns; }; I think the one-definition-per-line style is better than the `unsigned long foo,bar,zot,zit;' thing. Easier to read, easier to read subsequent patches and it leaves more room for a comment describing what the field does. - None of these fields have comments describing what they do ;) - __exit_signal() does apparently-unlocked 64-bit arith. Is there some implicit locking here or do we not care about the occasional race-induced inaccuracy? (ditto, lots of places, I expect) (Gee, there's shitloads of 64-bit stuff in there. Does it all _really_ need to be 64-bit on 32-bit?) - weight_s64() (what does this do?) looks too big to inline on 32-bit. - update_stats_enqueue() looks too big to inline even on 64-bit. - Overall, this change is tremendously huge for something which is supposedly ready-to-merge. Looks like a lot of that is the sched_entity conversion, but afaict there's quite a lot besides. - Should "4" in (sysctl_sched_features & 4) be enumerated? - Maybe even __check_preempt_curr_fair() is too porky to inline. - There really is an astonishing amount of 64-bit arith in here... - Some opportunities for useful comments have been missed ;) #define NICE_0_LOAD SCHED_LOAD_SCALE #define NICE_0_SHIFT SCHED_LOAD_SHIFT <wonders what these mean> - Should any of those new 64-bit arith functions in sched.c be pulled out and made generic? - update_curr_load() is huge, inlined and has several callsites? - lots more macros-which-dont-need-to-be-macros in sched.c: load_weight(), PRIO_TO_load_weight(), RTPRIO_TO_load_weight(), maybe others. People are more inclined to comment functions than they are macros, for some reason. - inc_load(), dec_load(), inc_nr_running(), dec_nr_running(): these will generate plenty of code on 32-bit and they're all inlined with multiple callsites. - overall, CFS takes sched.o from 41157 of .text up to 48781 on x86_64, which at 18% is rather a large bloat. Hopefully a lot of this is the new debug stuff. - On i386 sched.o went from 33755 up to 43660 which is 29% growth. Possibly acceptable, but why did it increase a lot more than the x86_64 version? All that 64-bit arith, I assume? - style (or the lack thereof): p->se.sum_wait_runtime = p->se.sum_sleep_runtime = 0; p->se.sleep_start = p->se.sleep_start_fair = p->se.block_start = 0; p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0; p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0; bit of an eyesore? - in sched_init() this looks funny: rq->ls.load_update_last = sched_clock(); rq->ls.load_update_start = sched_clock(); was it intended that these both get the same value? ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-26 3:02 ` Andrew Morton @ 2007-06-26 8:38 ` Ingo Molnar 2007-06-26 9:00 ` Andrew Morton 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-06-26 8:38 UTC (permalink / raw) To: Andrew Morton Cc: S.Çağlar Onur, linux-kernel, Linus Torvalds, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri * Andrew Morton <akpm@linux-foundation.org> wrote: > So I locally generated the diff to take -mm up to the above version of > CFS. thx. I released a diff against mm2: http://people.redhat.com/mingo/cfs-scheduler/sched-cfs-v2.6.22-rc4-mm2-v18.patch but indeed the -git diff serves you better if you updated -mm to Linus' latest. firstly, thanks a ton for your review feedback! > - sys_sched_yield_to() went away? I guess I missed that. yep. Nobody tried it and sent any feedback on it, it was causing patch-logistical complications both in -mm and for packagers that bundle CFS (the experimental-schedulers site has a CFS repo and Fedora rawhide started carrying CFS recently as well), and i dont really agree with adding yet another yield interface anyway. So we can and should do this independently of CFS. > - Curious. the simplification of task_tick_rt() seems to go only > halfway. Could do > > if (p->policy != SCHED_RR) > return; > > if (--p->time_slice) > return; > > /* stuff goes here */ yeah. I have fixed it in my v19 tree for it to look like you suggest. > - dud macro: > > #define is_rt_policy(p) ((p) == SCHED_FIFO || (p) == SCHED_RR) > > It evaluates its arg twice and could and should be coded in C. > > There are a bunch of other don't-need-to-be-implemented-as-a-macro > macros around there too. Generally, I suggest you review all the > patchset for macros-which-don't-need-to-be-macros. yep, fixed. (is a historic macro) > - Extraneous newline: > > enum cpu_idle_type > { fixed. (is a pre-existing enum) > - Style thing: > > struct sched_entity { > u64 sleep_start, sleep_start_fair; fixed. > - None of these fields have comments describing what they do ;) one of them has ;-) Will fill this in. > - __exit_signal() does apparently-unlocked 64-bit arith. Is there > some implicit locking here or do we not care about the occasional > race-induced inaccuracy? do you mean the tsk->se.sum_exec_runtime addition, etc? That runs with interrupts disabled so sum_sched_runtime is protected. > (ditto, lots of places, I expect) which places do you mean? > (Gee, there's shitloads of 64-bit stuff in there. Does it all > _really_ need to be 64-bit on 32-bit?) yes - CFS is fundamentally designed for 64-bit, with still pretty OK arithmetics performance for 32-bit. > - weight_s64() (what does this do?) looks too big to inline on 32-bit. ok, i've uninlined it. > - update_stats_enqueue() looks too big to inline even on 64-bit. done. > - Overall, this change is tremendously huge for something which is > supposedly ready-to-merge. [...] hey, that's not fair, your review comments just made it 10K larger ;-) > [...] Looks like a lot of that is the sched_entity conversion, but > afaict there's quite a lot besides. > > - Should "4" in > > (sysctl_sched_features & 4) > > be enumerated? yep, done. > - Maybe even __check_preempt_curr_fair() is too porky to inline. yep - undone. > - There really is an astonishing amount of 64-bit arith in here... > > - Some opportunities for useful comments have been missed ;) > > #define NICE_0_LOAD SCHED_LOAD_SCALE > #define NICE_0_SHIFT SCHED_LOAD_SHIFT > > <wonders what these mean> SCHED_LOAD_SCALE is the smpnice stuff. CFS reuses that and also makes it clear via this define that a nice-0 task has a 'load' contribution to the CPU as of NICE_0_LOAD. Sometimes, when doing smpnice load-balancing calculations we want to use 'SCHED_LOAD_SCALE', sometimes we want to stress it's NICE_0_LOAD. > - Should any of those new 64-bit arith functions in sched.c be pulled > out and made generic? yep, the plan is to put this all into reciprocal_div.h and to convert existing users of reciprocal_div to the cleaner stuff from Thomas. The patch wont get any smaller due to that though ;-) > - update_curr_load() is huge, inlined and has several callsites? this is a reasonable tradeoff i think - update_curr_load()'s slowpath is in __update_curr_load(). Anyway, it probably wont get inlined when the kernel is built with -Os and without forced-inlining. > - lots more macros-which-dont-need-to-be-macros in sched.c: > load_weight(), PRIO_TO_load_weight(), RTPRIO_TO_load_weight(), maybe > others. People are more inclined to comment functions than they are > macros, for some reason. these are mostly ancient macros. I fixed up some of them in my current tree. > - inc_load(), dec_load(), inc_nr_running(), dec_nr_running(): these will > generate plenty of code on 32-bit and they're all inlined with > multiple callsites. yep - i'll revisit the inlining picture. This is not really a primary worry i think because it's easy to tweak and people can already express their inlining preference via CONFIG_CC_OPTIMIZE_FOR_SIZE and CONFIG_FORCED_INLINING. > - overall, CFS takes sched.o from 41157 of .text up to 48781 on x86_64, > which at 18% is rather a large bloat. Hopefully a lot of this is > the new debug stuff. > - On i386 sched.o went from 33755 up to 43660 which is 29% growth. > Possibly acceptable, but why did it increase a lot more than the x86_64 > version? All that 64-bit arith, I assume? the main reason is the sched debugging stuff: text data bss dec hex filename 37570 2538 20 40128 9cc0 kernel/sched.o 30692 2426 20 33138 8172 kernel/sched-no_sched_debug.o i can make it depend on CONFIG_SCHEDSTATS, although i'd prefer it to be always on. > - style (or the lack thereof): > > p->se.sum_wait_runtime = p->se.sum_sleep_runtime = 0; > p->se.sleep_start = p->se.sleep_start_fair = p->se.block_start = 0; > p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0; > p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0; > > bit of an eyesore? fixed. (this heap grew gradually and now is/was an eyesore indeed.) > - in sched_init() this looks funny: > > rq->ls.load_update_last = sched_clock(); > rq->ls.load_update_start = sched_clock(); > > was it intended that these both get the same value? it doesnt really matter, i fixed them to be initialized to the same 'now' value. i've attached my current fixes. (Please dont apply it yet.) Ingo Not-Signed-off-by: Ingo Molnar <mingo@elte.hu> --- Makefile | 2 include/linux/sched.h | 120 +++++++++++++++++++++++++++++--------------------- kernel/exit.c | 2 kernel/sched.c | 58 ++++++++++++++++-------- kernel/sched_debug.c | 2 kernel/sched_fair.c | 61 ++++++++++++------------- kernel/sched_rt.c | 15 +++--- 7 files changed, 149 insertions(+), 111 deletions(-) Index: linux/Makefile =================================================================== --- linux.orig/Makefile +++ linux/Makefile @@ -1,7 +1,7 @@ VERSION = 2 PATCHLEVEL = 6 SUBLEVEL = 22 -EXTRAVERSION = -rc6-cfs-v18 +EXTRAVERSION = -rc6-cfs-v19 NAME = Holy Dancing Manatees, Batman! # *DOCUMENTATION* Index: linux/include/linux/sched.h =================================================================== --- linux.orig/include/linux/sched.h +++ linux/include/linux/sched.h @@ -528,31 +528,6 @@ struct signal_struct { #define SIGNAL_STOP_CONTINUED 0x00000004 /* SIGCONT since WCONTINUED reap */ #define SIGNAL_GROUP_EXIT 0x00000008 /* group exit in progress */ - -/* - * Priority of a process goes from 0..MAX_PRIO-1, valid RT - * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH - * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority - * values are inverted: lower p->prio value means higher priority. - * - * The MAX_USER_RT_PRIO value allows the actual maximum - * RT priority to be separate from the value exported to - * user-space. This allows kernel threads to set their - * priority to a value higher than any user task. Note: - * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO. - */ - -#define MAX_USER_RT_PRIO 100 -#define MAX_RT_PRIO MAX_USER_RT_PRIO - -#define MAX_PRIO (MAX_RT_PRIO + 40) -#define DEFAULT_PRIO (MAX_RT_PRIO + 20) - -#define rt_prio(prio) unlikely((prio) < MAX_RT_PRIO) -#define rt_task(p) rt_prio((p)->prio) -#define is_rt_policy(p) ((p) == SCHED_FIFO || (p) == SCHED_RR) -#define has_rt_policy(p) unlikely(is_rt_policy((p)->policy)) - /* * Some day this will be a full-fledged user tracking system.. */ @@ -646,8 +621,7 @@ static inline int sched_info_on(void) #endif } -enum cpu_idle_type -{ +enum cpu_idle_type { CPU_IDLE, CPU_NOT_IDLE, CPU_NEWLY_IDLE, @@ -843,30 +817,45 @@ struct load_weight { unsigned long weight, inv_weight; }; -/* CFS stats for a schedulable entity (task, task-group etc) */ +/* + * CFS stats for a schedulable entity (task, task-group etc) + * + * Current field usage histogram: + * + * 4 se->block_start + * 4 se->run_node + * 4 se->sleep_start + * 4 se->sleep_start_fair + * 6 se->load.weight + * 7 se->delta_fair + * 15 se->wait_runtime + */ struct sched_entity { - struct load_weight load; /* for nice- load-balancing purposes */ - int on_rq; - struct rb_node run_node; - unsigned long delta_exec; - s64 delta_fair; - - u64 wait_start_fair; - u64 wait_start; - u64 exec_start; - u64 sleep_start, sleep_start_fair; - u64 block_start; - u64 sleep_max; - u64 block_max; - u64 exec_max; - u64 wait_max; - u64 last_ran; - - s64 wait_runtime; - u64 sum_exec_runtime; - s64 fair_key; - s64 sum_wait_runtime, sum_sleep_runtime; - unsigned long wait_runtime_overruns, wait_runtime_underruns; + s64 wait_runtime; + s64 delta_fair; + struct load_weight load; /* for load-balancing */ + struct rb_node run_node; + int on_rq; + unsigned long delta_exec; + + u64 wait_start_fair; + u64 wait_start; + u64 exec_start; + u64 sleep_start; + u64 sleep_start_fair; + u64 block_start; + u64 sleep_max; + u64 block_max; + u64 exec_max; + u64 wait_max; + u64 last_ran; + + u64 sum_exec_runtime; + s64 fair_key; + s64 sum_wait_runtime; + s64 sum_sleep_runtime; + unsigned long wait_runtime_overruns; + unsigned long wait_runtime_underruns; }; struct task_struct { @@ -1126,6 +1115,37 @@ struct task_struct { #endif }; +/* + * Priority of a process goes from 0..MAX_PRIO-1, valid RT + * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH + * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority + * values are inverted: lower p->prio value means higher priority. + * + * The MAX_USER_RT_PRIO value allows the actual maximum + * RT priority to be separate from the value exported to + * user-space. This allows kernel threads to set their + * priority to a value higher than any user task. Note: + * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO. + */ + +#define MAX_USER_RT_PRIO 100 +#define MAX_RT_PRIO MAX_USER_RT_PRIO + +#define MAX_PRIO (MAX_RT_PRIO + 40) +#define DEFAULT_PRIO (MAX_RT_PRIO + 20) + +static inline int rt_prio(int prio) +{ + if (unlikely(prio < MAX_RT_PRIO)) + return 1; + return 0; +} + +static inline int rt_task(struct task_struct *p) +{ + return rt_prio(p->prio); +} + static inline pid_t process_group(struct task_struct *tsk) { return tsk->signal->pgrp; Index: linux/kernel/exit.c =================================================================== --- linux.orig/kernel/exit.c +++ linux/kernel/exit.c @@ -290,7 +290,7 @@ static void reparent_to_kthreadd(void) /* Set the exit signal to SIGCHLD so we signal init on exit */ current->exit_signal = SIGCHLD; - if (!has_rt_policy(current) && (task_nice(current) < 0)) + if (task_nice(current) < 0) set_user_nice(current, 0); /* cpus_allowed? */ /* rt_priority? */ Index: linux/kernel/sched.c =================================================================== --- linux.orig/kernel/sched.c +++ linux/kernel/sched.c @@ -106,6 +106,18 @@ unsigned long long __attribute__((weak)) #define MIN_TIMESLICE max(5 * HZ / 1000, 1) #define DEF_TIMESLICE (100 * HZ / 1000) +static inline int rt_policy(int policy) +{ + if (unlikely(policy == SCHED_FIFO) || unlikely(policy == SCHED_RR)) + return 1; + return 0; +} + +static inline int task_has_rt_policy(struct task_struct *p) +{ + return rt_policy(p->policy); +} + /* * This is the priority-queue data structure of the RT scheduling class: */ @@ -752,7 +764,7 @@ static void set_load_weight(struct task_ task_rq(p)->cfs.wait_runtime -= p->se.wait_runtime; p->se.wait_runtime = 0; - if (has_rt_policy(p)) { + if (task_has_rt_policy(p)) { p->se.load.weight = prio_to_weight[0] * 2; p->se.load.inv_weight = prio_to_wmult[0] >> 1; return; @@ -805,7 +817,7 @@ static inline int normal_prio(struct tas { int prio; - if (has_rt_policy(p)) + if (task_has_rt_policy(p)) prio = MAX_RT_PRIO-1 - p->rt_priority; else prio = __normal_prio(p); @@ -1476,17 +1488,24 @@ int fastcall wake_up_state(struct task_s */ static void __sched_fork(struct task_struct *p) { - p->se.wait_start_fair = p->se.wait_start = p->se.exec_start = 0; - p->se.sum_exec_runtime = 0; - p->se.delta_exec = 0; - p->se.delta_fair = 0; - - p->se.wait_runtime = 0; - - p->se.sum_wait_runtime = p->se.sum_sleep_runtime = 0; - p->se.sleep_start = p->se.sleep_start_fair = p->se.block_start = 0; - p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0; - p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0; + p->se.wait_start_fair = 0; + p->se.wait_start = 0; + p->se.exec_start = 0; + p->se.sum_exec_runtime = 0; + p->se.delta_exec = 0; + p->se.delta_fair = 0; + p->se.wait_runtime = 0; + p->se.sum_wait_runtime = 0; + p->se.sum_sleep_runtime = 0; + p->se.sleep_start = 0; + p->se.sleep_start_fair = 0; + p->se.block_start = 0; + p->se.sleep_max = 0; + p->se.block_max = 0; + p->se.exec_max = 0; + p->se.wait_max = 0; + p->se.wait_runtime_overruns = 0; + p->se.wait_runtime_underruns = 0; INIT_LIST_HEAD(&p->run_list); p->se.on_rq = 0; @@ -1799,7 +1818,7 @@ static void update_cpu_load(struct rq *t int i, scale; this_rq->nr_load_updates++; - if (sysctl_sched_features & 64) + if (unlikely(!(sysctl_sched_features & SCHED_FEAT_PRECISE_CPU_LOAD))) goto do_avg; /* Update delta_fair/delta_exec fields first */ @@ -3801,7 +3820,7 @@ void set_user_nice(struct task_struct *p * it wont have any effect on scheduling until the task is * SCHED_FIFO/SCHED_RR: */ - if (has_rt_policy(p)) { + if (task_has_rt_policy(p)) { p->static_prio = NICE_TO_PRIO(nice); goto out_unlock; } @@ -3999,14 +4018,14 @@ recheck: (p->mm && param->sched_priority > MAX_USER_RT_PRIO-1) || (!p->mm && param->sched_priority > MAX_RT_PRIO-1)) return -EINVAL; - if (is_rt_policy(policy) != (param->sched_priority != 0)) + if (rt_policy(policy) != (param->sched_priority != 0)) return -EINVAL; /* * Allow unprivileged RT tasks to decrease priority: */ if (!capable(CAP_SYS_NICE)) { - if (is_rt_policy(policy)) { + if (rt_policy(policy)) { unsigned long rlim_rtprio; unsigned long flags; @@ -6186,6 +6205,7 @@ int in_sched_functions(unsigned long add void __init sched_init(void) { + u64 now = sched_clock(); int highest_cpu = 0; int i, j; @@ -6206,8 +6226,8 @@ void __init sched_init(void) rq->nr_running = 0; rq->cfs.tasks_timeline = RB_ROOT; rq->clock = rq->cfs.fair_clock = 1; - rq->ls.load_update_last = sched_clock(); - rq->ls.load_update_start = sched_clock(); + rq->ls.load_update_last = now; + rq->ls.load_update_start = now; for (j = 0; j < CPU_LOAD_IDX_MAX; j++) rq->cpu_load[j] = 0; Index: linux/kernel/sched_debug.c =================================================================== --- linux.orig/kernel/sched_debug.c +++ linux/kernel/sched_debug.c @@ -157,7 +157,7 @@ static int sched_debug_show(struct seq_f u64 now = ktime_to_ns(ktime_get()); int cpu; - SEQ_printf(m, "Sched Debug Version: v0.03, cfs-v18, %s %.*s\n", + SEQ_printf(m, "Sched Debug Version: v0.03, cfs-v19, %s %.*s\n", init_utsname()->release, (int)strcspn(init_utsname()->version, " "), init_utsname()->version); Index: linux/kernel/sched_fair.c =================================================================== --- linux.orig/kernel/sched_fair.c +++ linux/kernel/sched_fair.c @@ -61,8 +61,24 @@ unsigned int sysctl_sched_stat_granulari */ unsigned int sysctl_sched_runtime_limit __read_mostly; +/* + * Debugging: various feature bits + */ +enum { + SCHED_FEAT_IGNORE_PREEMPTED = 1, + SCHED_FEAT_DISTRIBUTE = 2, + SCHED_FEAT_FAIR_SLEEPERS = 4, + SCHED_FEAT_SLEEPER_AVG = 32, + SCHED_FEAT_PRECISE_CPU_LOAD = 64, + SCHED_FEAT_START_DEBIT = 128, + SCHED_FEAT_SKIP_INITIAL = 256, +}; + unsigned int sysctl_sched_features __read_mostly = - 0 | 2 | 4 | 8 | 0 | 0 | 0 | 0; + SCHED_FEAT_DISTRIBUTE | + SCHED_FEAT_FAIR_SLEEPERS | + SCHED_FEAT_SLEEPER_AVG | + SCHED_FEAT_PRECISE_CPU_LOAD; extern struct sched_class fair_sched_class; @@ -145,7 +161,7 @@ static inline void __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) { if (cfs_rq->rb_leftmost == &se->run_node) - cfs_rq->rb_leftmost = NULL; + cfs_rq->rb_leftmost = rb_next(&se->run_node); rb_erase(&se->run_node, &cfs_rq->tasks_timeline); update_load_sub(&cfs_rq->load, se->load.weight); cfs_rq->nr_running--; @@ -258,7 +274,7 @@ __update_curr(struct cfs_rq *cfs_rq, str * Task already marked for preemption, do not burden * it with the cost of not having left the CPU yet: */ - if (unlikely(sysctl_sched_features & 1)) + if (unlikely(sysctl_sched_features & SCHED_FEAT_IGNORE_PREEMPTED)) if (unlikely(test_tsk_thread_flag(curtask, TIF_NEED_RESCHED))) return; @@ -305,7 +321,7 @@ update_stats_wait_start(struct cfs_rq *c se->wait_start = now; } -static inline s64 weight_s64(s64 calc, unsigned long weight, int shift) +static s64 weight_s64(s64 calc, unsigned long weight, int shift) { if (calc < 0) { calc = - calc * weight; @@ -317,7 +333,7 @@ static inline s64 weight_s64(s64 calc, u /* * Task is being enqueued - update stats: */ -static inline void +static void update_stats_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se, u64 now) { s64 key; @@ -438,7 +454,7 @@ static void distribute_fair_add(struct c struct sched_entity *curr = cfs_rq_curr(cfs_rq); s64 delta_fair = 0; - if (!(sysctl_sched_features & 2)) + if (!(sysctl_sched_features & SCHED_FEAT_DISTRIBUTE)) return; if (cfs_rq->nr_running) { @@ -469,7 +485,7 @@ __enqueue_sleeper(struct cfs_rq *cfs_rq, * Fix up delta_fair with the effect of us running * during the whole sleep period: */ - if (!(sysctl_sched_features & 32)) + if (sysctl_sched_features & SCHED_FEAT_SLEEPER_AVG) delta_fair = div64_s(delta_fair * load, load + se->load.weight); delta_fair = weight_s64(delta_fair, se->load.weight, NICE_0_SHIFT); @@ -495,7 +511,7 @@ enqueue_sleeper(struct cfs_rq *cfs_rq, s s64 delta_fair; if ((entity_is_task(se) && tsk->policy == SCHED_BATCH) || - !(sysctl_sched_features & 4)) + !(sysctl_sched_features & SCHED_FEAT_FAIR_SLEEPERS)) return; delta_fair = cfs_rq->fair_clock - se->sleep_start_fair; @@ -574,7 +590,7 @@ static void dequeue_entity(struct cfs_rq /* * Preempt the current task with a newly woken task if needed: */ -static inline void +static void __check_preempt_curr_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, struct sched_entity *curr, unsigned long granularity) { @@ -612,23 +628,6 @@ put_prev_entity(struct cfs_rq *cfs_rq, s int updated = 0; /* - * If the task is still waiting for the CPU (it just got - * preempted), update its position within the tree and - * start the wait period: - */ - if ((sysctl_sched_features & 16) && entity_is_task(prev)) { - struct task_struct *prevtask = task_of(prev); - - if (prev->on_rq && - test_tsk_thread_flag(prevtask, TIF_NEED_RESCHED)) { - - dequeue_entity(cfs_rq, prev, 0, now); - enqueue_entity(cfs_rq, prev, 0, now); - updated = 1; - } - } - - /* * If still on the runqueue then deactivate_task() * was not called and update_curr() has to be done: */ @@ -741,10 +740,8 @@ static void check_preempt_curr_fair(stru unsigned long gran; if (unlikely(rt_prio(p->prio))) { - if (sysctl_sched_features & 8) { - if (rt_prio(p->prio)) - update_curr(cfs_rq, rq_clock(rq)); - } + if (rt_prio(p->prio)) + update_curr(cfs_rq, rq_clock(rq)); resched_task(curr); return; } @@ -850,14 +847,14 @@ static void task_new_fair(struct rq *rq, * The first wait is dominated by the child-runs-first logic, * so do not credit it with that waiting time yet: */ - if (sysctl_sched_features & 256) + if (sysctl_sched_features & SCHED_FEAT_SKIP_INITIAL) p->se.wait_start_fair = 0; /* * The statistical average of wait_runtime is about * -granularity/2, so initialize the task with that: */ - if (sysctl_sched_features & 128) + if (sysctl_sched_features & SCHED_FEAT_START_DEBIT) p->se.wait_runtime = -(s64)(sysctl_sched_granularity / 2); __enqueue_entity(cfs_rq, se); Index: linux/kernel/sched_rt.c =================================================================== --- linux.orig/kernel/sched_rt.c +++ linux/kernel/sched_rt.c @@ -12,7 +12,7 @@ static inline void update_curr_rt(struct struct task_struct *curr = rq->curr; u64 delta_exec; - if (!has_rt_policy(curr)) + if (!task_has_rt_policy(curr)) return; delta_exec = now - curr->se.exec_start; @@ -179,13 +179,14 @@ static void task_tick_rt(struct rq *rq, if (p->policy != SCHED_RR) return; - if (!(--p->time_slice)) { - p->time_slice = static_prio_timeslice(p->static_prio); - set_tsk_need_resched(p); + if (--p->time_slice) + return; - /* put it at the end of the queue: */ - requeue_task_rt(rq, p); - } + p->time_slice = static_prio_timeslice(p->static_prio); + set_tsk_need_resched(p); + + /* put it at the end of the queue: */ + requeue_task_rt(rq, p); } /* ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-26 8:38 ` Ingo Molnar @ 2007-06-26 9:00 ` Andrew Morton 2007-06-26 9:38 ` Ingo Molnar 0 siblings, 1 reply; 43+ messages in thread From: Andrew Morton @ 2007-06-26 9:00 UTC (permalink / raw) To: Ingo Molnar Cc: S.Çağlar Onur, linux-kernel, Linus Torvalds, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri On Tue, 26 Jun 2007 10:38:13 +0200 Ingo Molnar <mingo@elte.hu> wrote: > > > - __exit_signal() does apparently-unlocked 64-bit arith. Is there > > some implicit locking here or do we not care about the occasional > > race-induced inaccuracy? > > do you mean the tsk->se.sum_exec_runtime addition, etc? That runs with > interrupts disabled so sum_sched_runtime is protected. > > > (ditto, lots of places, I expect) > > which places do you mean? I forget ;) There seemed to be rather a lot of 64-bit addition with no obvious locking in sight, that's all. > > ... > > (Gee, there's shitloads of 64-bit stuff in there. Does it all > > _really_ need to be 64-bit on 32-bit?) > > yes - CFS is fundamentally designed for 64-bit, with still pretty OK > arithmetics performance for 32-bit. It may have been designed for 64-bit, but was that the correct design? The cost on 32-bit appears to be pretty high. Perhaps a round of uninlining will help. > > - overall, CFS takes sched.o from 41157 of .text up to 48781 on x86_64, > > which at 18% is rather a large bloat. Hopefully a lot of this is > > the new debug stuff. > > > - On i386 sched.o went from 33755 up to 43660 which is 29% growth. > > Possibly acceptable, but why did it increase a lot more than the x86_64 > > version? All that 64-bit arith, I assume? > > the main reason is the sched debugging stuff: That would serve to explain the 18% growth on x86_64. But why did i386 grow by much more: 29%? I'd be suspecting all the new 64-bit arithmetic. ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-26 9:00 ` Andrew Morton @ 2007-06-26 9:38 ` Ingo Molnar 0 siblings, 0 replies; 43+ messages in thread From: Ingo Molnar @ 2007-06-26 9:38 UTC (permalink / raw) To: Andrew Morton Cc: S.??a??lar Onur, linux-kernel, Linus Torvalds, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri * Andrew Morton <akpm@linux-foundation.org> wrote: > > the main reason is the sched debugging stuff: > > That would serve to explain the 18% growth on x86_64. But why did > i386 grow by much more: 29%? I'd be suspecting all the new 64-bit > arithmetic. this is what i see on 32-bit: text data bss dec hex filename 28732 3905 24 32661 7f95 kernel/sched.o-vanilla 37986 2538 20 40544 9e60 kernel/sched.o-v18 31092 2426 20 33538 8302 kernel/sched.o-v18-no_sched_debug text is larger but data got smaller. While they are not equivalent in function, the two almost even out each other (and that's without any of the uninlining that is in v19). In fact, there's a 1.5K per CPU percpu data size win with CFS, which is not visible in this stat. So on dual-core the net cost should already be zero. > The cost on 32-bit appears to be pretty high. Perhaps a round of > uninlining will help. agreed, i've done one more round of uninlining. Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar 2007-06-22 22:09 ` S.Çağlar Onur @ 2007-06-22 23:08 ` Gene Heskett 2007-06-23 7:11 ` Ingo Molnar 2007-06-23 10:22 ` Antonino Ingargiola ` (4 subsequent siblings) 6 siblings, 1 reply; 43+ messages in thread From: Gene Heskett @ 2007-06-22 23:08 UTC (permalink / raw) To: Ingo Molnar Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri On Friday 22 June 2007, Ingo Molnar wrote: >i'm pleased to announce release -v18 of the CFS scheduler patchset. > >The rolled-up CFS patch against today's -git kernel, v2.6.22-rc5, >v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be downloaded from the >usual place: > > http://people.redhat.com/mingo/cfs-scheduler/ > >The biggest change in -v18 are various performance related improvements. >Thomas Gleixner has eliminated expensive 64-bit divisions by converting >the arithmetics to scaled math (without impacting the quality of >calculations). Srivatsa Vaddagiri and Dmitry Adamushko have continued >the abstraction and cleanup work. Srivatsa Vaddagiri and Christoph >Lameter fixed the NUMA balancing bug reported by Paul McKenney. There >were also a good number of other refinements to the CFS code. (No >reproducible behavioral regressions were reported against -v17 so far, >so the 'behavioral' bits are mostly unchanged.) > >Changes since -v17: > > - implement scaled math speedups for CFS. (Thomas Gleixner) > > - lots of core code updates, cleanups and streamlining. > (Srivatsa Vaddagiri, Dmitry Adamushko, me.) > > - bugfix: fix NUMA balancing. (Srivatsa Vaddagiri, Christoph Lameter, > Paul E. McKenney) > > - feature: SCHED_IDLE now also implies block-scheduler (CFQ) > idle-IO-priority. (suggested by Thomas Sattler, picked up from -ck) > > - build fix for ppc32. (reported, tested and confirmed fixed by > Art Haas) > > - ARM fix. (reported and debugged by Thomas Gleixner) > > - cleanup: implemented idle_sched_class in kernel/sched_idletask.c as a > way to separate out rq->idle handling out of the core scheduler. This > made a good deal of idle-task related special-cases go away. > > - debug: make the sysctls safer by introducing high and low limits. > > - cleanup: move some of the debug counters to under CONFIG_SCHEDSTATS. > > - speedup: various micro-optimizations > > - various other small updates. > >As usual, any sort of feedback, bugreport, fix and suggestion is more >than welcome! > Humm, problem methinks. Applying the patch, with 2.6.22-rc5 applied to 2.6.21 completed, from my script: now applying patch sched-cfs-v2.6.22-rc5-v18.patch patching file Documentation/kernel-parameters.txt patching file Documentation/sched-design-CFS.txt patching file Makefile patching file arch/i386/kernel/smpboot.c patching file arch/i386/kernel/tsc.c patching file arch/ia64/kernel/setup.c patching file arch/mips/kernel/smp.c patching file arch/sparc/kernel/smp.c patching file arch/sparc64/kernel/smp.c patching file block/cfq-iosched.c patching file fs/proc/array.c patching file fs/proc/base.c patching file include/asm-generic/bitops/sched.h patching file include/linux/hardirq.h patching file include/linux/sched.h patching file include/linux/topology.h patching file init/main.c patching file kernel/delayacct.c patching file kernel/exit.c patching file kernel/fork.c patching file kernel/posix-cpu-timers.c patching file kernel/sched.c patching file kernel/sched_debug.c patching file kernel/sched_fair.c patching file kernel/sched_idletask.c patching file kernel/sched_rt.c patching file kernel/sched_stats.h patching file kernel/softirq.c patching file kernel/sysctl.c The next patch would delete the file l/kernel/sched.c, which does not exist! Assume -R? [n] How to proceed? -- Cheers, Gene "There are four boxes to be used in defense of liberty: soap, ballot, jury, and ammo. Please use in that order." -Ed Howdershelt (Author) Even more amazing was the realization that God has Internet access. I wonder if He has a full newsfeed? -- Matt Welsh ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 23:08 ` Gene Heskett @ 2007-06-23 7:11 ` Ingo Molnar 2007-06-23 9:55 ` Gene Heskett 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-06-23 7:11 UTC (permalink / raw) To: Gene Heskett Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri * Gene Heskett <gene.heskett@gmail.com> wrote: > patching file kernel/softirq.c > patching file kernel/sysctl.c > The next patch would delete the file l/kernel/sched.c, > which does not exist! Assume -R? [n] > > How to proceed? oops - just ignore it, or re-download the patch (i fixed it). Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-23 7:11 ` Ingo Molnar @ 2007-06-23 9:55 ` Gene Heskett 0 siblings, 0 replies; 43+ messages in thread From: Gene Heskett @ 2007-06-23 9:55 UTC (permalink / raw) To: linux-kernel On Saturday 23 June 2007, Ingo Molnar wrote: >* Gene Heskett <gene.heskett@gmail.com> wrote: >> patching file kernel/softirq.c >> patching file kernel/sysctl.c >> The next patch would delete the file l/kernel/sched.c, >> which does not exist! Assume -R? [n] >> >> How to proceed? > >oops - just ignore it, or re-download the patch (i fixed it). > > Ingo answering n for all that, I note the build, at the end of the make bzImage, spits out this: MODPOST vmlinux WARNING: arch/i386/kernel/built-in.o(.text+0x845d): Section mismatch: reference to .init.text:amd_init_mtrr (between 'mtrr_bp_init' and 'mtrr_save_state') WARNING: arch/i386/kernel/built-in.o(.text+0x8462): Section mismatch: reference to .init.text:cyrix_init_mtrr (between 'mtrr_bp_init' and 'mtrr_save_state') WARNING: arch/i386/kernel/built-in.o(.text+0x8467): Section mismatch: reference to .init.text:centaur_init_mtrr (between 'mtrr_bp_init' and 'mtrr_save_state') WARNING: arch/i386/kernel/built-in.o(.text+0x9284): Section mismatch: reference to .init.text: (between 'get_mtrr_state' and 'generic_get_mtrr') WARNING: arch/i386/kernel/built-in.o(.text+0x9298): Section mismatch: reference to .init.text: (between 'get_mtrr_state' and 'generic_get_mtrr') WARNING: arch/i386/kernel/built-in.o(.text+0x92bc): Section mismatch: reference to .init.text: (between 'get_mtrr_state' and 'generic_get_mtrr') But then proceeds with the make modules stage. I believe I've seen references to this in other threads. Is It Serious? -- Cheers, Gene "There are four boxes to be used in defense of liberty: soap, ballot, jury, and ammo. Please use in that order." -Ed Howdershelt (Author) Debug is human, de-fix divine. ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar 2007-06-22 22:09 ` S.Çağlar Onur 2007-06-22 23:08 ` Gene Heskett @ 2007-06-23 10:22 ` Antonino Ingargiola 2007-06-23 17:25 ` Ingo Molnar 2007-06-23 13:24 ` Willy Tarreau ` (3 subsequent siblings) 6 siblings, 1 reply; 43+ messages in thread From: Antonino Ingargiola @ 2007-06-23 10:22 UTC (permalink / raw) To: Ingo Molnar; +Cc: linux-kernel Hi, 2007/6/23, Ingo Molnar <mingo@elte.hu>: > > i'm pleased to announce release -v18 of the CFS scheduler patchset. I'm running -v18 on 2.6.22-rc5, no problems so far. How can I change a task to SCHED_IDLE or SCHED_BATCH priority under CFS? Thanks, ~ Antonio ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-23 10:22 ` Antonino Ingargiola @ 2007-06-23 17:25 ` Ingo Molnar 2007-06-24 10:02 ` Antonino Ingargiola 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-06-23 17:25 UTC (permalink / raw) To: Antonino Ingargiola; +Cc: linux-kernel * Antonino Ingargiola <tritemio@gmail.com> wrote: > 2007/6/23, Ingo Molnar <mingo@elte.hu>: > > > >i'm pleased to announce release -v18 of the CFS scheduler patchset. > > I'm running -v18 on 2.6.22-rc5, no problems so far. How can I change a > task to SCHED_IDLE or SCHED_BATCH priority under CFS? pick up schedtool, and these are the choices it gives: -N for SCHED_NORMAL -F -p PRIO for SCHED_FIFO only as root -R -p PRIO for SCHED_RR only as root -B for SCHED_BATCH -I -p PRIO for SCHED_ISO -D for SCHED_IDLEPRIO then for example to start up something as SCHED_IDLE: schedtool -D -e ./somecommand.sh Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-23 17:25 ` Ingo Molnar @ 2007-06-24 10:02 ` Antonino Ingargiola 2007-06-24 11:07 ` Ingo Molnar 0 siblings, 1 reply; 43+ messages in thread From: Antonino Ingargiola @ 2007-06-24 10:02 UTC (permalink / raw) To: Ingo Molnar; +Cc: linux-kernel 2007/6/23, Ingo Molnar <mingo@elte.hu>: > > * Antonino Ingargiola <tritemio@gmail.com> wrote: > > > 2007/6/23, Ingo Molnar <mingo@elte.hu>: > > > > > >i'm pleased to announce release -v18 of the CFS scheduler patchset. > > > > I'm running -v18 on 2.6.22-rc5, no problems so far. How can I change a > > task to SCHED_IDLE or SCHED_BATCH priority under CFS? > > pick up schedtool, and these are the choices it gives: > > -N for SCHED_NORMAL > -F -p PRIO for SCHED_FIFO only as root > -R -p PRIO for SCHED_RR only as root > -B for SCHED_BATCH > -I -p PRIO for SCHED_ISO > -D for SCHED_IDLEPRIO > > then for example to start up something as SCHED_IDLE: > > schedtool -D -e ./somecommand.sh > Thank you very much! I was thinking that schedtool was suitable only for -ck. I've installed schedtool and it works fine. Anyway, I've discovered with great pleasure that CFS has also the SCHED_ISO priority. I may have missed something, but I don't remember to have read this in any of the CFS release notes :). For me this is a really useful feature. Thanks. Regards, ~ Antonio > Ingo > ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-24 10:02 ` Antonino Ingargiola @ 2007-06-24 11:07 ` Ingo Molnar 2007-06-25 7:27 ` Antonino Ingargiola 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-06-24 11:07 UTC (permalink / raw) To: Antonino Ingargiola; +Cc: linux-kernel * Antonino Ingargiola <tritemio@gmail.com> wrote: > Anyway, I've discovered with great pleasure that CFS has also the > SCHED_ISO priority. I may have missed something, but I don't remember > to have read this in any of the CFS release notes :). For me this is a > really useful feature. Thanks. well, it's only a hack and emulated: SCHED_ISO in CFS is recognized as a policy but it falls back to SCHED_NORMAL. Could you check how well this (i.e. SCHED_NORMAL) works for your workload, compared to SD's SCHED_ISO? If you'd like to increase the priority of a task, i'd suggest to use negative nice levels. (use the 'nice' option in /etc/security/limits.conf with a newer version of PAM to allow unprivileged users to use negative nice levels.) Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-24 11:07 ` Ingo Molnar @ 2007-06-25 7:27 ` Antonino Ingargiola 0 siblings, 0 replies; 43+ messages in thread From: Antonino Ingargiola @ 2007-06-25 7:27 UTC (permalink / raw) To: Ingo Molnar; +Cc: linux-kernel 2007/6/24, Ingo Molnar <mingo@elte.hu>: > > * Antonino Ingargiola <tritemio@gmail.com> wrote: > > > Anyway, I've discovered with great pleasure that CFS has also the > > SCHED_ISO priority. I may have missed something, but I don't remember > > to have read this in any of the CFS release notes :). For me this is a > > really useful feature. Thanks. > > well, it's only a hack and emulated: SCHED_ISO in CFS is recognized as a > policy but it falls back to SCHED_NORMAL. Could you check how well this > (i.e. SCHED_NORMAL) works for your workload, compared to SD's SCHED_ISO? To be fair, my workload is not really "critical". I'm used to skip-free audio listening (no matter what) since long time running my audio player with SCHED_ISO. Even in mainline the skips aren't so frequent, but still annoying. I'm using SCHED_ISO for the confidence it gives in providing skip-free audio. For my modest needs also CFS SCHED_NORMAL has been just fine (in these latest days). I'll report if I can find a more critical workload that can possibly stress CFS SCHED_NORMAL. Regards, ~ Antonio ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar ` (2 preceding siblings ...) 2007-06-23 10:22 ` Antonino Ingargiola @ 2007-06-23 13:24 ` Willy Tarreau 2007-06-24 15:52 ` Ingo Molnar 2007-06-26 20:17 ` Fortier,Vincent [Montreal] ` (2 subsequent siblings) 6 siblings, 1 reply; 43+ messages in thread From: Willy Tarreau @ 2007-06-23 13:24 UTC (permalink / raw) To: Ingo Molnar Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri Hi Ingo, On Sat, Jun 23, 2007 at 12:02:02AM +0200, Ingo Molnar wrote: > > i'm pleased to announce release -v18 of the CFS scheduler patchset. > > The rolled-up CFS patch against today's -git kernel, v2.6.22-rc5, > v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be downloaded from the > usual place: > > http://people.redhat.com/mingo/cfs-scheduler/ > > The biggest change in -v18 are various performance related improvements. > Thomas Gleixner has eliminated expensive 64-bit divisions by converting > the arithmetics to scaled math (without impacting the quality of > calculations). Srivatsa Vaddagiri and Dmitry Adamushko have continued > the abstraction and cleanup work. Srivatsa Vaddagiri and Christoph > Lameter fixed the NUMA balancing bug reported by Paul McKenney. There > were also a good number of other refinements to the CFS code. (No > reproducible behavioral regressions were reported against -v17 so far, > so the 'behavioral' bits are mostly unchanged.) Today I had a little time to try CFS again (last time it was -v9!). I ran it on top of 2.6.20.14, and simply tried ocbench again. You remember ? With -v9, I ran 64 processes which all progressed very smoothly. With -v18, it's not the case anymore. When I run 64 processes, only 7 of them show smooth rounds, while all the other ones are only updated once a second. Sometimes they only progress by one iteration, sometimes by a full round. Some are even updated once ever 2 seconds, because if I drag an xterm above them and quickly remove it, the xterm leaves a trace there for up to 2 seconds. Also, only one of my 2 CPUs is used. I see the rq vary between 1 and 5, with a permanent 50% idle... : procs memory swap io system cpu r b w swpd free buff cache si so bi bo in cs us sy id 1 0 0 0 874400 7864 90436 0 0 0 0 279 2204 50 0 50 3 0 0 0 874408 7864 90436 0 0 0 0 273 2122 50 1 50 1 0 0 0 874408 7864 90436 0 0 0 0 253 1660 49 1 50 3 0 0 0 874408 7864 90436 0 0 0 0 252 1977 50 0 50 2 0 0 0 874408 7864 90436 0 0 0 0 253 2274 49 1 50 3 0 0 0 874408 7864 90436 0 0 0 0 252 1846 49 1 50 1 0 0 0 874408 7864 90436 0 0 0 0 339 1782 49 1 50 I have no idea about what version brought that unexpected behaviour, but it's clearly something which needs to be tracked down. My scheduler is at 250 Hz, and here are the values I found in /proc/sys/kernel: root@pcw:kernel# grep '' sched_* sched_batch_wakeup_granularity_ns:40000000 sched_child_runs_first:1 sched_features:14 sched_granularity_ns:10000000 sched_runtime_limit_ns:40000000 sched_stat_granularity_ns:0 sched_wakeup_granularity_ns:4000000 I have tried to change each of them, with absolutely no effect. Seems really strange. Unfortunately, I have to leave right now. Maybe you can try on your side with this simple command : $ ./ocbench -x 8 -y 8 Sorry for not giving more information right now. Regards, Willy ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-23 13:24 ` Willy Tarreau @ 2007-06-24 15:52 ` Ingo Molnar 2007-06-24 17:08 ` Willy Tarreau 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-06-24 15:52 UTC (permalink / raw) To: Willy Tarreau Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri * Willy Tarreau <w@1wt.eu> wrote: > Today I had a little time to try CFS again (last time it was -v9!). I > ran it on top of 2.6.20.14, and simply tried ocbench again. You > remember ? With -v9, I ran 64 processes which all progressed very > smoothly. With -v18, it's not the case anymore. When I run 64 > processes, only 7 of them show smooth rounds, while all the other ones > are only updated once a second. Sometimes they only progress by one > iteration, sometimes by a full round. Some are even updated once ever > 2 seconds, because if I drag an xterm above them and quickly remove > it, the xterm leaves a trace there for up to 2 seconds. > > Also, only one of my 2 CPUs is used. I see the rq vary between 1 and > 5, with a permanent 50% idle... : > > procs memory swap io system cpu > r b w swpd free buff cache si so bi bo in cs us sy id > 1 0 0 0 874400 7864 90436 0 0 0 0 279 2204 50 0 50 > 3 0 0 0 874408 7864 90436 0 0 0 0 273 2122 50 1 50 > 1 0 0 0 874408 7864 90436 0 0 0 0 253 1660 49 1 50 > 3 0 0 0 874408 7864 90436 0 0 0 0 252 1977 50 0 50 > 2 0 0 0 874408 7864 90436 0 0 0 0 253 2274 49 1 50 > 3 0 0 0 874408 7864 90436 0 0 0 0 252 1846 49 1 50 > 1 0 0 0 874408 7864 90436 0 0 0 0 339 1782 49 1 50 > > I have no idea about what version brought that unexpected behaviour, > but it's clearly something which needs to be tracked down. hm, the two problems might be related. Could you try v17 perhaps? In v18 i have 'unified' all the sched.c's between the various kernel releases, maybe that brought in something unexpected on 2.6.20.14. (perhaps try v2.6.21.5 based cfs too?) > My scheduler is at 250 Hz, and here are the values I found in > /proc/sys/kernel: could you send me the file the cfs-debug-info.sh script produced. You can pick the script up from: http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh (i'd suggest to send it in private mail, output is large and detailed. If your kernel has no /proc/config.gz then please send me your .config file too.) Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-24 15:52 ` Ingo Molnar @ 2007-06-24 17:08 ` Willy Tarreau 2007-06-24 20:31 ` Ingo Molnar 0 siblings, 1 reply; 43+ messages in thread From: Willy Tarreau @ 2007-06-24 17:08 UTC (permalink / raw) To: Ingo Molnar Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri Hi Ingo, On Sun, Jun 24, 2007 at 05:52:14PM +0200, Ingo Molnar wrote: > > * Willy Tarreau <w@1wt.eu> wrote: > > > Today I had a little time to try CFS again (last time it was -v9!). I > > ran it on top of 2.6.20.14, and simply tried ocbench again. You > > remember ? With -v9, I ran 64 processes which all progressed very > > smoothly. With -v18, it's not the case anymore. When I run 64 > > processes, only 7 of them show smooth rounds, while all the other ones > > are only updated once a second. Sometimes they only progress by one > > iteration, sometimes by a full round. Some are even updated once ever > > 2 seconds, because if I drag an xterm above them and quickly remove > > it, the xterm leaves a trace there for up to 2 seconds. > > > > Also, only one of my 2 CPUs is used. I see the rq vary between 1 and > > 5, with a permanent 50% idle... : > > > > procs memory swap io system cpu > > r b w swpd free buff cache si so bi bo in cs us sy id > > 1 0 0 0 874400 7864 90436 0 0 0 0 279 2204 50 0 50 > > 3 0 0 0 874408 7864 90436 0 0 0 0 273 2122 50 1 50 > > 1 0 0 0 874408 7864 90436 0 0 0 0 253 1660 49 1 50 > > 3 0 0 0 874408 7864 90436 0 0 0 0 252 1977 50 0 50 > > 2 0 0 0 874408 7864 90436 0 0 0 0 253 2274 49 1 50 > > 3 0 0 0 874408 7864 90436 0 0 0 0 252 1846 49 1 50 > > 1 0 0 0 874408 7864 90436 0 0 0 0 339 1782 49 1 50 > > > > I have no idea about what version brought that unexpected behaviour, > > but it's clearly something which needs to be tracked down. > > hm, the two problems might be related. Could you try v17 perhaps? In v18 > i have 'unified' all the sched.c's between the various kernel releases, > maybe that brought in something unexpected on 2.6.20.14. (perhaps try > v2.6.21.5 based cfs too?) Well, forget this, I'm nuts. I'm sorry, but I did not set any of the -R and -S parameter on ocbench, which means that all the processes ran at full speed and did not sleep. The load distribution was not fair, but since they put a lot of stress on the X server, I think it might be one of the reasons for the unfairness. I got the same behaviour with -v17, -v9 and even 2.4 ! It told me something was wrong on my side ;-) I've retried with 50%/50% run/sleep, and it now works like a charm. It's perfectly smooth with both small and long run/sleep times (between 1 and 100 ms). I think that with X saturated, it might explain why I only had one CPU running at 100% ! Next time, I'll try to take a bit more time for such a test. > could you send me the file the cfs-debug-info.sh script produced. You > can pick the script up from: > > http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh OK I got it, but I've not run it since the problem was between the keyboard and the chair. If you want an output anyway, I can give it a run. Sorry again for the wrong alert. regards, willy ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-24 17:08 ` Willy Tarreau @ 2007-06-24 20:31 ` Ingo Molnar 0 siblings, 0 replies; 43+ messages in thread From: Ingo Molnar @ 2007-06-24 20:31 UTC (permalink / raw) To: Willy Tarreau Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri * Willy Tarreau <w@1wt.eu> wrote: > > > I have no idea about what version brought that unexpected > > > behaviour, but it's clearly something which needs to be tracked > > > down. > > > > hm, the two problems might be related. Could you try v17 perhaps? In > > v18 i have 'unified' all the sched.c's between the various kernel > > releases, maybe that brought in something unexpected on 2.6.20.14. > > (perhaps try v2.6.21.5 based cfs too?) > > Well, forget this, I'm nuts. I'm sorry, but I did not set any of the > -R and -S parameter on ocbench, which means that all the processes ran > at full speed and did not sleep. The load distribution was not fair, > but since they put a lot of stress on the X server, I think it might > be one of the reasons for the unfairness. I got the same behaviour > with -v17, -v9 and even 2.4 ! It told me something was wrong on my > side ;-) > > I've retried with 50%/50% run/sleep, and it now works like a charm. > It's perfectly smooth with both small and long run/sleep times > (between 1 and 100 ms). I think that with X saturated, it might > explain why I only had one CPU running at 100% ! ah, great! :-) My testbox needs a 90% / 10% ratio between sleep/run for an 8x8 matrix of ocbench tasks to not overload the X server. Once the overload happens X starts penalizing certain clients it finds abusive (i think), and that mechanism seems to be wall-clock based and it thus brings in alot of non-determinism and skews the clients. Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* RE: [patch] CFS scheduler, -v18 2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar ` (3 preceding siblings ...) 2007-06-23 13:24 ` Willy Tarreau @ 2007-06-26 20:17 ` Fortier,Vincent [Montreal] 2007-06-27 10:51 ` Ingo Molnar 2007-06-30 21:06 ` Willy Tarreau 2007-07-02 11:44 ` Vegard Nossum 6 siblings, 1 reply; 43+ messages in thread From: Fortier,Vincent [Montreal] @ 2007-06-26 20:17 UTC (permalink / raw) To: Ingo Molnar, linux-kernel > -----Message d'origine----- > De : linux-kernel-owner@vger.kernel.org > [mailto:linux-kernel-owner@vger.kernel.org] De la part de Ingo Molnar > Envoyé : 22 juin 2007 18:02 > > i'm pleased to announce release -v18 of the CFS scheduler patchset. > > The rolled-up CFS patch against today's -git kernel, > v2.6.22-rc5, v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be > downloaded from the usual place: > > http://people.redhat.com/mingo/cfs-scheduler/ > > ... Pre-built kernels for Fedora 7 & Fedora Core 6 (yum reporsitory available) and Debian Sarge/Etch using latest CFS v18 available at http://linux-dev.qc.ec.gc.ca > > As usual, any sort of feedback, bugreport, fix and suggestion > is more than welcome! > > Ingo - vin ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-26 20:17 ` Fortier,Vincent [Montreal] @ 2007-06-27 10:51 ` Ingo Molnar 0 siblings, 0 replies; 43+ messages in thread From: Ingo Molnar @ 2007-06-27 10:51 UTC (permalink / raw) To: Fortier,Vincent [Montreal]; +Cc: linux-kernel * Fortier,Vincent [Montreal] <Vincent.Fortier1@EC.GC.CA> wrote: > > -----Message d'origine----- > > De : linux-kernel-owner@vger.kernel.org > > [mailto:linux-kernel-owner@vger.kernel.org] De la part de Ingo Molnar > > Envoyé : 22 juin 2007 18:02 > > > > i'm pleased to announce release -v18 of the CFS scheduler patchset. > > > > The rolled-up CFS patch against today's -git kernel, > > v2.6.22-rc5, v2.6.22-rc4-mm2, v2.6.21.5 or v2.6.20.14 can be > > downloaded from the usual place: > > > > http://people.redhat.com/mingo/cfs-scheduler/ > > > > ... > > Pre-built kernels for Fedora 7 & Fedora Core 6 (yum reporsitory > available) and Debian Sarge/Etch using latest CFS v18 available at > http://linux-dev.qc.ec.gc.ca thanks Vincent! Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar ` (4 preceding siblings ...) 2007-06-26 20:17 ` Fortier,Vincent [Montreal] @ 2007-06-30 21:06 ` Willy Tarreau 2007-07-01 8:31 ` Ingo Molnar 2007-07-02 11:44 ` Vegard Nossum 6 siblings, 1 reply; 43+ messages in thread From: Willy Tarreau @ 2007-06-30 21:06 UTC (permalink / raw) To: Ingo Molnar Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri Ingo, I've accidentally discovered a problem with -v18. Some time ago, I wrote a small program to prevent my laptop from entering low-power mode, and noticed that after upgrading my laptop's kernel from 2.4.20.9+cfs-v6 to 2.4.20.14+cfs-v18, it completely freezes if I run this program. The program is trivial, it just sets its prio to nice +20 and forks a busy loop. I've added the ability to stop the loop after a user-definable number of iterations, and I can confirm that it unfreezes when the loop ends. I'm not even root when I run it. Everything freezes, including the frame-buffer. It's not 100% reproducible, I would say 90% only. Sometimes it requires a few seconds before freezing. It *seems* to me that running another task in parallel (such as vmstat) increases its chances to freeze. It seems like nicing to +19 does not cause any trouble. I've tried it on my dual athlon, and with 1 process I see occasional pauses of 1 or 2 seconds, and with 2 processes, I see fairly larger pauses. Here's the trivial program. Probably you'll find an obvious bug. Regards, Willy --- /* * cfs-freeze.c * Fork a busy loop running with idle prio. This often results * in complete freezes with CFS-v18. * * $ gcc -O2 -s -o cfs-freeze cfs-freeze.c * $ ./cfs-freeze */ #include <stdio.h> #include <stdlib.h> #include <unistd.h> #include <sched.h> #include <sys/time.h> #include <sys/resource.h> int main(int argc, char **argv) { struct sched_param sch; long long i; if (argc > 1) i = atoll(argv[1]); if (i <= 0) i = 4 * 1000 * 1000 * 1000ULL; sch.sched_priority = 0; sched_setscheduler(0, SCHED_OTHER, &sch); setpriority(PRIO_PROCESS, 0, 20); if (fork() == 0) while (i--); return 0; } --- ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-30 21:06 ` Willy Tarreau @ 2007-07-01 8:31 ` Ingo Molnar 2007-07-01 8:45 ` Ingo Molnar 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-07-01 8:31 UTC (permalink / raw) To: Willy Tarreau Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri * Willy Tarreau <w@1wt.eu> wrote: > Ingo, > > I've accidentally discovered a problem with -v18. > > Some time ago, I wrote a small program to prevent my laptop from > entering low-power mode, and noticed that after upgrading my laptop's > kernel from 2.4.20.9+cfs-v6 to 2.4.20.14+cfs-v18, it completely > freezes if I run this program. > > The program is trivial, it just sets its prio to nice +20 and forks a > busy loop. I've added the ability to stop the loop after a > user-definable number of iterations, and I can confirm that it > unfreezes when the loop ends. I'm not even root when I run it. hm, i tried your test-app and it causes no problems here. (which is not a surprise - your app starts a nice +19 busy loop, which is one of the common tests i do here too.) To further debug this, could you try to create a 'high priority shell' on a text console (i.e. not under X) that is SCHED_FIFO prio 98? Something like: chrt -f -p 98 $$ should do the trick. And then run this script: http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh before the test, during the test and after the test, from the high-prio shell session. (the shell runs at SCHED_FIFO, so the expectation would be for that it will be able to run during the test too) Then please send me the resulting 3 debug files. Thanks, Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-01 8:31 ` Ingo Molnar @ 2007-07-01 8:45 ` Ingo Molnar 2007-07-01 9:00 ` Willy Tarreau 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-07-01 8:45 UTC (permalink / raw) To: Willy Tarreau Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri Willy, could you check whether your current v18 CFS tree has the fix below included? I discovered it right after having released v18 so i updated the v18 files in place - but maybe you downloaded an early version? I thought it's relatively harmless, that it would only affect SCHED_IDLE tasks, but maybe it affects nice +19 tasks too on your box! Ingo Index: linux/kernel/sched_fair.c =================================================================== --- linux.orig/kernel/sched_fair.c +++ linux/kernel/sched_fair.c @@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_ s64 tmp; if (se->wait_runtime < 0) { - tmp = (0 - se->wait_runtime) << NICE_0_SHIFT; - key += (tmp * se->load.inv_weight) >> WMULT_SHIFT; + tmp = -se->wait_runtime; + key += (tmp * se->load.inv_weight) >> + (WMULT_SHIFT - NICE_0_SHIFT); } else { tmp = se->wait_runtime * se->load.weight; key -= tmp >> NICE_0_SHIFT; ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-01 8:45 ` Ingo Molnar @ 2007-07-01 9:00 ` Willy Tarreau 0 siblings, 0 replies; 43+ messages in thread From: Willy Tarreau @ 2007-07-01 9:00 UTC (permalink / raw) To: Ingo Molnar Cc: linux-kernel, Linus Torvalds, Andrew Morton, Mike Galbraith, Arjan van de Ven, Thomas Gleixner, Dmitry Adamushko, Srivatsa Vaddagiri Hi Ingo, On Sun, Jul 01, 2007 at 10:45:01AM +0200, Ingo Molnar wrote: > could you check whether your current v18 CFS tree has the fix below > included? I discovered it right after having released v18 so i updated > the v18 files in place - but maybe you downloaded an early version? I > thought it's relatively harmless, that it would only affect SCHED_IDLE > tasks, but maybe it affects nice +19 tasks too on your box! Good catch, it was the cause of the problem. I've just applied your fix below and rebuilt and the system behaves perfectly now. Thanks very much ! Willy ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar ` (5 preceding siblings ...) 2007-06-30 21:06 ` Willy Tarreau @ 2007-07-02 11:44 ` Vegard Nossum 2007-07-02 13:01 ` Dmitry Adamushko ` (2 more replies) 6 siblings, 3 replies; 43+ messages in thread From: Vegard Nossum @ 2007-07-02 11:44 UTC (permalink / raw) To: Ingo Molnar; +Cc: linux-kernel Hello, On 6/23/07, Ingo Molnar <mingo@elte.hu> wrote: > i'm pleased to announce release -v18 of the CFS scheduler patchset. > As usual, any sort of feedback, bugreport, fix and suggestion is more > than welcome! I have been running cfs-v18 for a couple of days now, and today I stumbled upon a rather strange problem. Consider the following short program: while(1) printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC); Running this in an xterm makes the xterm totally unresponsive. Ctrl-C takes about two seconds to terminate the program, during which the program will keep running. In fact, it seems that the longer it runs, the longer it takes to terminate (towards 5 seconds after running for a couple of minutes). This is rather surprising, as the rest of the system is quite responsive (even remarkably so). I think this is also in contrast with the expected behaviour, that Ctrl-C/program termination should be prioritized somehow. Some other observations: X.Org seems to be running at about 75% CPU on CPU 1, the xterm at about 45% on CPU 0, and a.out at about 20% on CPU 0. (HT processor) Killing with -2 or -9 from another terminal works immediately. Ctrl-Z takes the same time as Ctrl-C. Another thing to note is that simply looping with no output retains the expected responsiveness of the xterm. Printing i++ is somewhere halfway in between. Is this behaviour expected or even intended? My main point is that Ctrl-C is a safety fallback which suddenly doesn't work as usual. I might even go so far as to call it a regression. I'd also like to point out that Folding@Home seems to draw more CPU than it should. Or, at least, in top, it shows up as using 50% CPU even though other processes are demanding as much as they can get. The FAH program should be running with idle priority. I expect it to fall to near 0% when other programs are running at full speed, but it keeps trotting along. And I am pretty sure that this is not due to SMP/HT (I made sure to utilize both CPUs). Lastly, I'd like to mention that I got BUGs (soft lockups) with -v8, though it has not been reproducible with -v18, so I suppose it must have been fixed already. Otherwise, I am satisfied with the performance of CFS. Especially the desktop is noticably smoother. Thanks! Kind regards, Vegard Nossum ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 11:44 ` Vegard Nossum @ 2007-07-02 13:01 ` Dmitry Adamushko 2007-07-02 13:43 ` Vegard Nossum 2007-07-02 14:13 ` Bill Davidsen 2007-07-03 7:15 ` Ingo Molnar 2 siblings, 1 reply; 43+ messages in thread From: Dmitry Adamushko @ 2007-07-02 13:01 UTC (permalink / raw) To: Vegard Nossum; +Cc: Ingo Molnar, Linux Kernel On 02/07/07, Vegard Nossum <vegard.nossum@gmail.com> wrote: > Hello, > > On 6/23/07, Ingo Molnar <mingo@elte.hu> wrote: > > i'm pleased to announce release -v18 of the CFS scheduler patchset. > > > As usual, any sort of feedback, bugreport, fix and suggestion is more > > than welcome! > > I have been running cfs-v18 for a couple of days now, and today I > stumbled upon a rather strange problem. Consider the following short > program: > > while(1) > printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC); Is it running with a default (0) nice value? could you please run the following script when your application is running? As you have pointed out : "... In fact, it seems that the longer it runs, the longer it takes to terminate (towards 5 seconds after running for a couple of minutes ..." , please run the script a few times : say, before starting up your application, 10 sec. after it's got started, 1 minute, a few minute... http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh then send us the resulting files. TIA, -- Best regards, Dmitry Adamushko ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 13:01 ` Dmitry Adamushko @ 2007-07-02 13:43 ` Vegard Nossum 2007-07-02 15:50 ` Ingo Molnar 0 siblings, 1 reply; 43+ messages in thread From: Vegard Nossum @ 2007-07-02 13:43 UTC (permalink / raw) To: Dmitry Adamushko; +Cc: Ingo Molnar, Linux Kernel On 7/2/07, Dmitry Adamushko <dmitry.adamushko@gmail.com> wrote: > On 02/07/07, Vegard Nossum <vegard.nossum@gmail.com> wrote: > > I have been running cfs-v18 for a couple of days now, and today I > > stumbled upon a rather strange problem. Consider the following short > > program: > > > > while(1) > > printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC); > > Is it running with a default (0) nice value? Yes. > could you please run the following script when your application is running? > As you have pointed out : > > "... In fact, it seems that the longer it runs, the longer it takes to > terminate (towards 5 seconds after running for a couple of minutes > ..." , (I don't think it'll ever take more than 5 seconds, though.) > please run the script a few times : say, before starting up your > application, 10 sec. after it's got started, 1 minute, a few minute... > > http://people.redhat.com/mingo/cfs-scheduler/tools/cfs-debug-info.sh Resulting files at http://vegard.afraid.org:1104/pub/cfs/ cfs-debug-info-2007.07.02-15:18:13 Before running program cfs-debug-info-2007.07.02-15:19:51 ~10 secs after start cfs-debug-info-2007.07.02-15:20:54 ~1 minute after start cfs-debug-info-2007.07.02-15:25:52 ~5 minutes after start cfs-debug-info-2007.07.02-15:30:54 ~10 minutes after start a.out is my program, FahCore_78 is the f@h client. Hope this helps. Vegard ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 13:43 ` Vegard Nossum @ 2007-07-02 15:50 ` Ingo Molnar 2007-07-02 16:40 ` Vegard Nossum 0 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-07-02 15:50 UTC (permalink / raw) To: Vegard Nossum; +Cc: Dmitry Adamushko, Linux Kernel * Vegard Nossum <vegard.nossum@gmail.com> wrote: > Resulting files at > http://vegard.afraid.org:1104/pub/cfs/ > > cfs-debug-info-2007.07.02-15:18:13 Before running program > cfs-debug-info-2007.07.02-15:19:51 ~10 secs after start > cfs-debug-info-2007.07.02-15:20:54 ~1 minute after start > cfs-debug-info-2007.07.02-15:25:52 ~5 minutes after start > cfs-debug-info-2007.07.02-15:30:54 ~10 minutes after start > > a.out is my program, FahCore_78 is the f@h client. thx. As an initial matter, could you double-check whether your v18 kernel source has the patch below applied already? Ingo Index: linux/kernel/sched_fair.c =================================================================== --- linux.orig/kernel/sched_fair.c +++ linux/kernel/sched_fair.c @@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_ s64 tmp; if (se->wait_runtime < 0) { - tmp = (0 - se->wait_runtime) << NICE_0_SHIFT; - key += (tmp * se->load.inv_weight) >> WMULT_SHIFT; + tmp = -se->wait_runtime; + key += (tmp * se->load.inv_weight) >> + (WMULT_SHIFT - NICE_0_SHIFT); } else { tmp = se->wait_runtime * se->load.weight; key -= tmp >> NICE_0_SHIFT; ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 15:50 ` Ingo Molnar @ 2007-07-02 16:40 ` Vegard Nossum 2007-07-02 18:13 ` Ingo Molnar 2007-07-03 7:12 ` Mike Galbraith 0 siblings, 2 replies; 43+ messages in thread From: Vegard Nossum @ 2007-07-02 16:40 UTC (permalink / raw) To: Ingo Molnar; +Cc: Dmitry Adamushko, Linux Kernel On 7/2/07, Ingo Molnar <mingo@elte.hu> wrote: > thx. As an initial matter, could you double-check whether your v18 > kernel source has the patch below applied already? > > Ingo > > Index: linux/kernel/sched_fair.c > =================================================================== > --- linux.orig/kernel/sched_fair.c > +++ linux/kernel/sched_fair.c > @@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_ > s64 tmp; > > if (se->wait_runtime < 0) { > - tmp = (0 - se->wait_runtime) << NICE_0_SHIFT; > - key += (tmp * se->load.inv_weight) >> WMULT_SHIFT; > + tmp = -se->wait_runtime; > + key += (tmp * se->load.inv_weight) >> > + (WMULT_SHIFT - NICE_0_SHIFT); > } else { > tmp = se->wait_runtime * se->load.weight; > key -= tmp >> NICE_0_SHIFT; > > It does. Vegard ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 16:40 ` Vegard Nossum @ 2007-07-02 18:13 ` Ingo Molnar 2007-07-03 7:01 ` Vegard Nossum 2007-07-03 7:12 ` Mike Galbraith 1 sibling, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-07-02 18:13 UTC (permalink / raw) To: Vegard Nossum; +Cc: Dmitry Adamushko, Linux Kernel * Vegard Nossum <vegard.nossum@gmail.com> wrote: > > thx. As an initial matter, could you double-check whether your v18 > > kernel source has the patch below applied already? > > It does. ok. Does the xterm slowdown get any better if you do: echo 46 > /proc/sys/kernel/sched_features ? The default on v18 is: echo 14 > /proc/sys/kernel/sched_features Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 18:13 ` Ingo Molnar @ 2007-07-03 7:01 ` Vegard Nossum 0 siblings, 0 replies; 43+ messages in thread From: Vegard Nossum @ 2007-07-03 7:01 UTC (permalink / raw) To: Ingo Molnar; +Cc: Dmitry Adamushko, Linux Kernel On 7/2/07, Ingo Molnar <mingo@elte.hu> wrote: > ok. Does the xterm slowdown get any better if you do: > > echo 46 > /proc/sys/kernel/sched_features > > ? The default on v18 is: > > echo 14 > /proc/sys/kernel/sched_features No. The Ctrl-C still hangs between 1 and 3 seconds, again seemingly depending on how long the program runs before I terminate it. Vegard ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 16:40 ` Vegard Nossum 2007-07-02 18:13 ` Ingo Molnar @ 2007-07-03 7:12 ` Mike Galbraith 2007-07-03 7:22 ` Ingo Molnar 1 sibling, 1 reply; 43+ messages in thread From: Mike Galbraith @ 2007-07-03 7:12 UTC (permalink / raw) To: Vegard Nossum; +Cc: Ingo Molnar, Dmitry Adamushko, Linux Kernel On Mon, 2007-07-02 at 18:40 +0200, Vegard Nossum wrote: > On 7/2/07, Ingo Molnar <mingo@elte.hu> wrote: > > thx. As an initial matter, could you double-check whether your v18 > > kernel source has the patch below applied already? > > > > Ingo > > > > Index: linux/kernel/sched_fair.c > > =================================================================== > > --- linux.orig/kernel/sched_fair.c > > +++ linux/kernel/sched_fair.c > > @@ -342,8 +342,9 @@ update_stats_enqueue(struct cfs_rq *cfs_ > > s64 tmp; > > > > if (se->wait_runtime < 0) { > > - tmp = (0 - se->wait_runtime) << NICE_0_SHIFT; > > - key += (tmp * se->load.inv_weight) >> WMULT_SHIFT; > > + tmp = -se->wait_runtime; > > + key += (tmp * se->load.inv_weight) >> > > + (WMULT_SHIFT - NICE_0_SHIFT); > > } else { > > tmp = se->wait_runtime * se->load.weight; > > key -= tmp >> NICE_0_SHIFT; > > > > > > It does. Hi, This doesn't appear to be a CFS problem. I can reproduce the problem easily in virgin 2.6.22-rc7 by starting xterm-spam at nice -1 or better. As soon as xterm-spam can get enough CPU to keep the xterm fully busy, it's game over, the xterm freezes. The more accurate fairness of CFS to sleepers just tips the balance quicker. In mainline, the xterm has an unfair advantage and maintains it indefinitely... until you tip the scales just a wee bit, at which time it inverts. -Mike ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-03 7:12 ` Mike Galbraith @ 2007-07-03 7:22 ` Ingo Molnar 2007-07-03 8:08 ` Keith Packard 2007-07-04 12:11 ` Andi Kleen 0 siblings, 2 replies; 43+ messages in thread From: Ingo Molnar @ 2007-07-03 7:22 UTC (permalink / raw) To: Mike Galbraith Cc: Vegard Nossum, Dmitry Adamushko, Linux Kernel, Keith Packard [-- Attachment #1: Type: text/plain, Size: 1214 bytes --] * Mike Galbraith <efault@gmx.de> wrote: > This doesn't appear to be a CFS problem. I can reproduce the problem > easily in virgin 2.6.22-rc7 by starting xterm-spam at nice -1 or > better. As soon as xterm-spam can get enough CPU to keep the xterm > fully busy, it's game over, the xterm freezes. The more accurate > fairness of CFS to sleepers just tips the balance quicker. In > mainline, the xterm has an unfair advantage and maintains it > indefinitely... until you tip the scales just a wee bit, at which time > it inverts. ah. That indeed makes sense. It seems like the xterm doesnt process the Ctrl-C/Z keypresses _at all_ when it is 'spammed' with output. Normally, output 'spam' is throttled by the scroll buffer's overhead. But in Vegard's case, the printout involves a \r carriage return: printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC); which allows xterm-spam (attached) to easily flood the xterm (without any scrolling that would act as a throttle) and the xterm to flood Xorg. I suspect we need the help of an xterm/Xorg expert? (maybe Keith can give us further pointers? I can reproduce the problem on a T60 with i940 and Core2Duo running Fedora 7 + Xorg 7.1.) Ingo [-- Attachment #2: xterm-spam.c --] [-- Type: text/plain, Size: 101 bytes --] #include <time.h> main () { while(1) printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC); } ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-03 7:22 ` Ingo Molnar @ 2007-07-03 8:08 ` Keith Packard 2007-07-03 8:31 ` Ingo Molnar 2007-07-04 12:11 ` Andi Kleen 1 sibling, 1 reply; 43+ messages in thread From: Keith Packard @ 2007-07-03 8:08 UTC (permalink / raw) To: Ingo Molnar Cc: keith.packard, Mike Galbraith, Vegard Nossum, Dmitry Adamushko, Linux Kernel [-- Attachment #1: Type: text/plain, Size: 1439 bytes --] On Tue, 2007-07-03 at 09:22 +0200, Ingo Molnar wrote: > which allows xterm-spam (attached) to easily flood the xterm (without > any scrolling that would act as a throttle) and the xterm to flood Xorg. It's just an Xterm bug. Xterm will look for X input if it ever manages to fill the input buffer past 32768 bytes. If it manages to get more than 4096 bytes in one read, it will invoke sched_yield. and then check for input. Gotta love that sched_yield call. As it always processes all of the incoming data before trying to read again, there doesn't appear to be any way it can ever have more than 32768 characters in the buffer. And, as the kernel will not buffer more than 4095 bytes from a pty, there isn't any way it will ever read 4096 bytes. So, it sits there carefully reading every byte from the pty and painting them on the screen. You can 'fix' xterm with: $ xterm -xrm '*minBufSize: 4095' I hesitate to even suggest a patch to xterm that would solve this problem correctly. Note that xterm has kludges in several of the output processing steps which explicitly look for input (most vertical cursor motion, it seems), which is why any application which scrolls doesn't cause this problem. Do you need more reasons to switch to another terminal emulator? gnome-terminal has finally gotten reasonable; I expect rxvt or konsole would work just as well. -- keith.packard@intel.com [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-03 8:08 ` Keith Packard @ 2007-07-03 8:31 ` Ingo Molnar 0 siblings, 0 replies; 43+ messages in thread From: Ingo Molnar @ 2007-07-03 8:31 UTC (permalink / raw) To: Keith Packard Cc: Mike Galbraith, Vegard Nossum, Dmitry Adamushko, Linux Kernel * Keith Packard <keith.packard@intel.com> wrote: > On Tue, 2007-07-03 at 09:22 +0200, Ingo Molnar wrote: > > > which allows xterm-spam (attached) to easily flood the xterm > > (without any scrolling that would act as a throttle) and the xterm > > to flood Xorg. > > It's just an Xterm bug. > > Xterm will look for X input if it ever manages to fill the input > buffer past 32768 bytes. If it manages to get more than 4096 bytes in > one read, it will invoke sched_yield. and then check for input. Gotta > love that sched_yield call. > > As it always processes all of the incoming data before trying to read > again, there doesn't appear to be any way it can ever have more than > 32768 characters in the buffer. > > And, as the kernel will not buffer more than 4095 bytes from a pty, > there isn't any way it will ever read 4096 bytes. > > So, it sits there carefully reading every byte from the pty and > painting them on the screen. ah. Thanks for the explanation! > You can 'fix' xterm with: > > $ xterm -xrm '*minBufSize: 4095' indeed, that solves the xterm-spam Ctrl-C/Ctrl-Z problem here! > I hesitate to even suggest a patch to xterm that would solve this > problem correctly. Note that xterm has kludges in several of the > output processing steps which explicitly look for input (most vertical > cursor motion, it seems), which is why any application which scrolls > doesn't cause this problem. > > Do you need more reasons to switch to another terminal emulator? > gnome-terminal has finally gotten reasonable; I expect rxvt or konsole > would work just as well. yeah, i use gnome-terminal exclusively. But testers looking for CFS regressions do run every shell on the planet :-) gnome-terminal is also faster all around (at least on my box): $ (echo '#!/bin/bash' ; echo 'for ((i=0; i<100000; i++)); do echo $i; done') > 1.sh; chmod +x 1.sh; time xterm $HOME/1.sh; time gnome-terminal -x ./1.sh real 0m3.193s user 0m2.840s sys 0m0.460s real 0m2.495s user 0m2.430s sys 0m1.520s Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-03 7:22 ` Ingo Molnar 2007-07-03 8:08 ` Keith Packard @ 2007-07-04 12:11 ` Andi Kleen 1 sibling, 0 replies; 43+ messages in thread From: Andi Kleen @ 2007-07-04 12:11 UTC (permalink / raw) To: Ingo Molnar Cc: Mike Galbraith, Vegard Nossum, Dmitry Adamushko, Linux Kernel, Keith Packard Ingo Molnar <mingo@elte.hu> writes: > > ah. That indeed makes sense. It seems like the xterm doesnt process the > Ctrl-C/Z keypresses _at all_ when it is 'spammed' with output. Normally, > output 'spam' is throttled by the scroll buffer's overhead. But in > Vegard's case, the printout involves a \r carriage return: > > printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC); > > which allows xterm-spam (attached) to easily flood the xterm (without > any scrolling that would act as a throttle) and the xterm to flood Xorg. > > I suspect we need the help of an xterm/Xorg expert? (maybe Keith can > give us further pointers? I can reproduce the problem on a T60 with i940 > and Core2Duo running Fedora 7 + Xorg 7.1.) Xorg seems to have a couple of starvation issues. e.g. I found the Gantt view in icemon during a busy compile session can starve all other X clients for tenths of seconds. -Andi ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 11:44 ` Vegard Nossum 2007-07-02 13:01 ` Dmitry Adamushko @ 2007-07-02 14:13 ` Bill Davidsen 2007-07-03 7:15 ` Ingo Molnar 2 siblings, 0 replies; 43+ messages in thread From: Bill Davidsen @ 2007-07-02 14:13 UTC (permalink / raw) To: Vegard Nossum; +Cc: Ingo Molnar, linux-kernel Vegard Nossum wrote: > Hello, > > On 6/23/07, Ingo Molnar <mingo@elte.hu> wrote: >> i'm pleased to announce release -v18 of the CFS scheduler patchset. > >> As usual, any sort of feedback, bugreport, fix and suggestion is more >> than welcome! > > I have been running cfs-v18 for a couple of days now, and today I > stumbled upon a rather strange problem. Consider the following short > program: > > while(1) > printf("%ld\r", 1000 * clock() / CLOCKS_PER_SEC); > > Running this in an xterm makes the xterm totally unresponsive. Ctrl-C > takes about two seconds to terminate the program, during which the > program will keep running. In fact, it seems that the longer it runs, > the longer it takes to terminate (towards 5 seconds after running for > a couple of minutes). This is rather surprising, as the rest of the > system is quite responsive (even remarkably so). I think this is also > in contrast with the expected behaviour, that Ctrl-C/program > termination should be prioritized somehow. > This sounds as though it might be related to the issues I see with my "glitch1" script, posted here a while ago. With cfs-v18 the effect of having multiple xterms scrolling is obvious, occasionally they behave as if they were "owed" more CPU and get paid back all at once. I've seen this effect to one degree or another since cfs-v13, which did NOT show the effect. > Some other observations: X.Org seems to be running at about 75% CPU on > CPU 1, the xterm at about 45% on CPU 0, and a.out at about 20% on CPU > 0. (HT processor) > > Killing with -2 or -9 from another terminal works immediately. Ctrl-Z > takes the same time as Ctrl-C. > I think this is because the shell to read the keypress is getting high latency, rather than the process taking a long time to react. I have been wrong before... I read Ingo's reply to this, I'll gather the same information when the test machine is available later this morning and send it off to Ingo. > Another thing to note is that simply looping with no output retains > the expected responsiveness of the xterm. Printing i++ is somewhere > halfway in between. > See http://www.tmr.com/~public/source (note the tilde) for glitch1. > Is this behaviour expected or even intended? My main point is that > Ctrl-C is a safety fallback which suddenly doesn't work as usual. I > might even go so far as to call it a regression. > > I'd also like to point out that Folding@Home seems to draw more CPU > than it should. Or, at least, in top, it shows up as using 50% CPU > even though other processes are demanding as much as they can get. The > FAH program should be running with idle priority. I expect it to fall > to near 0% when other programs are running at full speed, but it keeps > trotting along. And I am pretty sure that this is not due to SMP/HT (I > made sure to utilize both CPUs). > > Lastly, I'd like to mention that I got BUGs (soft lockups) with -v8, > though it has not been reproducible with -v18, so I suppose it must > have been fixed already. > > Otherwise, I am satisfied with the performance of CFS. Especially the > desktop is noticably smoother. Thanks! > > Kind regards, > Vegard Nossum -- Bill Davidsen <davidsen@tmr.com> "We have more to fear from the bungling of the incompetent than from the machinations of the wicked." - from Slashdot ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-02 11:44 ` Vegard Nossum 2007-07-02 13:01 ` Dmitry Adamushko 2007-07-02 14:13 ` Bill Davidsen @ 2007-07-03 7:15 ` Ingo Molnar 2007-07-03 9:11 ` Vegard Nossum 2 siblings, 1 reply; 43+ messages in thread From: Ingo Molnar @ 2007-07-03 7:15 UTC (permalink / raw) To: Vegard Nossum; +Cc: linux-kernel * Vegard Nossum <vegard.nossum@gmail.com> wrote: > I'd also like to point out that Folding@Home seems to draw more CPU > than it should. Or, at least, in top, it shows up as using 50% CPU > even though other processes are demanding as much as they can get. The > FAH program should be running with idle priority. I expect it to fall > to near 0% when other programs are running at full speed, but it keeps > trotting along. And I am pretty sure that this is not due to SMP/HT (I > made sure to utilize both CPUs). your FAH processes are running with nice +19 - that should be enough to throttle them. With v18 you can also mark it SCHED_IDLE: schedtool -D $$ # mark the shell idle SCHED_IDLE gets inherited by child tasks so if you mark the shell that starts up FEH as SCHED_IDLE, all FEH threads should be SCHED_IDLE too. (or you can start it up via schedtool -D -e ... ) does it still get more CPU time than you'd expect it to get? A reniced or SCHED_IDLE task will 'fill in' any idle time that it senses, so in itself it's not an anomaly if a task gets 50% and FEH fills in the remaining 50%. Does it still get CPU time if you start two CPU hogs: for (( N=0; N < 2; N++ )); do ( while :; do :; done ) & done ? If yes, that would be an anomaly. > Otherwise, I am satisfied with the performance of CFS. Especially the > desktop is noticably smoother. Thanks! great! :-) Ingo ^ permalink raw reply [flat|nested] 43+ messages in thread
* Re: [patch] CFS scheduler, -v18 2007-07-03 7:15 ` Ingo Molnar @ 2007-07-03 9:11 ` Vegard Nossum 0 siblings, 0 replies; 43+ messages in thread From: Vegard Nossum @ 2007-07-03 9:11 UTC (permalink / raw) To: Ingo Molnar; +Cc: linux-kernel On 7/3/07, Ingo Molnar <mingo@elte.hu> wrote: > does it still get more CPU time than you'd expect it to get? A reniced > or SCHED_IDLE task will 'fill in' any idle time that it senses, so in > itself it's not an anomaly if a task gets 50% and FEH fills in the > remaining 50%. Does it still get CPU time if you start two CPU hogs: > > for (( N=0; N < 2; N++ )); do ( while :; do :; done ) & done > > ? If yes, that would be an anomaly. No, with this test, it gets the expected amount of CPU. Thanks for clearing it up! Vegard ^ permalink raw reply [flat|nested] 43+ messages in thread
[parent not found: <8yZun-1bO-5@gated-at.bofh.it>]
[parent not found: <8CszT-4hd-11@gated-at.bofh.it>]
[parent not found: <8CtPi-6qj-9@gated-at.bofh.it>]
[parent not found: <8Cus1-7eL-11@gated-at.bofh.it>]
[parent not found: <8CwtU-1Z5-3@gated-at.bofh.it>]
[parent not found: <8Cxgd-3fN-9@gated-at.bofh.it>]
[parent not found: <8CKQ8-7HN-13@gated-at.bofh.it>]
[parent not found: <8CKZV-7Ur-11@gated-at.bofh.it>]
[parent not found: <8CLCt-vX-3@gated-at.bofh.it>]
* Re: [patch] CFS scheduler, -v18 [not found] ` <8CLCt-vX-3@gated-at.bofh.it> @ 2007-07-05 13:29 ` Thomas Dickey [not found] ` <8CM5x-19K-3@gated-at.bofh.it> 1 sibling, 0 replies; 43+ messages in thread From: Thomas Dickey @ 2007-07-05 13:29 UTC (permalink / raw) To: Linux Kernel; +Cc: Keith Packard [-- Attachment #1: Type: text/plain, Size: 829 bytes --] On Tue, Jul 03, 2007 at 10:10:05AM +0200, Keith Packard wrote: > On Tue, 2007-07-03 at 09:22 +0200, Ingo Molnar wrote: > > > which allows xterm-spam (attached) to easily flood the xterm (without > > any scrolling that would act as a throttle) and the xterm to flood Xorg. ...deleting the offhand remarks ...here's a little content: > And, as the kernel will not buffer more than 4095 bytes from a pty, > there isn't any way it will ever read 4096 bytes. The sched_yield() call was added to address problems in an earlier 2.6.x kernel (and the report then stated that it could read more than 4096 because it would have to perform several reads to keep up with a burst of output). ...deleting the other offhand remarks -- Thomas E. Dickey http://invisible-island.net ftp://invisible-island.net [-- Attachment #2: Type: application/pgp-signature, Size: 228 bytes --] ^ permalink raw reply [flat|nested] 43+ messages in thread
[parent not found: <8CM5x-19K-3@gated-at.bofh.it>]
* Re: [patch] CFS scheduler, -v18 [not found] ` <8CM5x-19K-3@gated-at.bofh.it> @ 2007-07-05 13:39 ` Thomas Dickey 0 siblings, 0 replies; 43+ messages in thread From: Thomas Dickey @ 2007-07-05 13:39 UTC (permalink / raw) To: Linux Kernel; +Cc: Ingo Molnar [-- Attachment #1: Type: text/plain, Size: 782 bytes --] On Tue, Jul 03, 2007 at 10:40:07AM +0200, Ingo Molnar wrote: > yeah, i use gnome-terminal exclusively. But testers looking for CFS > regressions do run every shell on the planet :-) ...and people running older kernels get different results (no surprise) fwiw, I ran 'top' on 5 terminals with xterm-spam running concurrently on 2.6.15 (rxvt, pterm, xterm, konsole and gnome-terminal). For that case, gnome-terminal was definitely the slowest, and used the most CPU time (more than a factor of three slower than xterm). konsole was about 2.5, pterm was about the same as xterm, and rxvt about half the CPU (ymmv). > > gnome-terminal is also faster all around (at least on my box): -- Thomas E. Dickey http://invisible-island.net ftp://invisible-island.net [-- Attachment #2: Type: application/pgp-signature, Size: 228 bytes --] ^ permalink raw reply [flat|nested] 43+ messages in thread
end of thread, other threads:[~2007-07-05 13:58 UTC | newest]
Thread overview: 43+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2007-06-22 22:02 [patch] CFS scheduler, -v18 Ingo Molnar
2007-06-22 22:09 ` S.Çağlar Onur
2007-06-22 22:16 ` S.Çağlar Onur
2007-06-22 22:20 ` Ingo Molnar
2007-06-26 3:02 ` Andrew Morton
2007-06-26 8:38 ` Ingo Molnar
2007-06-26 9:00 ` Andrew Morton
2007-06-26 9:38 ` Ingo Molnar
2007-06-22 23:08 ` Gene Heskett
2007-06-23 7:11 ` Ingo Molnar
2007-06-23 9:55 ` Gene Heskett
2007-06-23 10:22 ` Antonino Ingargiola
2007-06-23 17:25 ` Ingo Molnar
2007-06-24 10:02 ` Antonino Ingargiola
2007-06-24 11:07 ` Ingo Molnar
2007-06-25 7:27 ` Antonino Ingargiola
2007-06-23 13:24 ` Willy Tarreau
2007-06-24 15:52 ` Ingo Molnar
2007-06-24 17:08 ` Willy Tarreau
2007-06-24 20:31 ` Ingo Molnar
2007-06-26 20:17 ` Fortier,Vincent [Montreal]
2007-06-27 10:51 ` Ingo Molnar
2007-06-30 21:06 ` Willy Tarreau
2007-07-01 8:31 ` Ingo Molnar
2007-07-01 8:45 ` Ingo Molnar
2007-07-01 9:00 ` Willy Tarreau
2007-07-02 11:44 ` Vegard Nossum
2007-07-02 13:01 ` Dmitry Adamushko
2007-07-02 13:43 ` Vegard Nossum
2007-07-02 15:50 ` Ingo Molnar
2007-07-02 16:40 ` Vegard Nossum
2007-07-02 18:13 ` Ingo Molnar
2007-07-03 7:01 ` Vegard Nossum
2007-07-03 7:12 ` Mike Galbraith
2007-07-03 7:22 ` Ingo Molnar
2007-07-03 8:08 ` Keith Packard
2007-07-03 8:31 ` Ingo Molnar
2007-07-04 12:11 ` Andi Kleen
2007-07-02 14:13 ` Bill Davidsen
2007-07-03 7:15 ` Ingo Molnar
2007-07-03 9:11 ` Vegard Nossum
[not found] <8yZun-1bO-5@gated-at.bofh.it>
[not found] ` <8CszT-4hd-11@gated-at.bofh.it>
[not found] ` <8CtPi-6qj-9@gated-at.bofh.it>
[not found] ` <8Cus1-7eL-11@gated-at.bofh.it>
[not found] ` <8CwtU-1Z5-3@gated-at.bofh.it>
[not found] ` <8Cxgd-3fN-9@gated-at.bofh.it>
[not found] ` <8CKQ8-7HN-13@gated-at.bofh.it>
[not found] ` <8CKZV-7Ur-11@gated-at.bofh.it>
[not found] ` <8CLCt-vX-3@gated-at.bofh.it>
2007-07-05 13:29 ` Thomas Dickey
[not found] ` <8CM5x-19K-3@gated-at.bofh.it>
2007-07-05 13:39 ` Thomas Dickey
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox