public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
* [QUESTION] 2.4.x nice level
@ 2001-04-02 22:13 BERECZ Szabolcs
  0 siblings, 0 replies; 15+ messages in thread
From: BERECZ Szabolcs @ 2001-04-02 22:13 UTC (permalink / raw)
  To: linux-kernel

Hi!

I just noticed, that a process with nice level 19, gets some processor
time, even if there is another process, which would use all of the
processor time.
for example, there is a setiathome running at nice level 19, and a
bladeenc at nice level 0. setiathome uses 14 percent, and bladeenc uses 84
percent of the processor. I think, setiathome should use max 2-3 percent.
the 14 percent is way too much for me.
or is there any other way to run a process only if nothing uses the
processor?
with kernel 2.2.16 it worked for me.
now I use 2.4.2-ac20

Bye,
Szabi



^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [QUESTION] 2.4.x nice level
@ 2001-04-02 23:04 Quim K Holland
  2001-04-03  3:02 ` LA Walsh
  0 siblings, 1 reply; 15+ messages in thread
From: Quim K Holland @ 2001-04-02 23:04 UTC (permalink / raw)
  To: szabi; +Cc: linux-kernel

>>>>> "BS" == BERECZ Szabolcs <szabi@inf.elte.hu> writes:

BS> ... a setiathome running at nice level 19, and a bladeenc at
BS> nice level 0. setiathome uses 14 percent, and bladeenc uses
BS> 84 percent of the processor. I think, setiathome should use
BS> max 2-3 percent.  the 14 percent is way too much for me.
BS> ...
BS> with kernel 2.2.16 it worked for me.
BS> now I use 2.4.2-ac20

Would it the case that bladeenc running on 2.4.2 spends more
time doing I/O?  I am not saying that the userland makes more I/O
requests, but if the same set of I/O requests are taking longer
to complete on 2.4.2, then while bladeenc is waiting for their
completion, it is not so surprising that the other process uses
the otherwise-idle CPU cycles.



------------------------------------------------------------
--== Sent via Deja.com ==--
http://www.deja.com/



^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [QUESTION] 2.4.x nice level
  2001-04-02 23:04 Quim K Holland
@ 2001-04-03  3:02 ` LA Walsh
  0 siblings, 0 replies; 15+ messages in thread
From: LA Walsh @ 2001-04-03  3:02 UTC (permalink / raw)
  To: Quim K Holland; +Cc: linux-kernel

Quim K Holland wrote:
> 
> >>>>> "BS" == BERECZ Szabolcs <szabi@inf.elte.hu> writes:
> 
> BS> ... a setiathome running at nice level 19, and a bladeenc at
> BS> nice level 0. setiathome uses 14 percent, and bladeenc uses
> BS> 84 percent of the processor. I think, setiathome should use
> BS> max 2-3 percent.  the 14 percent is way too much for me.
> BS> ...
> BS> with kernel 2.2.16 it worked for me.
> BS> now I use 2.4.2-ac20
---
	I was running 2 copies of setiathome on a 4 CPU server
@ work.  The two processes ran nice'd -19.  The builds we were 
running still took 20-30% longer as opposed to when setiathome wasn't
running (went from 45 minutes up to about an hour).  This machine
has 1G, so I don't think it was hurting from swapping.

I finally wrote a script that checked every 30 seconds -- if the
load on the machine climbed over '4', the script would SIGSTOP the
seti jobs.  Once the load on the machine fell below 2, it would 
send a SIGCONT to them.  

I was also running setiathome on my laptop for a short while --
but the fan kept coming on and the computer would get really hot.
So I stopped that.  Linux @ idle doesn't seem to ever kick on
the fan, but turn on a CPU cruching program and it sure seemed
to heat up the machine.  I still wonder how many kilo or mega watts
go to running dispersed computation programs.  Just one of those
things I may never know....

-l

-- 
The above thoughts and           | They may have nothing to do with
writings are my own.             | the opinions of my employer. :-)
L A Walsh                        | Trust Technology, Core Linux, SGI
law@sgi.com                      | Voice: (650) 933-5338

^ permalink raw reply	[flat|nested] 15+ messages in thread

* [QUESTION] 2.4.x nice level
@ 2001-04-04 16:12 SodaPop
  2001-04-10  3:37 ` george anzinger
  0 siblings, 1 reply; 15+ messages in thread
From: SodaPop @ 2001-04-04 16:12 UTC (permalink / raw)
  To: linux-kernel

I too have noticed that nicing processes does not work nearly as
effectively as I'd like it to.  I run on an underpowered machine,
and have had to stop running things such as seti because it steals too
much cpu time, even when maximally niced.

As an example, I can run mpg123 and a kernel build concurrently without
trouble; but if I add a single maximally niced seti process, mpg123 runs
out of gas and will start to skip while decoding.

Is there any way we can make nice levels stronger than they currently are
in 2.4?  Or is this perhaps a timeslice problem, where once seti gets cpu
time it runs longer than it should since it makes relatively few system
calls?

-dennis T


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [QUESTION] 2.4.x nice level
       [not found] ` <fa.dkui9av.1ulsbjm@ifi.uio.no>
@ 2001-04-05 17:24   ` Tor Arntsen
  0 siblings, 0 replies; 15+ messages in thread
From: Tor Arntsen @ 2001-04-05 17:24 UTC (permalink / raw)
  To: linux-kernel

LA Walsh <law@sgi.com> writes:
>	I was running 2 copies of setiathome on a 4 CPU server
>@ work.  The two processes ran nice'd -19.  The builds we were 
>running still took 20-30% longer as opposed to when setiathome wasn't
>running (went from 45 minutes up to about an hour).  This machine
>has 1G, so I don't think it was hurting from swapping.

It would be nice to have IRIX weightless processes on Linux.. 
setiathome on SGI computers don't affect anything else except
in extreme cases.

-Tor

^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [QUESTION] 2.4.x nice level
  2001-04-04 16:12 [QUESTION] 2.4.x nice level SodaPop
@ 2001-04-10  3:37 ` george anzinger
  2001-04-10 16:10   ` Rik van Riel
  0 siblings, 1 reply; 15+ messages in thread
From: george anzinger @ 2001-04-10  3:37 UTC (permalink / raw)
  To: SodaPop; +Cc: linux-kernel

SodaPop wrote:
> 
> I too have noticed that nicing processes does not work nearly as
> effectively as I'd like it to.  I run on an underpowered machine,
> and have had to stop running things such as seti because it steals too
> much cpu time, even when maximally niced.
> 
> As an example, I can run mpg123 and a kernel build concurrently without
> trouble; but if I add a single maximally niced seti process, mpg123 runs
> out of gas and will start to skip while decoding.
> 
> Is there any way we can make nice levels stronger than they currently are
> in 2.4?  Or is this perhaps a timeslice problem, where once seti gets cpu
> time it runs longer than it should since it makes relatively few system
> calls?
> 
In kernel/sched.c for HZ < 200 an adjustment of nice to tick is set up
to be nice>>2 (i.e. nice /4).  This gives the ratio of nice to time
slice.  Adjustments are made to make the MOST nice yield 1 jiffy, so
using this scale and remembering nice ranges from -19 to 20 the least
nice is 40/4 or 10 ticks.  This implies that if only two tasks are
running and they are most and least niced then one will get 1/11 of the
processor, the other 10/11 (about 10% and 90%).  If one is niced and the
other is not you get 1 and 5 for the time slices or 1/6 and 5/6 (17% and
83%).  

In 2.2.x systems the full range of nice was used one to one to give 1
and 39 or 40 or 2.5% and 97.5% for max nice to min.  For most nice to
normal you would get 1 and 20 or 4.7% and 95.3%.

The comments say the objective is to come up with a time slice of 50ms,
presumably for the normal nice value of zero.  After translating the
range this would be a value of 20 and, yep 20/4 give 5 jiffies or 50
ms.  Sure puts a crimp in the min to max range, however.

George

^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [QUESTION] 2.4.x nice level
  2001-04-10  3:37 ` george anzinger
@ 2001-04-10 16:10   ` Rik van Riel
  2001-04-10 16:39     ` george anzinger
  2001-04-11 10:34     ` [test-PATCH] " Rik van Riel
  0 siblings, 2 replies; 15+ messages in thread
From: Rik van Riel @ 2001-04-10 16:10 UTC (permalink / raw)
  To: george anzinger; +Cc: SodaPop, linux-kernel

On Mon, 9 Apr 2001, george anzinger wrote:
> SodaPop wrote:
> > 
> > I too have noticed that nicing processes does not work nearly as
> > effectively as I'd like it to.  I run on an underpowered machine,
> > and have had to stop running things such as seti because it steals too
> > much cpu time, even when maximally niced.

> In kernel/sched.c for HZ < 200 an adjustment of nice to tick is set up
> to be nice>>2 (i.e. nice /4).  This gives the ratio of nice to time
> slice.  Adjustments are made to make the MOST nice yield 1 jiffy, so
	[snip 2.4 nice scale is too limited]

I'll try to come up with a recalculation change that will make
this thing behave better, while still retaining the short time
slices for multiple normal-priority tasks and the cache footprint
schedule() and friends currently have...

[I've got some vague ideas ... give me a few hours to put them
into code ;)]

regards,

Rik
--
Virtual memory is like a game you can't win;
However, without VM there's truly nothing to lose...

		http://www.surriel.com/
http://www.conectiva.com/	http://distro.conectiva.com.br/


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [QUESTION] 2.4.x nice level
  2001-04-10 16:10   ` Rik van Riel
@ 2001-04-10 16:39     ` george anzinger
  2001-04-11 10:34     ` [test-PATCH] " Rik van Riel
  1 sibling, 0 replies; 15+ messages in thread
From: george anzinger @ 2001-04-10 16:39 UTC (permalink / raw)
  To: Rik van Riel, linux-kernel@vger.kernel.org

Rik van Riel wrote:
> 
> On Mon, 9 Apr 2001, george anzinger wrote:
> > SodaPop wrote:
> > >
> > > I too have noticed that nicing processes does not work nearly as
> > > effectively as I'd like it to.  I run on an underpowered machine,
> > > and have had to stop running things such as seti because it steals too
> > > much cpu time, even when maximally niced.
> 
> > In kernel/sched.c for HZ < 200 an adjustment of nice to tick is set up
> > to be nice>>2 (i.e. nice /4).  This gives the ratio of nice to time
> > slice.  Adjustments are made to make the MOST nice yield 1 jiffy, so
>         [snip 2.4 nice scale is too limited]
> 
> I'll try to come up with a recalculation change that will make
> this thing behave better, while still retaining the short time
> slices for multiple normal-priority tasks and the cache footprint
> schedule() and friends currently have...
> 
> [I've got some vague ideas ... give me a few hours to put them
> into code ;)]

You might check out this:

http://rtsched.sourceforge.net/

I did some work on leveling out the recalculation overhead.  I think, as
the code shows, that it can be done without dropping the run queue lock.

I wonder if the wave nature of the recalculation cycle is a problem.  By
this I mean after a recalculation tasks run for relatively long times
(50 ms today) but as the recalculation time approaches, the time reduces
to 10 ms.  Gets one to thinking about a way to come up with a more
uniform, over time, mix.

George

George

^ permalink raw reply	[flat|nested] 15+ messages in thread

* [test-PATCH] Re: [QUESTION] 2.4.x nice level
  2001-04-10 16:10   ` Rik van Riel
  2001-04-10 16:39     ` george anzinger
@ 2001-04-11 10:34     ` Rik van Riel
  2001-04-11 15:53       ` Rik van Riel
  2001-04-11 16:27       ` george anzinger
  1 sibling, 2 replies; 15+ messages in thread
From: Rik van Riel @ 2001-04-11 10:34 UTC (permalink / raw)
  To: george anzinger; +Cc: SodaPop, alexey, linux-kernel

On Tue, 10 Apr 2001, Rik van Riel wrote:

> I'll try to come up with a recalculation change that will make
> this thing behave better, while still retaining the short time
> slices for multiple normal-priority tasks and the cache footprint
> schedule() and friends currently have...

OK, here it is. It's nothing like montavista's singing-dancing
scheduler patch that does all, just a really minimal change that
should stretch the nice levels to yield the following CPU usage:

Nice    0    5   10   15   19
%CPU  100   56   25    6    1

Note that the code doesn't change the actual scheduling code,
just the recalculation. Care has also been taken to not increase
the cache footprint of the scheduling and recalculation code.

I'd love to hear some test results from people who are interested
in wider nice levels. How does this run on your system?  Can you
trigger bad behaviour in any way?

regards,

Rik
--
Virtual memory is like a game you can't win;
However, without VM there's truly nothing to lose...

		http://www.surriel.com/
http://www.conectiva.com/	http://distro.conectiva.com.br/



--- linux-2.4.3-ac4/kernel/sched.c.orig	Tue Apr 10 21:04:06 2001
+++ linux-2.4.3-ac4/kernel/sched.c	Wed Apr 11 06:18:46 2001
@@ -686,8 +686,26 @@
 		struct task_struct *p;
 		spin_unlock_irq(&runqueue_lock);
 		read_lock(&tasklist_lock);
-		for_each_task(p)
+		for_each_task(p) {
+		    if (p->nice <= 0) {
+			/* The normal case... */
 			p->counter = (p->counter >> 1) + NICE_TO_TICKS(p->nice);
+		    } else {
+			/*
+			 * Niced tasks get less CPU less often, leading to
+			 * the following distribution of CPU time:
+			 *
+			 * Nice    0    5   10   15   19
+			 * %CPU  100   56   25    6    1	
+			 */
+			short prio = 20 - p->nice;
+			p->nice_calc += prio;
+			if (p->nice_calc >= 20) {
+			    p->nice_calc -= 20;
+			    p->counter = (p->counter >> 1) + NICE_TO_TICKS(p->nice);
+			}
+		    }
+		}
 		read_unlock(&tasklist_lock);
 		spin_lock_irq(&runqueue_lock);
 	}
--- linux-2.4.3-ac4/include/linux/sched.h.orig	Tue Apr 10 21:04:13 2001
+++ linux-2.4.3-ac4/include/linux/sched.h	Wed Apr 11 06:26:47 2001
@@ -303,7 +303,8 @@
  * the goodness() loop in schedule().
  */
 	long counter;
-	long nice;
+	short nice_calc;
+	short nice;
 	unsigned long policy;
 	struct mm_struct *mm;
 	int has_cpu, processor;


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [test-PATCH] Re: [QUESTION] 2.4.x nice level
  2001-04-11 10:34     ` [test-PATCH] " Rik van Riel
@ 2001-04-11 15:53       ` Rik van Riel
  2001-04-12 22:51         ` Pozsar Balazs
  2001-04-11 16:27       ` george anzinger
  1 sibling, 1 reply; 15+ messages in thread
From: Rik van Riel @ 2001-04-11 15:53 UTC (permalink / raw)
  To: george anzinger; +Cc: SodaPop, alexey, linux-kernel

On Wed, 11 Apr 2001, Rik van Riel wrote:

> OK, here it is. It's nothing like montavista's singing-dancing
> scheduler patch that does all, just a really minimal change that
> should stretch the nice levels to yield the following CPU usage:
> 
> Nice    0    5   10   15   19
> %CPU  100   56   25    6    1

  PID USER     PRI  NI  SIZE SWAP  RSS SHARE STAT %CPU %MEM   TIME COMMAND
  980 riel      17   0   296    0  296   240 R    54.1  0.5  54:19 loop
 1005 riel      16   5   296    0  296   240 R N  27.0  0.5   0:34 loop
 1006 riel      17  10   296    0  296   240 R N  13.5  0.5   0:16 loop
 1007 riel      18  15   296    0  296   240 R N   4.5  0.5   0:05 loop
  987 riel      20  19   296    0  296   240 R N   0.4  0.5   0:25 loop

... is what I got when testing it here. It seems that nice levels
REALLY mean something with the patch applied ;)

You can get it at http://www.surriel.com/patches/2.4/2.4.3ac4-largenice

Since there seems to be quite a bit of demand for this feature,
please test it and try to make it break. If it doesn't break we
can try to put it in the kernel...

regards,

Rik
--
Virtual memory is like a game you can't win;
However, without VM there's truly nothing to lose...

		http://www.surriel.com/
http://www.conectiva.com/	http://distro.conectiva.com.br/


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [test-PATCH] Re: [QUESTION] 2.4.x nice level
  2001-04-11 10:34     ` [test-PATCH] " Rik van Riel
  2001-04-11 15:53       ` Rik van Riel
@ 2001-04-11 16:27       ` george anzinger
  2001-04-12 23:51         ` Pavel Machek
  1 sibling, 1 reply; 15+ messages in thread
From: george anzinger @ 2001-04-11 16:27 UTC (permalink / raw)
  To: Rik van Riel; +Cc: SodaPop, alexey, linux-kernel

One rule of optimization is to move any code you can outside the loop. 
Why isn't the nice_to_ticks calculation done when nice is changed
instead of EVERY recalc.?  I guess another way to ask this is, who needs
to see the original nice?  Would it be worth another task_struct entry
to move this calculation out of the loop?

George

Rik van Riel wrote:
> 
> On Tue, 10 Apr 2001, Rik van Riel wrote:
> 
> > I'll try to come up with a recalculation change that will make
> > this thing behave better, while still retaining the short time
> > slices for multiple normal-priority tasks and the cache footprint
> > schedule() and friends currently have...
> 
> OK, here it is. It's nothing like montavista's singing-dancing
> scheduler patch that does all, just a really minimal change that
> should stretch the nice levels to yield the following CPU usage:
> 
> Nice    0    5   10   15   19
> %CPU  100   56   25    6    1
> 
> Note that the code doesn't change the actual scheduling code,
> just the recalculation. Care has also been taken to not increase
> the cache footprint of the scheduling and recalculation code.
> 
> I'd love to hear some test results from people who are interested
> in wider nice levels. How does this run on your system?  Can you
> trigger bad behaviour in any way?
> 
> regards,
> 
> Rik
> --
> Virtual memory is like a game you can't win;
> However, without VM there's truly nothing to lose...
> 
>                 http://www.surriel.com/
> http://www.conectiva.com/       http://distro.conectiva.com.br/
> 
> --- linux-2.4.3-ac4/kernel/sched.c.orig Tue Apr 10 21:04:06 2001
> +++ linux-2.4.3-ac4/kernel/sched.c      Wed Apr 11 06:18:46 2001
> @@ -686,8 +686,26 @@
>                 struct task_struct *p;
>                 spin_unlock_irq(&runqueue_lock);
>                 read_lock(&tasklist_lock);
> -               for_each_task(p)
> +               for_each_task(p) {
> +                   if (p->nice <= 0) {
> +                       /* The normal case... */
>                         p->counter = (p->counter >> 1) + NICE_TO_TICKS(p->nice);
> +                   } else {
> +                       /*
> +                        * Niced tasks get less CPU less often, leading to
> +                        * the following distribution of CPU time:
> +                        *
> +                        * Nice    0    5   10   15   19
> +                        * %CPU  100   56   25    6    1
> +                        */
> +                       short prio = 20 - p->nice;
> +                       p->nice_calc += prio;
> +                       if (p->nice_calc >= 20) {
> +                           p->nice_calc -= 20;
> +                           p->counter = (p->counter >> 1) + NICE_TO_TICKS(p->nice);
> +                       }
> +                   }
> +               }
>                 read_unlock(&tasklist_lock);
>                 spin_lock_irq(&runqueue_lock);
>         }
> --- linux-2.4.3-ac4/include/linux/sched.h.orig  Tue Apr 10 21:04:13 2001
> +++ linux-2.4.3-ac4/include/linux/sched.h       Wed Apr 11 06:26:47 2001
> @@ -303,7 +303,8 @@
>   * the goodness() loop in schedule().
>   */
>         long counter;
> -       long nice;
> +       short nice_calc;
> +       short nice;
>         unsigned long policy;
>         struct mm_struct *mm;
>         int has_cpu, processor;

^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [test-PATCH] Re: [QUESTION] 2.4.x nice level
  2001-04-11 15:53       ` Rik van Riel
@ 2001-04-12 22:51         ` Pozsar Balazs
  0 siblings, 0 replies; 15+ messages in thread
From: Pozsar Balazs @ 2001-04-12 22:51 UTC (permalink / raw)
  To: Rik van Riel; +Cc: linux-kernel

On Wed, Apr 11, 2001 at 12:53:16PM -0300, Rik van Riel wrote:
> On Wed, 11 Apr 2001, Rik van Riel wrote:
> 
> > OK, here it is. It's nothing like montavista's singing-dancing
> > scheduler patch that does all, just a really minimal change that
> > should stretch the nice levels to yield the following CPU usage:
> > 
> > Nice    0    5   10   15   19
> > %CPU  100   56   25    6    1
> 
>   PID USER     PRI  NI  SIZE SWAP  RSS SHARE STAT %CPU %MEM   TIME COMMAND
>   980 riel      17   0   296    0  296   240 R    54.1  0.5  54:19 loop
>  1005 riel      16   5   296    0  296   240 R N  27.0  0.5   0:34 loop
>  1006 riel      17  10   296    0  296   240 R N  13.5  0.5   0:16 loop
>  1007 riel      18  15   296    0  296   240 R N   4.5  0.5   0:05 loop
>   987 riel      20  19   296    0  296   240 R N   0.4  0.5   0:25 loop

How does this scale to negative nice levels? Afaik it should, in some way.
(I don't mean that it's wrong in this state, i'm just asking).

regards,
Balazs.

^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [test-PATCH] Re: [QUESTION] 2.4.x nice level
  2001-04-11 16:27       ` george anzinger
@ 2001-04-12 23:51         ` Pavel Machek
  2001-04-16 14:18           ` Rik van Riel
  0 siblings, 1 reply; 15+ messages in thread
From: Pavel Machek @ 2001-04-12 23:51 UTC (permalink / raw)
  To: george anzinger; +Cc: Rik van Riel, SodaPop, alexey, linux-kernel

Hi!

> One rule of optimization is to move any code you can outside the loop. 
> Why isn't the nice_to_ticks calculation done when nice is changed
> instead of EVERY recalc.?  I guess another way to ask this is, who needs

This way change is localized very nicely, and it is "obviously right".

-- 
Philips Velo 1: 1"x4"x8", 300gram, 60, 12MB, 40bogomips, linux, mutt,
details at http://atrey.karlin.mff.cuni.cz/~pavel/velo/index.html.


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [test-PATCH] Re: [QUESTION] 2.4.x nice level
  2001-04-12 23:51         ` Pavel Machek
@ 2001-04-16 14:18           ` Rik van Riel
  2001-04-16 17:49             ` george anzinger
  0 siblings, 1 reply; 15+ messages in thread
From: Rik van Riel @ 2001-04-16 14:18 UTC (permalink / raw)
  To: Pavel Machek; +Cc: george anzinger, SodaPop, alexey, linux-kernel

On Thu, 12 Apr 2001, Pavel Machek wrote:

> > One rule of optimization is to move any code you can outside the loop. 
> > Why isn't the nice_to_ticks calculation done when nice is changed
> > instead of EVERY recalc.?  I guess another way to ask this is, who needs
> 
> This way change is localized very nicely, and it is "obviously right".

Except for two obvious things:

1. we need to load the nice level anyway
2. a shift takes less cycles than a load on most
   CPUs

Rik
--
Virtual memory is like a game you can't win;
However, without VM there's truly nothing to lose...

		http://www.surriel.com/
http://www.conectiva.com/	http://distro.conectiva.com.br/


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: [test-PATCH] Re: [QUESTION] 2.4.x nice level
  2001-04-16 14:18           ` Rik van Riel
@ 2001-04-16 17:49             ` george anzinger
  0 siblings, 0 replies; 15+ messages in thread
From: george anzinger @ 2001-04-16 17:49 UTC (permalink / raw)
  To: Rik van Riel; +Cc: Pavel Machek, SodaPop, alexey, linux-kernel

Rik van Riel wrote:
> 
> On Thu, 12 Apr 2001, Pavel Machek wrote:
> 
> > > One rule of optimization is to move any code you can outside the loop.
> > > Why isn't the nice_to_ticks calculation done when nice is changed
> > > instead of EVERY recalc.?  I guess another way to ask this is, who needs
> >
> > This way change is localized very nicely, and it is "obviously right".
> 
> Except for two obvious things:
> 
> 1. we need to load the nice level anyway
> 2. a shift takes less cycles than a load on most
>    CPUs
> 
Gosh, what am I missing here?  I think "top" and "ps" want to see the
"nice" value so it needs to be available and since the NICE_TO_TICK()
function looses information (i.e. is not reversible) we can not compute
it from ticks.  Still, yes we need to load something, but is it nice? 
Why not the result of the NICE_TO_TICK()?  

A shift and a subtract are fast, yes, but this loop runs over all tasks
(not just the run list).  This loop can put a real dent in preemption
times AND the notion of turning on interrupts while it is done can run
into some interesting race conditions.  (This is why the MontaVista
scheduler does the loop without dropping the lock, AFTER optimizing the
h... out of it.)

What am I missing?

George

^ permalink raw reply	[flat|nested] 15+ messages in thread

end of thread, other threads:[~2001-04-16 17:52 UTC | newest]

Thread overview: 15+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2001-04-04 16:12 [QUESTION] 2.4.x nice level SodaPop
2001-04-10  3:37 ` george anzinger
2001-04-10 16:10   ` Rik van Riel
2001-04-10 16:39     ` george anzinger
2001-04-11 10:34     ` [test-PATCH] " Rik van Riel
2001-04-11 15:53       ` Rik van Riel
2001-04-12 22:51         ` Pozsar Balazs
2001-04-11 16:27       ` george anzinger
2001-04-12 23:51         ` Pavel Machek
2001-04-16 14:18           ` Rik van Riel
2001-04-16 17:49             ` george anzinger
     [not found] <fa.j9vo8pv.1rj8up9@ifi.uio.no>
     [not found] ` <fa.dkui9av.1ulsbjm@ifi.uio.no>
2001-04-05 17:24   ` Tor Arntsen
  -- strict thread matches above, loose matches on Subject: below --
2001-04-02 23:04 Quim K Holland
2001-04-03  3:02 ` LA Walsh
2001-04-02 22:13 BERECZ Szabolcs

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox