public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Ingo Molnar <mingo@elte.hu>
To: Satoru Takeuchi <takeuchi_satoru@jp.fujitsu.com>
Cc: surya.prabhakar@wipro.com, kernel@kolivas.org,
	linux-kernel@vger.kernel.org, torvalds@linux-foundation.org,
	akpm@linux-foundation.org, npiggin@suse.de, efault@gmx.de,
	arjan@infradead.org, tglx@linutronix.de, wli@holomorphy.com
Subject: Re: [TEST RESULT]massive_intr.c -- cfs/vanilla/sd-0.40
Date: Mon, 16 Apr 2007 10:47:25 +0200	[thread overview]
Message-ID: <20070416084724.GA10912@elte.hu> (raw)
In-Reply-To: <877iscsp1x.wl%takeuchi_satoru@jp.fujitsu.com>


* Satoru Takeuchi <takeuchi_satoru@jp.fujitsu.com> wrote:

> > btw., other schedulers might work better with some more test-time: 
> > i'd suggest to use 60 seconds (./massive_intr 10 60) [or maybe more, 
> > using more threads] to see long-term fairness effects.
> 
> I tested CFS with massive_intr. I did long term, many CPUs, and many 
> processes cases.
> 
> Test environment
> ================
> 
>  - kernel:         2.6.21-rc6-CFS
>  - run time:       300 secs
>  - # of CPU:       1 or 4
>  - # of processes: 200 or 800
> 
> Result
> ======
> 
>   +---------+-----------+-------+------+------+--------+
>   |   # of  |   # of    | avg   | max  | min  |  stdev |
>   |   CPUs  | processes | (*1)  | (*2) | (*3) |  (*4)  |
>   +---------+-----------+-------+------+------+--------+
>   | 1(i386) |           | 117.9 |  123 |  115 |    1.2 |
>   +---------|       200 +-------+------+------+--------+
>   |         |           | 750.2 |  767 |  735 |   10.6 |
>   | 4(ia64) +-----------+-------+------+------+--------+
>   |         |   800(*5) | 187.3 |  189 |  186 |    0.8 |
>   +---------+-----------+-------+------+------+--------+
> 
>   *1) average number of loops among all processes
>   *2) maximum number of loops among all processes
>   *3) minimum number of loops among all processes
>   *4) standard deviation
>   *5) Its # of processes per CPU is equal to first test case.
> 
> Pretty good! CFS seems to be fair in any situation.

thanks for testing this! Indeed the min-max values and standard 
deviation look all pretty healthy. (They in fact seem to be better than 
the other patch of mine against upstream that you tested, correct?)

[ And there's also another nice little detail in your feedback: CFS
  actually builds, boots and works fine on ia64 too ;-) ]

	Ingo

  reply	other threads:[~2007-04-16  8:47 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2007-04-14  8:49 [TEST RESULT]massive_intr.c -- cfs/vanilla/sd-0.40 surya.prabhakar
2007-04-14  9:06 ` Willy Tarreau
2007-04-14 12:02 ` Ingo Molnar
2007-04-14 19:55   ` William Lee Irwin III
2007-04-16  8:26   ` Satoru Takeuchi
2007-04-16  8:47     ` Ingo Molnar [this message]
2007-04-16 10:13       ` Satoru Takeuchi
2007-04-16 10:22         ` Ingo Molnar

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20070416084724.GA10912@elte.hu \
    --to=mingo@elte.hu \
    --cc=akpm@linux-foundation.org \
    --cc=arjan@infradead.org \
    --cc=efault@gmx.de \
    --cc=kernel@kolivas.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=npiggin@suse.de \
    --cc=surya.prabhakar@wipro.com \
    --cc=takeuchi_satoru@jp.fujitsu.com \
    --cc=tglx@linutronix.de \
    --cc=torvalds@linux-foundation.org \
    --cc=wli@holomorphy.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox