public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Jens Axboe <axboe@suse.de>
To: Davide Libenzi <davidel@xmailserver.org>
Cc: Matthias Hanisch <mjh@vr-web.de>,
	Mikael Pettersson <mikpe@csd.uu.se>,
	Linus Torvalds <torvalds@transmeta.com>,
	lkml <linux-kernel@vger.kernel.org>
Subject: Re: 2.5.2-pre performance degradation on an old 486
Date: Sun, 6 Jan 2002 11:21:29 +0100	[thread overview]
Message-ID: <20020106112129.D8673@suse.de> (raw)
In-Reply-To: <Pine.LNX.4.10.10201050904070.1001-100000@pingu.franken.de> <Pine.LNX.4.40.0201051506170.1607-100000@blue1.dev.mcafeelabs.com>
In-Reply-To: <Pine.LNX.4.40.0201051506170.1607-100000@blue1.dev.mcafeelabs.com>

On Sat, Jan 05 2002, Davide Libenzi wrote:
> > > (*) 100MHz 486DX4, 28MB ram, no L2 cache, two old and slow IDE disks,
> > > small custom no-nonsense RedHat 7.2, kernels compiled with gcc 2.95.3.
> >
> > Is this ISA (maybe it has something to do with ISA bouncing)? Mine is:
> >
> > 486 DX/2 ISA, Adaptec 1542, two slow scsi disks and a self-made
> > slackware-based system.
> >
> > Can you also backout the scheduler changes to verify this? I have a
> > backout patch for 2.5.2-pre6, if you don't want to do this for yourself.
> 
> There should be some part of the kernel that assume a certain scheduler
> behavior. There was a guy that reported a bad  hdparm  performance and i
> tried it. By running  hdparm -t  my system has a context switch of 20-30
> and an irq load of about 100-110.
> The scheduler itself, even if you code it in visual basic, cannot make
> this with such loads.
> Did you try to profile the kernel ?

Davide,

If this is caused by ISA bounce problems, then you should be able to
reproduce by doing something ala

[ drivers/ide/ide-dma.c ]

ide_toggle_bounce()
{
	...

+	addr = BLK_BOUNCE_ISA;
	blk_queue_bounce_limit(&drive->queue, addr);
}

pseudo-diff, just add the addr = line. Now compare performance with and
without your scheduler changes.

-- 
Jens Axboe


  reply	other threads:[~2002-01-06 10:22 UTC|newest]

Thread overview: 18+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2002-01-05  0:51 2.5.2-pre performance degradation on an old 486 Mikael Pettersson
2002-01-05  8:25 ` Matthias Hanisch
2002-01-05 23:10   ` Davide Libenzi
2002-01-06 10:21     ` Jens Axboe [this message]
2002-01-06 10:33       ` Andre Hedrick
2002-01-06 23:59       ` [patch] 2.5.2 scheduler code for 2.4.18-pre1 ( was 2.5.2-pre performance degradation on an old 486 ) Davide Libenzi
2002-01-07  1:38         ` Andrea Arcangeli
2002-01-07 14:35           ` J.A. Magallon
2002-01-07 14:37             ` Andrea Arcangeli
2002-01-07  7:32         ` Jens Axboe
2002-01-07 18:10           ` Davide Libenzi
2002-01-07  7:22     ` 2.5.2-pre performance degradation on an old 486 Matthias Hanisch
2002-01-07 16:43       ` Linus Torvalds
2002-01-07 18:31       ` Davide Libenzi
2002-01-07 21:43         ` Matthias Hanisch
2002-01-07 22:17           ` Davide Libenzi
  -- strict thread matches above, loose matches on Subject: below --
2002-01-07 18:01 Mikael Pettersson
2002-01-07 18:36 ` Davide Libenzi

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20020106112129.D8673@suse.de \
    --to=axboe@suse.de \
    --cc=davidel@xmailserver.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=mikpe@csd.uu.se \
    --cc=mjh@vr-web.de \
    --cc=torvalds@transmeta.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox