public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Russell Cattelan <cattelan@thebarn.com>
To: Chris Allen <chris@cjx.com>
Cc: xfs@oss.sgi.com
Subject: Re: XFS stack space crashes - current status?
Date: Wed, 02 Aug 2006 16:32:21 -0500	[thread overview]
Message-ID: <44D119E5.3070504@thebarn.com> (raw)
In-Reply-To: <44D0A296.9020307@cjx.com>

Chris Allen wrote:
> I have a box running XFS over md (raid5) over Fedora core5 2.6.17-1 
> kernel.
>
> The box contains 16x750GB SATA drives combined into a single 11TB raid5
> partition using md, and this partition contains a single XFS filesystem.
>
> I can consistently crash the box within about ten minutes with a simple
> perl script that spawns 25 processes each of which loop writing random
> files to the filesystem.
Ya md with raid5 and XFS is not real happy with 4k stacks.
I never bothered to spend the time to track down who might be
the worst offenders.
It's not really XFS that is a problem here but the combination
of all the drivers you have stacked up.


You might try turning on 8k stacks and all the stack debugging routines
that will dump stack when you over a preset thread hold.

Which scsi driver are you using?



>
> The only message I get on the console is something like this:
>
> do_IRQ: stack overflow: 492
> <c0406460>
>
> Once crashed, the box requires a hard reboot to rescue it (and needs 
> to resync
> the RAID array).
>
>
> As the box is to be used for a production upload fileserver receiving 
> several hundred
> simultaneous uploads, I would most likely be seeing this problem lots.
>
> So..... questions:
>
> 1. How much is known about this problem? Seeing as it is 100% 
> reproducible,
> is there any active development underway to fix it?
>
> 2. I have seen postings that say compiling a kernel with 8K stacks 
> will fix the
> problem. Is this the case? Or will I be able to trigger it again by 
> running 100 or
> 200 simultaneous writes?
>
> 3. Any suggestions as to what I should try? At present it looks like I 
> am stuck between
> finding a fix for XFS and splitting the box into 2 or 3 EXT3 
> partitions (which I really don't
> want to do). I have tried ReiserFS (max FS size is 8TB even though the 
> FAQ says 16), and
> JFS (jfs_fsck segfaults which doesn't fill me with confidence).
>
>
> Many thanks for any suggestions,
>
> Chris Allen.
>
>
>

      parent reply	other threads:[~2006-08-02 21:59 UTC|newest]

Thread overview: 3+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2006-08-02 13:03 XFS stack space crashes - current status? Chris Allen
2006-08-02 18:59 ` Eric Sandeen
2006-08-02 21:32 ` Russell Cattelan [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=44D119E5.3070504@thebarn.com \
    --to=cattelan@thebarn.com \
    --cc=chris@cjx.com \
    --cc=xfs@oss.sgi.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox