public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Stefan Priebe - Profihost AG <s.priebe@profihost.ag>
To: Dave Chinner <david@fromorbit.com>
Cc: Christoph Hellwig <hch@infradead.org>,
	"xfs-masters@oss.sgi.com" <xfs-masters@oss.sgi.com>,
	"xfs@oss.sgi.com" <xfs@oss.sgi.com>
Subject: Re: [xfs-masters] xfs deadlock in stable kernel 3.0.4
Date: Wed, 21 Sep 2011 09:40:03 +0200	[thread overview]
Message-ID: <4E7994D3.5020103@profihost.ag> (raw)
In-Reply-To: <20110921021133.GM15688@dastard>

Am 21.09.2011 04:11, schrieb Dave Chinner:

> How much memory does your test machine have? The performance will be
> vastly different if there is enough RAM to hold the working set of
> inodes and page cache (~20GB all up), and that could be one of the
> factors contributing to the problems.
The livesystems which crash within hours have between 48GB and 64GB RAM. 
But my testing system has only 8GB.

> The above xfs_info output is from your 160GB SSD - what's the output
> from the 1TB device?

The 1TB device is now doing something else and does not have XFS on it 
anymore. But here are the layouts of two livesystems.

xfs_info /dev/sda6
meta-data=/dev/root              isize=256    agcount=4, agsize=35767872 
blks
          =                       sectsz=512   attr=2
data     =                       bsize=4096   blocks=143071488, imaxpct=25
          =                       sunit=64     swidth=512 blks
naming   =version 2              bsize=4096   ascii-ci=0
log      =internal               bsize=4096   blocks=69888, version=2
          =                       sectsz=512   sunit=64 blks, lazy-count=1
realtime =none                   extsz=4096   blocks=0, rtextents=0

xfs_info /dev/sda6
meta-data=/dev/root              isize=256    agcount=4, agsize=35768000 
blks
          =                       sectsz=512   attr=2
data     =                       bsize=4096   blocks=143071774, imaxpct=25
          =                       sunit=64     swidth=512 blks
naming   =version 2              bsize=4096   ascii-ci=0
log      =internal               bsize=4096   blocks=32768, version=2
          =                       sectsz=512   sunit=64 blks, lazy-count=1
realtime =none                   extsz=4096   blocks=0, rtextents=0


> Also, what phase do you see it hanging in? the random stat phase is
> terribly slow on spinning disks, so if I can avoid that it woul dbe
> nice....
Creating or deleting files. never in the stat phase.

Stefan

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

  reply	other threads:[~2011-09-21  7:40 UTC|newest]

Thread overview: 49+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2011-09-10 12:23 xfs deadlock in stable kernel 3.0.4 Stefan Priebe
2011-09-12 15:21 ` Christoph Hellwig
2011-09-12 16:46   ` Stefan Priebe
2011-09-12 20:05     ` Christoph Hellwig
2011-09-13  6:04       ` Stefan Priebe - Profihost AG
2011-09-13 19:31         ` Stefan Priebe - Profihost AG
2011-09-13 20:50         ` Christoph Hellwig
2011-09-13 21:52           ` [xfs-masters] " Alex Elder
2011-09-13 21:58             ` Alex Elder
2011-09-13 22:26               ` Christoph Hellwig
2011-09-14  7:26           ` Stefan Priebe - Profihost AG
2011-09-14  7:48             ` Stefan Priebe - Profihost AG
2011-09-14  8:49               ` Stefan Priebe - Profihost AG
2011-09-14 14:30                 ` Christoph Hellwig
2011-09-14 14:30               ` Christoph Hellwig
2011-09-14 16:06                 ` Stefan Priebe - Profihost AG
2011-09-18  9:14                 ` Stefan Priebe - Profihost AG
2011-09-18 20:04                   ` Christoph Hellwig
2011-09-19 10:54                     ` Stefan Priebe - Profihost AG
2011-09-18 23:02                   ` Dave Chinner
2011-09-20  0:47                     ` Stefan Priebe
2011-09-20  1:01                       ` Stefan Priebe
2011-09-20 10:09                     ` Stefan Priebe - Profihost AG
2011-09-20 16:02                       ` Christoph Hellwig
2011-09-20 17:23                         ` Stefan Priebe - Profihost AG
2011-09-20 17:24                           ` Christoph Hellwig
2011-09-20 17:35                             ` Stefan Priebe - Profihost AG
2011-09-20 22:30                               ` Christoph Hellwig
2011-09-21  2:11                                 ` [xfs-masters] " Dave Chinner
2011-09-21  7:40                                   ` Stefan Priebe - Profihost AG [this message]
2011-09-21 11:42                                     ` Dave Chinner
2011-09-21 11:55                                       ` Stefan Priebe - Profihost AG
2011-09-21 12:26                                       ` Christoph Hellwig
2011-09-21 13:42                                         ` Stefan Priebe
2011-09-21 16:48                                         ` Stefan Priebe - Profihost AG
2011-09-21 17:26                                           ` Stefan Priebe - Profihost AG
2011-09-21 19:01                                         ` Stefan Priebe - Profihost AG
2011-09-21 23:07                                         ` Dave Chinner
2011-09-22 14:14                                           ` Christoph Hellwig
2011-09-22 21:49                                             ` Dave Chinner
2011-09-22 22:01                                               ` Christoph Hellwig
2011-09-23  5:28                                                 ` Stefan Priebe - Profihost AG
2011-09-22  0:53                                       ` Dave Chinner
2011-09-22  5:27                                         ` Stefan Priebe - Profihost AG
2011-09-22  7:52                                           ` Stefan Priebe - Profihost AG
2011-09-21  7:36                                 ` Stefan Priebe - Profihost AG
2011-09-21 11:39                                   ` Christoph Hellwig
2011-09-21 13:39                                     ` Stefan Priebe
2011-09-21 14:17                                       ` Christoph Hellwig

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4E7994D3.5020103@profihost.ag \
    --to=s.priebe@profihost.ag \
    --cc=david@fromorbit.com \
    --cc=hch@infradead.org \
    --cc=xfs-masters@oss.sgi.com \
    --cc=xfs@oss.sgi.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox