public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: David Chinner <dgc@sgi.com>
To: xfs-masters@oss.sgi.com
Cc: nathans@sgi.com, xfs@oss.sgi.com
Subject: Re: [xfs-masters] [BUG]: soft lock detected
Date: Wed, 16 Aug 2006 16:18:33 +1000	[thread overview]
Message-ID: <20060816061833.GH51703024@melbourne.sgi.com> (raw)
In-Reply-To: <OFE45AA35E.CBCEE6A2-ON482571CB.00362A71-482571CB.003716BA@cn.ibm.com>

On Tue, Aug 15, 2006 at 06:04:21PM +0800, Yi CDL Yang wrote:
> 
> Hi,
> 
> When I stress XFS filesystem, I find a bug, I can regenerate it on
> 2.6.18-rc3 and 2.6.18-rc4, my steps is:
> # mount /dev/sda5 /mnt/sda5
> #su oneuser
> $ mkdir /mnt/sda5/xfstest
> $ cd /mnt/sda5
> $ bonnie++ -d xfstest -s 2048 -r 512

Is that single threaded?

> After a while, kernel will output the following debug information:
> 
> BUG: soft lockup detected on CPU#0!
> Call Trace:
> [C0000001C7E5EEA0] [D000000000973018] .xfs_icsb_disable_counter+0x90/0x1ac
> [xfs]
> [C0000001C7E5EF60] [D000000000973274] .xfs_icsb_balance_counter+0x70/0x294
> [xfs]
> [C0000001C7E5F010] [D000000000973870]
> .xfs_icsb_modify_counters_int+0x188/0x1f4 [xfs]

We take spinlocks in these functions - but unless you've got lots of
CPUs they aren't taken for very long. We haven't seen these reports on
large CPU count machines, so I'm not sure off the top of my head
what would cause this.

FWIW, what type of machine and how many CPUs do you have?

> /////////////////////
> BUG: soft lockup detected on CPU#2!
> Call Trace:
> --- Exception: 901 at .xfs_alloc_fix_freelist+0x7c/0x4c4 [xfs]
>     LR = .xfs_alloc_vextent+0x2f0/0x494 [xfs]
> [C0000001C10CAF00] [0000000000000000] 0x0 (unreliable)
> [C0000001C10CB040] [D000000000933298] .xfs_alloc_vextent+0x2f0/0x494 [xfs]
> [C0000001C10CB110] [D000000000942F9C] .xfs_bmapi+0xd18/0x1834 [xfs]
> [C0000001C10CB390] [D0000000009688F8] .xfs_iomap_write_allocate+0x264/0x470

And we don't even hold spinlocks in that function. We do nothing that
would hold off interrupts or the scheduler in these functions....

> According to these information, I can't find the reason of the problem, for
> soft lockup, I think
> only preemption disabling or interrupt disabling can result in this, but
> the above functions don't
> run such an operation, I don't know what is your idea?

Spinlocks disable preemption so that could cause it, but I cannot see
how that second trace is at all valid....

Cheers,

Dave.
-- 
Dave Chinner
Principal Engineer
SGI Australian Software Group

           reply	other threads:[~2006-08-16  6:20 UTC|newest]

Thread overview: expand[flat|nested]  mbox.gz  Atom feed
 [parent not found: <OFE45AA35E.CBCEE6A2-ON482571CB.00362A71-482571CB.003716BA@cn.ibm.com>]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20060816061833.GH51703024@melbourne.sgi.com \
    --to=dgc@sgi.com \
    --cc=nathans@sgi.com \
    --cc=xfs-masters@oss.sgi.com \
    --cc=xfs@oss.sgi.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox