public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
* [2.6.33.3] scheduling while atomic (inode reclaim races still?)
@ 2010-05-20  8:31 Simon Kirby
  2010-05-20  9:03 ` Christoph Hellwig
  0 siblings, 1 reply; 2+ messages in thread
From: Simon Kirby @ 2010-05-20  8:31 UTC (permalink / raw)
  To: xfs

This started happening on host about 10 days after upgrading to 2.6.33.3
(in hopes that it fixed all of the reclaim issues present in 2.6.33.2). 
I don't see any fixes in 2.6.33.4 relevant to this particular issue.

After this first error, the kernel logs just filled with repeat
occurrences with different backtraces.

BUG: scheduling while atomic: nfsd/29671/0x00000250
Modules linked in: aoe xt_MARK ipmi_devintf ipmi_si ipmi_msghandler e1000e bnx2
Pid: 29671, comm: nfsd Not tainted 2.6.33.3-hw #1
Call Trace:
 [<ffffffff81298f5e>] ? xfs_iflush+0x2ee/0x350
 [<ffffffff81045691>] __schedule_bug+0x61/0x70
 [<ffffffff81656f68>] schedule+0x588/0xa00
 [<ffffffff8106aee2>] ? bit_waitqueue+0x12/0xc0
 [<ffffffff8165773f>] schedule_timeout+0x18f/0x270
 [<ffffffff8105cd00>] ? process_timeout+0x0/0x10
 [<ffffffff8165697f>] io_schedule_timeout+0x8f/0xf0
 [<ffffffff810bed88>] balance_dirty_pages_ratelimited_nr+0x178/0x3a0
 [<ffffffff810b65d3>] generic_file_buffered_write+0x193/0x230
 [<ffffffff812befb6>] xfs_write+0x7c6/0x8f0
 [<ffffffff81294ec0>] ? xfs_iget+0x4f0/0x660
 [<ffffffff812bab50>] ? xfs_file_aio_write+0x0/0x60
 [<ffffffff812baba6>] xfs_file_aio_write+0x56/0x60
 [<ffffffff810eea6b>] do_sync_readv_writev+0xcb/0x110
 [<ffffffff81204616>] ? exportfs_decode_fh+0xe6/0x270
 [<ffffffff812081d0>] ? nfsd_acceptable+0x0/0x120
 [<ffffffff810ee88e>] ? rw_copy_check_uvector+0x7e/0x130
 [<ffffffff810ef15f>] do_readv_writev+0xcf/0x1f0
 [<ffffffff81208362>] ? nfsd_setuser_and_check_port+0x72/0x80
 [<ffffffff81208cac>] ? nfsd_permission+0xec/0x160
 [<ffffffff810ef2c0>] vfs_writev+0x40/0x60
 [<ffffffff81209ece>] nfsd_vfs_write+0xde/0x420
 [<ffffffff810ed1bd>] ? dentry_open+0x4d/0xb0
 [<ffffffff8120a8ce>] ? nfsd_open+0x16e/0x200
 [<ffffffff8120ad0a>] nfsd_write+0xea/0x100
 [<ffffffff812104cb>] ? nfsd_cache_lookup+0x2bb/0x3e0
 [<ffffffff81212b7f>] nfsd3_proc_write+0xaf/0x140
 [<ffffffff81204b4b>] nfsd_dispatch+0xbb/0x260
 [<ffffffff81618eaf>] svc_process+0x4af/0x820
 [<ffffffff81205190>] ? nfsd+0x0/0x160
 [<ffffffff8120526d>] nfsd+0xdd/0x160
 [<ffffffff8106aac6>] kthread+0x96/0xb0
 [<ffffffff8100ace4>] kernel_thread_helper+0x4/0x10
 [<ffffffff8106aa30>] ? kthread+0x0/0xb0
 [<ffffffff8100ace0>] ? kernel_thread_helper+0x0/0x10
BUG: scheduling while atomic: nfsd/29671/0x00000250
Modules linked in: aoe xt_MARK ipmi_devintf ipmi_si ipmi_msghandler e1000e bnx2
Pid: 29671, comm: nfsd Not tainted 2.6.33.3-hw #1
Call Trace:
 [<ffffffff81045691>] __schedule_bug+0x61/0x70
 [<ffffffff81656f68>] schedule+0x588/0xa00
 [<ffffffff8105cb34>] ? try_to_del_timer_sync+0xa4/0xd0
 [<ffffffff8165773f>] schedule_timeout+0x18f/0x270
 [<ffffffff8105cd00>] ? process_timeout+0x0/0x10
 [<ffffffff8165697f>] io_schedule_timeout+0x8f/0xf0
 [<ffffffff810bed88>] balance_dirty_pages_ratelimited_nr+0x178/0x3a0
 [<ffffffff810b65d3>] generic_file_buffered_write+0x193/0x230
 [<ffffffff812befb6>] xfs_write+0x7c6/0x8f0
 [<ffffffff81294ec0>] ? xfs_iget+0x4f0/0x660
 [<ffffffff812bab50>] ? xfs_file_aio_write+0x0/0x60
 [<ffffffff812baba6>] xfs_file_aio_write+0x56/0x60
 [<ffffffff810eea6b>] do_sync_readv_writev+0xcb/0x110
 [<ffffffff81204616>] ? exportfs_decode_fh+0xe6/0x270
 [<ffffffff812081d0>] ? nfsd_acceptable+0x0/0x120
 [<ffffffff810ee88e>] ? rw_copy_check_uvector+0x7e/0x130
 [<ffffffff810ef15f>] do_readv_writev+0xcf/0x1f0
 [<ffffffff81208362>] ? nfsd_setuser_and_check_port+0x72/0x80
 [<ffffffff81208cac>] ? nfsd_permission+0xec/0x160
 [<ffffffff810ef2c0>] vfs_writev+0x40/0x60
 [<ffffffff81209ece>] nfsd_vfs_write+0xde/0x420
 [<ffffffff810ed1bd>] ? dentry_open+0x4d/0xb0
 [<ffffffff8120a8ce>] ? nfsd_open+0x16e/0x200
 [<ffffffff8120ad0a>] nfsd_write+0xea/0x100
 [<ffffffff812104cb>] ? nfsd_cache_lookup+0x2bb/0x3e0
 [<ffffffff81212b7f>] nfsd3_proc_write+0xaf/0x140
 [<ffffffff81204b4b>] nfsd_dispatch+0xbb/0x260
 [<ffffffff81618eaf>] svc_process+0x4af/0x820
 [<ffffffff81205190>] ? nfsd+0x0/0x160
 [<ffffffff8120526d>] nfsd+0xdd/0x160
 [<ffffffff8106aac6>] kthread+0x96/0xb0
 [<ffffffff8100ace4>] kernel_thread_helper+0x4/0x10
 [<ffffffff8106aa30>] ? kthread+0x0/0xb0
 [<ffffffff8100ace0>] ? kernel_thread_helper+0x0/0x10
BUG: scheduling while atomic: nfsd/29671/0x00000250
Modules linked in: aoe xt_MARK ipmi_devintf ipmi_si ipmi_msghandler e1000e bnx2
Pid: 29671, comm: nfsd Not tainted 2.6.33.3-hw #1
Call Trace:
 [<ffffffff81045691>] __schedule_bug+0x61/0x70
 [<ffffffff81656f68>] schedule+0x588/0xa00
 [<ffffffff810bd448>] ? __alloc_pages_nodemask+0x108/0x6e0
 [<ffffffff8165773f>] schedule_timeout+0x18f/0x270
 [<ffffffff8105cd00>] ? process_timeout+0x0/0x10
 [<ffffffff81626f99>] svc_recv+0x539/0x8b0
 [<ffffffff81047f30>] ? default_wake_function+0x0/0x10
 [<ffffffff81205190>] ? nfsd+0x0/0x160
 [<ffffffff8120522d>] nfsd+0x9d/0x160
 [<ffffffff8106aac6>] kthread+0x96/0xb0
 [<ffffffff8100ace4>] kernel_thread_helper+0x4/0x10
 [<ffffffff8106aa30>] ? kthread+0x0/0xb0
 [<ffffffff8100ace0>] ? kernel_thread_helper+0x0/0x10

(followed by several million more "scheduling while atomic" errors with
different backtraces)

Simon-

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 2+ messages in thread

* Re: [2.6.33.3] scheduling while atomic (inode reclaim races still?)
  2010-05-20  8:31 [2.6.33.3] scheduling while atomic (inode reclaim races still?) Simon Kirby
@ 2010-05-20  9:03 ` Christoph Hellwig
  0 siblings, 0 replies; 2+ messages in thread
From: Christoph Hellwig @ 2010-05-20  9:03 UTC (permalink / raw)
  To: Simon Kirby; +Cc: xfs

On Thu, May 20, 2010 at 01:31:04AM -0700, Simon Kirby wrote:
> This started happening on host about 10 days after upgrading to 2.6.33.3
> (in hopes that it fixed all of the reclaim issues present in 2.6.33.2). 
> I don't see any fixes in 2.6.33.4 relevant to this particular issue.
> 
> After this first error, the kernel logs just filled with repeat
> occurrences with different backtraces.
> 
> BUG: scheduling while atomic: nfsd/29671/0x00000250

The bug unfortunately doesn't tell us what caused the "atomic"
situation.  But given that XFS doesn't take any spinlock in that path
it seems like this must be caused by nfsd.

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2010-05-20  9:01 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2010-05-20  8:31 [2.6.33.3] scheduling while atomic (inode reclaim races still?) Simon Kirby
2010-05-20  9:03 ` Christoph Hellwig

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox