* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
[not found] ` <201103152102.p2FL2NN8006070@demeter2.kernel.org>
@ 2011-03-15 21:24 ` Andrew Morton
2011-03-15 23:54 ` Mehmet Giritli
0 siblings, 1 reply; 10+ messages in thread
From: Andrew Morton @ 2011-03-15 21:24 UTC (permalink / raw)
To: linux-fsdevel, Nick Piggin, Al Viro
Cc: bugzilla-daemon, Mehmet Giritli, Ian Kent
(switched to email. Please respond via emailed reply-to-all, not via the
bugzilla web interface).
Seems that we have a nasty involving autofs, nfs and the VFS.
Mehmet, the kernel should have printed some diagnostics prior to doing
the BUG() call:
if (dentry->d_count != 0) {
printk(KERN_ERR
"BUG: Dentry %p{i=%lx,n=%s}"
" still in use (%d)"
" [unmount of %s %s]\n",
dentry,
dentry->d_inode ?
dentry->d_inode->i_ino : 0UL,
dentry->d_name.name,
dentry->d_count,
dentry->d_sb->s_type->name,
dentry->d_sb->s_id);
BUG();
}
Please find those in the log and email them to use - someone might find
it useful.
On Tue, 15 Mar 2011 21:02:23 GMT
bugzilla-daemon@bugzilla.kernel.org wrote:
> https://bugzilla.kernel.org/show_bug.cgi?id=30882
>
>
>
>
>
> --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> Here is that crash happening again, the system was NOT running overclocked or
> anything...
>
> [ 1860.156122] ------------[ cut here ]------------
> [ 1860.156124] kernel BUG at fs/dcache.c:943!
> [ 1860.156126] invalid opcode: 0000 [#1] SMP
> [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> [ 1860.156128] CPU 3
> [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> nvidia(P)
> [ 1860.156137]
> [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> shrink_dcache_for_umount_subtree+0x268/0x270
> [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> 000000000003ffff
> [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> ffffffff8174c9f8
> [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> 0000000000000006
> [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> ffff88023a07f5e0
> [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> ffff880211d38740
> [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> knlGS:00000000f74186c0
> [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> 00000000000006e0
> [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> 0000000000000000
> [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> 0000000000000400
> [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> ffff880211fd5640)
> [ 1860.156162] Stack:
> [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> ffff88020c05cc00
> [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> ffffffff810e96a4
> [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> ffffffff810d5d15
> [ 1860.156169] Call Trace:
> [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> [ 1860.156201] RIP [<ffffffff810e9648>]
> shrink_dcache_for_umount_subtree+0x268/0x270
> [ 1860.156204] RSP <ffff8800be82fe08>
> [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
>
> --
> Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> ------- You are receiving this mail because: -------
> You are on the CC list for the bug.
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-15 21:24 ` [Bug 30882] Automatic process group scheduling causes crashes after a while Andrew Morton
@ 2011-03-15 23:54 ` Mehmet Giritli
2011-03-16 2:32 ` Ian Kent
0 siblings, 1 reply; 10+ messages in thread
From: Mehmet Giritli @ 2011-03-15 23:54 UTC (permalink / raw)
To: Andrew Morton
Cc: linux-fsdevel, Nick Piggin, Al Viro, bugzilla-daemon,
Mehmet Giritli, Ian Kent
The missing piece is as follows:
Mar 15 22:37:38 mordor kernel: [ 1860.156114] BUG: Dentry
ffff88023f96e600{i=25f56f,n=} still in use (1) [unmount of nfs 0:f]
(sorry for the inconvenience Andrew)
On Tue, 2011-03-15 at 14:24 -0700, Andrew Morton wrote:
> (switched to email. Please respond via emailed reply-to-all, not via the
> bugzilla web interface).
>
> Seems that we have a nasty involving autofs, nfs and the VFS.
>
> Mehmet, the kernel should have printed some diagnostics prior to doing
> the BUG() call:
>
> if (dentry->d_count != 0) {
> printk(KERN_ERR
> "BUG: Dentry %p{i=%lx,n=%s}"
> " still in use (%d)"
> " [unmount of %s %s]\n",
> dentry,
> dentry->d_inode ?
> dentry->d_inode->i_ino : 0UL,
> dentry->d_name.name,
> dentry->d_count,
> dentry->d_sb->s_type->name,
> dentry->d_sb->s_id);
> BUG();
> }
>
> Please find those in the log and email them to use - someone might find
> it useful.
>
>
> On Tue, 15 Mar 2011 21:02:23 GMT
> bugzilla-daemon@bugzilla.kernel.org wrote:
>
> > https://bugzilla.kernel.org/show_bug.cgi?id=30882
> >
> >
> >
> >
> >
> > --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> > Here is that crash happening again, the system was NOT running overclocked or
> > anything...
> >
> > [ 1860.156122] ------------[ cut here ]------------
> > [ 1860.156124] kernel BUG at fs/dcache.c:943!
> > [ 1860.156126] invalid opcode: 0000 [#1] SMP
> > [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> > [ 1860.156128] CPU 3
> > [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> > xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> > xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> > nvidia(P)
> > [ 1860.156137]
> > [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> > Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> > [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> > shrink_dcache_for_umount_subtree+0x268/0x270
> > [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> > [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> > 000000000003ffff
> > [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> > ffffffff8174c9f8
> > [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> > 0000000000000006
> > [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> > ffff88023a07f5e0
> > [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> > ffff880211d38740
> > [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> > knlGS:00000000f74186c0
> > [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> > 00000000000006e0
> > [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> > 0000000000000000
> > [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> > 0000000000000400
> > [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> > ffff880211fd5640)
> > [ 1860.156162] Stack:
> > [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> > ffff88020c05cc00
> > [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> > ffffffff810e96a4
> > [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> > ffffffff810d5d15
> > [ 1860.156169] Call Trace:
> > [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> > [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> > [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> > [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> > [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> > [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> > [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> > [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> > [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> > 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> > eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> > [ 1860.156201] RIP [<ffffffff810e9648>]
> > shrink_dcache_for_umount_subtree+0x268/0x270
> > [ 1860.156204] RSP <ffff8800be82fe08>
> > [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
> >
> > --
> > Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> > ------- You are receiving this mail because: -------
> > You are on the CC list for the bug.
>
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-15 23:54 ` Mehmet Giritli
@ 2011-03-16 2:32 ` Ian Kent
2011-03-16 14:27 ` Mehmet Giritli
0 siblings, 1 reply; 10+ messages in thread
From: Ian Kent @ 2011-03-16 2:32 UTC (permalink / raw)
To: mgiritli
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon, Mehmet Giritli
On Wed, 2011-03-16 at 01:54 +0200, Mehmet Giritli wrote:
> The missing piece is as follows:
>
> Mar 15 22:37:38 mordor kernel: [ 1860.156114] BUG: Dentry
> ffff88023f96e600{i=25f56f,n=} still in use (1) [unmount of nfs 0:f]
This might be the same problem I saw and described in rc1.
However, for me the fs in the BUG() report was autofs.
Hopefully that just means my autofs setup is different.
At the moment I believe a dentry leak Al Viro spotted is the cause.
Please try this patch.
autofs4 - fix dentry leak in autofs4_expire_direct()
From: Ian Kent <raven@themaw.net>
There is a missing dput() when returning from autofs4_expire_direct()
when we see that the dentry is already a pending mount.
Signed-off-by: Ian Kent <raven@themaw.net>
---
fs/autofs4/expire.c | 7 +++----
1 files changed, 3 insertions(+), 4 deletions(-)
diff --git a/fs/autofs4/expire.c b/fs/autofs4/expire.c
index c896dd6..c403abc 100644
--- a/fs/autofs4/expire.c
+++ b/fs/autofs4/expire.c
@@ -290,10 +290,8 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
spin_lock(&sbi->fs_lock);
ino = autofs4_dentry_ino(root);
/* No point expiring a pending mount */
- if (ino->flags & AUTOFS_INF_PENDING) {
- spin_unlock(&sbi->fs_lock);
- return NULL;
- }
+ if (ino->flags & AUTOFS_INF_PENDING)
+ goto out;
if (!autofs4_direct_busy(mnt, root, timeout, do_now)) {
struct autofs_info *ino = autofs4_dentry_ino(root);
ino->flags |= AUTOFS_INF_EXPIRING;
@@ -301,6 +299,7 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
spin_unlock(&sbi->fs_lock);
return root;
}
+out:
spin_unlock(&sbi->fs_lock);
dput(root);
>
> (sorry for the inconvenience Andrew)
>
> On Tue, 2011-03-15 at 14:24 -0700, Andrew Morton wrote:
> > (switched to email. Please respond via emailed reply-to-all, not via the
> > bugzilla web interface).
> >
> > Seems that we have a nasty involving autofs, nfs and the VFS.
> >
> > Mehmet, the kernel should have printed some diagnostics prior to doing
> > the BUG() call:
> >
> > if (dentry->d_count != 0) {
> > printk(KERN_ERR
> > "BUG: Dentry %p{i=%lx,n=%s}"
> > " still in use (%d)"
> > " [unmount of %s %s]\n",
> > dentry,
> > dentry->d_inode ?
> > dentry->d_inode->i_ino : 0UL,
> > dentry->d_name.name,
> > dentry->d_count,
> > dentry->d_sb->s_type->name,
> > dentry->d_sb->s_id);
> > BUG();
> > }
> >
> > Please find those in the log and email them to use - someone might find
> > it useful.
> >
> >
> > On Tue, 15 Mar 2011 21:02:23 GMT
> > bugzilla-daemon@bugzilla.kernel.org wrote:
> >
> > > https://bugzilla.kernel.org/show_bug.cgi?id=30882
> > >
> > >
> > >
> > >
> > >
> > > --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> > > Here is that crash happening again, the system was NOT running overclocked or
> > > anything...
> > >
> > > [ 1860.156122] ------------[ cut here ]------------
> > > [ 1860.156124] kernel BUG at fs/dcache.c:943!
> > > [ 1860.156126] invalid opcode: 0000 [#1] SMP
> > > [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> > > [ 1860.156128] CPU 3
> > > [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> > > xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> > > xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> > > nvidia(P)
> > > [ 1860.156137]
> > > [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> > > Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> > > [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> > > [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> > > 000000000003ffff
> > > [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> > > ffffffff8174c9f8
> > > [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> > > 0000000000000006
> > > [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> > > ffff88023a07f5e0
> > > [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> > > ffff880211d38740
> > > [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> > > knlGS:00000000f74186c0
> > > [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > > [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> > > 00000000000006e0
> > > [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> > > 0000000000000000
> > > [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> > > 0000000000000400
> > > [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> > > ffff880211fd5640)
> > > [ 1860.156162] Stack:
> > > [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> > > ffff88020c05cc00
> > > [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> > > ffffffff810e96a4
> > > [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> > > ffffffff810d5d15
> > > [ 1860.156169] Call Trace:
> > > [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> > > [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> > > [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> > > [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> > > [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> > > [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> > > [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> > > [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> > > [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> > > 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> > > eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> > > [ 1860.156201] RIP [<ffffffff810e9648>]
> > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > [ 1860.156204] RSP <ffff8800be82fe08>
> > > [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
> > >
> > > --
> > > Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> > > ------- You are receiving this mail because: -------
> > > You are on the CC list for the bug.
> >
>
>
^ permalink raw reply related [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-16 2:32 ` Ian Kent
@ 2011-03-16 14:27 ` Mehmet Giritli
2011-03-16 15:21 ` Ian Kent
0 siblings, 1 reply; 10+ messages in thread
From: Mehmet Giritli @ 2011-03-16 14:27 UTC (permalink / raw)
To: Ian Kent
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon, Mehmet Giritli
Ian,
I am having much more frequent crashes now. I havent been able to
cleanly reboot my machine yet and I have tried three times so far. Init
scripts fail to unmount the file systems and I have to reboot manually
On Wed, 2011-03-16 at 10:32 +0800, Ian Kent wrote:
> On Wed, 2011-03-16 at 01:54 +0200, Mehmet Giritli wrote:
> > The missing piece is as follows:
> >
> > Mar 15 22:37:38 mordor kernel: [ 1860.156114] BUG: Dentry
> > ffff88023f96e600{i=25f56f,n=} still in use (1) [unmount of nfs 0:f]
>
> This might be the same problem I saw and described in rc1.
> However, for me the fs in the BUG() report was autofs.
> Hopefully that just means my autofs setup is different.
>
> At the moment I believe a dentry leak Al Viro spotted is the cause.
> Please try this patch.
>
> autofs4 - fix dentry leak in autofs4_expire_direct()
>
> From: Ian Kent <raven@themaw.net>
>
> There is a missing dput() when returning from autofs4_expire_direct()
> when we see that the dentry is already a pending mount.
>
> Signed-off-by: Ian Kent <raven@themaw.net>
> ---
>
> fs/autofs4/expire.c | 7 +++----
> 1 files changed, 3 insertions(+), 4 deletions(-)
>
>
> diff --git a/fs/autofs4/expire.c b/fs/autofs4/expire.c
> index c896dd6..c403abc 100644
> --- a/fs/autofs4/expire.c
> +++ b/fs/autofs4/expire.c
> @@ -290,10 +290,8 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> spin_lock(&sbi->fs_lock);
> ino = autofs4_dentry_ino(root);
> /* No point expiring a pending mount */
> - if (ino->flags & AUTOFS_INF_PENDING) {
> - spin_unlock(&sbi->fs_lock);
> - return NULL;
> - }
> + if (ino->flags & AUTOFS_INF_PENDING)
> + goto out;
> if (!autofs4_direct_busy(mnt, root, timeout, do_now)) {
> struct autofs_info *ino = autofs4_dentry_ino(root);
> ino->flags |= AUTOFS_INF_EXPIRING;
> @@ -301,6 +299,7 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> spin_unlock(&sbi->fs_lock);
> return root;
> }
> +out:
> spin_unlock(&sbi->fs_lock);
> dput(root);
>
>
> >
> > (sorry for the inconvenience Andrew)
> >
> > On Tue, 2011-03-15 at 14:24 -0700, Andrew Morton wrote:
> > > (switched to email. Please respond via emailed reply-to-all, not via the
> > > bugzilla web interface).
> > >
> > > Seems that we have a nasty involving autofs, nfs and the VFS.
> > >
> > > Mehmet, the kernel should have printed some diagnostics prior to doing
> > > the BUG() call:
> > >
> > > if (dentry->d_count != 0) {
> > > printk(KERN_ERR
> > > "BUG: Dentry %p{i=%lx,n=%s}"
> > > " still in use (%d)"
> > > " [unmount of %s %s]\n",
> > > dentry,
> > > dentry->d_inode ?
> > > dentry->d_inode->i_ino : 0UL,
> > > dentry->d_name.name,
> > > dentry->d_count,
> > > dentry->d_sb->s_type->name,
> > > dentry->d_sb->s_id);
> > > BUG();
> > > }
> > >
> > > Please find those in the log and email them to use - someone might find
> > > it useful.
> > >
> > >
> > > On Tue, 15 Mar 2011 21:02:23 GMT
> > > bugzilla-daemon@bugzilla.kernel.org wrote:
> > >
> > > > https://bugzilla.kernel.org/show_bug.cgi?id=30882
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> > > > Here is that crash happening again, the system was NOT running overclocked or
> > > > anything...
> > > >
> > > > [ 1860.156122] ------------[ cut here ]------------
> > > > [ 1860.156124] kernel BUG at fs/dcache.c:943!
> > > > [ 1860.156126] invalid opcode: 0000 [#1] SMP
> > > > [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> > > > [ 1860.156128] CPU 3
> > > > [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> > > > xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> > > > xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> > > > nvidia(P)
> > > > [ 1860.156137]
> > > > [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> > > > Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> > > > [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> > > > [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> > > > 000000000003ffff
> > > > [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> > > > ffffffff8174c9f8
> > > > [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> > > > 0000000000000006
> > > > [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> > > > ffff88023a07f5e0
> > > > [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> > > > ffff880211d38740
> > > > [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> > > > knlGS:00000000f74186c0
> > > > [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > > > [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> > > > 00000000000006e0
> > > > [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> > > > 0000000000000000
> > > > [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> > > > 0000000000000400
> > > > [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> > > > ffff880211fd5640)
> > > > [ 1860.156162] Stack:
> > > > [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> > > > ffff88020c05cc00
> > > > [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> > > > ffffffff810e96a4
> > > > [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> > > > ffffffff810d5d15
> > > > [ 1860.156169] Call Trace:
> > > > [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> > > > [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> > > > [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> > > > [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> > > > [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> > > > [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> > > > [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> > > > [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> > > > [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> > > > 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> > > > eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> > > > [ 1860.156201] RIP [<ffffffff810e9648>]
> > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > [ 1860.156204] RSP <ffff8800be82fe08>
> > > > [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
> > > >
> > > > --
> > > > Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> > > > ------- You are receiving this mail because: -------
> > > > You are on the CC list for the bug.
> > >
> >
> >
>
>
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-16 14:27 ` Mehmet Giritli
@ 2011-03-16 15:21 ` Ian Kent
2011-03-16 15:29 ` Mehmet Giritli
0 siblings, 1 reply; 10+ messages in thread
From: Ian Kent @ 2011-03-16 15:21 UTC (permalink / raw)
To: mgiritli
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon, Mehmet Giritli
On Wed, 2011-03-16 at 16:27 +0200, Mehmet Giritli wrote:
> Ian,
>
> I am having much more frequent crashes now. I havent been able to
> cleanly reboot my machine yet and I have tried three times so far. Init
> scripts fail to unmount the file systems and I have to reboot manually
What do your autofs maps look like?
>
> On Wed, 2011-03-16 at 10:32 +0800, Ian Kent wrote:
> > On Wed, 2011-03-16 at 01:54 +0200, Mehmet Giritli wrote:
> > > The missing piece is as follows:
> > >
> > > Mar 15 22:37:38 mordor kernel: [ 1860.156114] BUG: Dentry
> > > ffff88023f96e600{i=25f56f,n=} still in use (1) [unmount of nfs 0:f]
> >
> > This might be the same problem I saw and described in rc1.
> > However, for me the fs in the BUG() report was autofs.
> > Hopefully that just means my autofs setup is different.
> >
> > At the moment I believe a dentry leak Al Viro spotted is the cause.
> > Please try this patch.
> >
> > autofs4 - fix dentry leak in autofs4_expire_direct()
> >
> > From: Ian Kent <raven@themaw.net>
> >
> > There is a missing dput() when returning from autofs4_expire_direct()
> > when we see that the dentry is already a pending mount.
> >
> > Signed-off-by: Ian Kent <raven@themaw.net>
> > ---
> >
> > fs/autofs4/expire.c | 7 +++----
> > 1 files changed, 3 insertions(+), 4 deletions(-)
> >
> >
> > diff --git a/fs/autofs4/expire.c b/fs/autofs4/expire.c
> > index c896dd6..c403abc 100644
> > --- a/fs/autofs4/expire.c
> > +++ b/fs/autofs4/expire.c
> > @@ -290,10 +290,8 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > spin_lock(&sbi->fs_lock);
> > ino = autofs4_dentry_ino(root);
> > /* No point expiring a pending mount */
> > - if (ino->flags & AUTOFS_INF_PENDING) {
> > - spin_unlock(&sbi->fs_lock);
> > - return NULL;
> > - }
> > + if (ino->flags & AUTOFS_INF_PENDING)
> > + goto out;
> > if (!autofs4_direct_busy(mnt, root, timeout, do_now)) {
> > struct autofs_info *ino = autofs4_dentry_ino(root);
> > ino->flags |= AUTOFS_INF_EXPIRING;
> > @@ -301,6 +299,7 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > spin_unlock(&sbi->fs_lock);
> > return root;
> > }
> > +out:
> > spin_unlock(&sbi->fs_lock);
> > dput(root);
> >
> >
> > >
> > > (sorry for the inconvenience Andrew)
> > >
> > > On Tue, 2011-03-15 at 14:24 -0700, Andrew Morton wrote:
> > > > (switched to email. Please respond via emailed reply-to-all, not via the
> > > > bugzilla web interface).
> > > >
> > > > Seems that we have a nasty involving autofs, nfs and the VFS.
> > > >
> > > > Mehmet, the kernel should have printed some diagnostics prior to doing
> > > > the BUG() call:
> > > >
> > > > if (dentry->d_count != 0) {
> > > > printk(KERN_ERR
> > > > "BUG: Dentry %p{i=%lx,n=%s}"
> > > > " still in use (%d)"
> > > > " [unmount of %s %s]\n",
> > > > dentry,
> > > > dentry->d_inode ?
> > > > dentry->d_inode->i_ino : 0UL,
> > > > dentry->d_name.name,
> > > > dentry->d_count,
> > > > dentry->d_sb->s_type->name,
> > > > dentry->d_sb->s_id);
> > > > BUG();
> > > > }
> > > >
> > > > Please find those in the log and email them to use - someone might find
> > > > it useful.
> > > >
> > > >
> > > > On Tue, 15 Mar 2011 21:02:23 GMT
> > > > bugzilla-daemon@bugzilla.kernel.org wrote:
> > > >
> > > > > https://bugzilla.kernel.org/show_bug.cgi?id=30882
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> > > > > Here is that crash happening again, the system was NOT running overclocked or
> > > > > anything...
> > > > >
> > > > > [ 1860.156122] ------------[ cut here ]------------
> > > > > [ 1860.156124] kernel BUG at fs/dcache.c:943!
> > > > > [ 1860.156126] invalid opcode: 0000 [#1] SMP
> > > > > [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> > > > > [ 1860.156128] CPU 3
> > > > > [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> > > > > xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> > > > > xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> > > > > nvidia(P)
> > > > > [ 1860.156137]
> > > > > [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> > > > > Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> > > > > [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> > > > > [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> > > > > 000000000003ffff
> > > > > [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> > > > > ffffffff8174c9f8
> > > > > [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> > > > > 0000000000000006
> > > > > [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> > > > > ffff88023a07f5e0
> > > > > [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> > > > > ffff880211d38740
> > > > > [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> > > > > knlGS:00000000f74186c0
> > > > > [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > > > > [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> > > > > 00000000000006e0
> > > > > [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> > > > > 0000000000000000
> > > > > [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> > > > > 0000000000000400
> > > > > [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> > > > > ffff880211fd5640)
> > > > > [ 1860.156162] Stack:
> > > > > [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> > > > > ffff88020c05cc00
> > > > > [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> > > > > ffffffff810e96a4
> > > > > [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> > > > > ffffffff810d5d15
> > > > > [ 1860.156169] Call Trace:
> > > > > [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> > > > > [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> > > > > [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> > > > > [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> > > > > [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> > > > > [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> > > > > [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> > > > > [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> > > > > [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> > > > > 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> > > > > eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> > > > > [ 1860.156201] RIP [<ffffffff810e9648>]
> > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > [ 1860.156204] RSP <ffff8800be82fe08>
> > > > > [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
> > > > >
> > > > > --
> > > > > Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> > > > > ------- You are receiving this mail because: -------
> > > > > You are on the CC list for the bug.
> > > >
> > >
> > >
> >
> >
>
>
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-16 15:21 ` Ian Kent
@ 2011-03-16 15:29 ` Mehmet Giritli
2011-03-16 15:44 ` Ian Kent
0 siblings, 1 reply; 10+ messages in thread
From: Mehmet Giritli @ 2011-03-16 15:29 UTC (permalink / raw)
To: Ian Kent
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon, Mehmet Giritli
On Wed, 2011-03-16 at 23:21 +0800, Ian Kent wrote:
> On Wed, 2011-03-16 at 16:27 +0200, Mehmet Giritli wrote:
> > Ian,
> >
> > I am having much more frequent crashes now. I havent been able to
> > cleanly reboot my machine yet and I have tried three times so far. Init
> > scripts fail to unmount the file systems and I have to reboot manually
>
> What do your autofs maps look like?
>
>
Here is the contents of my auto.misc:
gollum-media -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/mnt/media
gollum-distfiles -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/usr/portage/distfiles
gollum-www -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/var/www
gollum-WebDav -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/var/dav
> >
> > On Wed, 2011-03-16 at 10:32 +0800, Ian Kent wrote:
> > > On Wed, 2011-03-16 at 01:54 +0200, Mehmet Giritli wrote:
> > > > The missing piece is as follows:
> > > >
> > > > Mar 15 22:37:38 mordor kernel: [ 1860.156114] BUG: Dentry
> > > > ffff88023f96e600{i=25f56f,n=} still in use (1) [unmount of nfs 0:f]
> > >
> > > This might be the same problem I saw and described in rc1.
> > > However, for me the fs in the BUG() report was autofs.
> > > Hopefully that just means my autofs setup is different.
> > >
> > > At the moment I believe a dentry leak Al Viro spotted is the cause.
> > > Please try this patch.
> > >
> > > autofs4 - fix dentry leak in autofs4_expire_direct()
> > >
> > > From: Ian Kent <raven@themaw.net>
> > >
> > > There is a missing dput() when returning from autofs4_expire_direct()
> > > when we see that the dentry is already a pending mount.
> > >
> > > Signed-off-by: Ian Kent <raven@themaw.net>
> > > ---
> > >
> > > fs/autofs4/expire.c | 7 +++----
> > > 1 files changed, 3 insertions(+), 4 deletions(-)
> > >
> > >
> > > diff --git a/fs/autofs4/expire.c b/fs/autofs4/expire.c
> > > index c896dd6..c403abc 100644
> > > --- a/fs/autofs4/expire.c
> > > +++ b/fs/autofs4/expire.c
> > > @@ -290,10 +290,8 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > > spin_lock(&sbi->fs_lock);
> > > ino = autofs4_dentry_ino(root);
> > > /* No point expiring a pending mount */
> > > - if (ino->flags & AUTOFS_INF_PENDING) {
> > > - spin_unlock(&sbi->fs_lock);
> > > - return NULL;
> > > - }
> > > + if (ino->flags & AUTOFS_INF_PENDING)
> > > + goto out;
> > > if (!autofs4_direct_busy(mnt, root, timeout, do_now)) {
> > > struct autofs_info *ino = autofs4_dentry_ino(root);
> > > ino->flags |= AUTOFS_INF_EXPIRING;
> > > @@ -301,6 +299,7 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > > spin_unlock(&sbi->fs_lock);
> > > return root;
> > > }
> > > +out:
> > > spin_unlock(&sbi->fs_lock);
> > > dput(root);
> > >
> > >
> > > >
> > > > (sorry for the inconvenience Andrew)
> > > >
> > > > On Tue, 2011-03-15 at 14:24 -0700, Andrew Morton wrote:
> > > > > (switched to email. Please respond via emailed reply-to-all, not via the
> > > > > bugzilla web interface).
> > > > >
> > > > > Seems that we have a nasty involving autofs, nfs and the VFS.
> > > > >
> > > > > Mehmet, the kernel should have printed some diagnostics prior to doing
> > > > > the BUG() call:
> > > > >
> > > > > if (dentry->d_count != 0) {
> > > > > printk(KERN_ERR
> > > > > "BUG: Dentry %p{i=%lx,n=%s}"
> > > > > " still in use (%d)"
> > > > > " [unmount of %s %s]\n",
> > > > > dentry,
> > > > > dentry->d_inode ?
> > > > > dentry->d_inode->i_ino : 0UL,
> > > > > dentry->d_name.name,
> > > > > dentry->d_count,
> > > > > dentry->d_sb->s_type->name,
> > > > > dentry->d_sb->s_id);
> > > > > BUG();
> > > > > }
> > > > >
> > > > > Please find those in the log and email them to use - someone might find
> > > > > it useful.
> > > > >
> > > > >
> > > > > On Tue, 15 Mar 2011 21:02:23 GMT
> > > > > bugzilla-daemon@bugzilla.kernel.org wrote:
> > > > >
> > > > > > https://bugzilla.kernel.org/show_bug.cgi?id=30882
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> > > > > > Here is that crash happening again, the system was NOT running overclocked or
> > > > > > anything...
> > > > > >
> > > > > > [ 1860.156122] ------------[ cut here ]------------
> > > > > > [ 1860.156124] kernel BUG at fs/dcache.c:943!
> > > > > > [ 1860.156126] invalid opcode: 0000 [#1] SMP
> > > > > > [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> > > > > > [ 1860.156128] CPU 3
> > > > > > [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> > > > > > xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> > > > > > xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> > > > > > nvidia(P)
> > > > > > [ 1860.156137]
> > > > > > [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> > > > > > Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> > > > > > [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> > > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > > [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> > > > > > [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> > > > > > 000000000003ffff
> > > > > > [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> > > > > > ffffffff8174c9f8
> > > > > > [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> > > > > > 0000000000000006
> > > > > > [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> > > > > > ffff88023a07f5e0
> > > > > > [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> > > > > > ffff880211d38740
> > > > > > [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> > > > > > knlGS:00000000f74186c0
> > > > > > [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > > > > > [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> > > > > > 00000000000006e0
> > > > > > [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> > > > > > 0000000000000000
> > > > > > [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> > > > > > 0000000000000400
> > > > > > [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> > > > > > ffff880211fd5640)
> > > > > > [ 1860.156162] Stack:
> > > > > > [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> > > > > > ffff88020c05cc00
> > > > > > [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> > > > > > ffffffff810e96a4
> > > > > > [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> > > > > > ffffffff810d5d15
> > > > > > [ 1860.156169] Call Trace:
> > > > > > [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> > > > > > [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> > > > > > [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> > > > > > [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> > > > > > [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> > > > > > [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> > > > > > [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> > > > > > [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> > > > > > [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> > > > > > 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> > > > > > eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> > > > > > [ 1860.156201] RIP [<ffffffff810e9648>]
> > > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > > [ 1860.156204] RSP <ffff8800be82fe08>
> > > > > > [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
> > > > > >
> > > > > > --
> > > > > > Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> > > > > > ------- You are receiving this mail because: -------
> > > > > > You are on the CC list for the bug.
> > > > >
> > > >
> > > >
> > >
> > >
> >
> >
>
>
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-16 15:29 ` Mehmet Giritli
@ 2011-03-16 15:44 ` Ian Kent
2011-03-16 16:48 ` Mehmet Giritli
2011-03-16 17:47 ` Mehmet Giritli
0 siblings, 2 replies; 10+ messages in thread
From: Ian Kent @ 2011-03-16 15:44 UTC (permalink / raw)
To: mgiritli
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon, Mehmet Giritli
On Wed, 2011-03-16 at 17:29 +0200, Mehmet Giritli wrote:
> On Wed, 2011-03-16 at 23:21 +0800, Ian Kent wrote:
> > On Wed, 2011-03-16 at 16:27 +0200, Mehmet Giritli wrote:
> > > Ian,
> > >
> > > I am having much more frequent crashes now. I havent been able to
> > > cleanly reboot my machine yet and I have tried three times so far. Init
> > > scripts fail to unmount the file systems and I have to reboot manually
> >
> > What do your autofs maps look like?
> >
> >
>
> Here is the contents of my auto.misc:
>
> gollum-media -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/mnt/media
> gollum-distfiles -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/usr/portage/distfiles
> gollum-www -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/var/www
> gollum-WebDav -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/var/dav
What, that's it, and your only using "/misc /etc/auto.misc" in the
master map and your having problems.
Are the crashes always the same?
How have you established that the BUG()s are in fact due to automount
umounting mounts and that the BUG()s correspond to NFS mounts previously
mounted by autofs?
Is there any noise at all in the syslog?
Are you sure your using a kernel with the dentry leak patch?
What sort of automounting load is happening on the machine, ie.
frequency or mounts and umounts and what timeout are you using?
The dentry leak patch got rid of the BUG()s I was seeing but by that
time I did have a couple of other patches. I still don't think the other
patches made much difference for this particular case.
>
> > >
> > > On Wed, 2011-03-16 at 10:32 +0800, Ian Kent wrote:
> > > > On Wed, 2011-03-16 at 01:54 +0200, Mehmet Giritli wrote:
> > > > > The missing piece is as follows:
> > > > >
> > > > > Mar 15 22:37:38 mordor kernel: [ 1860.156114] BUG: Dentry
> > > > > ffff88023f96e600{i=25f56f,n=} still in use (1) [unmount of nfs 0:f]
> > > >
> > > > This might be the same problem I saw and described in rc1.
> > > > However, for me the fs in the BUG() report was autofs.
> > > > Hopefully that just means my autofs setup is different.
> > > >
> > > > At the moment I believe a dentry leak Al Viro spotted is the cause.
> > > > Please try this patch.
> > > >
> > > > autofs4 - fix dentry leak in autofs4_expire_direct()
> > > >
> > > > From: Ian Kent <raven@themaw.net>
> > > >
> > > > There is a missing dput() when returning from autofs4_expire_direct()
> > > > when we see that the dentry is already a pending mount.
> > > >
> > > > Signed-off-by: Ian Kent <raven@themaw.net>
> > > > ---
> > > >
> > > > fs/autofs4/expire.c | 7 +++----
> > > > 1 files changed, 3 insertions(+), 4 deletions(-)
> > > >
> > > >
> > > > diff --git a/fs/autofs4/expire.c b/fs/autofs4/expire.c
> > > > index c896dd6..c403abc 100644
> > > > --- a/fs/autofs4/expire.c
> > > > +++ b/fs/autofs4/expire.c
> > > > @@ -290,10 +290,8 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > > > spin_lock(&sbi->fs_lock);
> > > > ino = autofs4_dentry_ino(root);
> > > > /* No point expiring a pending mount */
> > > > - if (ino->flags & AUTOFS_INF_PENDING) {
> > > > - spin_unlock(&sbi->fs_lock);
> > > > - return NULL;
> > > > - }
> > > > + if (ino->flags & AUTOFS_INF_PENDING)
> > > > + goto out;
> > > > if (!autofs4_direct_busy(mnt, root, timeout, do_now)) {
> > > > struct autofs_info *ino = autofs4_dentry_ino(root);
> > > > ino->flags |= AUTOFS_INF_EXPIRING;
> > > > @@ -301,6 +299,7 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > > > spin_unlock(&sbi->fs_lock);
> > > > return root;
> > > > }
> > > > +out:
> > > > spin_unlock(&sbi->fs_lock);
> > > > dput(root);
> > > >
> > > >
> > > > >
> > > > > (sorry for the inconvenience Andrew)
> > > > >
> > > > > On Tue, 2011-03-15 at 14:24 -0700, Andrew Morton wrote:
> > > > > > (switched to email. Please respond via emailed reply-to-all, not via the
> > > > > > bugzilla web interface).
> > > > > >
> > > > > > Seems that we have a nasty involving autofs, nfs and the VFS.
> > > > > >
> > > > > > Mehmet, the kernel should have printed some diagnostics prior to doing
> > > > > > the BUG() call:
> > > > > >
> > > > > > if (dentry->d_count != 0) {
> > > > > > printk(KERN_ERR
> > > > > > "BUG: Dentry %p{i=%lx,n=%s}"
> > > > > > " still in use (%d)"
> > > > > > " [unmount of %s %s]\n",
> > > > > > dentry,
> > > > > > dentry->d_inode ?
> > > > > > dentry->d_inode->i_ino : 0UL,
> > > > > > dentry->d_name.name,
> > > > > > dentry->d_count,
> > > > > > dentry->d_sb->s_type->name,
> > > > > > dentry->d_sb->s_id);
> > > > > > BUG();
> > > > > > }
> > > > > >
> > > > > > Please find those in the log and email them to use - someone might find
> > > > > > it useful.
> > > > > >
> > > > > >
> > > > > > On Tue, 15 Mar 2011 21:02:23 GMT
> > > > > > bugzilla-daemon@bugzilla.kernel.org wrote:
> > > > > >
> > > > > > > https://bugzilla.kernel.org/show_bug.cgi?id=30882
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> > > > > > > Here is that crash happening again, the system was NOT running overclocked or
> > > > > > > anything...
> > > > > > >
> > > > > > > [ 1860.156122] ------------[ cut here ]------------
> > > > > > > [ 1860.156124] kernel BUG at fs/dcache.c:943!
> > > > > > > [ 1860.156126] invalid opcode: 0000 [#1] SMP
> > > > > > > [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> > > > > > > [ 1860.156128] CPU 3
> > > > > > > [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> > > > > > > xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> > > > > > > xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> > > > > > > nvidia(P)
> > > > > > > [ 1860.156137]
> > > > > > > [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> > > > > > > Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> > > > > > > [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> > > > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > > > [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> > > > > > > [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> > > > > > > 000000000003ffff
> > > > > > > [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> > > > > > > ffffffff8174c9f8
> > > > > > > [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> > > > > > > 0000000000000006
> > > > > > > [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> > > > > > > ffff88023a07f5e0
> > > > > > > [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> > > > > > > ffff880211d38740
> > > > > > > [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> > > > > > > knlGS:00000000f74186c0
> > > > > > > [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > > > > > > [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> > > > > > > 00000000000006e0
> > > > > > > [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> > > > > > > 0000000000000000
> > > > > > > [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> > > > > > > 0000000000000400
> > > > > > > [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> > > > > > > ffff880211fd5640)
> > > > > > > [ 1860.156162] Stack:
> > > > > > > [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> > > > > > > ffff88020c05cc00
> > > > > > > [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> > > > > > > ffffffff810e96a4
> > > > > > > [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> > > > > > > ffffffff810d5d15
> > > > > > > [ 1860.156169] Call Trace:
> > > > > > > [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> > > > > > > [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> > > > > > > [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> > > > > > > [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> > > > > > > [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> > > > > > > [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> > > > > > > [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> > > > > > > [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> > > > > > > [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> > > > > > > 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> > > > > > > eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> > > > > > > [ 1860.156201] RIP [<ffffffff810e9648>]
> > > > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > > > [ 1860.156204] RSP <ffff8800be82fe08>
> > > > > > > [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
> > > > > > >
> > > > > > > --
> > > > > > > Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> > > > > > > ------- You are receiving this mail because: -------
> > > > > > > You are on the CC list for the bug.
> > > > > >
> > > > >
> > > > >
> > > >
> > > >
> > >
> > >
> >
> >
>
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-16 15:44 ` Ian Kent
@ 2011-03-16 16:48 ` Mehmet Giritli
2011-03-16 17:47 ` Mehmet Giritli
1 sibling, 0 replies; 10+ messages in thread
From: Mehmet Giritli @ 2011-03-16 16:48 UTC (permalink / raw)
To: Ian Kent
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon, Mehmet Giritli
On Wed, 2011-03-16 at 23:44 +0800, Ian Kent wrote:
> On Wed, 2011-03-16 at 17:29 +0200, Mehmet Giritli wrote:
> > On Wed, 2011-03-16 at 23:21 +0800, Ian Kent wrote:
> > > On Wed, 2011-03-16 at 16:27 +0200, Mehmet Giritli wrote:
> > > > Ian,
> > > >
> > > > I am having much more frequent crashes now. I havent been able to
> > > > cleanly reboot my machine yet and I have tried three times so far. Init
> > > > scripts fail to unmount the file systems and I have to reboot manually
> > >
> > > What do your autofs maps look like?
> > >
> > >
> >
> > Here is the contents of my auto.misc:
> >
> > gollum-media -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/mnt/media
> > gollum-distfiles -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/usr/portage/distfiles
> > gollum-www -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/var/www
> > gollum-WebDav -rsize=8192,wsize=8192,soft,timeo=10,rw gollum.giritli.eu:/var/dav
>
> What, that's it, and your only using "/misc /etc/auto.misc" in the
> master map and your having problems.
yes
>
> Are the crashes always the same?
identical
> How have you established that the BUG()s are in fact due to automount
> umounting mounts and that the BUG()s correspond to NFS mounts previously
> mounted by autofs?
I havent established anything. However, thats the only way I mount nfs
and my file manager hangs, init scripts hang when trying to unmount...
> Is there any noise at all in the syslog?
nothing unusual
> Are you sure your using a kernel with the dentry leak patch?
yes
> What sort of automounting load is happening on the machine, ie.
> frequency or mounts and umounts and what timeout are you using?
from auto.master:
/mnt/autofs /etc/auto.misc --timeout=300 --ghost
Not very much. Lets say 2-3 times every hour for each mount point.
> The dentry leak patch got rid of the BUG()s I was seeing but by that
> time I did have a couple of other patches. I still don't think the other
> patches made much difference for this particular case.
>
> >
> > > >
> > > > On Wed, 2011-03-16 at 10:32 +0800, Ian Kent wrote:
> > > > > On Wed, 2011-03-16 at 01:54 +0200, Mehmet Giritli wrote:
> > > > > > The missing piece is as follows:
> > > > > >
> > > > > > Mar 15 22:37:38 mordor kernel: [ 1860.156114] BUG: Dentry
> > > > > > ffff88023f96e600{i=25f56f,n=} still in use (1) [unmount of nfs 0:f]
> > > > >
> > > > > This might be the same problem I saw and described in rc1.
> > > > > However, for me the fs in the BUG() report was autofs.
> > > > > Hopefully that just means my autofs setup is different.
> > > > >
> > > > > At the moment I believe a dentry leak Al Viro spotted is the cause.
> > > > > Please try this patch.
> > > > >
> > > > > autofs4 - fix dentry leak in autofs4_expire_direct()
> > > > >
> > > > > From: Ian Kent <raven@themaw.net>
> > > > >
> > > > > There is a missing dput() when returning from autofs4_expire_direct()
> > > > > when we see that the dentry is already a pending mount.
> > > > >
> > > > > Signed-off-by: Ian Kent <raven@themaw.net>
> > > > > ---
> > > > >
> > > > > fs/autofs4/expire.c | 7 +++----
> > > > > 1 files changed, 3 insertions(+), 4 deletions(-)
> > > > >
> > > > >
> > > > > diff --git a/fs/autofs4/expire.c b/fs/autofs4/expire.c
> > > > > index c896dd6..c403abc 100644
> > > > > --- a/fs/autofs4/expire.c
> > > > > +++ b/fs/autofs4/expire.c
> > > > > @@ -290,10 +290,8 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > > > > spin_lock(&sbi->fs_lock);
> > > > > ino = autofs4_dentry_ino(root);
> > > > > /* No point expiring a pending mount */
> > > > > - if (ino->flags & AUTOFS_INF_PENDING) {
> > > > > - spin_unlock(&sbi->fs_lock);
> > > > > - return NULL;
> > > > > - }
> > > > > + if (ino->flags & AUTOFS_INF_PENDING)
> > > > > + goto out;
> > > > > if (!autofs4_direct_busy(mnt, root, timeout, do_now)) {
> > > > > struct autofs_info *ino = autofs4_dentry_ino(root);
> > > > > ino->flags |= AUTOFS_INF_EXPIRING;
> > > > > @@ -301,6 +299,7 @@ struct dentry *autofs4_expire_direct(struct super_block *sb,
> > > > > spin_unlock(&sbi->fs_lock);
> > > > > return root;
> > > > > }
> > > > > +out:
> > > > > spin_unlock(&sbi->fs_lock);
> > > > > dput(root);
> > > > >
> > > > >
> > > > > >
> > > > > > (sorry for the inconvenience Andrew)
> > > > > >
> > > > > > On Tue, 2011-03-15 at 14:24 -0700, Andrew Morton wrote:
> > > > > > > (switched to email. Please respond via emailed reply-to-all, not via the
> > > > > > > bugzilla web interface).
> > > > > > >
> > > > > > > Seems that we have a nasty involving autofs, nfs and the VFS.
> > > > > > >
> > > > > > > Mehmet, the kernel should have printed some diagnostics prior to doing
> > > > > > > the BUG() call:
> > > > > > >
> > > > > > > if (dentry->d_count != 0) {
> > > > > > > printk(KERN_ERR
> > > > > > > "BUG: Dentry %p{i=%lx,n=%s}"
> > > > > > > " still in use (%d)"
> > > > > > > " [unmount of %s %s]\n",
> > > > > > > dentry,
> > > > > > > dentry->d_inode ?
> > > > > > > dentry->d_inode->i_ino : 0UL,
> > > > > > > dentry->d_name.name,
> > > > > > > dentry->d_count,
> > > > > > > dentry->d_sb->s_type->name,
> > > > > > > dentry->d_sb->s_id);
> > > > > > > BUG();
> > > > > > > }
> > > > > > >
> > > > > > > Please find those in the log and email them to use - someone might find
> > > > > > > it useful.
> > > > > > >
> > > > > > >
> > > > > > > On Tue, 15 Mar 2011 21:02:23 GMT
> > > > > > > bugzilla-daemon@bugzilla.kernel.org wrote:
> > > > > > >
> > > > > > > > https://bugzilla.kernel.org/show_bug.cgi?id=30882
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > > --- Comment #4 from Mehmet Giritli <mehmet@giritli.eu> 2011-03-15 21:02:22 ---
> > > > > > > > Here is that crash happening again, the system was NOT running overclocked or
> > > > > > > > anything...
> > > > > > > >
> > > > > > > > [ 1860.156122] ------------[ cut here ]------------
> > > > > > > > [ 1860.156124] kernel BUG at fs/dcache.c:943!
> > > > > > > > [ 1860.156126] invalid opcode: 0000 [#1] SMP
> > > > > > > > [ 1860.156127] last sysfs file: /sys/devices/platform/it87.552/fan3_input
> > > > > > > > [ 1860.156128] CPU 3
> > > > > > > > [ 1860.156129] Modules linked in: iptable_mangle iptable_nat nf_nat ipt_LOG
> > > > > > > > xt_tcpudp nf_conntrack_ipv4 nf_defrag_ipv4 xt_state xt_mac iptable_filter
> > > > > > > > xt_multiport xt_mark xt_conntrack xt_connmark nf_conntrack ip_tables x_tables
> > > > > > > > nvidia(P)
> > > > > > > > [ 1860.156137]
> > > > > > > > [ 1860.156139] Pid: 7388, comm: umount.nfs Tainted: P 2.6.38-rc8 #9
> > > > > > > > Gigabyte Technology Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
> > > > > > > > [ 1860.156142] RIP: 0010:[<ffffffff810e9648>] [<ffffffff810e9648>]
> > > > > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > > > > [ 1860.156147] RSP: 0018:ffff8800be82fe08 EFLAGS: 00010296
> > > > > > > > [ 1860.156149] RAX: 0000000000000065 RBX: ffff88023f96e600 RCX:
> > > > > > > > 000000000003ffff
> > > > > > > > [ 1860.156150] RDX: ffffffff8161f888 RSI: 0000000000000046 RDI:
> > > > > > > > ffffffff8174c9f8
> > > > > > > > [ 1860.156151] RBP: ffff88023f96e600 R08: 0000000000012c37 R09:
> > > > > > > > 0000000000000006
> > > > > > > > [ 1860.156152] R10: 0000000000000000 R11: 0000000000000000 R12:
> > > > > > > > ffff88023a07f5e0
> > > > > > > > [ 1860.156154] R13: ffff88023f96e65c R14: ffff8800be82ff18 R15:
> > > > > > > > ffff880211d38740
> > > > > > > > [ 1860.156155] FS: 00007f3428cb2700(0000) GS:ffff8800bfac0000(0000)
> > > > > > > > knlGS:00000000f74186c0
> > > > > > > > [ 1860.156156] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> > > > > > > > [ 1860.156157] CR2: 00007f7c97da1000 CR3: 00000000bea08000 CR4:
> > > > > > > > 00000000000006e0
> > > > > > > > [ 1860.156159] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> > > > > > > > 0000000000000000
> > > > > > > > [ 1860.156160] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
> > > > > > > > 0000000000000400
> > > > > > > > [ 1860.156161] Process umount.nfs (pid: 7388, threadinfo ffff8800be82e000, task
> > > > > > > > ffff880211fd5640)
> > > > > > > > [ 1860.156162] Stack:
> > > > > > > > [ 1860.156163] ffff88020c05ce50 0000000000000000 ffff88023fc07128
> > > > > > > > ffff88020c05cc00
> > > > > > > > [ 1860.156165] ffff88023f96e6c0 ffff8800be82ff28 ffff88023f96e300
> > > > > > > > ffffffff810e96a4
> > > > > > > > [ 1860.156167] ffff88023f49f480 ffff88020c05cc00 ffffffff8146d4a0
> > > > > > > > ffffffff810d5d15
> > > > > > > > [ 1860.156169] Call Trace:
> > > > > > > > [ 1860.156172] [<ffffffff810e96a4>] ? shrink_dcache_for_umount+0x54/0x60
> > > > > > > > [ 1860.156174] [<ffffffff810d5d15>] ? generic_shutdown_super+0x25/0x100
> > > > > > > > [ 1860.156176] [<ffffffff810d5e79>] ? kill_anon_super+0x9/0x40
> > > > > > > > [ 1860.156179] [<ffffffff81179aed>] ? nfs_kill_super+0xd/0x20
> > > > > > > > [ 1860.156181] [<ffffffff810d5f13>] ? deactivate_locked_super+0x43/0x70
> > > > > > > > [ 1860.156183] [<ffffffff810ef4d8>] ? release_mounts+0x68/0x90
> > > > > > > > [ 1860.156185] [<ffffffff810efa54>] ? sys_umount+0x314/0x3d0
> > > > > > > > [ 1860.156187] [<ffffffff8100243b>] ? system_call_fastpath+0x16/0x1b
> > > > > > > > [ 1860.156188] Code: 8b 0a 31 d2 48 85 f6 74 07 48 8b 96 a8 00 00 00 48 05 50
> > > > > > > > 02 00 00 48 89 de 48 c7 c7 40 3a 52 81 48 89 04 24 31 c0 e8 a1 bc 35 00 <0f> 0b
> > > > > > > > eb fe 0f 0b eb fe 55 53 48 89 fb 48 8d 7f 68 48 83 ec 08
> > > > > > > > [ 1860.156201] RIP [<ffffffff810e9648>]
> > > > > > > > shrink_dcache_for_umount_subtree+0x268/0x270
> > > > > > > > [ 1860.156204] RSP <ffff8800be82fe08>
> > > > > > > > [ 1860.156205] ---[ end trace ee03486c16c108a7 ]---
> > > > > > > >
> > > > > > > > --
> > > > > > > > Configure bugmail: https://bugzilla.kernel.org/userprefs.cgi?tab=email
> > > > > > > > ------- You are receiving this mail because: -------
> > > > > > > > You are on the CC list for the bug.
> > > > > > >
> > > > > >
> > > > > >
> > > > >
> > > > >
> > > >
> > > >
> > >
> > >
> >
> >
> > --
> > To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in
> > the body of a message to majordomo@vger.kernel.org
> > More majordomo info at http://vger.kernel.org/majordomo-info.html
>
>
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-16 15:44 ` Ian Kent
2011-03-16 16:48 ` Mehmet Giritli
@ 2011-03-16 17:47 ` Mehmet Giritli
2011-03-17 2:36 ` Ian Kent
1 sibling, 1 reply; 10+ messages in thread
From: Mehmet Giritli @ 2011-03-16 17:47 UTC (permalink / raw)
To: Ian Kent
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon
Ian,
Well, here is a copy-paste of the crash that I am seeing *everytime at
shutdown* now that I am running with your patch. It is pretty much
identical to the older ones..except that I see this at every time,
instead of just rare crashes...
Now I am going to revert your patch and check for changes in the
frequency of the crashes...
Mar 16 18:49:27 mordor kernel: [ 5696.670114] BUG: Dentry
ffff88015007f300{i=2,n=donkey} still in use (1) [unmount of nfs 0:f]
Mar 16 18:49:27 mordor kernel: [ 5696.670134] ------------[ cut
here ]------------
Mar 16 18:49:27 mordor kernel: [ 5696.670187] kernel BUG at
fs/dcache.c:943!
Mar 16 18:49:27 mordor kernel: [ 5696.670237] invalid opcode: 0000 [#1]
SMP
Mar 16 18:49:27 mordor kernel: [ 5696.670369] last sysfs
file: /sys/devices/platform/it87.552/pwm1_enable
Mar 16 18:49:27 mordor kernel: [ 5696.670421] CPU 2
Mar 16 18:49:27 mordor kernel: [ 5696.670466] Modules linked in: ipt_LOG
xt_tcpudp xt_state xt_mac iptable_filter iptable_nat nf_nat
nf_conntrack_ipv4 nf_defrag_ipv4 iptable_mangle xt_multiport xt_mark
xt_conntrack xt_connmark nf_conntr$
Mar 16 18:49:27 mordor kernel: [ 5696.671003]
Mar 16 18:49:27 mordor kernel: [ 5696.671003] Pid: 21015, comm:
umount.nfs Tainted: P 2.6.38-gentoo #3 Gigabyte Technology
Co., Ltd. GA-790FXTA-UD5/GA-790FXTA-UD5
Mar 16 18:49:27 mordor kernel: [ 5696.671003] RIP:
0010:[<ffffffff810e95c8>] [<ffffffff810e95c8>]
shrink_dcache_for_umount_subtree+0x268/0x270
Mar 16 18:49:27 mordor kernel: [ 5696.671003] RSP: 0018:ffff880219903e08
EFLAGS: 00010296
Mar 16 18:49:27 mordor kernel: [ 5696.671003] RAX: 0000000000000066 RBX:
ffff88015007f300 RCX: 000000000003ffff
Mar 16 18:49:27 mordor kernel: [ 5696.671003] RDX: ffffffff81623888 RSI:
0000000000000046 RDI: ffffffff817509f8
Mar 16 18:49:27 mordor kernel: [ 5696.671003] RBP: ffff88023a1480c0 R08:
000000000001fa7c R09: 0000000000000006
Mar 16 18:49:27 mordor kernel: [ 5696.671003] R10: 0000000000000000 R11:
0000000000000000 R12: ffff88015007f3a0
Mar 16 18:49:27 mordor kernel: [ 5696.671003] R13: ffff88023a14811c R14:
ffff880219903f18 R15: ffff8802107d9640
Mar 16 18:49:27 mordor kernel: [ 5696.671003] FS:
00007ff9ec925700(0000) GS:ffff8800bfa80000(0000) knlGS:0000000000000000
Mar 16 18:49:27 mordor kernel: [ 5696.671003] CS: 0010 DS: 0000 ES:
0000 CR0: 000000008005003b
Mar 16 18:49:27 mordor kernel: [ 5696.671003] CR2: 00007fd326cc2bc8 CR3:
0000000232272000 CR4: 00000000000006e0
Mar 16 18:49:27 mordor kernel: [ 5696.671003] DR0: 0000000000000000 DR1:
0000000000000000 DR2: 0000000000000000
Mar 16 18:49:27 mordor kernel: [ 5696.671003] DR3: 0000000000000000 DR6:
00000000ffff0ff0 DR7: 0000000000000400
Mar 16 18:49:27 mordor kernel: [ 5696.671003] Process umount.nfs (pid:
21015, threadinfo ffff880219902000, task ffff88023ca16780)
Mar 16 18:49:27 mordor kernel: [ 5696.671003] Stack:
Mar 16 18:49:27 mordor kernel: [ 5696.671003] ffff880219af1650
0000000000000000 ffff88023fc07128 ffff880219af1400
Mar 16 18:49:27 mordor kernel: [ 5696.671003] ffff88023a1486c0
ffff880219903f28 ffff88023a03a180 ffffffff810e9624
Mar 16 18:49:27 mordor kernel: [ 5696.671003] ffff88023f4f2480
ffff880219af1400 ffffffff81471460 ffffffff810d5ce5
Mar 16 18:49:27 mordor kernel: [ 5696.671003] Call Trace:
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff810e9624>] ?
shrink_dcache_for_umount+0x54/0x60
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff810d5ce5>] ?
generic_shutdown_super+0x25/0x100
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff810d5e49>] ?
kill_anon_super+0x9/0x40
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff81179acd>] ?
nfs_kill_super+0xd/0x20
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff810d5ee3>] ?
deactivate_locked_super+0x43/0x70
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff810ef4d0>] ?
release_mounts+0x70/0x90
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff810efa44>] ?
sys_umount+0x314/0x3d0
Mar 16 18:49:27 mordor kernel: [ 5696.671003] [<ffffffff8100243b>] ?
system_call_fastpath+0x16/0x1b
Mar 16 18:49:27 mordor kernel: [ 5696.671003] Code: 8b 0a 31 d2 48 85 f6
74 07 48 8b 96 a8 00 00 00 48 05 50 02 00 00 48 89 de 48 c7 c7 e0 7f 52
81 48 89 04 24 31 c0 e8 f1 f0 35 00 <0f> 0b eb fe 0f 0b eb fe 55 53 48
89 fb 48 8d 7f 68 48$
Mar 16 18:49:27 mordor kernel: [ 5696.678048] RIP [<ffffffff810e95c8>]
shrink_dcache_for_umount_subtree+0x268/0x270
Mar 16 18:49:27 mordor kernel: [ 5696.678048] RSP <ffff880219903e08>
Mar 16 18:49:27 mordor kernel: [ 5696.678552] ---[ end trace
24269d237584cd43 ]---
Mar 16 18:49:28 mordor mountd[4362]: Caught signal 15, un-registering
and exiting.
Mar 16 18:49:28 mordor kernel: [ 5697.256066] nfsd: last server has
exited, flushing export cache
Mar 16 18:49:28 mordor kernel: [ 5697.272903] nfsd: last server has
exited, flushing export cache
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [Bug 30882] Automatic process group scheduling causes crashes after a while
2011-03-16 17:47 ` Mehmet Giritli
@ 2011-03-17 2:36 ` Ian Kent
0 siblings, 0 replies; 10+ messages in thread
From: Ian Kent @ 2011-03-17 2:36 UTC (permalink / raw)
To: mgiritli
Cc: Andrew Morton, linux-fsdevel, Nick Piggin, Al Viro,
bugzilla-daemon
On Wed, 2011-03-16 at 19:47 +0200, Mehmet Giritli wrote:
> Well, here is a copy-paste of the crash that I am seeing *everytime at
> shutdown* now that I am running with your patch. It is pretty much
> identical to the older ones..except that I see this at every time,
> instead of just rare crashes...
That's odd because, with the maps your using, the patch changes code
that's never executed.
^ permalink raw reply [flat|nested] 10+ messages in thread
end of thread, other threads:[~2011-03-17 2:36 UTC | newest]
Thread overview: 10+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
[not found] <bug-30882-27@https.bugzilla.kernel.org/>
[not found] ` <201103152102.p2FL2NN8006070@demeter2.kernel.org>
2011-03-15 21:24 ` [Bug 30882] Automatic process group scheduling causes crashes after a while Andrew Morton
2011-03-15 23:54 ` Mehmet Giritli
2011-03-16 2:32 ` Ian Kent
2011-03-16 14:27 ` Mehmet Giritli
2011-03-16 15:21 ` Ian Kent
2011-03-16 15:29 ` Mehmet Giritli
2011-03-16 15:44 ` Ian Kent
2011-03-16 16:48 ` Mehmet Giritli
2011-03-16 17:47 ` Mehmet Giritli
2011-03-17 2:36 ` Ian Kent
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).