From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752819Ab1AJJFn (ORCPT ); Mon, 10 Jan 2011 04:05:43 -0500 Received: from mx1.redhat.com ([209.132.183.28]:52177 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751881Ab1AJJFl (ORCPT ); Mon, 10 Jan 2011 04:05:41 -0500 Subject: Re: [Cluster-devel] 2.6.37 - GFS2 trouble From: Steven Whitehouse To: Nikola Ciprich Cc: cluster-devel@redhat.com, Linux kernel list , nikola.ciprich@linuxbox.cz In-Reply-To: <20110109210642.GG2400@nik-comp.lan> References: <20110109210642.GG2400@nik-comp.lan> Content-Type: text/plain; charset="UTF-8" Organization: Red Hat UK Ltd Date: Mon, 10 Jan 2011 09:06:38 +0000 Message-ID: <1294650398.2450.3.camel@dolmen> Mime-Version: 1.0 Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, On Sun, 2011-01-09 at 22:06 +0100, Nikola Ciprich wrote: > Hello, > I wanted to try 2.6.37 on my cluster, but all tasks trying to access GFS2-mounted > partition got stuck. Kernel started spitting following messages: > > Jan 5 22:16:46 vbox1 [ 3001.948125] INFO: task gfs2_quotad:12532 blocked for more than 120 seconds. > Jan 5 22:16:46 vbox1 [ 3001.948137] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. > Jan 5 22:16:46 vbox1 [ 3001.948141] gfs2_quotad D ffffffff8140a4c0 0 12532 2 0x00000080 > Jan 5 22:16:46 vbox1 [ 3001.948149] ffff88040465fc30 0000000000000046 ffff88040465fb20 00000000000116c0 > Jan 5 22:16:46 vbox1 [ 3001.948156] ffff8804054b48d8 0000000000000007 ffff8804054b4530 ffff88042fd45c40 > Jan 5 22:16:46 vbox1 [ 3001.948163] ffff88040465ffd8 0000000000000000 000000000465fb50 ffffffff8136cce3 > Jan 5 22:16:46 vbox1 [ 3001.948170] Call Trace: > Jan 5 22:16:46 vbox1 [ 3001.948184] [] ? _raw_spin_unlock+0x13/0x40 > Jan 5 22:16:46 vbox1 [ 3001.948199] [] ? dlm_put_lockspace+0x28/0x30 [dlm] > Jan 5 22:16:46 vbox1 [ 3001.948208] [] ? dlm_lock+0x86/0x180 [dlm] > Jan 5 22:16:46 vbox1 [ 3001.948228] [] ? gdlm_bast+0x0/0x50 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948233] [] ? update_curr+0xb2/0x170 > Jan 5 22:16:46 vbox1 [ 3001.948239] [] ? get_parent_ip+0x11/0x50 > Jan 5 22:16:46 vbox1 [ 3001.948243] [] ? sub_preempt_count+0x9d/0xd0 > Jan 5 22:16:46 vbox1 [ 3001.948249] [] ? _raw_spin_unlock_irqrestore+0x1d/0x50 > Jan 5 22:16:46 vbox1 [ 3001.948260] [] gfs2_glock_holder_wait+0x9/0x10 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948265] [] __wait_on_bit+0x55/0x80 > Jan 5 22:16:46 vbox1 [ 3001.948275] [] ? gfs2_glock_holder_wait+0x0/0x10 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948285] [] ? gfs2_glock_holder_wait+0x0/0x10 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948291] [] out_of_line_wait_on_bit+0x78/0x90 > Jan 5 22:16:46 vbox1 [ 3001.948296] [] ? wake_bit_function+0x0/0x30 > Jan 5 22:16:46 vbox1 [ 3001.948301] [] ? get_parent_ip+0x11/0x50 > Jan 5 22:16:46 vbox1 [ 3001.948311] [] gfs2_glock_wait+0x42/0x50 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948323] [] gfs2_glock_nq+0x28d/0x3a0 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948336] [] gfs2_statfs_sync+0x59/0x1a0 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948350] [] ? gfs2_statfs_sync+0x51/0x1a0 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948354] [] ? sub_preempt_count+0x9d/0xd0 > Jan 5 22:16:46 vbox1 [ 3001.948368] [] quotad_check_timeo+0x57/0x90 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948381] [] gfs2_quotad+0x207/0x240 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948386] [] ? autoremove_wake_function+0x0/0x40 > Jan 5 22:16:46 vbox1 [ 3001.948391] [] ? _raw_spin_unlock_irqrestore+0x1d/0x50 > Jan 5 22:16:46 vbox1 [ 3001.948404] [] ? gfs2_quotad+0x0/0x240 [gfs2] > Jan 5 22:16:46 vbox1 [ 3001.948409] [] kthread+0x96/0xa0 > Jan 5 22:16:46 vbox1 [ 3001.948416] [] kernel_thread_helper+0x4/0x10 > Jan 5 22:16:46 vbox1 [ 3001.948420] [] ? kthread+0x0/0xa0 > Jan 5 22:16:46 vbox1 [ 3001.948425] [] ? kernel_thread_helper+0x0/0x10 > > Could somebody with insight have a look at this? Is it some known problem? > If I could somehow help to debug, I'll be happy to do so. > cheers > nik > > Quotad is often the first victim of any slow down or problem on GFS2 since it runs periodically (even if you are not using quotas, it also looks after statfs too). So I can't tell directly from that information what is wrong. I'd suggest taking glock dumps (via debugfs) as a first step. Grabbing backtraces of the glock_workqueue and other processes using gfs2 is probably the next step, and if that fails to point the way, the gfs2 tracepoints are the next thing to look at, Steve.