From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from cuda.sgi.com (cuda1.sgi.com [192.48.157.11]) by oss.sgi.com (8.14.3/8.14.3/SuSE Linux 0.8) with ESMTP id oB7MOFv7234234 for ; Tue, 7 Dec 2010 16:24:15 -0600 Received: from mail.internode.on.net (localhost [127.0.0.1]) by cuda.sgi.com (Spam Firewall) with ESMTP id 0DBC914262B4 for ; Tue, 7 Dec 2010 14:26:01 -0800 (PST) Received: from mail.internode.on.net (bld-mail12.adl6.internode.on.net [150.101.137.97]) by cuda.sgi.com with ESMTP id M7tIUP3CZbkVnxYi for ; Tue, 07 Dec 2010 14:26:01 -0800 (PST) Date: Wed, 8 Dec 2010 09:25:58 +1100 From: Dave Chinner Subject: Re: kernel panic-xfs errors Message-ID: <20101207222558.GC29333@dastard> References: <30397503.post@talk.nabble.com> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: <30397503.post@talk.nabble.com> List-Id: XFS Filesystem from SGI List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Sender: xfs-bounces@oss.sgi.com Errors-To: xfs-bounces@oss.sgi.com To: blacknred Cc: xfs@oss.sgi.com On Tue, Dec 07, 2010 at 07:42:56AM -0800, blacknred wrote: > > Hi..... > > I get a kernel panic on my HP Proliant Server. > > here's trace: > > BUG: unable to handle kernel NULL pointer dereference at virtual address > 00000052 > printing eip: > *pde = 2c731001 > Oops: 0000 [#1] > SMP > > CPU: 2 > EIP: 0060:[] Tainted: GF VLI ^^^^^^^^^^^ You've done a forced module load. No guarantee your kernel is in any sane shape if you've done that.... > EFLAGS: 00010272 (2.6.33.3-85.fc13.x86_64 #1) > EIP is at do_page_fault+0x245/0x617 > eax: ec5ee000 ebx: 00000000 ecx: eb5de084 edx: 0000000e > esi: 00013103 edi: ec5de0b3 ebp: 00000023 esp: ec5de024 > ds: 008b es: 008b ss: 0078 > Process bm (pid: 3210, ti=ec622000 task=ec5e3450 task.ti=ec6ee000) > Stack: 00000000 00000000 ecd5e0a4 00000024 00000093 f7370000 00000007 > 00000000 > ed6ef0a4 c0639569 00000000 0000000f 0000000b 00000000 00000000 > 00000000 > 00015106 c0629b9d 00000014 c0305b83 00000000 ec3d40f7 0000000e > 00013006 > Call Trace: > [] do_page_fault+0x0/0x607 > [] error_code+0x49/0x50 > [] do_page_fault+0x204/00x607 > [] elv_next_request+0x137/0x234 > [] do_cciss_request+0x397/0x3a3 [cciss] > [] do_page_fault+0x0/0x607 > [] error_code+0x49/0x40 > [] do_page_fault+0x215/0x607 > [] deadline_set_request+0x26/0x57 > [] do_page_fault+0x0/0x607 > [] error_code+0x39/0x40 > [] __down+0x2b/0xbb > [] default_wake_function+0x0/0xc > [] __down_failed+0x7/0xc > [] .text.lock.xfs_buf+0x17/0x5f [xfs] > [] xfs_buf_read_flags+0x48/0x76 [xfs] > [] xfs_trans_read_buf+0x1bb/0x2c0 [xfs] > [] xfs_btree_read_bufl+0x96/0xb3 [xfs] > [] xfs_bmbt_lookup+0x135/0x478 [xfs] > [] xfs_bmap_add_extent+0xd2b/0x1e30 [xfs] > [] xfs_alloc_update+0x3a/0xbc [xfs] > [] xfs_alloc_fixup_trees+0x217/0x29a [xfs] > [] xfs_trans_log_buf+0x49/0x6c [xfs] > [] xfs_alloc_search_busy+0x20/0xae [xfs] > [] xfs_iext_bno_to_ext+0xd8/0x191 [xfs] > [] kmem_zone_zalloc+0x1d/0x41 [xfs] > [] xfs_bmapi+0x15fe/0x2016 [xfs] > [] xfs_iext_bno_to_ext+0x48/0x191 [xfs] > [] xfs_bmap_search_multi_extents+0x8a/0xc5 [xfs] > [] xfs_iomap_write_allocate+0x29c/0x469 [xfs] > [] lock_timer_base+0x15/0x2f > [] del_timer+0x41/0x47 > [] xfs_iomap+0x409/0x71d [xfs] > [] xfs_map_blocks+0x29/0x52 [xfs] > [] xfs_page_state_convert+0x37b/0xd2e [xfs] > [] xfs_bmap_add_extent+0x1dcf/0x1e30 [xfs] > [] xfs_bmap_search_multi_extents+0x8a/0xc5 [xfs] > [] xfs_bmapi+0x272/0x2017 [xfs] > [] xfs_bmapi+0x1853/0x2017 [xfs] > [] find_get_pages_tag+0x40/0x75 > [] xfs_vm_writepage+0x8f/0xd2 [xfs] > [] mpage_writepages+0x1b7/0x310 > [] xfs_vm_writepage+0x0/0xc4 [xfs] > [] do_writepages+0x20/0x42 > [] __writeback_single_inode+0x180/0x2af > [] write_inode_now+0x67/0xa7 > [] file_fsync+0xf/0x6c > [] moddw_ioctl+0x420/0x679 [mod_dw] > [] __cond_resched+0x16/0x54 > [] do_ioctl+0x47/0x5d > [] vfs_ioctl+0x47b/0x4d3 > [] sys_ioctl+0x48/0x4f > [] sysenter_past_esp+0x46/0x79 Strange failure. Hmmm - i386 arch and fedora - are you running with 4k stacks? If so, maybe it blew the stack... > > dmesg shows: > XFS: bad magic number > XFS: SB validate failed > > I rebooted the server, now xfs_repair comes clean. > > But the server has hung again after an hour. No panic this time, checked > dmesg output and it again > shows same > XFS: bad magic number > XFS: SB validate failed > messages.. Any thoughts?? What does this give you before and after the failure: # dd if= bs=512 count=1 | od -c Cheers, Dave. -- Dave Chinner david@fromorbit.com _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs