public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
* xfs_repair segfault
@ 2013-10-01 19:57 Viet Nguyen
  2013-10-01 20:19 ` Dave Chinner
  0 siblings, 1 reply; 23+ messages in thread
From: Viet Nguyen @ 2013-10-01 19:57 UTC (permalink / raw)
  To: xfs


[-- Attachment #1.1: Type: text/plain, Size: 1466 bytes --]

Hi,

I have a corrupted xfs partition that segfaults when I run xfs_repair, at
the same place every time.

I'm using the latest version of xfs_repair that I am aware of: xfs_repair
version 3.2.0-alpha1

I simply run it as so: xfs_repair -P /dev/sda1

Here's a sample of the last few lines that are spit out:
correcting nextents for inode 8637985
correcting nblocks for inode 8637985, was 198 - counted 0
correcting nextents for inode 8637985, was 1 - counted 0
data fork in regular inode 8637987 claims used block 7847452695
correcting nextents for inode 8637987
correcting nblocks for inode 8637987, was 198 - counted 0
correcting nextents for inode 8637987, was 1 - counted 0
data fork in regular inode 8637999 claims used block 11068974204
correcting nextents for inode 8637999
correcting nblocks for inode 8637999, was 200 - counted 0
correcting nextents for inode 8637999, was 1 - counted 0
data fork in regular inode 8638002 claims used block 11873152787
correcting nextents for inode 8638002
correcting nblocks for inode 8638002, was 201 - counted 0
correcting nextents for inode 8638002, was 1 - counted 0
imap claims a free inode 8638005 is in use, correcting imap and clearing
inode
cleared inode 8638005
imap claims a free inode 8638011 is in use, correcting imap and clearing
inode
cleared inode 8638011
Segmentation fault (core dumped)

It crashes after attempting to clear that same inode every time.

Any advice you can give me on this?

Thanks,
Viet

[-- Attachment #1.2: Type: text/html, Size: 1899 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 23+ messages in thread
* xfs_repair segfault
@ 2015-03-09 15:50 Rui Gomes
  2015-03-09 15:55 ` Carsten Aulbert
  2015-03-09 16:14 ` Eric Sandeen
  0 siblings, 2 replies; 23+ messages in thread
From: Rui Gomes @ 2015-03-09 15:50 UTC (permalink / raw)
  To: xfs; +Cc: Ómar Hermannsson


[-- Attachment #1.1: Type: text/plain, Size: 698 bytes --]

Hello Guys, 

I have a XFS filesystem that went rough, the file system got corrupted without any hardware failure or power outage what is strange enough. 
But after trying to run xfs_repair it segmented fault, this system was originally CentOs6.5, we upgraded to Centos7 using xfsprogs-3.2.0-0.10.alpha2.el7.x86_64, and run the newer version of the xfs_repair, same result segmentation fault. 


Full output and GDB Backtrace in the attachment, do you guys have any advice how can we get xfs_repair to do a clean run? 



Regards 

------------------------------- 
Rui Gomes 
CTO 


RVX - Reykjavik Visual Effects 
Seljavegur 2, 
101 Reykjavik 
Iceland 


Tel: + 354 527 3330 
Mob: + 354 663 3360 

[-- Attachment #1.2.1: Type: text/html, Size: 7086 bytes --]

[-- Attachment #1.2.2: rvx-logo.png --]
[-- Type: image/png, Size: 2226 bytes --]

[-- Attachment #2: gdb.txt --]
[-- Type: text/plain, Size: 6507 bytes --]

Starting program: /usr/sbin/xfs_repair -n -P -m 500000000000000 /dev/sdb1
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib64/libthread_db.so.1".
[New Thread 0x7fffb0ec9700 (LWP 16151)]
[New Thread 0x7fffb06c8700 (LWP 16152)]
[New Thread 0x7fffafec7700 (LWP 16153)]
[New Thread 0x7fffaf6c6700 (LWP 16154)]
[New Thread 0x7fffaeec5700 (LWP 16155)]
[New Thread 0x7fffae6c4700 (LWP 16156)]
[New Thread 0x7fffadec3700 (LWP 16157)]
[New Thread 0x7fffad6c2700 (LWP 16158)]
[New Thread 0x7fffacec1700 (LWP 16159)]
[New Thread 0x7fffac6c0700 (LWP 16160)]
[New Thread 0x7fffabebf700 (LWP 16161)]
[New Thread 0x7fffab6be700 (LWP 16162)]
[New Thread 0x7fffaaebd700 (LWP 16163)]
[New Thread 0x7fffaa6bc700 (LWP 16164)]
[New Thread 0x7fffa9ebb700 (LWP 16165)]
[New Thread 0x7fffa96ba700 (LWP 16166)]
[New Thread 0x7fffa8eb9700 (LWP 16167)]
[New Thread 0x7fffa86b8700 (LWP 16168)]
[New Thread 0x7fffa7eb7700 (LWP 16169)]
[New Thread 0x7fffa76b6700 (LWP 16170)]
[New Thread 0x7fffa6eb5700 (LWP 16171)]
[New Thread 0x7fffa66b4700 (LWP 16172)]
[New Thread 0x7fffa5eb3700 (LWP 16173)]
[New Thread 0x7fffa56b2700 (LWP 16174)]
[New Thread 0x7fffa4eb1700 (LWP 16175)]
[New Thread 0x7fffa46b0700 (LWP 16176)]
[New Thread 0x7fffa3eaf700 (LWP 16177)]
[New Thread 0x7fffa36ae700 (LWP 16178)]
[New Thread 0x7fffa2ead700 (LWP 16179)]
[New Thread 0x7fffa26ac700 (LWP 16180)]
[New Thread 0x7fffa1eab700 (LWP 16181)]
[New Thread 0x7fffa16aa700 (LWP 16182)]
[Thread 0x7fffac6c0700 (LWP 16160) exited]
[Thread 0x7fffa86b8700 (LWP 16168) exited]
[Thread 0x7fffacec1700 (LWP 16159) exited]
[Thread 0x7fffa76b6700 (LWP 16170) exited]
[Thread 0x7fffa56b2700 (LWP 16174) exited]
[Thread 0x7fffa6eb5700 (LWP 16171) exited]
[Thread 0x7fffa96ba700 (LWP 16166) exited]
[Thread 0x7fffa9ebb700 (LWP 16165) exited]
[Thread 0x7fffaa6bc700 (LWP 16164) exited]
[Thread 0x7fffabebf700 (LWP 16161) exited]
[Thread 0x7fffab6be700 (LWP 16162) exited]
[Thread 0x7fffa4eb1700 (LWP 16175) exited]
[Thread 0x7fffa16aa700 (LWP 16182) exited]
[Thread 0x7fffa8eb9700 (LWP 16167) exited]
[Thread 0x7fffa5eb3700 (LWP 16173) exited]
[Thread 0x7fffa2ead700 (LWP 16179) exited]
[Thread 0x7fffae6c4700 (LWP 16156) exited]
[Thread 0x7fffadec3700 (LWP 16157) exited]
[Thread 0x7fffa7eb7700 (LWP 16169) exited]
[Thread 0x7fffaeec5700 (LWP 16155) exited]
[Thread 0x7fffad6c2700 (LWP 16158) exited]
[Thread 0x7fffa1eab700 (LWP 16181) exited]
[Thread 0x7fffb0ec9700 (LWP 16151) exited]
[Thread 0x7fffafec7700 (LWP 16153) exited]
[Thread 0x7fffa26ac700 (LWP 16180) exited]
[Thread 0x7fffb06c8700 (LWP 16152) exited]
[Thread 0x7fffaaebd700 (LWP 16163) exited]
[Thread 0x7fffa3eaf700 (LWP 16177) exited]
[Thread 0x7fffa66b4700 (LWP 16172) exited]
[Thread 0x7fffaf6c6700 (LWP 16154) exited]
[Thread 0x7fffa46b0700 (LWP 16176) exited]
[Thread 0x7fffa36ae700 (LWP 16178) exited]

Program received signal SIGABRT, Aborted.
0x00007ffff74275c9 in __GI_raise (sig=sig@entry=6) at ../nptl/sysdeps/unix/sysv/linux/raise.c:56
56	  return INLINE_SYSCALL (tgkill, 3, pid, selftid, sig);
#0  0x00007ffff74275c9 in __GI_raise (sig=sig@entry=6) at ../nptl/sysdeps/unix/sysv/linux/raise.c:56
#1  0x00007ffff7428cd8 in __GI_abort () at abort.c:90
#2  0x00007ffff7467db7 in __libc_message (do_abort=do_abort@entry=2, fmt=fmt@entry=0x7ffff756f561 "*** %s ***: %s terminated\n") at ../sysdeps/unix/sysv/linux/libc_fatal.c:196
#3  0x00007ffff74ff9c7 in __GI___fortify_fail (msg=msg@entry=0x7ffff756f507 "buffer overflow detected") at fortify_fail.c:31
#4  0x00007ffff74fdb90 in __GI___chk_fail () at chk_fail.c:28
#5  0x0000000000414ea8 in memmove (__len=18446744073709551615, __src=0x1e562094, __dest=0x7fffffffd8f0) at /usr/include/bits/string3.h:57
#6  process_sf_dir2 (dirname=0x46b0e2 "", repair=<synthetic pointer>, parent=0x7fffffffdc20, dino_dirty=0x7fffffffdc18, ino_discovery=1, dip=0x1e562000, ino=260256256, mp=0x1e562091) at dir2.c:992
#7  process_dir2 (mp=mp@entry=0x7fffffffe020, ino=ino@entry=260256256, dip=dip@entry=0x1e562000, ino_discovery=ino_discovery@entry=1, dino_dirty=dino_dirty@entry=0x7fffffffdc18, dirname=dirname@entry=0x46b0e2 "", 
    parent=parent@entry=0x7fffffffdc20, blkmap=0x0) at dir2.c:1988
#8  0x000000000041189f in process_dinode_int (mp=mp@entry=0x7fffffffe020, dino=dino@entry=0x1e562000, agno=agno@entry=0, ino=ino@entry=260256256, was_free=<optimized out>, dirty=dirty@entry=0x7fffffffdc18, 
    used=used@entry=0x7fffffffdc14, verify_mode=verify_mode@entry=0, uncertain=uncertain@entry=0, ino_discovery=ino_discovery@entry=1, check_dups=check_dups@entry=0, extra_attr_check=extra_attr_check@entry=1, 
    isa_dir=isa_dir@entry=0x7fffffffdc1c, parent=parent@entry=0x7fffffffdc20) at dinode.c:2881
#9  0x00000000004124ce in process_dinode (mp=mp@entry=0x7fffffffe020, dino=dino@entry=0x1e562000, agno=agno@entry=0, ino=ino@entry=260256256, was_free=<optimized out>, dirty=dirty@entry=0x7fffffffdc18, used=used@entry=0x7fffffffdc14, 
    ino_discovery=ino_discovery@entry=1, check_dups=check_dups@entry=0, extra_attr_check=extra_attr_check@entry=1, isa_dir=isa_dir@entry=0x7fffffffdc1c, parent=parent@entry=0x7fffffffdc20) at dinode.c:2989
#10 0x000000000040b96f in process_inode_chunk (mp=mp@entry=0x7fffffffe020, agno=agno@entry=0, first_irec=first_irec@entry=0x7fff9c55b580, ino_discovery=ino_discovery@entry=1, check_dups=check_dups@entry=0, 
    extra_attr_check=extra_attr_check@entry=1, bogus=bogus@entry=0x7fffffffdca4, num_inos=<optimized out>) at dino_chunks.c:772
#11 0x000000000040cddd in process_aginodes (mp=0x7fffffffe020, pf_args=pf_args@entry=0x0, agno=agno@entry=0, ino_discovery=ino_discovery@entry=1, check_dups=check_dups@entry=0, extra_attr_check=extra_attr_check@entry=1)
    at dino_chunks.c:1025
#12 0x000000000041964e in process_ag_func (wq=0x7fffffffdd90, agno=0, arg=0x0) at phase3.c:77
#13 0x00000000004265da in prefetch_ag_range (work=0x7fffffffdd90, start_ag=<optimized out>, end_ag=32, dirs_only=false, func=0x419600 <process_ag_func>) at prefetch.c:907
#14 0x000000000042666c in do_inode_prefetch (mp=mp@entry=0x7fffffffe020, stride=0, func=func@entry=0x419600 <process_ag_func>, check_cache=check_cache@entry=false, dirs_only=dirs_only@entry=false) at prefetch.c:970
#15 0x000000000041975d in process_ags (mp=0x7fffffffe020) at phase3.c:85
#16 phase3 (mp=mp@entry=0x7fffffffe020) at phase3.c:121
#17 0x000000000040388e in main (argc=<optimized out>, argv=<optimized out>) at xfs_repair.c:785
A debugging session is active.

	Inferior 1 [process 16147] will be killed.

Quit anyway? (y or n) 

[-- Attachment #3: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 23+ messages in thread
* xfs_repair segfault
@ 2007-04-03 19:11 James W. Abendschan
  2007-04-04  0:45 ` Barry Naujok
  0 siblings, 1 reply; 23+ messages in thread
From: James W. Abendschan @ 2007-04-03 19:11 UTC (permalink / raw)
  To: xfs

Hi there -- I have a 6.9TB XFS volume that is acting up
after a power failure (I understand XFS + no UPS + PC
hardware == badness.  Not my decision.)

The machine is a dual proc x86 (intel xeon 5130) w/ 8GB RAM
running a custom 2.6.18 kernel on top of Ubuntu 6.06.

Since xfs_check can't repair volumes of this size without
scads of memory, I've been using xfs_repair to correct
power-related problems before.

Unfortunately, for some reason xfs_repair is segfaulting:

# ulimit -c unlimited
# xfs_repair -v /dev/md1
Phase 1 - find and verify superblock...
Phase 2 - using internal log
        - zero log...
zero_log: head block 8 tail block 8
        - scan filesystem freespace and inode maps...
        - found root inode chunk
Phase 3 - for each AG...
        - scan and clear agi unlinked lists...
        - process known inodes and perform inode discovery...
        - agno = 0
        - agno = 1
        - agno = 2
        - agno = 3
        - agno = 4
        - agno = 5
        - agno = 6
        - agno = 7
        - agno = 8
        - agno = 9
        - agno = 10
        - agno = 11
        - agno = 12
        - agno = 13
        - agno = 14
        - agno = 15
        - agno = 16
        - agno = 17
        - agno = 18
        - agno = 19
        - agno = 20
        - agno = 21
        - agno = 22
        - agno = 23
        - agno = 24
        - agno = 25
        - agno = 26
        - agno = 27
        - agno = 28
        - agno = 29
        - agno = 30
        - agno = 31
        - process newly discovered inodes...
Phase 4 - check for duplicate blocks...
        - setting up duplicate extent list...
        - clear lost+found (if it exists) ...
        - clearing existing "lost+found" inode
Segmentation fault      (core dumped)


gdb doesn't show anything useful (I don't know how to interpret
the I/O error) :


# gdb /sbin/xfs_repair core
GNU gdb 6.4-debian
Copyright 2005 Free Software Foundation, Inc.
GDB is free software, covered by the GNU General Public License, and you are
welcome to change it and/or distribute copies of it under certain conditions.
Type "show copying" to see the conditions.
There is absolutely no warranty for GDB.  Type "show warranty" for details.
This GDB was configured as "i486-linux-gnu"...(no debugging symbols found)
Using host libthread_db library "/lib/tls/i686/cmov/libthread_db.so.1".

(no debugging symbols found)
Core was generated by `xfs_repair -v /dev/md1'.
Program terminated with signal 11, Segmentation fault.

warning: Can't read pathname for load map: Input/output error.
Reading symbols from /lib/libuuid.so.1...(no debugging symbols found)...done.
Loaded symbols for /lib/libuuid.so.1
Reading symbols from /lib/tls/i686/cmov/libc.so.6...(no debugging symbols found)
Loaded symbols for /lib/tls/i686/cmov/libc.so.6
Reading symbols from /lib/ld-linux.so.2...(no debugging symbols found)...done.
Loaded symbols for /lib/ld-linux.so.2

#0  0x08052f42 in ?? ()
(gdb) bt
#0  0x08052f42 in ?? ()
#1  0x000088e9 in ?? ()
#2  0x00000800 in ?? ()
#3  0x00000080 in ?? ()
#4  0x00000000 in ?? ()


What's the next step?

Thanks,
James

^ permalink raw reply	[flat|nested] 23+ messages in thread

end of thread, other threads:[~2015-03-09 20:14 UTC | newest]

Thread overview: 23+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2013-10-01 19:57 xfs_repair segfault Viet Nguyen
2013-10-01 20:19 ` Dave Chinner
2013-10-01 21:12   ` Viet Nguyen
2013-10-02 10:42     ` Dave Chinner
2013-10-04 17:51       ` Viet Nguyen
2013-10-04 21:43         ` Dave Chinner
2013-10-07 20:09           ` Viet Nguyen
2013-10-08 20:23             ` Dave Chinner
2013-10-09 18:59               ` Viet Nguyen
2013-10-09 20:15                 ` Dave Chinner
2013-10-10 21:13               ` Viet Nguyen
  -- strict thread matches above, loose matches on Subject: below --
2015-03-09 15:50 Rui Gomes
2015-03-09 15:55 ` Carsten Aulbert
2015-03-09 16:11   ` Rui Gomes
2015-03-09 16:14 ` Eric Sandeen
2015-03-09 16:24   ` Rui Gomes
2015-03-09 17:34     ` Eric Sandeen
2015-03-09 17:50       ` Rui Gomes
2015-03-09 18:18         ` Eric Sandeen
2015-03-09 18:24           ` Rui Gomes
2015-03-09 20:13             ` Eric Sandeen
2007-04-03 19:11 James W. Abendschan
2007-04-04  0:45 ` Barry Naujok

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox