* Filessystem corruptions while using rsnapshot @ 2009-09-09 13:30 Markus Trippelsdorf 2009-09-09 14:40 ` Eric Sandeen 0 siblings, 1 reply; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-09 13:30 UTC (permalink / raw) To: linux-ext4 I'm using rsnapshot (http://rsnapshot.org/) to automatically backup my root filesystem (btrfs) to a second harddrive running ext4. Rsnapshot uses rsync and a massive amount of hard links to keep multiple backups instantly available. It seems that the sheer number of hard links overwhelms ext4 and results in problems that require manual fsck.ext4 runs to bring the fs back to normal. For example this is output from fsck.ext4: Problem in HTREE directory inode <some number> ... node <increasing number> has invalid depth (2) ... node <increasing number> has bad mxhash ... node <increasing number> not referenced This output is repeated ad nauseam while <increasing number> increases at each round. The bug is very simple to reproduce here. Just run rsnapshot several times per day and you will eventually hit the problem. # tune2fs -l /dev/sda1 tune2fs 1.41.9 (22-Aug-2009) Filesystem volume name: <none> Last mounted on: /var Filesystem UUID: 46e99ee2-615d-4ce8-9641-a8c0118fdaa3 Filesystem magic number: 0xEF53 Filesystem revision #: 1 (dynamic) Filesystem features: has_journal ext_attr resize_inode dir_index filetype needs_recovery extent flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isize Filesystem flags: signed_directory_hash Default mount options: (none) Filesystem state: clean Errors behavior: Continue Filesystem OS type: Linux Inode count: 6545408 Block count: 26165860 Reserved block count: 1308292 Free blocks: 18575443 Free inodes: 5761975 First block: 0 Block size: 4096 Fragment size: 4096 Reserved GDT blocks: 1017 Blocks per group: 32768 Fragments per group: 32768 Inodes per group: 8192 Inode blocks per group: 512 RAID stride: 32732 Flex block group size: 16 Filesystem created: Mon Jun 8 12:08:51 2009 Last mount time: Wed Sep 9 15:03:04 2009 Last write time: Wed Sep 9 15:03:04 2009 Mount count: 1 Maximum mount count: 5 Last checked: Wed Sep 9 15:02:21 2009 Check interval: 15552000 (6 months) Next check after: Mon Mar 8 14:02:21 2010 Lifetime writes: 434 GB Reserved blocks uid: 0 (user root) Reserved blocks gid: 0 (group root) First inode: 11 Inode size: 256 Required extra isize: 28 Desired extra isize: 28 Journal inode: 8 Default directory hash: half_md4 Directory Hash Seed: 28e527ba-4f7f-4588-a956-c87042f237e6 Journal backup: inode blocks This machine is running the latest git kernel (2.6.31-rc9). -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 13:30 Filessystem corruptions while using rsnapshot Markus Trippelsdorf @ 2009-09-09 14:40 ` Eric Sandeen 2009-09-09 14:54 ` Markus Trippelsdorf ` (2 more replies) 0 siblings, 3 replies; 13+ messages in thread From: Eric Sandeen @ 2009-09-09 14:40 UTC (permalink / raw) To: Markus Trippelsdorf; +Cc: linux-ext4 Markus Trippelsdorf wrote: > I'm using rsnapshot (http://rsnapshot.org/) to automatically backup my > root filesystem (btrfs) to a second harddrive running ext4. Rsnapshot > uses rsync and a massive amount of hard links to keep multiple backups > instantly available. > It seems that the sheer number of hard links overwhelms ext4 and results > in problems that require manual fsck.ext4 runs to bring the fs back to > normal. > > For example this is output from fsck.ext4: > > Problem in HTREE directory inode <some number> > ... node <increasing number> has invalid depth (2) > ... node <increasing number> has bad mxhash > ... node <increasing number> not referenced > > This output is repeated ad nauseam while <increasing number> increases > at each round. > > The bug is very simple to reproduce here. Just run rsnapshot several > times per day and you will eventually hit the problem. Could you provide a bzip2'd e2image -r of a corrupted filesystem for analysis? Thanks, -Eric ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 14:40 ` Eric Sandeen @ 2009-09-09 14:54 ` Markus Trippelsdorf 2009-09-09 17:20 ` Markus Trippelsdorf 2009-09-17 20:33 ` Markus Trippelsdorf 2 siblings, 0 replies; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-09 14:54 UTC (permalink / raw) To: Eric Sandeen; +Cc: linux-ext4 On Wed, Sep 09, 2009 at 09:40:41AM -0500, Eric Sandeen wrote: > Markus Trippelsdorf wrote: > > I'm using rsnapshot (http://rsnapshot.org/) to automatically backup my > > root filesystem (btrfs) to a second harddrive running ext4. Rsnapshot > > uses rsync and a massive amount of hard links to keep multiple backups > > instantly available. > > It seems that the sheer number of hard links overwhelms ext4 and results > > in problems that require manual fsck.ext4 runs to bring the fs back to > > normal. > > > > For example this is output from fsck.ext4: > > > > Problem in HTREE directory inode <some number> > > ... node <increasing number> has invalid depth (2) > > ... node <increasing number> has bad mxhash > > ... node <increasing number> not referenced > > > > This output is repeated ad nauseam while <increasing number> increases > > at each round. > > > > The bug is very simple to reproduce here. Just run rsnapshot several > > times per day and you will eventually hit the problem. > > Could you provide a bzip2'd e2image -r of a corrupted filesystem for > analysis? Sure, I will post a link the next time I'll hit the problem... -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 14:40 ` Eric Sandeen 2009-09-09 14:54 ` Markus Trippelsdorf @ 2009-09-09 17:20 ` Markus Trippelsdorf 2009-09-09 17:21 ` Eric Sandeen 2009-09-17 20:33 ` Markus Trippelsdorf 2 siblings, 1 reply; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-09 17:20 UTC (permalink / raw) To: Eric Sandeen; +Cc: linux-ext4 On Wed, Sep 09, 2009 at 09:40:41AM -0500, Eric Sandeen wrote: > Markus Trippelsdorf wrote: > > I'm using rsnapshot (http://rsnapshot.org/) to automatically backup my > > root filesystem (btrfs) to a second harddrive running ext4. Rsnapshot > > uses rsync and a massive amount of hard links to keep multiple backups > > instantly available. > > It seems that the sheer number of hard links overwhelms ext4 and results > > in problems that require manual fsck.ext4 runs to bring the fs back to > > normal. > > > > For example this is output from fsck.ext4: > > > > Problem in HTREE directory inode <some number> > > ... node <increasing number> has invalid depth (2) > > ... node <increasing number> has bad mxhash > > ... node <increasing number> not referenced > > > > This output is repeated ad nauseam while <increasing number> increases > > at each round. > > > > The bug is very simple to reproduce here. Just run rsnapshot several > > times per day and you will eventually hit the problem. > > Could you provide a bzip2'd e2image -r of a corrupted filesystem for > analysis? OK I've uploaded the file (~80 MB): http://www.2shared.com/file/7681344/d19d7154/disc_e2image.html Hope it helps. -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 17:20 ` Markus Trippelsdorf @ 2009-09-09 17:21 ` Eric Sandeen 2009-09-09 21:29 ` Markus Trippelsdorf 0 siblings, 1 reply; 13+ messages in thread From: Eric Sandeen @ 2009-09-09 17:21 UTC (permalink / raw) To: Markus Trippelsdorf; +Cc: linux-ext4 Markus Trippelsdorf wrote: > On Wed, Sep 09, 2009 at 09:40:41AM -0500, Eric Sandeen wrote: >> Markus Trippelsdorf wrote: ... >>> The bug is very simple to reproduce here. Just run rsnapshot several >>> times per day and you will eventually hit the problem. >> Could you provide a bzip2'd e2image -r of a corrupted filesystem for >> analysis? > > OK I've uploaded the file (~80 MB): > http://www.2shared.com/file/7681344/d19d7154/disc_e2image.html > > Hope it helps. Thanks, can't guarantee that the post-mortem will lead to the anwer but it's worth a look. -Eric ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 17:21 ` Eric Sandeen @ 2009-09-09 21:29 ` Markus Trippelsdorf 2009-09-09 21:35 ` Eric Sandeen 0 siblings, 1 reply; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-09 21:29 UTC (permalink / raw) To: Eric Sandeen; +Cc: linux-ext4 On Wed, Sep 09, 2009 at 12:21:21PM -0500, Eric Sandeen wrote: > Markus Trippelsdorf wrote: > > On Wed, Sep 09, 2009 at 09:40:41AM -0500, Eric Sandeen wrote: > >> Markus Trippelsdorf wrote: > > ... > > >>> The bug is very simple to reproduce here. Just run rsnapshot several > >>> times per day and you will eventually hit the problem. > >> Could you provide a bzip2'd e2image -r of a corrupted filesystem for > >> analysis? > > > > OK I've uploaded the file (~80 MB): > > http://www.2shared.com/file/7681344/d19d7154/disc_e2image.html > > > > Hope it helps. > > Thanks, can't guarantee that the post-mortem will lead to the anwer but > it's worth a look. I also took a look myself and it turned out that my maildir, which contains the LKML messages of the last three months, is the root of the problem. I can now reproduce the bug, by simply running (e.g.): cp -al /var/backup/hourly.3/localhost/home/markus/.maildir/lkml/. /var/tmp/debug where /var is mounted as: "/dev/sda1 on /var type ext4 (rw,noatime,commit=60)" and /var/backup/hourly.3/localhost/ is a previous backup produced by rsnapshot. -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 21:29 ` Markus Trippelsdorf @ 2009-09-09 21:35 ` Eric Sandeen 2009-09-09 21:42 ` Markus Trippelsdorf 0 siblings, 1 reply; 13+ messages in thread From: Eric Sandeen @ 2009-09-09 21:35 UTC (permalink / raw) To: Markus Trippelsdorf; +Cc: linux-ext4 Markus Trippelsdorf wrote: > On Wed, Sep 09, 2009 at 12:21:21PM -0500, Eric Sandeen wrote: >> Markus Trippelsdorf wrote: >>> On Wed, Sep 09, 2009 at 09:40:41AM -0500, Eric Sandeen wrote: >>>> Markus Trippelsdorf wrote: >> ... >> >>>>> The bug is very simple to reproduce here. Just run rsnapshot several >>>>> times per day and you will eventually hit the problem. >>>> Could you provide a bzip2'd e2image -r of a corrupted filesystem for >>>> analysis? >>> OK I've uploaded the file (~80 MB): >>> http://www.2shared.com/file/7681344/d19d7154/disc_e2image.html >>> >>> Hope it helps. >> Thanks, can't guarantee that the post-mortem will lead to the anwer but >> it's worth a look. > > I also took a look myself and it turned out that my maildir, which > contains the LKML messages of the last three months, is the root of the > problem. I can now reproduce the bug, by simply running (e.g.): > > cp -al /var/backup/hourly.3/localhost/home/markus/.maildir/lkml/. /var/tmp/debug > > where /var is mounted as: "/dev/sda1 on /var type ext4 (rw,noatime,commit=60)" > and /var/backup/hourly.3/localhost/ is a previous backup produced by rsnapshot. From the original post it wasn't quite clear; do you hit runtime corruption, indicated by the kernel logs when the fs is mounted, and a subsequent fsck also finds errors? Just for completeness, if there are errors in the system logs, can you post them here as well? Thanks, -Eric ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 21:35 ` Eric Sandeen @ 2009-09-09 21:42 ` Markus Trippelsdorf 2009-09-09 21:46 ` Eric Sandeen 0 siblings, 1 reply; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-09 21:42 UTC (permalink / raw) To: Eric Sandeen; +Cc: linux-ext4 On Wed, Sep 09, 2009 at 04:35:45PM -0500, Eric Sandeen wrote: > Markus Trippelsdorf wrote: > > On Wed, Sep 09, 2009 at 12:21:21PM -0500, Eric Sandeen wrote: > >> Markus Trippelsdorf wrote: > >>> On Wed, Sep 09, 2009 at 09:40:41AM -0500, Eric Sandeen wrote: > >>>> Markus Trippelsdorf wrote: > >> ... > >> > >>>>> The bug is very simple to reproduce here. Just run rsnapshot several > >>>>> times per day and you will eventually hit the problem. > >>>> Could you provide a bzip2'd e2image -r of a corrupted filesystem for > >>>> analysis? > >>> OK I've uploaded the file (~80 MB): > >>> http://www.2shared.com/file/7681344/d19d7154/disc_e2image.html > >>> > >>> Hope it helps. > >> Thanks, can't guarantee that the post-mortem will lead to the anwer but > >> it's worth a look. > > > > I also took a look myself and it turned out that my maildir, which > > contains the LKML messages of the last three months, is the root of the > > problem. I can now reproduce the bug, by simply running (e.g.): > > > > cp -al /var/backup/hourly.3/localhost/home/markus/.maildir/lkml/. /var/tmp/debug > > > > where /var is mounted as: "/dev/sda1 on /var type ext4 (rw,noatime,commit=60)" > > and /var/backup/hourly.3/localhost/ is a previous backup produced by rsnapshot. > > From the original post it wasn't quite clear; do you hit runtime > corruption, indicated by the kernel logs when the fs is mounted, and a > subsequent fsck also finds errors? > > Just for completeness, if there are errors in the system logs, can you > post them here as well? There are no errors in the system logs. I only learned about this bug when a regular boot-time fsck was running that printed the problems mentionted above. -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 21:42 ` Markus Trippelsdorf @ 2009-09-09 21:46 ` Eric Sandeen 2009-09-09 21:50 ` Markus Trippelsdorf 0 siblings, 1 reply; 13+ messages in thread From: Eric Sandeen @ 2009-09-09 21:46 UTC (permalink / raw) To: Markus Trippelsdorf; +Cc: linux-ext4 Markus Trippelsdorf wrote: > On Wed, Sep 09, 2009 at 04:35:45PM -0500, Eric Sandeen wrote: ... >> Just for completeness, if there are errors in the system logs, can you >> post them here as well? > > There are no errors in the system logs. > I only learned about this bug when a regular boot-time fsck was running that > printed the problems mentionted above. Hm, but there should be no regular boot-time fsck unless you have your forced-fsck interval set pretty low; a normal boot-time fsck will simply replay the log. Was that it; did you just happen to catch this on the 21st boot or whatever the interval may be? Otherwise, the fs was flagged with errors, and -that- forced the full fsck at boot time, and hopefully something previous was logged somewhere ... -Eric ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 21:46 ` Eric Sandeen @ 2009-09-09 21:50 ` Markus Trippelsdorf [not found] ` <4AA82C62.40305@redhat.com> 0 siblings, 1 reply; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-09 21:50 UTC (permalink / raw) To: Eric Sandeen; +Cc: linux-ext4 On Wed, Sep 09, 2009 at 04:46:12PM -0500, Eric Sandeen wrote: > Markus Trippelsdorf wrote: > > On Wed, Sep 09, 2009 at 04:35:45PM -0500, Eric Sandeen wrote: > > ... > > >> Just for completeness, if there are errors in the system logs, can you > >> post them here as well? > > > > There are no errors in the system logs. > > I only learned about this bug when a regular boot-time fsck was running that > > printed the problems mentionted above. > > Hm, but there should be no regular boot-time fsck unless you have your > forced-fsck interval set pretty low; a normal boot-time fsck will simply > replay the log. Was that it; did you just happen to catch this on the > 21st boot or whatever the interval may be? Yes, sorry I didn't make myself clear, I just happen to catch this on the 21st boot... After that I booted from an USB stick and ran fsck by hand from there. -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
[parent not found: <4AA82C62.40305@redhat.com>]
* Re: Filessystem corruptions while using rsnapshot [not found] ` <4AA82C62.40305@redhat.com> @ 2009-09-10 1:11 ` Markus Trippelsdorf 0 siblings, 0 replies; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-10 1:11 UTC (permalink / raw) To: Eric Sandeen; +Cc: linux-ext4 On Wed, Sep 09, 2009 at 05:29:54PM -0500, Eric Sandeen wrote: > Markus Trippelsdorf wrote: > > On Wed, Sep 09, 2009 at 04:46:12PM -0500, Eric Sandeen wrote: > >> Markus Trippelsdorf wrote: > >>> On Wed, Sep 09, 2009 at 04:35:45PM -0500, Eric Sandeen wrote: > >> ... > >> > >>>> Just for completeness, if there are errors in the system logs, can you > >>>> post them here as well? > >>> There are no errors in the system logs. > >>> I only learned about this bug when a regular boot-time fsck was running that > >>> printed the problems mentionted above. > >> Hm, but there should be no regular boot-time fsck unless you have your > >> forced-fsck interval set pretty low; a normal boot-time fsck will simply > >> replay the log. Was that it; did you just happen to catch this on the > >> 21st boot or whatever the interval may be? > > > > Yes, sorry I didn't make myself clear, I just happen to catch this on > > the 21st boot... > > After that I booted from an USB stick and ran fsck by hand from there. > > Are you certain there are no errors logged? The fs -is- flagged with > errors: > > # file disc_e2image > disc_e2image: Linux rev 1.0 ext4 filesystem data (needs journal > recovery) (errors) (extents) (large files) (huge files) > > At what point did you make that image? I did the cp -al; after that I ran "tune2fs -C 99 /dev/sda1" to force a e2fsck -p at next boot. I rebooted, the fs became marked with errors by e2fsck -p. Then I ran "e2image -r /dev/sda1 - | bzip2 > disc_e2image.bz2" on the mounted fs. BTW the image shows the errors nicely: # fsck.ext4 -n disc_e2image e2fsck 1.41.9 (22-Aug-2009) Warning: skipping journal recovery because doing a read-only filesystem check. disc_e2image contains a file system with errors, check forced. Pass 1: Checking inodes, blocks, and sizes Pass 2: Checking directory structure Problem in HTREE directory inode 2409285: node (1) has invalid depth (2) Problem in HTREE directory inode 2409285: node (1) has bad max hash Problem in HTREE directory inode 2409285: node (1) not referenced Problem in HTREE directory inode 2409285: node (2) has invalid depth (2) Problem in HTREE directory inode 2409285: node (2) has bad max hash Problem in HTREE directory inode 2409285: node (2) not referenced Problem in HTREE directory inode 2409285: node (3) has invalid depth (2) Problem in HTREE directory inode 2409285: node (3) has bad max hash Problem in HTREE directory inode 2409285: node (3) not referenced Problem in HTREE directory inode 2409285: node (4) has invalid depth (2) Problem in HTREE directory inode 2409285: node (4) has bad max hash Problem in HTREE directory inode 2409285: node (4) not referenced Problem in HTREE directory inode 2409285: node (5) has invalid depth (2) Problem in HTREE directory inode 2409285: node (5) has bad max hash Problem in HTREE directory inode 2409285: node (5) not referenced Problem in HTREE directory inode 2409285: node (6) has invalid depth (2) Problem in HTREE directory inode 2409285: node (6) has bad max hash Problem in HTREE directory inode 2409285: node (6) not referenced Problem in HTREE directory inode 2409285: node (7) has invalid depth (2) Problem in HTREE directory inode 2409285: node (7) has bad max hash Problem in HTREE directory inode 2409285: node (7) not referenced Problem in HTREE directory inode 2409285: node (8) has invalid depth (2) Problem in HTREE directory inode 2409285: node (8) has bad max hash Problem in HTREE directory inode 2409285: node (8) not referenced . . . Problem in HTREE directory inode 2409285: node (1026) has invalid depth (2) Problem in HTREE directory inode 2409285: node (1026) has bad max hash Problem in HTREE directory inode 2409285: node (1026) not referenced Problem in HTREE directory inode 2409285: node (1027) has invalid depth (2) Problem in HTREE directory inode 2409285: node (1027) has bad max hash Problem in HTREE directory inode 2409285: node (1027) not referenced Problem in HTREE directory inode 2409285: node (1028) has invalid depth (2) Problem in HTREE directory inode 2409285: node (1028) has bad max hash Problem in HTREE directory inode 2409285: node (1028) not referenced Invalid HTREE directory inode 2409285 (/backup/hourly.2/localhost/home/markus/.maildir/lkml/cur). Clear HTree index? no Pass 3: Checking directory connectivity Pass 4: Checking reference counts Pass 5: Checking group summary information Free blocks count wrong (18497534, counted=18497453). Fix? no disc_e2image: ********** WARNING: Filesystem still has errors ********** disc_e2image: 809496/6545408 files (0.3% non-contiguous), 7668326/26165860 blocks -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-09 14:40 ` Eric Sandeen 2009-09-09 14:54 ` Markus Trippelsdorf 2009-09-09 17:20 ` Markus Trippelsdorf @ 2009-09-17 20:33 ` Markus Trippelsdorf 2009-09-17 20:39 ` Eric Sandeen 2 siblings, 1 reply; 13+ messages in thread From: Markus Trippelsdorf @ 2009-09-17 20:33 UTC (permalink / raw) To: Eric Sandeen; +Cc: linux-ext4 On Wed, Sep 09, 2009 at 09:40:41AM -0500, Eric Sandeen wrote: > Markus Trippelsdorf wrote: > > I'm using rsnapshot (http://rsnapshot.org/) to automatically backup my > > root filesystem (btrfs) to a second harddrive running ext4. Rsnapshot > > uses rsync and a massive amount of hard links to keep multiple backups > > instantly available. > > It seems that the sheer number of hard links overwhelms ext4 and results > > in problems that require manual fsck.ext4 runs to bring the fs back to > > normal. > > > > For example this is output from fsck.ext4: > > > > Problem in HTREE directory inode <some number> > > ... node <increasing number> has invalid depth (2) > > ... node <increasing number> has bad mxhash > > ... node <increasing number> not referenced > > > > This output is repeated ad nauseam while <increasing number> increases > > at each round. > > > > The bug is very simple to reproduce here. Just run rsnapshot several > > times per day and you will eventually hit the problem. > > Could you provide a bzip2'd e2image -r of a corrupted filesystem for > analysis? Eric, I'm sorry that I may have wasted your time with this. It turned out that a bad RAM module was the most likely cause of these corruptions. (I've found out only today, after I got btrfs csum errors in my log, see: http://thread.gmane.org/gmane.comp.file-systems.btrfs/3613) So please accept my apologies. Thanks. -- Markus ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Filessystem corruptions while using rsnapshot 2009-09-17 20:33 ` Markus Trippelsdorf @ 2009-09-17 20:39 ` Eric Sandeen 0 siblings, 0 replies; 13+ messages in thread From: Eric Sandeen @ 2009-09-17 20:39 UTC (permalink / raw) To: Markus Trippelsdorf; +Cc: linux-ext4 Markus Trippelsdorf wrote: ... > Eric, I'm sorry that I may have wasted your time with this. > It turned out that a bad RAM module was the most likely cause of these > corruptions. (I've found out only today, after I got btrfs csum errors > in my log, see: http://thread.gmane.org/gmane.comp.file-systems.btrfs/3613) > > So please accept my apologies. > Thanks. That's ok, I've been so swamped I haven't had much time at all to look at it, and was feeling bad. So this is actually good news on several fronts. ;) Thanks for the followup! -Eric ^ permalink raw reply [flat|nested] 13+ messages in thread
end of thread, other threads:[~2009-09-17 20:39 UTC | newest]
Thread overview: 13+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-09-09 13:30 Filessystem corruptions while using rsnapshot Markus Trippelsdorf
2009-09-09 14:40 ` Eric Sandeen
2009-09-09 14:54 ` Markus Trippelsdorf
2009-09-09 17:20 ` Markus Trippelsdorf
2009-09-09 17:21 ` Eric Sandeen
2009-09-09 21:29 ` Markus Trippelsdorf
2009-09-09 21:35 ` Eric Sandeen
2009-09-09 21:42 ` Markus Trippelsdorf
2009-09-09 21:46 ` Eric Sandeen
2009-09-09 21:50 ` Markus Trippelsdorf
[not found] ` <4AA82C62.40305@redhat.com>
2009-09-10 1:11 ` Markus Trippelsdorf
2009-09-17 20:33 ` Markus Trippelsdorf
2009-09-17 20:39 ` Eric Sandeen
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).