* lost with degraded RAID1
@ 2014-01-29 19:16 Johan Kröckel
2014-01-30 3:28 ` Duncan
` (3 more replies)
0 siblings, 4 replies; 25+ messages in thread
From: Johan Kröckel @ 2014-01-29 19:16 UTC (permalink / raw)
To: linux-btrfs
My situation:
Former btrfs-RAID1 on two luks encrypted partitions (bunkerA and bunkerB).
Disk holding bunkerB died online.
Now I started rebalancing bunkerA to single, but in between had the
idea to go over a reboot ("maybe the disk reappears"?). So stopped
rebalancing, rebooted.
Now the damaged disk disappeared completely. No sign of it in /proc/diskstats.
So, i wanted to finish the rebalancing to single.
Label: bunker uuid: 7f954a85-7566-4251-832c-44f2d3de2211
Total devices 2 FS bytes used 1.58TiB
devid 1 size 1.82TiB used 1.58TiB path
devid 2 size 1.82TiB used 1.59TiB path /dev/mapper/bunkerA
btrfs filesystem df /mnt
Data, RAID1: total=1.58TiB, used=1.57TiB
Data, single: total=11.00GiB, used=10.00GiB
System, RAID1: total=8.00MiB, used=240.00KiB
Metadata, RAID1: total=3.00GiB, used=1.61GiB
The 11 GiB single Data was balanced in this way while having no second
actually functioning device in the array before reboot!
But now, I can't mount bunkerA degraded,RW because degraded
filesystems are not allowed to be mounted RW (?). The consequence is,
that I felt unable to do anything other than mounting it RO and back
it up (all data is fine) to another filesystem. I don't have a new
substitute disk, yet, so I couldn't test whether I can add a new
device to a RO-mounted filesystem. Adding a loop device didn't work.
What are my options now? What further information could be useful? I
cant believe, that I have all my data in front of me and have to start
over again with a new filesystem because of security checks and no
enforcement option to mount it RW.
Thanks
Johan Kröckel
^ permalink raw reply [flat|nested] 25+ messages in thread* Re: lost with degraded RAID1 2014-01-29 19:16 lost with degraded RAID1 Johan Kröckel @ 2014-01-30 3:28 ` Duncan 2014-01-30 11:53 ` Johan Kröckel ` (2 subsequent siblings) 3 siblings, 0 replies; 25+ messages in thread From: Duncan @ 2014-01-30 3:28 UTC (permalink / raw) To: linux-btrfs Johan Kröckel posted on Wed, 29 Jan 2014 20:16:09 +0100 as excerpted: [btrfs raid1 on a pair of luks-encrypted partitions, one went south and a balance back to single was started on the good one, interrupted by a reboot.] > But now, I can't mount bunkerA degraded,RW because degraded filesystems > are not allowed to be mounted RW (?). AFAIK that shouldn't be the case. Degraded should allow the RW mount -- I know it did some kernels ago when I tried it then, and if it changed, it's news to me too, in which case I need to do some reevaluation here. What I think /might/ have happened is that there's some other damage to the filesystem unrelated to the missing device in the raid1, which forces it read-only as soon as btrfs finds that damage. However, mount -o remount,rw,degraded should still work, I /think/. > The consequence is, > that I felt unable to do anything other than mounting it RO and back it > up (all data is fine) to another filesystem. I don't have a new > substitute disk, yet, so I couldn't test whether I can add a new device > to a RO-mounted filesystem. Adding a loop device didn't work. Well, given that btrfs isn't yet entirely stable, and both mkfs.btrfs and the btrfs wiki at btrfs.wiki.kernel.org warn to keep tested backups in case something goes wrong with the btrfs your testing and you lose that copy... you should have already had that backup, and wouldn't need to make it, only perhaps update a few files that changed since your last backup that you were willing to lose the changes too if it came to it, but since you can still mount ro, you might as well take the chance to update the backup given that you can. > What are my options now? What further information could be useful? I > cant believe, that I have all my data in front of me and have to start > over again with a new filesystem because of security checks and no > enforcement option to mount it RW. As I said, to the best of my (non-dev btrfs user and list regular) knowledge, mount -o degraded,rw should work. If it's not, there's likely something else going on triggering the ro remount, and seeing any related dmesg output would presumably help. One other thing that might help is the skip_balance mount option, to avoid restarting the in-process balance immediately. See if that lets you mount degraded,rw without immediately forcing back to ro. You can then try resuming the balance. If that fails, try canceling the balance and then starting a new balance using balance filters (see the balance questions in the FAQ on the wiki, which link to the balance-filters page). Meanwhile, at least you have ro access to all the data and can take that backup that you apparently ignored the warnings about making, previously. -- Duncan - List replies preferred. No HTML msgs. "Every nonfree program has a lord, a master -- and if you use the program, he is your master." Richard Stallman ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: Re: lost with degraded RAID1 2014-01-29 19:16 lost with degraded RAID1 Johan Kröckel 2014-01-30 3:28 ` Duncan @ 2014-01-30 11:53 ` Johan Kröckel 2014-01-30 17:57 ` Chris Murphy ` (2 more replies) 2014-01-30 17:33 ` Chris Murphy 2014-01-31 6:40 ` Johan Kröckel 3 siblings, 3 replies; 25+ messages in thread From: Johan Kröckel @ 2014-01-30 11:53 UTC (permalink / raw) To: linux-btrfs [Answer from Duncan, 1i5t5.duncan@DOMAIN.HIDDEN (Thanks for the try)] [AFAIK that shouldn't be the case. Degraded should allow the RW mount -- I know it did some kernels ago when I tried it then, and if it changed, it's news to me too, in which case I need to do some reevaluation here.] What I think /might/ have happened is that there's some other damage to the filesystem unrelated to the missing device in the raid1, which forces it read-only as soon as btrfs finds that damage. However, mount -o remount,rw,degraded should still work, I /think/.] http://www.spinics.net/lists/linux-btrfs/msg20164.html [you should have already had that backup, and wouldn't need to make it, only perhaps update a few files that changed since your last backup that you were willing to lose the changes too if it came to it, but since you can still mount ro, you might as well take the chance to update the backup given that you can.] I have backups on other disks and from. Thats not my problem. That was just as a notice, to show that the data IS totally fine. It was/is RAID1, so why shouldnt it be. [As I said, to the best of my (non-dev btrfs user and list regular) knowledge, mount -o degraded,rw should work.] No, it doesnt. Syslog says: "Jan 30 12:44:02 fortknox kernel: [756677.795661] Btrfs: too many missing devices, writeable mount is not allowed" [One other thing that might help is the skip_balance mount option, to avoid restarting the in-process balance immediately.] No, same in syslog. [If that fails, try canceling the balance and then starting a new balance using balance filters...] The balance was already canceled AND its not possible to balance a ro filesystem. [Meanwhile, at least you have ro access to all the data and can take that backup that you apparently ignored the warnings about making, previously.] Yes, i can and no, i did not. ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-30 11:53 ` Johan Kröckel @ 2014-01-30 17:57 ` Chris Murphy [not found] ` <CABgvyo9FYGSYpj+jL1oqCvtNUqsC8HZ+z=x-Gz7naWoEcCKWpQ@mail.gmail.com> 2014-01-30 22:18 ` Duncan 2014-01-31 2:19 ` Chris Murphy 2 siblings, 1 reply; 25+ messages in thread From: Chris Murphy @ 2014-01-30 17:57 UTC (permalink / raw) To: Johan Kröckel; +Cc: linux-btrfs On Jan 30, 2014, at 4:53 AM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > > [As I said, to the best of my (non-dev btrfs user and list regular) > knowledge, mount -o degraded,rw should work.] > > No, it doesnt. > Syslog says: "Jan 30 12:44:02 fortknox kernel: [756677.795661] Btrfs: > too many missing devices, writeable mount is not allowed" Is this encrypted Btrfs volume used for rootfs? Or is it only for data? If it's only for data, then make sure the volume (all subvolumes) are umounted, then mount with this: -o degraded,recovery,skip_balance Do you still get too many missing devices message? Vaguely related questions (for everyone) are: - If this is rootfs, by default grub options mount ro. So when we use rootflags=degraded,recovery that's effectively -o ro,degraded,recovery. Upon fstab being read, rootfs is remounted rw. Does the previous ro,degraded,recovery repair get written to the volume upon fstab remount rw? This is somewhat important to understand, and whether the grub default of ro mount, and the remount rw switch at fstab time is sufficiently detecting and repairing Btrfs problems (with and without rootflags=recovery). - Is an online conversion from profile X to profile Y expected to work when the volume is mounted degraded? If it's not expected to work, or isn't a good idea, then maybe balance needs to be disabled on degraded mounts? Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
[parent not found: <CABgvyo9FYGSYpj+jL1oqCvtNUqsC8HZ+z=x-Gz7naWoEcCKWpQ@mail.gmail.com>]
* Re: lost with degraded RAID1 [not found] ` <CABgvyo9FYGSYpj+jL1oqCvtNUqsC8HZ+z=x-Gz7naWoEcCKWpQ@mail.gmail.com> @ 2014-01-30 19:32 ` Chris Murphy 0 siblings, 0 replies; 25+ messages in thread From: Chris Murphy @ 2014-01-30 19:32 UTC (permalink / raw) To: Johan Kröckel; +Cc: Btrfs BTRFS On Jan 30, 2014, at 11:44 AM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > 2014-01-30 Chris Murphy <lists@colorremedies.com>: >> Is this encrypted Btrfs volume used for rootfs? Or is it only for data? If it's only for data, then make sure the volume (all subvolumes) are umounted, then mount with this: >> >> -o degraded,recovery,skip_balance >> >> Do you still get too many missing devices message? > > root@fortknox:/# mount -odegraded,recovery,skip_balance > /dev/mapper/bunkerA /mnt > mount: wrong fs type, bad option, bad superblock on /dev/mapper/bunkerA, > missing codepage or helper program, or other error > In some cases useful info is found in syslog - try > dmesg | tail or so > > root@fortknox:/# dmesg|tail > [781630.598155] btrfs: bdev (null) errs: wr 1375454, rd 186442, flush > 0, corrupt 0, gen 0 That's a lot of read and write errors. I wonder what that's all about. Is btrfs tracking the resulting read/write failures due to the missing device? So at this point it's worth a btrfs check /dev/mapper/bunkerA. This is read only, does not make repairs. And also I'm curious what you get for btrfs-show-super -a /dev/mapper/bunkerA. The volume shouldn't be mounted for the first one, and can be either mounted or not for the 2nd. Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-30 11:53 ` Johan Kröckel 2014-01-30 17:57 ` Chris Murphy @ 2014-01-30 22:18 ` Duncan 2014-01-31 3:00 ` Chris Murphy 2014-01-31 2:19 ` Chris Murphy 2 siblings, 1 reply; 25+ messages in thread From: Duncan @ 2014-01-30 22:18 UTC (permalink / raw) To: linux-btrfs Johan Kröckel posted on Thu, 30 Jan 2014 12:53:44 +0100 as excerpted: > [Answer from Duncan, 1i5t5.duncan@DOMAIN.HIDDEN (Thanks for the try)] > > [AFAIK that shouldn't be the case. Degraded should allow the RW mount > -- I know it did some kernels ago when I tried it then, and if it > changed, it's news to me too, in which case I need to do some > reevaluation here.] > > What I think /might/ have happened is that there's some other damage to > the filesystem unrelated to the missing device in the raid1, which > forces it read-only as soon as btrfs finds that damage. However, mount > -o remount,rw,degraded should still work, I /think/.] > > http://www.spinics.net/lists/linux-btrfs/msg20164.html Thanks. I had seen that message go by but obviously missed all the implications. IMO the idea of the patch is correct as a default, but there should be a way to override, since AFAIK ro-mount will I believe at times prevent undegrading as well, certainly when an admin's choice to clear the degraded state is reduced redundancy, which is what you're doing. IOW, I guess I don't agree with that patch as it was apparently committed. There needs to be a force option as well. Meanwhile, back to the current situation. Given that you found the patch preventing what you were trying to do, what happens if you simply find that commit in the git repo, revert and rebuild? Assuming there are no further commits building on that one, a revert and rebuild should at least allow you to complete the half-completed balance. And regardless of whether btrfs policy is to prevent that entirely in the future or not, definitely letting you start the balance before a reboot and not letting you finish it afterward is a bug. It should either be prevented entirely, or allowed to finish after a reboot if it was allowed to start. -- Duncan - List replies preferred. No HTML msgs. "Every nonfree program has a lord, a master -- and if you use the program, he is your master." Richard Stallman ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-30 22:18 ` Duncan @ 2014-01-31 3:00 ` Chris Murphy 2014-01-31 5:58 ` Duncan 0 siblings, 1 reply; 25+ messages in thread From: Chris Murphy @ 2014-01-31 3:00 UTC (permalink / raw) To: Btrfs BTRFS On Jan 30, 2014, at 3:18 PM, Duncan <1i5t5.duncan@cox.net> wrote: > IOW, I guess I don't agree with that patch as it was apparently > committed. There needs to be a force option as well. -o degraded is the force option. I think the problem here is that there's sufficient damage to the one remaining device, that it cannot be mounted rw. It's sort of a chicken - egg problem, the single device available has a file system that's sufficiently damaged that it undamaged metadata to get a rw mount. Since there are too few devices for that, it fails to mount rw. I'm not seeing it as any different from a single device volume with data/metadata profile single, with sufficient damage to cause it to not mount rw. If -o recovery can't fix it, I think it's done for. So something Johan sent to me but didn't make the list (I've asked him to repost) is that his attempts to mount degraded,recovery,skip_balance, result in a: /dev/mapper/bunkerA /mnt mount: wrong fs type, bad option, bad superblock He gets other errors also, and he has the results of btrfs check that might be more revealing, but to say the least it looks pretty nasty. Another question for Johann is what exact balance command was to go back to single? Was there -dconvert and -mconvert? Both are required to go from raid1/raid to single/DUP or single/single, and actually get rid of the 2nd device. And looking at the man page, I'm not sure how we do that conversion and specify which multiple device we're dropping: [filesystem] balance start [options] <path> With a missing device, presumably this is obvious, but… Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-31 3:00 ` Chris Murphy @ 2014-01-31 5:58 ` Duncan 2014-01-31 6:10 ` Chris Murphy 0 siblings, 1 reply; 25+ messages in thread From: Duncan @ 2014-01-31 5:58 UTC (permalink / raw) To: linux-btrfs Chris Murphy posted on Thu, 30 Jan 2014 20:00:39 -0700 as excerpted: > Another question for Johann is what exact balance command was to go back > to single? Was there -dconvert and -mconvert? Both are required to go > from raid1/raid to single/DUP or single/single, and actually get rid of > the 2nd device. And looking at the man page, I'm not sure how we do that > conversion and specify which multiple device we're dropping: > > [filesystem] balance start [options] <path> > > With a missing device, presumably this is obvious, but… Of course the normal no-missing-device usage of a balance converting back to single would be one of: 1) Multi-device raidN converting back to single with all devices present, presumably either to get more space (from raid1/5/6/10), or to get back /some/ device-loss tolerance when converting back from raid0 (presumably a data-only conversion in that case, with metadata kept as the default multi-device raid1). 2) Single-device, converting metadata-only to single from dup, either for space reasons, or because the physical device is SSD (in which case single-device metadata default is used due to some devices doing hardware/ firmware dedup with the effect of dup metadata on such a device thus not entirely predictable anyway), possibly an ssd that does deduping in hardware, so dup mode's not particularly useful and is somewhat unpredictable in any case. IOW, just because it's a conversion to single mode doesn't mean we're dropping a device, and a rebalance to single mode wasn't in fact designed to drop a device (that's what device delete is for), so it doesn't really need a way to specify a device to drop. If one is missing, a rebalance will obviously rebalance to existing devices, that's obvious, but otherwise, balance isn't the command for that, device delete is. But your question to Johann remains valid, since we don't know for sure that he was doing the "correct" full balance -dconvert -mconvert, and so far I've simply assumed he was. -- Duncan - List replies preferred. No HTML msgs. "Every nonfree program has a lord, a master -- and if you use the program, he is your master." Richard Stallman ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-31 5:58 ` Duncan @ 2014-01-31 6:10 ` Chris Murphy 2014-01-31 6:13 ` Chris Murphy 0 siblings, 1 reply; 25+ messages in thread From: Chris Murphy @ 2014-01-31 6:10 UTC (permalink / raw) To: Duncan, Johan Kröckel; +Cc: linux-btrfs On Jan 30, 2014, at 10:58 PM, Duncan <1i5t5.duncan@cox.net> wrote: > > IOW, just because it's a conversion to single mode doesn't mean we're > dropping a device, and a rebalance to single mode wasn't in fact designed > to drop a device (that's what device delete is for) Ahh yes, of course. So for two device raid1/raid1 going to two device single/DUP, we'd expect to see both drives still used something close to round robin. And then we'd do a device delete which would migrate chunks. I see a future optimization here :-) to avoid much of the writing this two step technique involves. I'm also seeing many "Error reading 1647012864000, -1" with different block addresses (same -1 though), and also "1659900002304failed to load free space cache for block group" also with different numbers. Maybe hundreds of these. I'm not sure if this is due to the missing device, and it's reporting missing meta data? Or if the working device also has some problem, which depending on the configuration might implicate a single SATA controller. Johan, can you post a full dmesg should be posted somewhere? And also smartctl -x results for the working drive? Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-31 6:10 ` Chris Murphy @ 2014-01-31 6:13 ` Chris Murphy 2014-01-31 7:37 ` Duncan 0 siblings, 1 reply; 25+ messages in thread From: Chris Murphy @ 2014-01-31 6:13 UTC (permalink / raw) To: Duncan, Johan Kröckel; +Cc: linux-btrfs On Jan 30, 2014, at 11:10 PM, Chris Murphy <lists@colorremedies.com> wrote: > > I'm also seeing many "Error reading 1647012864000, -1" with different block addresses (same -1 though), and also "1659900002304failed to load free space cache for block group" also with different numbers. Maybe hundreds of these. I'm not sure if this is due to the missing device, and it's reporting missing meta data? Or if the working device also has some problem, which depending on the configuration might implicate a single SATA controller. Sorry I'm being impatient since I'm away snowboarding tomorrow. The above refers to parts of Johan's btrfs check output which hasn't yet made it to the list. Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-31 6:13 ` Chris Murphy @ 2014-01-31 7:37 ` Duncan 0 siblings, 0 replies; 25+ messages in thread From: Duncan @ 2014-01-31 7:37 UTC (permalink / raw) To: linux-btrfs Chris Murphy posted on Thu, 30 Jan 2014 23:13:36 -0700 as excerpted: > On Jan 30, 2014, at 11:10 PM, Chris Murphy <lists@colorremedies.com> > wrote: >> >> I'm also seeing many "Error reading 1647012864000, -1" with different >> block addresses (same -1 though), and also "1659900002304failed to load >> free space cache for block group" also with different numbers. Maybe >> hundreds of these. I'm not sure if this is due to the missing device, >> and it's reporting missing meta data? Or if the working device also has >> some problem, which depending on the configuration might implicate a >> single SATA controller. > > Sorry I'm being impatient since I'm away snowboarding tomorrow. The > above refers to parts of Johan's btrfs check output which hasn't yet > made it to the list. Break a leg! =:^) (If you've done much drama/acting you may know the reference; there's a traditional superstition that wishing someone .... luck is a hex and they'll forget their lines or worse! So you wish them to break a leg instead! =:^) I'm assuming the error reading errors are due to the missing device, with the round-robin trying to read the bad device 50% of the time, but if it was full raid1, then the existing copy should be found and read instead, assuming of course that the checksum verifies, so despite all the alarming looking noise, I believe it's harmless... assuming it was indeed full raid1 both data and metadata. -- Duncan - List replies preferred. No HTML msgs. "Every nonfree program has a lord, a master -- and if you use the program, he is your master." Richard Stallman ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-30 11:53 ` Johan Kröckel 2014-01-30 17:57 ` Chris Murphy 2014-01-30 22:18 ` Duncan @ 2014-01-31 2:19 ` Chris Murphy 2 siblings, 0 replies; 25+ messages in thread From: Chris Murphy @ 2014-01-31 2:19 UTC (permalink / raw) To: Johan Kröckel; +Cc: linux-btrfs On Jan 30, 2014, at 4:53 AM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > http://www.spinics.net/lists/linux-btrfs/msg20164.html and > Syslog says: "Jan 30 12:44:02 fortknox kernel: [756677.795661] Btrfs: > too many missing devices, writeable mount is not allowed" By the way, the cited patch says "Btrfs: too many missing devices, writeable remount is not allowed" The patch says "remount" but your error says "mount". I'm not sure you have the right patch. Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-29 19:16 lost with degraded RAID1 Johan Kröckel 2014-01-30 3:28 ` Duncan 2014-01-30 11:53 ` Johan Kröckel @ 2014-01-30 17:33 ` Chris Murphy 2014-01-30 17:58 ` Hugo Mills [not found] ` <CABgvyo822bOAHeA1GH28MPaBAU+Zdi72MD_uwL+dhopt+nwMig@mail.gmail.com> 2014-01-31 6:40 ` Johan Kröckel 3 siblings, 2 replies; 25+ messages in thread From: Chris Murphy @ 2014-01-30 17:33 UTC (permalink / raw) To: Btrfs BTRFS On Jan 29, 2014, at 12:16 PM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > My situation: > > Former btrfs-RAID1 on two luks encrypted partitions (bunkerA and bunkerB). > Disk holding bunkerB died online. > Now I started rebalancing bunkerA to single, You're doing an online conversion of a degraded raid1 volume into single? Does anyone know if this is expected or intended to work? Obviously the tools allow you to start it, but for some reason to me it doesn't seem like a good idea. I'd just leave it -o degraded until I get a replacement drive. Deciding to do a conversion on a degraded volume doesn't seem like proper timing. > > Now the damaged disk disappeared completely. No sign of it in /proc/diskstats. I don't know what you mean by it's disappeared complete. Are you talking about the physical block device? Or the dm logical block device? What do you get for: lsblk blkid btrfs device scan --all-devices btrfs fi show > But now, I can't mount bunkerA degraded,RW because degraded > filesystems are not allowed to be mounted RW (?) Have you tried? What errors do you get in user space and dmesg? Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-30 17:33 ` Chris Murphy @ 2014-01-30 17:58 ` Hugo Mills 2014-01-30 18:25 ` Chris Murphy [not found] ` <CABgvyo822bOAHeA1GH28MPaBAU+Zdi72MD_uwL+dhopt+nwMig@mail.gmail.com> 1 sibling, 1 reply; 25+ messages in thread From: Hugo Mills @ 2014-01-30 17:58 UTC (permalink / raw) To: Chris Murphy; +Cc: Btrfs BTRFS [-- Attachment #1: Type: text/plain, Size: 980 bytes --] On Thu, Jan 30, 2014 at 10:33:21AM -0700, Chris Murphy wrote: > > On Jan 29, 2014, at 12:16 PM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > > > My situation: > > > > Former btrfs-RAID1 on two luks encrypted partitions (bunkerA and bunkerB). > > Disk holding bunkerB died online. > > Now I started rebalancing bunkerA to single, > > > You're doing an online conversion of a degraded raid1 volume into single? Does anyone know if this is expected or intended to work? I don't see why not. One suggested method of recovering RAID from a degraded situation is to rebalance over just the remaining devices (space permitting, of course). Hugo. -- === Hugo Mills: hugo@... carfax.org.uk | darksatanic.net | lug.org.uk === PGP key: 65E74AC0 from wwwkeys.eu.pgp.net or http://www.carfax.org.uk --- Everything simple is false. Everything which is --- complex is unusable. [-- Attachment #2: Digital signature --] [-- Type: application/pgp-signature, Size: 811 bytes --] ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-30 17:58 ` Hugo Mills @ 2014-01-30 18:25 ` Chris Murphy 2014-02-03 20:55 ` Johan Kröckel 0 siblings, 1 reply; 25+ messages in thread From: Chris Murphy @ 2014-01-30 18:25 UTC (permalink / raw) To: Hugo Mills; +Cc: Btrfs BTRFS On Jan 30, 2014, at 10:58 AM, Hugo Mills <hugo@carfax.org.uk> wrote: > On Thu, Jan 30, 2014 at 10:33:21AM -0700, Chris Murphy wrote: >> >> On Jan 29, 2014, at 12:16 PM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: >> >>> My situation: >>> >>> Former btrfs-RAID1 on two luks encrypted partitions (bunkerA and bunkerB). >>> Disk holding bunkerB died online. >>> Now I started rebalancing bunkerA to single, >> >> >> You're doing an online conversion of a degraded raid1 volume into single? Does anyone know if this is expected or intended to work? > > I don't see why not. One suggested method of recovering RAID from a > degraded situation is to rebalance over just the remaining devices > (space permitting, of course). Right but that's not a conversion. That's a regular balance on a degraded mount, with multiple remaining devices: e.g. a 4 disk raid1, drive fails, mount -o degraded, delete missing, then balance will replicate any missing 2nd copies onto three drives. The bigger problem at the moment is that -o degraded isn't working for Johan. The too many missing devices message seems like a bug and with limited information it may even be whatever that bug is, that cause the conversion to fail. Some 11GB were converted prior to the failure. Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-30 18:25 ` Chris Murphy @ 2014-02-03 20:55 ` Johan Kröckel 2014-02-03 21:08 ` Chris Murphy 0 siblings, 1 reply; 25+ messages in thread From: Johan Kröckel @ 2014-02-03 20:55 UTC (permalink / raw) Cc: Btrfs BTRFS 2014-01-30 Chris Murphy <lists@colorremedies.com>: > > On Jan 30, 2014, at 10:58 AM, Hugo Mills <hugo@carfax.org.uk> wrote: > >> On Thu, Jan 30, 2014 at 10:33:21AM -0700, Chris Murphy wrote: >>> You're doing an online conversion of a degraded raid1 volume into single? Does anyone know if this is expected or intended to work? >> >> I don't see why not. One suggested method of recovering RAID from a >> degraded situation is to rebalance over just the remaining devices >> (space permitting, of course). > > Right but that's not a conversion. That's a regular balance on a degraded mount, with multiple remaining devices: e.g. a 4 disk raid1, drive fails, mount -o degraded, delete missing, then balance will replicate any missing 2nd copies onto three drives. > > The bigger problem at the moment is that -o degraded isn't working for Johan. The too many missing devices message seems like a bug and with limited information it may even be whatever that bug is, that cause the conversion to fail. Some 11GB were converted prior to the failure. Which usefull information can provide. On the weekend I was at the server and found out, that the vanishing of the drive at reboot was strange behavior of the bios. So the drive is online again. but the filesystem is still showing strange behavior, but now I can mount it rw. > Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-02-03 20:55 ` Johan Kröckel @ 2014-02-03 21:08 ` Chris Murphy 2014-02-03 21:31 ` Johan Kröckel 0 siblings, 1 reply; 25+ messages in thread From: Chris Murphy @ 2014-02-03 21:08 UTC (permalink / raw) To: Johan Kröckel; +Cc: Btrfs BTRFS On Feb 3, 2014, at 1:55 PM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > 2014-01-30 Chris Murphy <lists@colorremedies.com>: >> >> On Jan 30, 2014, at 10:58 AM, Hugo Mills <hugo@carfax.org.uk> wrote: >> >>> On Thu, Jan 30, 2014 at 10:33:21AM -0700, Chris Murphy wrote: >>>> You're doing an online conversion of a degraded raid1 volume into single? Does anyone know if this is expected or intended to work? >>> >>> I don't see why not. One suggested method of recovering RAID from a >>> degraded situation is to rebalance over just the remaining devices >>> (space permitting, of course). >> >> Right but that's not a conversion. That's a regular balance on a degraded mount, with multiple remaining devices: e.g. a 4 disk raid1, drive fails, mount -o degraded, delete missing, then balance will replicate any missing 2nd copies onto three drives. >> >> The bigger problem at the moment is that -o degraded isn't working for Johan. The too many missing devices message seems like a bug and with limited information it may even be whatever that bug is, that cause the conversion to fail. Some 11GB were converted prior to the failure. > Which usefull information can provide. On the weekend I was at the > server and found out, that the vanishing of the drive at reboot was > strange behavior of the bios. So the drive is online again. but the > filesystem is still showing strange behavior, but now I can mount it > rw. I'd like to see btrfs fi df results for the volume. And new btrfs check. And then a backup if needed, and then a scrub to see if that fixes anything broken between them. I'm not sure what happens if a new generation object is broken and the old generation is OK, what scrub will do? Maybe it just reports it, I'm not sure. If you want you could do a btrfs scrub -r which is read only and just reports what the problems are. You also have an incomplete balance, right? So it's possible some things might not be fixable if the conversion to single was successful. You'll need to decide if you want to reconvert back to data/metadata raid1/raid from whatever you're at now. Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-02-03 21:08 ` Chris Murphy @ 2014-02-03 21:31 ` Johan Kröckel 2014-02-07 11:34 ` Johan Kröckel 0 siblings, 1 reply; 25+ messages in thread From: Johan Kröckel @ 2014-02-03 21:31 UTC (permalink / raw) Cc: Btrfs BTRFS State is: I wont use this filesystem again. I have a backup. So I am interested to give the necessary information for debuging it and afterwards format it and create a new one. I already did fscks and btrfschk --repair and pushed the output to txt-files but they are more than 4 mb in size. So I will post excerpts: ============file: btrfsck.out=========================== Checking filesystem on /dev/mapper/bunkerA UUID: 7f954a85-7566-4251-832c-44f2d3de2211 42 parent transid verify failed on 1887688011776 wanted 121037 found 88533 parent transid verify failed on 1888518615040 wanted 121481 found 90267 parent transid verify failed on 1681394102272 wanted 110919 found 91024 parent transid verify failed on 1888522838016 wanted 121486 found 90270 parent transid verify failed on 1888398331904 wanted 121062 found 89987 leaf parent key incorrect 1887867330560 bad block 1887867330560 leaf parent key incorrect 1888120320000 bad block 1888120320000 leaf parent key incorrect 1888124637184 bad block 1888124637184 leaf parent key incorrect 1888131444736 bad block 1888131444736 [...and so on for 4MB] bad block 1888513552384 leaf parent key incorrect 1888513642496 bad block 1888513642496 leaf parent key incorrect 1888513654784 bad block 1888513654784 leaf parent key incorrect 1888514023424 bad block 1888514023424 btrfsck: cmds-check.c:2212: check_owner_ref: Assertion `!(rec->is_root)' failed. ================file: smartctl-before-btrfschk-repair============== smartctl 5.41 2011-06-09 r3365 [x86_64-linux-3.12-0.bpo.1-amd64] (local build) Copyright (C) 2002-11 by Bruce Allen, http://smartmontools.sourceforge.net === START OF READ SMART DATA SECTION === SMART Attributes Data Structure revision number: 10 Vendor Specific SMART Attributes with Thresholds: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 1 Raw_Read_Error_Rate 0x000f 118 099 006 Pre-fail Always - 172055696 3 Spin_Up_Time 0x0003 093 093 000 Pre-fail Always - 0 4 Start_Stop_Count 0x0032 100 100 020 Old_age Always - 7 5 Reallocated_Sector_Ct 0x0033 100 100 010 Pre-fail Always - 0 7 Seek_Error_Rate 0x000f 069 060 030 Pre-fail Always - 9085642 9 Power_On_Hours 0x0032 097 097 000 Old_age Always - 2769 10 Spin_Retry_Count 0x0013 100 100 097 Pre-fail Always - 0 12 Power_Cycle_Count 0x0032 100 100 020 Old_age Always - 7 184 End-to-End_Error 0x0032 100 100 099 Old_age Always - 0 187 Reported_Uncorrect 0x0032 100 100 000 Old_age Always - 0 188 Command_Timeout 0x0032 100 100 000 Old_age Always - 0 189 High_Fly_Writes 0x003a 083 083 000 Old_age Always - 17 190 Airflow_Temperature_Cel 0x0022 077 071 045 Old_age Always - 23 (Min/Max 22/23) 191 G-Sense_Error_Rate 0x0032 100 100 000 Old_age Always - 0 192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always - 5 193 Load_Cycle_Count 0x0032 100 100 000 Old_age Always - 7 194 Temperature_Celsius 0x0022 023 040 000 Old_age Always - 23 (0 20 0 0) 197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always - 0 198 Offline_Uncorrectable 0x0010 100 100 000 Old_age Offline - 0 199 UDMA_CRC_Error_Count 0x003e 200 200 000 Old_age Always - 0 =================file:btrfsck-repair.out======================== enabling repair mode Checking filesystem on /dev/mapper/bunkerA UUID: 7f954a85-7566-4251-832c-44f2d3de2211 ify failed on 1887688011776 wanted 121037 found 88533 parent transid verify failed on 1888518615040 wanted 121481 found 90267 parent transid verify failed on 1681394102272 wanted 110919 found 91024 parent transid verify failed on 1888522838016 wanted 121486 found 90270 parent transid verify failed on 1888398331904 wanted 121062 found 89987 leaf parent key incorrect 1887867330560 bad block 1887867330560 [...and so on for 4MB] bad block 1888513642496 leaf parent key incorrect 1888513654784 bad block 1888513654784 leaf parent key incorrect 1888514023424 bad block 1888514023424 btrfsck: cmds-check.c:2212: check_owner_ref: Assertion `!(rec->is_root)' failed. ==============file:smartctl-after-btrfschk-repair================== smartctl 5.41 2011-06-09 r3365 [x86_64-linux-3.12-0.bpo.1-amd64] (local build) Copyright (C) 2002-11 by Bruce Allen, http://smartmontools.sourceforge.net === START OF READ SMART DATA SECTION === SMART Attributes Data Structure revision number: 10 Vendor Specific SMART Attributes with Thresholds: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 1 Raw_Read_Error_Rate 0x000f 118 099 006 Pre-fail Always - 178377016 3 Spin_Up_Time 0x0003 093 093 000 Pre-fail Always - 0 4 Start_Stop_Count 0x0032 100 100 020 Old_age Always - 7 5 Reallocated_Sector_Ct 0x0033 100 100 010 Pre-fail Always - 0 7 Seek_Error_Rate 0x000f 069 060 030 Pre-fail Always - 9087571 9 Power_On_Hours 0x0032 097 097 000 Old_age Always - 2769 10 Spin_Retry_Count 0x0013 100 100 097 Pre-fail Always - 0 12 Power_Cycle_Count 0x0032 100 100 020 Old_age Always - 7 184 End-to-End_Error 0x0032 100 100 099 Old_age Always - 0 187 Reported_Uncorrect 0x0032 100 100 000 Old_age Always - 0 188 Command_Timeout 0x0032 100 100 000 Old_age Always - 0 189 High_Fly_Writes 0x003a 083 083 000 Old_age Always - 17 190 Airflow_Temperature_Cel 0x0022 077 071 045 Old_age Always - 23 (Min/Max 22/23) 191 G-Sense_Error_Rate 0x0032 100 100 000 Old_age Always - 0 192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age Always - 5 193 Load_Cycle_Count 0x0032 100 100 000 Old_age Always - 7 194 Temperature_Celsius 0x0022 023 040 000 Old_age Always - 23 (0 20 0 0) 197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always - 0 198 Offline_Uncorrectable 0x0010 100 100 000 Old_age Offline - 0 199 UDMA_CRC_Error_Count 0x003e 200 200 000 Old_age Always - 0 2014-02-03 Chris Murphy <lists@colorremedies.com>: > > On Feb 3, 2014, at 1:55 PM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > >> 2014-01-30 Chris Murphy <lists@colorremedies.com>: >>> >>> On Jan 30, 2014, at 10:58 AM, Hugo Mills <hugo@carfax.org.uk> wrote: >>> >>>> On Thu, Jan 30, 2014 at 10:33:21AM -0700, Chris Murphy wrote: >>>>> You're doing an online conversion of a degraded raid1 volume into single? Does anyone know if this is expected or intended to work? >>>> >>>> I don't see why not. One suggested method of recovering RAID from a >>>> degraded situation is to rebalance over just the remaining devices >>>> (space permitting, of course). >>> >>> Right but that's not a conversion. That's a regular balance on a degraded mount, with multiple remaining devices: e.g. a 4 disk raid1, drive fails, mount -o degraded, delete missing, then balance will replicate any missing 2nd copies onto three drives. >>> >>> The bigger problem at the moment is that -o degraded isn't working for Johan. The too many missing devices message seems like a bug and with limited information it may even be whatever that bug is, that cause the conversion to fail. Some 11GB were converted prior to the failure. >> Which usefull information can provide. On the weekend I was at the >> server and found out, that the vanishing of the drive at reboot was >> strange behavior of the bios. So the drive is online again. but the >> filesystem is still showing strange behavior, but now I can mount it >> rw. > > I'd like to see btrfs fi df results for the volume. And new btrfs check. And then a backup if needed, and then a scrub to see if that fixes anything broken between them. I'm not sure what happens if a new generation object is broken and the old generation is OK, what scrub will do? Maybe it just reports it, I'm not sure. If you want you could do a btrfs scrub -r which is read only and just reports what the problems are. > > You also have an incomplete balance, right? So it's possible some things might not be fixable if the conversion to single was successful. You'll need to decide if you want to reconvert back to data/metadata raid1/raid from whatever you're at now. > > > Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-02-03 21:31 ` Johan Kröckel @ 2014-02-07 11:34 ` Johan Kröckel 2014-02-07 17:43 ` Chris Murphy 0 siblings, 1 reply; 25+ messages in thread From: Johan Kröckel @ 2014-02-07 11:34 UTC (permalink / raw) Cc: Btrfs BTRFS Is there anything else I should do with this setup or may I nuke the two partitions and reuse them? 2014-02-03 Johan Kröckel <johan.kroeckel@gmail.com>: > State is: I wont use this filesystem again. I have a backup. So I am > interested to give the necessary information for debuging it and > afterwards format it and create a new one. I already did fscks and > btrfschk --repair and pushed the output to txt-files but they are more > than 4 mb in size. > > So I will post excerpts: > > ============file: btrfsck.out=========================== > Checking filesystem on /dev/mapper/bunkerA > UUID: 7f954a85-7566-4251-832c-44f2d3de2211 > 42 > parent transid verify failed on 1887688011776 wanted 121037 found 88533 > parent transid verify failed on 1888518615040 wanted 121481 found 90267 > parent transid verify failed on 1681394102272 wanted 110919 found 91024 > parent transid verify failed on 1888522838016 wanted 121486 found 90270 > parent transid verify failed on 1888398331904 wanted 121062 found 89987 > leaf parent key incorrect 1887867330560 > bad block 1887867330560 > leaf parent key incorrect 1888120320000 > bad block 1888120320000 > leaf parent key incorrect 1888124637184 > bad block 1888124637184 > leaf parent key incorrect 1888131444736 > bad block 1888131444736 > > [...and so on for 4MB] > > bad block 1888513552384 > leaf parent key incorrect 1888513642496 > bad block 1888513642496 > leaf parent key incorrect 1888513654784 > bad block 1888513654784 > leaf parent key incorrect 1888514023424 > bad block 1888514023424 > btrfsck: cmds-check.c:2212: check_owner_ref: Assertion `!(rec->is_root)' failed. > > ================file: smartctl-before-btrfschk-repair============== > smartctl 5.41 2011-06-09 r3365 [x86_64-linux-3.12-0.bpo.1-amd64] (local build) > Copyright (C) 2002-11 by Bruce Allen, http://smartmontools.sourceforge.net > > === START OF READ SMART DATA SECTION === > SMART Attributes Data Structure revision number: 10 > Vendor Specific SMART Attributes with Thresholds: > ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE > UPDATED WHEN_FAILED RAW_VALUE > 1 Raw_Read_Error_Rate 0x000f 118 099 006 Pre-fail > Always - 172055696 > 3 Spin_Up_Time 0x0003 093 093 000 Pre-fail > Always - 0 > 4 Start_Stop_Count 0x0032 100 100 020 Old_age > Always - 7 > 5 Reallocated_Sector_Ct 0x0033 100 100 010 Pre-fail > Always - 0 > 7 Seek_Error_Rate 0x000f 069 060 030 Pre-fail > Always - 9085642 > 9 Power_On_Hours 0x0032 097 097 000 Old_age > Always - 2769 > 10 Spin_Retry_Count 0x0013 100 100 097 Pre-fail > Always - 0 > 12 Power_Cycle_Count 0x0032 100 100 020 Old_age > Always - 7 > 184 End-to-End_Error 0x0032 100 100 099 Old_age > Always - 0 > 187 Reported_Uncorrect 0x0032 100 100 000 Old_age > Always - 0 > 188 Command_Timeout 0x0032 100 100 000 Old_age > Always - 0 > 189 High_Fly_Writes 0x003a 083 083 000 Old_age > Always - 17 > 190 Airflow_Temperature_Cel 0x0022 077 071 045 Old_age > Always - 23 (Min/Max 22/23) > 191 G-Sense_Error_Rate 0x0032 100 100 000 Old_age > Always - 0 > 192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age > Always - 5 > 193 Load_Cycle_Count 0x0032 100 100 000 Old_age > Always - 7 > 194 Temperature_Celsius 0x0022 023 040 000 Old_age > Always - 23 (0 20 0 0) > 197 Current_Pending_Sector 0x0012 100 100 000 Old_age > Always - 0 > 198 Offline_Uncorrectable 0x0010 100 100 000 Old_age > Offline - 0 > 199 UDMA_CRC_Error_Count 0x003e 200 200 000 Old_age > Always - 0 > > =================file:btrfsck-repair.out======================== > enabling repair mode > Checking filesystem on /dev/mapper/bunkerA > UUID: 7f954a85-7566-4251-832c-44f2d3de2211 > ify failed on 1887688011776 wanted 121037 found 88533 > parent transid verify failed on 1888518615040 wanted 121481 found 90267 > parent transid verify failed on 1681394102272 wanted 110919 found 91024 > parent transid verify failed on 1888522838016 wanted 121486 found 90270 > parent transid verify failed on 1888398331904 wanted 121062 found 89987 > leaf parent key incorrect 1887867330560 > bad block 1887867330560 > > [...and so on for 4MB] > > bad block 1888513642496 > leaf parent key incorrect 1888513654784 > bad block 1888513654784 > leaf parent key incorrect 1888514023424 > bad block 1888514023424 > btrfsck: cmds-check.c:2212: check_owner_ref: Assertion `!(rec->is_root)' failed. > > ==============file:smartctl-after-btrfschk-repair================== > smartctl 5.41 2011-06-09 r3365 [x86_64-linux-3.12-0.bpo.1-amd64] (local build) > Copyright (C) 2002-11 by Bruce Allen, http://smartmontools.sourceforge.net > > === START OF READ SMART DATA SECTION === > SMART Attributes Data Structure revision number: 10 > Vendor Specific SMART Attributes with Thresholds: > ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE > UPDATED WHEN_FAILED RAW_VALUE > 1 Raw_Read_Error_Rate 0x000f 118 099 006 Pre-fail > Always - 178377016 > 3 Spin_Up_Time 0x0003 093 093 000 Pre-fail > Always - 0 > 4 Start_Stop_Count 0x0032 100 100 020 Old_age > Always - 7 > 5 Reallocated_Sector_Ct 0x0033 100 100 010 Pre-fail > Always - 0 > 7 Seek_Error_Rate 0x000f 069 060 030 Pre-fail > Always - 9087571 > 9 Power_On_Hours 0x0032 097 097 000 Old_age > Always - 2769 > 10 Spin_Retry_Count 0x0013 100 100 097 Pre-fail > Always - 0 > 12 Power_Cycle_Count 0x0032 100 100 020 Old_age > Always - 7 > 184 End-to-End_Error 0x0032 100 100 099 Old_age > Always - 0 > 187 Reported_Uncorrect 0x0032 100 100 000 Old_age > Always - 0 > 188 Command_Timeout 0x0032 100 100 000 Old_age > Always - 0 > 189 High_Fly_Writes 0x003a 083 083 000 Old_age > Always - 17 > 190 Airflow_Temperature_Cel 0x0022 077 071 045 Old_age > Always - 23 (Min/Max 22/23) > 191 G-Sense_Error_Rate 0x0032 100 100 000 Old_age > Always - 0 > 192 Power-Off_Retract_Count 0x0032 100 100 000 Old_age > Always - 5 > 193 Load_Cycle_Count 0x0032 100 100 000 Old_age > Always - 7 > 194 Temperature_Celsius 0x0022 023 040 000 Old_age > Always - 23 (0 20 0 0) > 197 Current_Pending_Sector 0x0012 100 100 000 Old_age > Always - 0 > 198 Offline_Uncorrectable 0x0010 100 100 000 Old_age > Offline - 0 > 199 UDMA_CRC_Error_Count 0x003e 200 200 000 Old_age > Always - 0 > > 2014-02-03 Chris Murphy <lists@colorremedies.com>: >> >> On Feb 3, 2014, at 1:55 PM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: >> >>> 2014-01-30 Chris Murphy <lists@colorremedies.com>: >>>> >>>> On Jan 30, 2014, at 10:58 AM, Hugo Mills <hugo@carfax.org.uk> wrote: >>>> >>>>> On Thu, Jan 30, 2014 at 10:33:21AM -0700, Chris Murphy wrote: >>>>>> You're doing an online conversion of a degraded raid1 volume into single? Does anyone know if this is expected or intended to work? >>>>> >>>>> I don't see why not. One suggested method of recovering RAID from a >>>>> degraded situation is to rebalance over just the remaining devices >>>>> (space permitting, of course). >>>> >>>> Right but that's not a conversion. That's a regular balance on a degraded mount, with multiple remaining devices: e.g. a 4 disk raid1, drive fails, mount -o degraded, delete missing, then balance will replicate any missing 2nd copies onto three drives. >>>> >>>> The bigger problem at the moment is that -o degraded isn't working for Johan. The too many missing devices message seems like a bug and with limited information it may even be whatever that bug is, that cause the conversion to fail. Some 11GB were converted prior to the failure. >>> Which usefull information can provide. On the weekend I was at the >>> server and found out, that the vanishing of the drive at reboot was >>> strange behavior of the bios. So the drive is online again. but the >>> filesystem is still showing strange behavior, but now I can mount it >>> rw. >> >> I'd like to see btrfs fi df results for the volume. And new btrfs check. And then a backup if needed, and then a scrub to see if that fixes anything broken between them. I'm not sure what happens if a new generation object is broken and the old generation is OK, what scrub will do? Maybe it just reports it, I'm not sure. If you want you could do a btrfs scrub -r which is read only and just reports what the problems are. >> >> You also have an incomplete balance, right? So it's possible some things might not be fixable if the conversion to single was successful. You'll need to decide if you want to reconvert back to data/metadata raid1/raid from whatever you're at now. >> >> >> Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-02-07 11:34 ` Johan Kröckel @ 2014-02-07 17:43 ` Chris Murphy 2014-02-08 11:09 ` Johan Kröckel 0 siblings, 1 reply; 25+ messages in thread From: Chris Murphy @ 2014-02-07 17:43 UTC (permalink / raw) To: Johan Kröckel; +Cc: Btrfs BTRFS On Feb 7, 2014, at 4:34 AM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > Is there anything else I should do with this setup or may I nuke the > two partitions and reuse them? Well I'm pretty sure once you run 'btrfs check --repair' that you've hit the end of the road. Possibly btrfs restore can still extract some files, it might be worth testing whether that works. Otherwise blow it away. I'd say test with 3.14-rc2 with a new file system and see if you can reproduce the sequence that caused this problem in the first place. If it's reproducible, I think there's a bug here somewhere. Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-02-07 17:43 ` Chris Murphy @ 2014-02-08 11:09 ` Johan Kröckel 2014-02-09 5:40 ` Duncan 0 siblings, 1 reply; 25+ messages in thread From: Johan Kröckel @ 2014-02-08 11:09 UTC (permalink / raw) To: Chris Murphy; +Cc: Btrfs BTRFS Ok, I did nuke it now and created the fs again using 3.12 kernel. So far so good. Runs fine. Finally, I know its kind of offtopic, but can some help me interpreting this (I think this is the error in the smart-log which started the whole mess)? Error 1 occurred at disk power-on lifetime: 2576 hours (107 days + 8 hours) When the command that caused the error occurred, the device was active or idle. After command completion occurred, registers were: ER ST SC SN CL CH DH -- -- -- -- -- -- -- 04 71 00 ff ff ff 0f Device Fault; Error: ABRT at LBA = 0x0fffffff = 268435455 Commands leading to the command that caused the error were: CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name -- -- -- -- -- -- -- -- ---------------- -------------------- 61 00 08 ff ff ff 4f 00 5d+04:53:11.169 WRITE FPDMA QUEUED 61 00 08 80 18 00 40 00 5d+04:52:45.129 WRITE FPDMA QUEUED 61 00 08 ff ff ff 4f 00 5d+04:52:44.701 WRITE FPDMA QUEUED 61 00 08 ff ff ff 4f 00 5d+04:52:44.700 WRITE FPDMA QUEUED 61 00 08 ff ff ff 4f 00 5d+04:52:44.679 WRITE FPDMA QUEUED 2014-02-07 Chris Murphy <lists@colorremedies.com>: > > On Feb 7, 2014, at 4:34 AM, Johan Kröckel <johan.kroeckel@gmail.com> wrote: > >> Is there anything else I should do with this setup or may I nuke the >> two partitions and reuse them? > > Well I'm pretty sure once you run 'btrfs check --repair' that you've hit the end of the road. Possibly btrfs restore can still extract some files, it might be worth testing whether that works. > > Otherwise blow it away. I'd say test with 3.14-rc2 with a new file system and see if you can reproduce the sequence that caused this problem in the first place. If it's reproducible, I think there's a bug here somewhere. > > > Chris Murphy ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-02-08 11:09 ` Johan Kröckel @ 2014-02-09 5:40 ` Duncan 2014-02-10 14:05 ` Johan Kröckel 0 siblings, 1 reply; 25+ messages in thread From: Duncan @ 2014-02-09 5:40 UTC (permalink / raw) To: linux-btrfs Johan Kröckel posted on Sat, 08 Feb 2014 12:09:46 +0100 as excerpted: > Ok, I did nuke it now and created the fs again using 3.12 kernel. So far > so good. Runs fine. > Finally, I know its kind of offtopic, but can some help me interpreting > this (I think this is the error in the smart-log which started the whole > mess)? > > Error 1 occurred at disk power-on lifetime: 2576 hours (107 days + 8 > hours) > When the command that caused the error occurred, the device was > active or idle. > > After command completion occurred, registers were: > ER ST SC SN CL CH DH > -- -- -- -- -- -- -- > 04 71 00 ff ff ff 0f > Device Fault; Error: ABRT at LBA = 0x0fffffff = 268435455 I'm no SMART expert, but that LBA number is incredibly suspicious. With standard 512-byte sectors that's the 128 GiB boundary, the old 28-bit LBA limit (LBA28, introduced with ATA-1 in 1994, modern drives are LBA48, introduced in 2003 with ATA-6 and offering an addressing capacity of 128 PiB, according to wikipedia's article on LBA). It looks like something flipped back to LBA28, and when a continuing operation happened to write past that value... it triggered the abort you see in the SMART log. Double-check your BIOS to be sure it didn't somehow revert to the old LBA28 compatibility mode or some such, and the drives, to make sure they aren't "clipped" to LBA28 compatibility mode as well. -- Duncan - List replies preferred. No HTML msgs. "Every nonfree program has a lord, a master -- and if you use the program, he is your master." Richard Stallman ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-02-09 5:40 ` Duncan @ 2014-02-10 14:05 ` Johan Kröckel 0 siblings, 0 replies; 25+ messages in thread From: Johan Kröckel @ 2014-02-10 14:05 UTC (permalink / raw) Cc: Btrfs BTRFS Thanks, that explains something. There was indeed a BIOS-Problem (the drive vanished was disabled in BIOS suddenly and was only useable again after reactivating it in the BIOS again). So should have been a BIOS-problem. 2014-02-09 Duncan <1i5t5.duncan@cox.net>: > Johan Kröckel posted on Sat, 08 Feb 2014 12:09:46 +0100 as excerpted: > >> Ok, I did nuke it now and created the fs again using 3.12 kernel. So far >> so good. Runs fine. >> Finally, I know its kind of offtopic, but can some help me interpreting >> this (I think this is the error in the smart-log which started the whole >> mess)? >> >> Error 1 occurred at disk power-on lifetime: 2576 hours (107 days + 8 >> hours) >> When the command that caused the error occurred, the device was >> active or idle. >> >> After command completion occurred, registers were: >> ER ST SC SN CL CH DH >> -- -- -- -- -- -- -- >> 04 71 00 ff ff ff 0f >> Device Fault; Error: ABRT at LBA = 0x0fffffff = 268435455 > > I'm no SMART expert, but that LBA number is incredibly suspicious. With > standard 512-byte sectors that's the 128 GiB boundary, the old 28-bit LBA > limit (LBA28, introduced with ATA-1 in 1994, modern drives are LBA48, > introduced in 2003 with ATA-6 and offering an addressing capacity of 128 > PiB, according to wikipedia's article on LBA). > > It looks like something flipped back to LBA28, and when a continuing > operation happened to write past that value... it triggered the abort you > see in the SMART log. > > Double-check your BIOS to be sure it didn't somehow revert to the old > LBA28 compatibility mode or some such, and the drives, to make sure they > aren't "clipped" to LBA28 compatibility mode as well. > > -- > Duncan - List replies preferred. No HTML msgs. > "Every nonfree program has a lord, a master -- > and if you use the program, he is your master." Richard Stallman > > -- > To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 25+ messages in thread
[parent not found: <CABgvyo822bOAHeA1GH28MPaBAU+Zdi72MD_uwL+dhopt+nwMig@mail.gmail.com>]
* Fwd: lost with degraded RAID1 [not found] ` <CABgvyo822bOAHeA1GH28MPaBAU+Zdi72MD_uwL+dhopt+nwMig@mail.gmail.com> @ 2014-01-31 6:28 ` Johan Kröckel 0 siblings, 0 replies; 25+ messages in thread From: Johan Kröckel @ 2014-01-31 6:28 UTC (permalink / raw) To: linux-btrfs 2014-01-30 Chris Murphy <lists@colorremedies.com>: >> Now the damaged disk disappeared completely. No sign of it in /proc/diskstats. > > I don't know what you mean by it's disappeared complete. Are you talking about the physical block device? Or the dm logical block device? It failed and doesn't appear anywhere. I don't have physical access to the server at the moment. > What do you get for: > lsblk > blkid > btrfs device scan --all-devices > btrfs fi show root@fortknox:/# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 232,9G 0 disk └─sda1 8:1 0 232,9G 0 part / sdb 8:16 0 1,8T 0 disk └─sdb1 8:17 0 1,8T 0 part └─bunkerA (dm-0) 253:0 0 1,8T 0 crypt sdc 8:32 0 2,7T 0 disk └─sdc1 8:33 0 2,7T 0 part /USBPLATTE loop1 7:1 0 500G 0 loop root@fortknox:/# blkid /dev/sda1: UUID="d055ffe7-80ed-455e-8117-4a2c8bef3ed5" TYPE="ext4" /dev/sdc1: UUID="812c784b-df5a-48a6-98ff-5b755e29b563" TYPE="ext4" /dev/sdb1: UUID="947206e2-488e-4dc4-b243-912901512ed4" TYPE="crypto_LUKS" /dev/mapper/bunkerA: LABEL="bunker" UUID="7f954a85-7566-4251-832c-44f2d3de2211" UUID_SUB="6c66c3c1-df8f-4b29-9acb-bd923a69b3ee" TYPE="btrfs" root@fortknox:/# btrfs device scan --all-devices Scanning for Btrfs filesystems root@fortknox:/# btrfs fi show Label: 'bunker' uuid: 7f954a85-7566-4251-832c-44f2d3de2211 Total devices 2 FS bytes used 1.58TiB devid 2 size 1.82TiB used 1.59TiB path /dev/mapper/bunkerA *** Some devices missing Btrfs v3.12 >> But now, I can't mount bunkerA degraded,RW because degraded >> filesystems are not allowed to be mounted RW (?) > > Have you tried? What errors do you get in user space and dmesg? root@fortknox:/# umount /mnt umount: /mnt: not mounted root@fortknox:/# mount -odegraded,rw /dev/mapper/bunkerA /mnt mount: wrong fs type, bad option, bad superblock on /dev/mapper/bunkerA, missing codepage or helper program, or other error In some cases useful info is found in syslog - try dmesg | tail or so root@fortknox:/# dmesg|tail [756751.574065] btrfs: bdev (null) errs: wr 1375454, rd 186442, flush 0, corrupt 0, gen 0 [756751.837133] Btrfs: too many missing devices, writeable mount is not allowed [756751.865278] btrfs: open_ctree failed [781332.902000] btrfs: device label bunker devid 2 transid 121487 /dev/dm-0 [781630.561889] btrfs: device label bunker devid 2 transid 121487 /dev/mapper/bunkerA [781630.564673] btrfs: allowing degraded mounts [781630.564684] btrfs: disk space caching is enabled [781630.598155] btrfs: bdev (null) errs: wr 1375454, rd 186442, flush 0, corrupt 0, gen 0 [781630.816989] Btrfs: too many missing devices, writeable mount is not allowed [781630.842901] btrfs: open_ctree failed ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: lost with degraded RAID1 2014-01-29 19:16 lost with degraded RAID1 Johan Kröckel ` (2 preceding siblings ...) 2014-01-30 17:33 ` Chris Murphy @ 2014-01-31 6:40 ` Johan Kröckel 3 siblings, 0 replies; 25+ messages in thread From: Johan Kröckel @ 2014-01-31 6:40 UTC (permalink / raw) To: linux-btrfs Sorry, I failed using this list in conjunction with gmail and send most answers to respectively duncan and chris directly. So I sum it up here again: > What do you get for: > lsblk > blkid > btrfs device scan --all-devices > btrfs fi show root@fortknox:/# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 232,9G 0 disk └─sda1 8:1 0 232,9G 0 part / sdb 8:16 0 1,8T 0 disk └─sdb1 8:17 0 1,8T 0 part └─bunkerA (dm-0) 253:0 0 1,8T 0 crypt sdc 8:32 0 2,7T 0 disk └─sdc1 8:33 0 2,7T 0 part /USBPLATTE loop1 7:1 0 500G 0 loop root@fortknox:/# blkid /dev/sda1: UUID="d055ffe7-80ed-455e-8117-4a2c8bef3ed5" TYPE="ext4" /dev/sdc1: UUID="812c784b-df5a-48a6-98ff-5b755e29b563" TYPE="ext4" /dev/sdb1: UUID="947206e2-488e-4dc4-b243-912901512ed4" TYPE="crypto_LUKS" /dev/mapper/bunkerA: LABEL="bunker" UUID="7f954a85-7566-4251-832c-44f2d3de2211" UUID_SUB="6c66c3c1-df8f-4b29-9acb-bd923a69b3ee" TYPE="btrfs" root@fortknox:/# btrfs device scan --all-devices Scanning for Btrfs filesystems root@fortknox:/# btrfs fi show Label: 'bunker' uuid: 7f954a85-7566-4251-832c-44f2d3de2211 Total devices 2 FS bytes used 1.58TiB devid 2 size 1.82TiB used 1.59TiB path /dev/mapper/bunkerA *** Some devices missing Btrfs v3.12 >> But now, I can't mount bunkerA degraded,RW because degraded >> filesystems are not allowed to be mounted RW (?) > > Have you tried? What errors do you get in user space and dmesg? root@fortknox:/# umount /mnt umount: /mnt: not mounted root@fortknox:/# mount -odegraded,rw /dev/mapper/bunkerA /mnt mount: wrong fs type, bad option, bad superblock on /dev/mapper/bunkerA, missing codepage or helper program, or other error In some cases useful info is found in syslog - try dmesg | tail or so root@fortknox:/# dmesg|tail [756751.574065] btrfs: bdev (null) errs: wr 1375454, rd 186442, flush 0, corrupt 0, gen 0 [756751.837133] Btrfs: too many missing devices, writeable mount is not allowed [756751.865278] btrfs: open_ctree failed [781332.902000] btrfs: device label bunker devid 2 transid 121487 /dev/dm-0 [781630.561889] btrfs: device label bunker devid 2 transid 121487 /dev/mapper/bunkerA [781630.564673] btrfs: allowing degraded mounts [781630.564684] btrfs: disk space caching is enabled [781630.598155] btrfs: bdev (null) errs: wr 1375454, rd 186442, flush 0, corrupt 0, gen 0 [781630.816989] Btrfs: too many missing devices, writeable mount is not allowed [781630.842901] btrfs: open_ctree failed root@fortknox:/# mount -odegraded,recovery,skip_balance /dev/mapper/bunkerA /mnt mount: wrong fs type, bad option, bad superblock on /dev/mapper/bunkerA, missing codepage or helper program, or other error In some cases useful info is found in syslog - try dmesg | tail or so root@fortknox:/# dmesg|tail [781630.598155] btrfs: bdev (null) errs: wr 1375454, rd 186442, flush 0, corrupt 0, gen 0 [781630.816989] Btrfs: too many missing devices, writeable mount is not allowed [781630.842901] btrfs: open_ctree failed [781781.311741] btrfs: device label bunker devid 2 transid 121487 /dev/mapper/bunkerA [781781.314204] btrfs: allowing degraded mounts [781781.314219] btrfs: enabling auto recovery [781781.314227] btrfs: disk space caching is enabled [781781.348118] btrfs: bdev (null) errs: wr 1375454, rd 186442, flush 0, corrupt 0, gen 0 [781781.572222] Btrfs: too many missing devices, writeable mount is not allowed [781781.597722] btrfs: open_ctree failed But (i think) read and write errors are not increasing. I dont know whether this is because of the missing drive or because the filesystem has never been mounted writeable since a point in time. root@fortknox:~# btrfs-show-super -a /dev/mapper/bunkerA superblock: bytenr=65536, device=/dev/mapper/bunkerA --------------------------------------------------------- csum 0x40e23443 [match] bytenr 65536 flags 0x1 magic _BHRfS_M [match] fsid 7f954a85-7566-4251-832c-44f2d3de2211 label bunker generation 121487 root 1888523018240 sys_array_size 129 chunk_root_generation 121477 root_level 2 chunk_root 3304906031104 chunk_root_level 1 log_root 0 log_root_transid 0 log_root_level 0 total_bytes 4000791559680 bytes_used 1743686074368 sectorsize 4096 nodesize 4096 leafsize 4096 stripesize 4096 root_dir 6 num_devices 2 compat_flags 0x0 compat_ro_flags 0x0 incompat_flags 0x1 csum_type 0 csum_size 4 cache_generation 121487 uuid_tree_generation 121487 dev_item.uuid 6c66c3c1-df8f-4b29-9acb-bd923a69b3ee dev_item.fsid 7f954a85-7566-4251-832c-44f2d3de2211 [match] dev_item.type 0 dev_item.total_bytes 2000395771392 dev_item.bytes_used 1746986336256 dev_item.io_align 4096 dev_item.io_width 4096 dev_item.sector_size 4096 dev_item.devid 2 dev_item.dev_group 0 dev_item.seek_speed 0 dev_item.bandwidth 0 dev_item.generation 0 superblock: bytenr=67108864, device=/dev/mapper/bunkerA --------------------------------------------------------- csum 0xe0831c8d [match] bytenr 67108864 flags 0x1 magic _BHRfS_M [match] fsid 7f954a85-7566-4251-832c-44f2d3de2211 label bunker generation 121487 root 1888523018240 sys_array_size 129 chunk_root_generation 121477 root_level 2 chunk_root 3304906031104 chunk_root_level 1 log_root 0 log_root_transid 0 log_root_level 0 total_bytes 4000791559680 bytes_used 1743686074368 sectorsize 4096 nodesize 4096 leafsize 4096 stripesize 4096 root_dir 6 num_devices 2 compat_flags 0x0 compat_ro_flags 0x0 incompat_flags 0x1 csum_type 0 csum_size 4 cache_generation 121487 uuid_tree_generation 121487 dev_item.uuid 6c66c3c1-df8f-4b29-9acb-bd923a69b3ee dev_item.fsid 7f954a85-7566-4251-832c-44f2d3de2211 [match] dev_item.type 0 dev_item.total_bytes 2000395771392 dev_item.bytes_used 1746986336256 dev_item.io_align 4096 dev_item.io_width 4096 dev_item.sector_size 4096 dev_item.devid 2 dev_item.dev_group 0 dev_item.seek_speed 0 dev_item.bandwidth 0 dev_item.generation 0 superblock: bytenr=274877906944, device=/dev/mapper/bunkerA --------------------------------------------------------- csum 0x1d044abc [match] bytenr 274877906944 flags 0x1 magic _BHRfS_M [match] fsid 7f954a85-7566-4251-832c-44f2d3de2211 label bunker generation 121487 root 1888523018240 sys_array_size 129 chunk_root_generation 121477 root_level 2 chunk_root 3304906031104 chunk_root_level 1 log_root 0 log_root_transid 0 log_root_level 0 total_bytes 4000791559680 bytes_used 1743686074368 sectorsize 4096 nodesize 4096 leafsize 4096 stripesize 4096 root_dir 6 num_devices 2 compat_flags 0x0 compat_ro_flags 0x0 incompat_flags 0x1 csum_type 0 csum_size 4 cache_generation 121487 uuid_tree_generation 121487 dev_item.uuid 6c66c3c1-df8f-4b29-9acb-bd923a69b3ee dev_item.fsid 7f954a85-7566-4251-832c-44f2d3de2211 [match] dev_item.type 0 dev_item.total_bytes 2000395771392 dev_item.bytes_used 1746986336256 dev_item.io_align 4096 dev_item.io_width 4096 dev_item.sector_size 4096 dev_item.devid 2 dev_item.dev_group 0 dev_item.seek_speed 0 dev_item.bandwidth 0 dev_item.generation 0 root@fortknox:~# btrfsck /dev/mapper/bunkerA warning, device 1 is missing warning devid 1 not found already Checking filesystem on /dev/mapper/bunkerA UUID: 7f954a85-7566-4251-832c-44f2d3de2211 checking extents checking free space cache Error reading 1647012864000, -1 Error reading 1647819816960, -1 Error reading 1647965962240, -1 Error reading 1647006314496, -1 Error reading 2778306510848, -1 Error reading 1655481655296, -1 Error reading 1655482179584, -1 Error reading 1655493296128, -1 Error reading 1788748169216, -1 Error reading 1644927381504, -1 Error reading 1671399948288, -1 Error reading 1646999830528, -1 Error reading 1647013462016, -1 Error reading 1738282172416, -1 Error reading 1648014381056, -1 Error reading 1655474946048, -1 Error reading 2778315161600, -1 failed to load free space cache for block group 1644867616768failed to load free space cache for block group 1645941358592failed to load free space cache for block group 1647015100416failed to load free space cache for block group 1654531293184failed to load free space cache for block group 1657752518656failed to load free space cache for block group 1658826260480failed to load free space cache for block group 1659900002304failed to load free space cache for block group 1667416195072failed to load free space cache for block group 1670637420544failed to load free space cache for block group 1676006129664failed to load free space cache for block group 1681374838784failed to load free space cache for block group 1695333482496failed to load free space cache for block group 1698554707968failed to load free space cache for block group 1699628449792failed to load free space cache for block group 1700702191616failed to load free space cache for block group 1703923417088failed to load free space cache for block group Error reading 2778315685888, -1 Error reading 2778315948032, -1 Error reading 2778316210176, -1 Error reading 2778316472320, -1 Error reading 2778317520896, -1 Error reading 2778318307328, -1 Error reading 2778318569472, -1 Error reading 2778320142336, -1 Error reading 2778320404480, -1 Error reading 2778320666624, -1 Error reading 1937998594048, -1 Error reading 1646997389312, -1 Error reading 1794117312512, -1 Error reading 1647971205120, -1 Error reading 1655480606720, -1 Error reading 1655481917440, -1 Error reading 1968063332352, -1 1715734577152failed to load free space cache for block group 1726471995392failed to load free space cache for block group 1742578122752failed to load free space cache for block group 1745799348224failed to load free space cache for block group 1752241799168failed to load free space cache for block group 1753315540992failed to load free space cache for block group 1758684250112failed to load free space cache for block group 1761905475584failed to load free space cache for block group 1762979217408failed to load free space cache for block group 1766200442880failed to load free space cache for block group 1774790377472failed to load free space cache for block group 1775864119296failed to load free space cache for block group 1780159086592failed to load free space cache for block group 1781232828416failed to load free space cache for block group 1782306570240failed to load free space cache for block group 1783380312064failed to load free space cache for block group 1786601537536failed to load free space cache forError reading 1655482441728, -1 Error reading 1655482949632, -1 Error reading 1975579820032, -1 Error reading 1646999044096, -1 Error reading 2099058638848, -1 Error reading 1655506403328, -1 Error reading 1661504389120, -1 Error reading 1747946557440, -1 Error reading 1752241340416, -1 Error reading 1766200131584, -1 Error reading 1769420718080, -1 Error reading 1776937304064, -1 Error reading 1778010673152, -1 Error reading 1778010935296, -1 Error reading 1784452825088, -1 Error reading 1786600423424, -1 Error reading 1787673284608, -1 block group 1787675279360failed to load free space cache for block group 1788749021184failed to load free space cache for block group 1789822763008failed to load free space cache for block group 1790896504832failed to load free space cache for block group 1791970246656failed to load free space cache for block group 1794117730304failed to load free space cache for block group 1795191472128failed to load free space cache for block group 1796265213952failed to load free space cache for block group 1797338955776failed to load free space cache for block group 1798412697600failed to load free space cache for block group 1799486439424failed to load free space cache for block group 1800560181248failed to load free space cache for block group 1801633923072failed to load free space cache for block group 1802707664896failed to load free space cache for block group 1803781406720failed to load free space cache for block group 1804855148544failed to load free space cache for block group 1805928890368failed to load free spError reading 1787674857472, -1 Error reading 1768347648000, -1 Error reading 1645348323328, -1 Error reading 1671399424000, -1 Error reading 1647011557376, -1 Error reading 1661503520768, -1 Error reading 1811296747520, -1 Error reading 1992759660544, -1 Error reading 2778269958144, -1 Error reading 2099059163136, -1 Error reading 3065128484864, -1 Error reading 1720028954624, -1 Error reading 1778011197440, -1 Error reading 1920818544640, -1 Error reading 1937998856192, -1 Error reading 1947662532608, -1 Error reading 2474869772288, -1 ace cache for block group 1807002632192failed to load free space cache for block group 1808076374016failed to load free space cache for block group 1810223857664failed to load free space cache for block group 1813445083136failed to load free space cache for block group 1815592566784failed to load free space cache for block group 1816666308608failed to load free space cache for block group 1818813792256failed to load free space cache for block group 1820961275904failed to load free space cache for block group 1827403726848failed to load free space cache for block group 1828477468672failed to load free space cache for block group 1829551210496failed to load free space cache for block group 1830624952320failed to load free space cache for block group 1832772435968failed to load free space cache for block group 1834919919616failed to load free space cache for block group 1835993661440failed to load free space cache for block group 1843509854208failed to load free space cache for block group 1844583596032failed toError reading 1721102303232, -1 Error reading 1957325766656, -1 Error reading 2164558200832, -1 Error reading 1912229044224, -1 Error reading 3099322155008, -1 Error reading 2778243129344, -1 Error reading 2150599700480, -1 Error reading 2163484598272, -1 Error reading 1722176700416, -1 Error reading 3099324702720, -1 Error reading 2099058900992, -1 Error reading 2778206109696, -1 Error reading 2778271531008, -1 Error reading 1646997127168, -1 Error reading 2499565780992, -1 Error reading 2099059609600, -1 load free space cache for block group 1845657337856failed to load free space cache for block group 1846731079680failed to load free space cache for block group 1849952305152failed to load free space cache for block group 1851026046976failed to load free space cache for block group 1854247272448failed to load free space cache for block group 1857468497920failed to load free space cache for block group 1858542239744failed to load free space cache for block group 1859615981568failed to load free space cache for block group 1863910948864failed to load free space cache for block group 1866058432512failed to load free space cache for block group 1872500883456failed to load free space cache for block group 1874648367104failed to load free space cache for block group 1880017076224failed to load free space cache for block group 1882164559872failed to load free space cache for block group 1887533268992failed to load free space cache for block group 1888607010816failed to load free space cache for block group 188968075Error reading 2080806600704, -1 Error reading 1671398899712, -1 Error reading 2257973387264, -1 Error reading 2173148192768, -1 Error reading 2500639551488, -1 Error reading 3222112632832, -1 Error reading 1941220032512, -1 Error reading 2049667981312, -1 Error reading 1646987771904, -1 Error reading 1743650672640, -1 Error reading 1943366107136, -1 Error reading 2778275590144, -1 Error reading 1644910018560, -1 Error reading 2778272055296, -1 Error reading 2778272317440, -1 Error reading 3211456638976, -1 Error reading 1749019860992, -1 2640failed to load free space cache for block group 1890754494464failed to load free space cache for block group 1891828236288failed to load free space cache for block group 1893975719936failed to load free space cache for block group 1895049461760failed to load free space cache for block group 1909008105472failed to load free space cache for block group 1911155589120failed to load free space cache for block group 1912229330944failed to load free space cache for block group 1917598040064failed to load free space cache for block group 1919745523712failed to load free space cache for block group 1925114232832failed to load free space cache for block group 1931556683776failed to load free space cache for block group 1936925392896failed to load free space cache for block group 1937999134720failed to load free space cache for block group 1939072876544failed to load free space cache for block group 1940146618368failed to load free space cache for block group 1944441585664failed to load free space cache for block grError reading 2147378200576, -1 Error reading 1959473307648, -1 Error reading 2109797539840, -1 Error reading 2398634049536, -1 Error reading 3075524984832, -1 Error reading 2873227935744, -1 Error reading 2139862138880, -1 Error reading 2895542550528, -1 Error reading 2165631942656, -1 Error reading 2778217775104, -1 Error reading 2778272841728, -1 Error reading 2235425173504, -1 Error reading 2778277281792, -1 Error reading 2150599438336, -1 Error reading 2313807921152, -1 Error reading 1945514999808, -1 Error reading 2538220527616, -1 oup 1945515327488failed to load free space cache for block group 1946589069312failed to load free space cache for block group 1953031520256failed to load free space cache for block group 1956252745728failed to load free space cache for block group 1958400229376failed to load free space cache for block group 1959473971200failed to load free space cache for block group 1960547713024failed to load free space cache for block group 1966990163968failed to load free space cache for block group 1971285131264failed to load free space cache for block group 1974506356736failed to load free space cache for block group 1978801324032failed to load free space cache for block group 1983096291328failed to load free space cache for block group 1985243774976failed to load free space cache for block group 1987391258624failed to load free space cache for block group 1990612484096failed to load free space cache for block group 1991686225920failed to load free space cache for block group 2005644869632failed to load free space cacheError reading 2645594722304, -1 Error reading 1782306279424, -1 Error reading 2944094769152, -1 Error reading 1647003238400, -1 Error reading 2778243915776, -1 Error reading 2778210566144, -1 Error reading 2778244177920, -1 Error reading 2273005338624, -1 Error reading 2256900136960, -1 Error reading 3203940081664, -1 Error reading 2280522493952, -1 Error reading 3257540214784, -1 Error reading 2596202586112, -1 Error reading 2240794132480, -1 Error reading 2778308608000, -1 Error reading 2849604960256, -1 Error reading 3005297262592, -1 for block group 2008866095104failed to load free space cache for block group 2028193447936failed to load free space cache for block group 2030340931584failed to load free space cache for block group 2035709640704failed to load free space cache for block group 2038930866176failed to load free space cache for block group 2043225833472failed to load free space cache for block group 2046447058944failed to load free space cache for block group 2047520800768failed to load free space cache for block group 2048594542592failed to load free space cache for block group 2057184477184failed to load free space cache for block group 2061479444480failed to load free space cache for block group 2064700669952failed to load free space cache for block group 2065774411776failed to load free space cache for block group 2068995637248failed to load free space cache for block group 2071143120896failed to load free space cache for block group 2072216862720failed to load free space cache for block group 2082954280960failed to load freError reading 1647005265920, -1 Error reading 3051468955648, -1 Error reading 2273006305280, -1 Error reading 3099324964864, -1 Error reading 2276227534848, -1 Error reading 2247236485120, -1 Error reading 2908661415936, -1 Error reading 3060058935296, -1 Error reading 2778244440064, -1 Error reading 2987044610048, -1 Error reading 3337841999872, -1 Error reading 1751167737856, -1 Error reading 1647006646272, -1 Error reading 1825255927808, -1 Error reading 3099325227008, -1 Error reading 3133073588224, -1 Error reading 2842089160704, -1 e space cache for block group 2084028022784failed to load free space cache for block group 2085101764608failed to load free space cache for block group 2090470473728failed to load free space cache for block group 2092617957376failed to load free space cache for block group 2093691699200failed to load free space cache for block group 2096912924672failed to load free space cache for block group 2097986666496failed to load free space cache for block group 2102281633792failed to load free space cache for block group 2108724084736failed to load free space cache for block group 2109797826560failed to load free space cache for block group 2110871568384failed to load free space cache for block group 2113019052032failed to load free space cache for block group 2116240277504failed to load free space cache for block group 2121608986624failed to load free space cache for block group 2123756470272failed to load free space cache for block group 2136641372160failed to load free space cache for block group 2146305048576faileError reading 3203940343808, -1 Error reading 2300923604992, -1 Error reading 2842089422848, -1 Error reading 3209145024512, -1 Error reading 2945168691200, -1 Error reading 2924767465472, -1 Error reading 2917251407872, -1 Error reading 2353536950272, -1 Error reading 1809149812736, -1 Error reading 3101692198912, -1 Error reading 2778244702208, -1 Error reading 3101692461056, -1 Error reading 3101692723200, -1 Error reading 2778267860992, -1 Error reading 3099325489152, -1 Error reading 2598350065664, -1 d to load free space cache for block group 2148452532224failed to load free space cache for block group 2149526274048failed to load free space cache for block group 2154894983168failed to load free space cache for block group 2155968724992failed to load free space cache for block group 2157042466816failed to load free space cache for block group 2159189950464failed to load free space cache for block group 2162411175936failed to load free space cache for block group 2163484917760failed to load free space cache for block group 2164558659584failed to load free space cache for block group 2167779885056failed to load free space cache for block group 2168853626880failed to load free space cache for block group 2169927368704failed to load free space cache for block group 2172074852352failed to load free space cache for block group 2173148594176failed to load free space cache for block group 2175296077824failed to load free space cache for block group 2181738528768failed to load free space cache for block group 21849Error reading 1647008669696, -1 Error reading 2778218299392, -1 Error reading 2778245226496, -1 Error reading 3099326013440, -1 Error reading 2275153801216, -1 Error reading 3078046679040, -1 Error reading 1646993993728, -1 Error reading 3101693247488, -1 Error reading 2999929532416, -1 Error reading 2972012052480, -1 Error reading 2778245488640, -1 Error reading 2778206633984, -1 Error reading 3099327062016, -1 Error reading 3099328110592, -1 Error reading 3099328372736, -1 Error reading 2368569409536, -1 Error reading 2160263286784, -1 59754240failed to load free space cache for block group 2193549688832failed to load free space cache for block group 2194623430656failed to load free space cache for block group 2196770914304failed to load free space cache for block group 2197844656128failed to load free space cache for block group 2199992139776failed to load free space cache for block group 2207508332544failed to load free space cache for block group 2212877041664failed to load free space cache for block group 2215024525312failed to load free space cache for block group 2216098267136failed to load free space cache for block group 2217172008960failed to load free space cache for block group 2218245750784failed to load free space cache for block group 2219319492608failed to load free space cache for block group 2220393234432failed to load free space cache for block group 2225761943552failed to load free space cache for block group 2228983169024failed to load free space cache for block group 2230056910848failed to load free space cache for blocError reading 3099329159168, -1 Error reading 3099330207744, -1 Error reading 2813098323968, -1 Error reading 2778216988672, -1 Error reading 1646990393344, -1 Error reading 2111944982528, -1 Error reading 2203213070336, -1 Error reading 1648066887680, -1 Error reading 2502787137536, -1 Error reading 2777428246528, -1 Error reading 2217171730432, -1 Error reading 2283743698944, -1 Error reading 2778245750784, -1 Error reading 2155968266240, -1 Error reading 2510303260672, -1 Error reading 2711092965376, -1 Error reading 2788402348032, -1 k group 2232204394496failed to load free space cache for block group 2233278136320failed to load free space cache for block group 2234351878144failed to load free space cache for block group 2235425619968failed to load free space cache for block group 2236499361792failed to load free space cache for block group 2237573103616failed to load free space cache for block group 2238646845440failed to load free space cache for block group 2240794329088failed to load free space cache for block group 2244015554560failed to load free space cache for block group 2246163038208failed to load free space cache for block group 2247236780032failed to load free space cache for block group 2248310521856failed to load free space cache for block group 2249384263680failed to load free space cache for block group 2250458005504failed to load free space cache for block group 2252605489152failed to load free space cache for block group 2253679230976failed to load free space cache for block group 2254752972800failed to load free space cError reading 2543589195776, -1 Error reading 2970937540608, -1 Error reading 2996708245504, -1 Error reading 1701765120000, -1 Error reading 2778246012928, -1 Error reading 2623046090752, -1 Error reading 2777427148800, -1 Error reading 2917250883584, -1 Error reading 2917251145728, -1 Error reading 3013888114688, -1 Error reading 1788747644928, -1 Error reading 2365347962880, -1 Error reading 3054690205696, -1 Error reading 2918324613120, -1 Error reading 2777427722240, -1 Error reading 2777428508672, -1 Error reading 2961274667008, -1 ache for block group 2255826714624failed to load free space cache for block group 2256900456448failed to load free space cache for block group 2257974198272failed to load free space cache for block group 2259047940096failed to load free space cache for block group 2260121681920failed to load free space cache for block group 2263342907392failed to load free space cache for block group 2264416649216failed to load free space cache for block group 2271932841984failed to load free space cache for block group 2273006583808failed to load free space cache for block group 2274080325632failed to load free space cache for block group 2275154067456failed to load free space cache for block group 2277301551104failed to load free space cache for block group 2278375292928failed to load free space cache for block group 2282670260224failed to load free space cache for block group 2283744002048failed to load free space cache for block group 2284817743872failed to load free space cache for block group 2299850129408failed to loadError reading 3033215467520, -1 Error reading 3076165009408, -1 Error reading 2778280181760, -1 Error reading 3099331256320, -1 Error reading 1738281648128, -1 Error reading 3099340431360, -1 Error reading 2777431654400, -1 Error reading 2973085794304, -1 Error reading 2777426886656, -1 Error reading 2923693670400, -1 Error reading 3099340693504, -1 Error reading 3099340955648, -1 Error reading 2537146613760, -1 Error reading 3099341479936, -1 Error reading 3020330565632, -1 Error reading 3025699323904, -1 Error reading 2975232884736, -1 free space cache for block group 2312735031296failed to load free space cache for block group 2319177482240failed to load free space cache for block group 2320251224064failed to load free space cache for block group 2321324965888failed to load free space cache for block group 2326693675008failed to load free space cache for block group 2327767416832failed to load free space cache for block group 2328841158656failed to load free space cache for block group 2352463478784failed to load free space cache for block group 2354610962432failed to load free space cache for block group 2355684704256failed to load free space cache for block group 2364274638848failed to load free space cache for block group 2376085798912failed to load free space cache for block group 2378233282560failed to load free space cache for block group 2384675733504failed to load free space cache for block group 2386823217152failed to load free space cache for block group 2397560635392failed to load free space cache for block group 2408298053632fError reading 1791968215040, -1 Error reading 1647009984512, -1 Error reading 2929062400000, -1 Error reading 2149525815296, -1 Error reading 1646985674752, -1 Error reading 3029994242048, -1 Error reading 2778207158272, -1 Error reading 3079048380416, -1 Error reading 3058985140224, -1 Error reading 3074017714176, -1 Error reading 2778224590848, -1 Error reading 3099342528512, -1 Error reading 3099342790656, -1 Error reading 2157042008064, -1 Error reading 2778225901568, -1 Error reading 3209144762368, -1 ailed to load free space cache for block group 2420109213696failed to load free space cache for block group 2427625406464failed to load free space cache for block group 2439436566528failed to load free space cache for block group 2440510308352failed to load free space cache for block group 2441584050176failed to load free space cache for block group 2451247726592failed to load free space cache for block group 2473796304896failed to load free space cache for block group 2478091272192failed to load free space cache for block group 2479165014016failed to load free space cache for block group 2498492366848failed to load free space cache for block group 2499566108672failed to load free space cache for block group 2509229785088failed to load free space cache for block group 2515672236032failed to load free space cache for block group 2530704621568failed to load free space cache for block group 2534999588864failed to load free space cache for block group 2542515781632failed to load free space cache for block group 2Error reading 2778230095872, -1 Error reading 2777433219072, -1 Error reading 2778207420416, -1 Error reading 1655478509568, -1 Error reading 1646986199040, -1 Error reading 3217720803328, -1 Error reading 3222111846400, -1 Error reading 2778305724416, -1 Error reading 2778229309440, -1 Error reading 2778270220288, -1 Error reading 2778277560320, -1 Error reading 2882891501568, -1 Error reading 2778219085824, -1 Error reading 2778226163712, -1 Error reading 1646986723328, -1 Error reading 1646986985472, -1 Error reading 1646987247616, -1 546810748928failed to load free space cache for block group 2570433069056failed to load free space cache for block group 2571506810880failed to load free space cache for block group 2573654294528failed to load free space cache for block group 2582244229120failed to load free space cache for block group 2585465454592failed to load free space cache for block group 2595129131008failed to load free space cache for block group 2597276614656failed to load free space cache for block group 2620898934784failed to load free space cache for block group 2621972676608failed to load free space cache for block group 2632710094848failed to load free space cache for block group 2661701124096failed to load free space cache for block group 2686397186048failed to load free space cache for block group 2688544669696failed to load free space cache for block group 2718609440768failed to load free space cache for block group 2752969179136failed to load free space cache for block group 2758337888256failed to load free space cache for Error reading 3099343052800, -1 Error reading 1728619188224, -1 Error reading 3079048642560, -1 Error reading 1647012601856, -1 Error reading 1786600161280, -1 Error reading 1655494344704, -1 Error reading 2778308083712, -1 Error reading 2778310180864, -1 Error reading 1644921614336, -1 Error reading 2778312540160, -1 Error reading 1683453624320, -1 Error reading 1797337710592, -1 Error reading 1645331808256, -1 Error reading 3209129295872, -1 Error reading 2998855598080, -1 Error reading 2787328196608, -1 Error reading 2831352135680, -1 block group 2759411630080failed to load free space cache for block group 2773370273792failed to load free space cache for block group 2774444015616failed to load free space cache for block group 2775517757440failed to load free space cache for block group 2776591499264failed to load free space cache for block group 2777665241088failed to load free space cache for block group 2778738982912failed to load free space cache for block group 2779812724736failed to load free space cache for block group 2780886466560failed to load free space cache for block group 2786255175680failed to load free space cache for block group 2789476401152failed to load free space cache for block group 2795918852096failed to load free space cache for block group 2807730012160failed to load free space cache for block group 2810951237632failed to load free space cache for block group 2812024979456failed to load free space cache for block group 2841016008704failed to load free space cache for block group 2847458459648failed to load free spaError reading 1645349634048, -1 Error reading 2883965411328, -1 Error reading 2837794250752, -1 Error reading 2970937802752, -1 Error reading 3099362435072, -1 Error reading 1646997995520, -1 Error reading 2877522681856, -1 Error reading 3222112894976, -1 Error reading 2778232455168, -1 Error reading 3066863624192, -1 ce cache for block group 2860343361536failed to load free space cache for block group 2862490845184failed to load free space cache for block group 2865712070656failed to load free space cache for block group 2873228263424failed to load free space cache for block group 2876449488896failed to load free space cache for block group 2882891939840failed to load free space cache for block group 2883965681664failed to load free space cache for block group 2885039423488failed to load free space cache for block group 2886113165312failed to load free space cache for block group 2889334390784failed to load free space cache for block group 2890408132608free space inode generation (0) did not match free space cache generation (20905) Error reading 1661505437696, -1 Error reading 2165631680512, -1 Error reading 2778314375168, -1 Error reading 2778314899456, -1 Error reading 1646988296192, -1 Error reading 2318103281664, -1 Error reading 2778318045184, -1 Error reading 3066863886336, -1 Error reading 3099360862208, -1 Error reading 1646989869056, -1 Error reading 3066864148480, -1 Error reading 1647005597696, -1 Error reading 3099361124352, -1 Error reading 1646992752640, -1 Error reading 1655480868864, -1 Error reading 1644921085952, -1 Error reading 2891481677824, -1 failed to load free space cache for block group 2894703099904failed to load free space cache for block group 2900071809024failed to load free space cache for block group 2902219292672failed to load free space cache for block group 2903293034496failed to load free space cache for block group 2910809227264failed to load free space cache for block group 2912956710912failed to load free space cache for block group 2914030452736failed to load free space cache for block group 2915104194560failed to load free space cache for block group 2917251678208failed to load free space cache for block group 2921546645504failed to load free space cache for block group 2922620387328failed to load free space cache for block group 2924767870976failed to load free space cache for block group 2925841612800failed to load free space cache for block group 2926915354624failed to load free space cache for block group 2930136580096failed to load free space cache for block group 2931210321920failed to load free space cache for block group Error reading 2916176310272, -1 Error reading 3038584225792, -1 Error reading 3046100369408, -1 Error reading 1647972515840, -1 Error reading 2923693408256, -1 Error reading 1645349109760, -1 Error reading 3058345299968, -1 Error reading 1988464431104, -1 Error reading 3058346348544, -1 Error reading 1647004549120, -1 Error reading 2951610437632, -1 Error reading 3099360337920, -1 Error reading 2994560761856, -1 Error reading 2997781913600, -1 Error reading 2348168204288, -1 Error reading 3066864934912, -1 Error reading 1644895338496, -1 2936579031040failed to load free space cache for block group 2940873998336failed to load free space cache for block group 2944095223808failed to load free space cache for block group 2948390191104failed to load free space cache for block group 2950537674752failed to load free space cache for block group 2954832642048failed to load free space cache for block group 2956980125696failed to load free space cache for block group 2964496318464failed to load free space cache for block group 2965570060288failed to load free space cache for block group 2969865027584failed to load free space cache for block group 2973086253056failed to load free space cache for block group 2975233736704failed to load free space cache for block group 2977381220352failed to load free space cache for block group 2980602445824failed to load free space cache for block group 2988118638592failed to load free space cache for block group 2990266122240failed to load free space cache for block group 2993487347712failed to load free space cache forError reading 3066865197056, -1 Error reading 1646989344768, -1 Error reading 2999929008128, -1 Error reading 2887186579456, -1 Error reading 1655477182464, -1 Error reading 1645349371904, -1 Error reading 3099361386496, -1 Error reading 1671399686144, -1 Error reading 3005297000448, -1 Error reading 1644915576832, -1 Error reading 1645348847616, -1 Error reading 2379306696704, -1 Error reading 3061920546816, -1 Error reading 3099362172928, -1 Error reading 1647976710144, -1 Error reading 1644915838976, -1 Error reading 1644926586880, -1 block group 2994561089536failed to load free space cache for block group 2996708573184failed to load free space cache for block group 2998856056832failed to load free space cache for block group 2999929798656failed to load free space cache for block group 3002077282304failed to load free space cache for block group 3004224765952failed to load free space cache for block group 3005298507776failed to load free space cache for block group 3006372249600failed to load free space cache for block group 3009593475072failed to load free space cache for block group 3013888442368failed to load free space cache for block group 3021404635136failed to load free space cache for block group 3022478376960failed to load free space cache for block group 3036437020672failed to load free space cache for block group 3037510762496failed to load free space cache for block group 3047174438912failed to load free space cache for block group 3066501791744failed to load free space cache for block group 3098714046464failed to load free spError reading 1648000040960, -1 Error reading 1669556535296, -1 Error reading 1671400472576, -1 Error reading 1644917489664, -1 Error reading 1644922470400, -1 Error reading 1644929810432, -1 Error reading 1644925353984, -1 Error reading 1645326041088, -1 Error reading 1647002632192, -1 Error reading 1647004286976, -1 Error reading 1648030711808, -1 Error reading 1647968583680, -1 Error reading 1648054566912, -1 Error reading 1648064790528, -1 Error reading 1648074096640, -1 Error reading 1647985623040, -1 Error reading 1655474683904, -1 ace cache for block group 3099787788288failed to load free space cache for block group 3100861530112failed to load free space cache for block group 3169581006848failed to load free space cache for block group 3171728490496failed to load free space cache for block group 3221120614400failed to load free space cache for block group 3222194356224failed to load free space cache for block group 3308135645184failed to load free space cache for block group 3311356870656failed to load free space cache for block group 3312430612480failed to load free space cache for block group 3313504354304failed to load free space cache for block group 3314578096128failed to load free space cache for block group 3315651837952failed to load free space cache for block group 3316725579776failed to load free space cache for block group 3317799321600failed to load free space cache for block group 3319946805248failed to load free space cache for block group 3321020547072failed to load free space cache for block group 3322094288896failed toError reading 1655477985280, -1 Error reading 1655479296000, -1 Error reading 1655499849728, -1 Error reading 1655506927616, -1 Error reading 1671703769088, -1 Error reading 1676003966976, -1 Error reading 1677069975552, -1 Error reading 1699627622400, -1 Error reading 1655479820288, -1 checking fs roots checking csums checking root refs load free space cache for block group 3323168030720failed to load free space cache for block group 3324241772544failed to load free space cache for block group 3325315514368failed to load free space cache for block group 3326389256192failed to load free space cache for block group 3327462998016failed to load free space cache for block group 3328536739840failed to load free space cache for block group 3329610481664failed to load free space cache for block group 3330684223488failed to load free space cache for block group 3350011576320failed to load free space cache for block group 3454164533248found 23759468894 bytes used err is 0 total csum bytes: 1533072864 total tree bytes: 1728692224 total fs tree bytes: 44019712 total extent tree bytes: 47996928 btree space waste bytes: 46886467 file data blocks allocated: 4140676993024 referenced 1741813084160 Btrfs v3.12 ^ permalink raw reply [flat|nested] 25+ messages in thread
end of thread, other threads:[~2014-02-10 14:06 UTC | newest]
Thread overview: 25+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2014-01-29 19:16 lost with degraded RAID1 Johan Kröckel
2014-01-30 3:28 ` Duncan
2014-01-30 11:53 ` Johan Kröckel
2014-01-30 17:57 ` Chris Murphy
[not found] ` <CABgvyo9FYGSYpj+jL1oqCvtNUqsC8HZ+z=x-Gz7naWoEcCKWpQ@mail.gmail.com>
2014-01-30 19:32 ` Chris Murphy
2014-01-30 22:18 ` Duncan
2014-01-31 3:00 ` Chris Murphy
2014-01-31 5:58 ` Duncan
2014-01-31 6:10 ` Chris Murphy
2014-01-31 6:13 ` Chris Murphy
2014-01-31 7:37 ` Duncan
2014-01-31 2:19 ` Chris Murphy
2014-01-30 17:33 ` Chris Murphy
2014-01-30 17:58 ` Hugo Mills
2014-01-30 18:25 ` Chris Murphy
2014-02-03 20:55 ` Johan Kröckel
2014-02-03 21:08 ` Chris Murphy
2014-02-03 21:31 ` Johan Kröckel
2014-02-07 11:34 ` Johan Kröckel
2014-02-07 17:43 ` Chris Murphy
2014-02-08 11:09 ` Johan Kröckel
2014-02-09 5:40 ` Duncan
2014-02-10 14:05 ` Johan Kröckel
[not found] ` <CABgvyo822bOAHeA1GH28MPaBAU+Zdi72MD_uwL+dhopt+nwMig@mail.gmail.com>
2014-01-31 6:28 ` Fwd: " Johan Kröckel
2014-01-31 6:40 ` Johan Kröckel
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).