* non-fresh data unavailable bug @ 2010-01-14 15:10 Brett Russ 2010-01-14 19:24 ` Michael Evans 0 siblings, 1 reply; 4+ messages in thread From: Brett Russ @ 2010-01-14 15:10 UTC (permalink / raw) To: linux-raid Slightly related to my last message here Re:non-fresh behavior, we have seen cases where the following happens: * healthy 2 disk raid1 (disks A & B) incurs a problem with disk B * disk B is removed, unit is now degraded * replacement disk C is added; recovery from A to C begins * during recovery, disk A incurs a brief lapse in connectivity. At this point C is still up yet only has a partial copy of the data. * a subsequent assemble operation on the raid1 results in disk A being kicked out as non-fresh, yet C is allowed in. This presents quite a data-unavailability problem and basically requires recognizing the situation and hand assembling the array with disk A (only) first, then adding C back in. Unfortunately this situation is hard to reproduce and we don't have a dump of the 'mdadm --examine' output for it yet. Any thoughts on this while we try to get a better reproduction case? Thanks, Brett ^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: non-fresh data unavailable bug 2010-01-14 15:10 non-fresh data unavailable bug Brett Russ @ 2010-01-14 19:24 ` Michael Evans 2010-01-15 15:36 ` Brett Russ 0 siblings, 1 reply; 4+ messages in thread From: Michael Evans @ 2010-01-14 19:24 UTC (permalink / raw) To: Brett Russ; +Cc: linux-raid On Thu, Jan 14, 2010 at 7:10 AM, Brett Russ <bruss@netezza.com> wrote: > Slightly related to my last message here Re:non-fresh behavior, we have seen > cases where the following happens: > * healthy 2 disk raid1 (disks A & B) incurs a problem with disk B > * disk B is removed, unit is now degraded > * replacement disk C is added; recovery from A to C begins > * during recovery, disk A incurs a brief lapse in connectivity. At this > point C is still up yet only has a partial copy of the data. > * a subsequent assemble operation on the raid1 results in disk A being > kicked out as non-fresh, yet C is allowed in. > > This presents quite a data-unavailability problem and basically requires > recognizing the situation and hand assembling the array with disk A (only) > first, then adding C back in. Unfortunately this situation is hard to > reproduce and we don't have a dump of the 'mdadm --examine' output for it > yet. > > Any thoughts on this while we try to get a better reproduction case? > > Thanks, > Brett > > > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html > I believe the desired and logical behavior here is to refuse running an incomplete array unless explicitly forced to do so. Incremental assembly might be what you're seeing. The only way to access the data from those devices, presuming that without the device that had the hiccup your array is incomplete, would be to force assembly with the older device included and hope. I very much recommend running it read-only until you can determine which assembly pattern produces the most viable results. -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: non-fresh data unavailable bug 2010-01-14 19:24 ` Michael Evans @ 2010-01-15 15:36 ` Brett Russ 2010-01-18 3:32 ` Neil Brown 0 siblings, 1 reply; 4+ messages in thread From: Brett Russ @ 2010-01-15 15:36 UTC (permalink / raw) To: linux-raid On 01/14/2010 02:24 PM, Michael Evans wrote: > On Thu, Jan 14, 2010 at 7:10 AM, Brett Russ<bruss@netezza.com> wrote: >> Slightly related to my last message here Re:non-fresh behavior, we have seen >> cases where the following happens: >> * healthy 2 disk raid1 (disks A& B) incurs a problem with disk B >> * disk B is removed, unit is now degraded >> * replacement disk C is added; recovery from A to C begins >> * during recovery, disk A incurs a brief lapse in connectivity. At this >> point C is still up yet only has a partial copy of the data. >> * a subsequent assemble operation on the raid1 results in disk A being >> kicked out as non-fresh, yet C is allowed in. > > I believe the desired and logical behavior here is to refuse running > an incomplete array unless explicitly forced to do so. Incremental > assembly might be what you're seeing. This brings up a good point. I didn't mention that the assemble in the last step above was forced. Thus, the "bug" I'm reporting is that under duress, mdadm/md chose to assemble the array with a partially recovered (but "newer") member instead of the older member which was the recovery *source* for the newer member. What I think should happen is members that are *destinations* for recovery should *never* receive a higher event count, timestamp, or any other marking than the recovery sources. By definition they are incomplete and can't be trusted, thus they should never trump a complete member during assemble. I would assume the code already does this but perhaps there is a hole. One other piece of information that may be relevant--we're using 2 member RAID1 units with one member marked write-mostly. At this time, I don't have the specifics for which member (A or B) was the write-mostly member in the example above, but I can find that out. > I very much recommend running it read-only until you can determine which > assembly pattern produces the most viable results. Good tip. We were able to manually recover the array in the case outlined above, now we're looking back to fixing the kernel to prevent it happening again. Thanks, Brett ^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: non-fresh data unavailable bug 2010-01-15 15:36 ` Brett Russ @ 2010-01-18 3:32 ` Neil Brown 0 siblings, 0 replies; 4+ messages in thread From: Neil Brown @ 2010-01-18 3:32 UTC (permalink / raw) To: Brett Russ; +Cc: linux-raid On Fri, 15 Jan 2010 10:36:39 -0500 Brett Russ <bruss@netezza.com> wrote: > On 01/14/2010 02:24 PM, Michael Evans wrote: > > On Thu, Jan 14, 2010 at 7:10 AM, Brett Russ<bruss@netezza.com> wrote: > >> Slightly related to my last message here Re:non-fresh behavior, we have seen > >> cases where the following happens: > >> * healthy 2 disk raid1 (disks A& B) incurs a problem with disk B > >> * disk B is removed, unit is now degraded > >> * replacement disk C is added; recovery from A to C begins > >> * during recovery, disk A incurs a brief lapse in connectivity. At this > >> point C is still up yet only has a partial copy of the data. > >> * a subsequent assemble operation on the raid1 results in disk A being > >> kicked out as non-fresh, yet C is allowed in. > > > > I believe the desired and logical behavior here is to refuse running > > an incomplete array unless explicitly forced to do so. Incremental > > assembly might be what you're seeing. > > This brings up a good point. I didn't mention that the assemble in the > last step above was forced. Thus, the "bug" I'm reporting is that under > duress, mdadm/md chose to assemble the array with a partially recovered > (but "newer") member instead of the older member which was the recovery > *source* for the newer member. > > What I think should happen is members that are *destinations* for > recovery should *never* receive a higher event count, timestamp, or any > other marking than the recovery sources. By definition they are > incomplete and can't be trusted, thus they should never trump a complete > member during assemble. I would assume the code already does this but > perhaps there is a hole. > > One other piece of information that may be relevant--we're using 2 > member RAID1 units with one member marked write-mostly. At this time, I > don't have the specifics for which member (A or B) was the write-mostly > member in the example above, but I can find that out. > > > I very much recommend running it read-only until you can determine which > > assembly pattern produces the most viable results. > > Good tip. We were able to manually recover the array in the case > outlined above, now we're looking back to fixing the kernel to prevent > it happening again. > Thanks for the report. It sounds like a real problem. I'm travelling at the moment so reproducing it would be a challenge. If you are able to, can you report the output of mdadm -E /dev/list-of-devices at the key points in the process, and also add "-v" to any mdadm --assemble command you use, and report the output? Thanks, NeilBrown ^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2010-01-18 3:32 UTC | newest] Thread overview: 4+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2010-01-14 15:10 non-fresh data unavailable bug Brett Russ 2010-01-14 19:24 ` Michael Evans 2010-01-15 15:36 ` Brett Russ 2010-01-18 3:32 ` Neil Brown
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).