linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Namhyung Kim <namhyung@gmail.com>
To: NeilBrown <neilb@suse.de>
Cc: linux-raid@vger.kernel.org
Subject: Re: [md PATCH 01/36] md: beginnings of bad block management.
Date: Sat, 23 Jul 2011 00:03:45 +0900	[thread overview]
Message-ID: <878vrq5pda.fsf@gmail.com> (raw)
In-Reply-To: <20110721025847.8422.29443.stgit@notabene.brown> (NeilBrown's message of "Thu, 21 Jul 2011 12:58:47 +1000")

NeilBrown <neilb@suse.de> writes:

> This the first step in allowing md to track bad-blocks per-device so
> that we can fail individual blocks rather than the whole device.
>
> This patch just adds a data structure for recording bad blocks, with
> routines to add, remove, search the list.
>
> Signed-off-by: NeilBrown <neilb@suse.de>
> ---
>
>  drivers/md/md.c |  457 +++++++++++++++++++++++++++++++++++++++++++++++++++++++
>  drivers/md/md.h |   49 ++++++
>  2 files changed, 502 insertions(+), 4 deletions(-)
>
> diff --git a/drivers/md/md.c b/drivers/md/md.c
> index 2a32050..220fadb 100644
> --- a/drivers/md/md.c
> +++ b/drivers/md/md.c
> @@ -1952,6 +1952,10 @@ static void unbind_rdev_from_array(mdk_rdev_t * rdev)
>  	sysfs_remove_link(&rdev->kobj, "block");
>  	sysfs_put(rdev->sysfs_state);
>  	rdev->sysfs_state = NULL;
> +	kfree(rdev->badblocks.page);
> +	rdev->badblocks.count = 0;
> +	rdev->badblocks.page = NULL;
> +	rdev->badblocks.active_page = NULL;
>  	/* We need to delay this, otherwise we can deadlock when
>  	 * writing to 'remove' to "dev/state".  We also need
>  	 * to delay it due to rcu usage.
> @@ -2778,7 +2782,7 @@ static struct kobj_type rdev_ktype = {
>  	.default_attrs	= rdev_default_attrs,
>  };
>  
> -void md_rdev_init(mdk_rdev_t *rdev)
> +int md_rdev_init(mdk_rdev_t *rdev)
>  {
>  	rdev->desc_nr = -1;
>  	rdev->saved_raid_disk = -1;
> @@ -2794,6 +2798,20 @@ void md_rdev_init(mdk_rdev_t *rdev)
>  
>  	INIT_LIST_HEAD(&rdev->same_set);
>  	init_waitqueue_head(&rdev->blocked_wait);
> +
> +	/* Add space to store bad block list.
> +	 * This reserves the space even on arrays where it cannot
> +	 * be used - I wonder if that matters
> +	 */
> +	rdev->badblocks.count = 0;
> +	rdev->badblocks.shift = 0;
> +	rdev->badblocks.page = kmalloc(PAGE_SIZE, GFP_KERNEL);
> +	rdev->badblocks.active_page = rdev->badblocks.page;
> +	spin_lock_init(&rdev->badblocks.lock);
> +	if (rdev->badblocks.page == NULL)
> +		return -ENOMEM;
> +
> +	return 0;
>  }
>  EXPORT_SYMBOL_GPL(md_rdev_init);
>  /*
> @@ -2819,8 +2837,11 @@ static mdk_rdev_t *md_import_device(dev_t newdev, int super_format, int super_mi
>  		return ERR_PTR(-ENOMEM);
>  	}
>  
> -	md_rdev_init(rdev);
> -	if ((err = alloc_disk_sb(rdev)))
> +	err = md_rdev_init(rdev);
> +	if (err)
> +		goto abort_free;
> +	err = alloc_disk_sb(rdev);
> +	if (err)
>  		goto abort_free;
>  
>  	err = lock_rdev(rdev, newdev, super_format == -2);
> @@ -7324,6 +7345,436 @@ void md_wait_for_blocked_rdev(mdk_rdev_t *rdev, mddev_t *mddev)
>  }
>  EXPORT_SYMBOL(md_wait_for_blocked_rdev);
>  
> +
> +/* Bad block management.
> + * We can record which blocks on each device are 'bad' and so just
> + * fail those blocks, or that stripe, rather than the whole device.
> + * Entries in the bad-block table are 64bits wide.  This comprises:
> + * Length of bad-range, in sectors: 0-511 for lengths 1-512
> + * Start of bad-range, sector offset, 54 bits (allows 8 exbibytes)
> + *  A 'shift' can be set so that larger blocks are tracked and
> + *  consequently larger devices can be covered.
> + * 'Acknowledged' flag - 1 bit. - the most significant bit.
> + */
> +/* Locking of the bad-block table is a two-layer affair.
> + * Read access through ->active_page only requires an rcu_readlock.
> + * However if ->active_page is found to be NULL, the table
> + * should be accessed through ->page which requires an irq-spinlock.
> + * Updating the page requires setting ->active_page to NULL,
> + * synchronising with rcu, then updating ->page under the same
> + * irq-spinlock.
> + * We always set or clear bad blocks from process context, but
> + * might look-up bad blocks from interrupt/bh context.
> + *

Empty line.

If the locking is complex, it'd be better defining separate functions to
deal with it, IMHO. Please see below.


> + */
> +/* When looking for a bad block we specify a range and want to
> + * know if any block in the range is bad.  So we binary-search
> + * to the last range that starts at-or-before the given endpoint,
> + * (or "before the sector after the target range")
> + * then see if it ends after the given start.
> + * We return
> + *  0 if there are no known bad blocks in the range
> + *  1 if there are known bad block which are all acknowledged
> + * -1 if there are bad blocks which have not yet been acknowledged in metadata.
> + * plus the start/length of the first bad section we overlap.
> + */
> +int md_is_badblock(struct badblocks *bb, sector_t s, int sectors,
> +		   sector_t *first_bad, int *bad_sectors)
> +{
> +	int hi;
> +	int lo = 0;
> +	u64 *p;
> +	int rv = 0;
> +	int havelock = 0;
> +	sector_t target = s + sectors;
> +	unsigned long uninitialized_var(flags);
> +
> +	if (bb->shift > 0) {
> +		/* round the start down, and the end up */
> +		s >>= bb->shift;
> +		target += (1<<bb->shift) - 1;
> +		target >>= bb->shift;
> +		sectors = target - s;
> +	}
> +	/* 'target' is now the first block after the bad range */
> +
> +	rcu_read_lock();
> +	p = rcu_dereference(bb->active_page);
> +	if (!p) {
> +		spin_lock_irqsave(&bb->lock, flags);
> +		p = bb->page;
> +		havelock = 1;
> +	}

Maybe something like:

      p = md_read_lock_bb(bb, &havelock, &flags);


> +	hi = bb->count;
> +
> +	/* Binary search between lo and hi for 'target'
> +	 * i.e. for the last range that starts before 'target'
> +	 */
> +	/* INVARIANT: ranges before 'lo' and at-or-after 'hi'
> +	 * are known not to be the last range before target.
> +	 * VARIANT: hi-lo is the number of possible
> +	 * ranges, and decreases until it reaches 1
> +	 */
> +	while (hi - lo > 1) {
> +		int mid = (lo + hi) / 2;
> +		sector_t a = BB_OFFSET(p[mid]);
> +		if (a < target)
> +			/* This could still be the one, earlier ranges
> +			 * could not. */
> +			lo = mid;
> +		else
> +			/* This and later ranges are definitely out. */
> +			hi = mid;
> +	}
> +	/* 'lo' might be the last that started before target, but 'hi' isn't */
> +	if (hi > lo) {
> +		/* need to check all range that end after 's' to see if
> +		 * any are unacknowledged.
> +		 */
> +		while (lo >= 0 &&
> +		       BB_OFFSET(p[lo]) + BB_LEN(p[lo]) > s) {
> +			if (BB_OFFSET(p[lo]) < target) {
> +				/* starts before the end, and finishes after
> +				 * the start, so they must overlap
> +				 */
> +				if (rv != -1 && BB_ACK(p[lo]))
> +					rv = 1;
> +				else
> +					rv = -1;
> +				*first_bad = BB_OFFSET(p[lo]);
> +				*bad_sectors = BB_LEN(p[lo]);
> +			}
> +			lo--;
> +		}
> +	}
> +
> +	if (havelock)
> +		spin_unlock_irqrestore(&bb->lock, flags);
> +	rcu_read_unlock();

And
        md_read_unlock_bb(bb, havelock, flags);


> +	return rv;
> +}
> +EXPORT_SYMBOL_GPL(md_is_badblock);
> +
> +/*
> + * Add a range of bad blocks to the table.
> + * This might extend the table, or might contract it
> + * if two adjacent ranges can be merged.
> + * We binary-search to find the 'insertion' point, then
> + * decide how best to handle it.
> + */
> +static int md_set_badblocks(struct badblocks *bb, sector_t s, int sectors,
> +			    int acknowledged)
> +{
> +	u64 *p;
> +	int lo, hi;
> +	int rv = 1;
> +
> +	if (bb->shift < 0)
> +		/* badblocks are disabled */
> +		return 0;
> +
> +	if (bb->shift) {
> +		/* round the start down, and the end up */
> +		sector_t next = s + sectors;
> +		s >>= bb->shift;
> +		next += (1<<bb->shift) - 1;
> +		next >>= bb->shift;
> +		sectors = next - s;
> +	}
> +
> +	while (1) {
> +		rcu_assign_pointer(bb->active_page, NULL);
> +		synchronize_rcu();
> +		spin_lock_irq(&bb->lock);
> +		if (bb->active_page == NULL)
> +			break;
> +		/* someone else just unlocked, better retry */
> +		spin_unlock_irq(&bb->lock);
> +	}

        md_write_lock_bb(bb);


> +	/* now have exclusive access to the page */
> +
> +	p = bb->page;
> +	lo = 0;
> +	hi = bb->count;
> +	/* Find the last range that starts at-or-before 's' */
> +	while (hi - lo > 1) {
> +		int mid = (lo + hi) / 2;
> +		sector_t a = BB_OFFSET(p[mid]);
> +		if (a <= s)
> +			lo = mid;
> +		else
> +			hi = mid;
> +	}
> +	if (hi > lo && BB_OFFSET(p[lo]) > s)
> +		hi = lo;
> +
> +	if (hi > lo) {
> +		/* we found a range that might merge with the start
> +		 * of our new range
> +		 */
> +		sector_t a = BB_OFFSET(p[lo]);
> +		sector_t e = a + BB_LEN(p[lo]);
> +		int ack = BB_ACK(p[lo]);
> +		if (e >= s) {
> +			/* Yes, we can merge with a previous range */
> +			if (s == a && s + sectors >= e)
> +				/* new range covers old */
> +				ack = acknowledged;
> +			else
> +				ack = ack && acknowledged;
> +
> +			if (e < s + sectors)
> +				e = s + sectors;
> +			if (e - a <= BB_MAX_LEN) {
> +				p[lo] = BB_MAKE(a, e-a, ack);
> +				s = e;
> +			} else {
> +				/* does not all fit in one range,
> +				 * make p[lo] maximal
> +				 */
> +				if (BB_LEN(p[lo]) != BB_MAX_LEN)
> +					p[lo] = BB_MAKE(a, BB_MAX_LEN, ack);
> +				s = a + BB_MAX_LEN;
> +			}
> +			sectors = e - s;
> +		}
> +	}
> +	if (sectors && hi < bb->count) {
> +		/* 'hi' points to the first range that starts after 's'.
> +		 * Maybe we can merge with the start of that range */
> +		sector_t a = BB_OFFSET(p[hi]);
> +		sector_t e = a + BB_LEN(p[hi]);
> +		int ack = BB_ACK(p[hi]);
> +		if (a <= s + sectors) {
> +			/* merging is possible */
> +			if (e <= s + sectors) {
> +				/* full overlap */
> +				e = s + sectors;
> +				ack = acknowledged;
> +			} else
> +				ack = ack && acknowledged;
> +
> +			a = s;
> +			if (e - a <= BB_MAX_LEN) {
> +				p[hi] = BB_MAKE(a, e-a, ack);
> +				s = e;
> +			} else {
> +				p[hi] = BB_MAKE(a, BB_MAX_LEN, ack);
> +				s = a + BB_MAX_LEN;
> +			}
> +			sectors = e - s;
> +			lo = hi;
> +			hi++;
> +		}
> +	}
> +	if (sectors == 0 && hi < bb->count) {
> +		/* we might be able to combine lo and hi */
> +		/* Note: 's' is at the end of 'lo' */
> +		sector_t a = BB_OFFSET(p[hi]);
> +		int lolen = BB_LEN(p[lo]);
> +		int hilen = BB_LEN(p[hi]);
> +		int newlen = lolen + hilen - (s - a);
> +		if (s >= a && newlen < BB_MAX_LEN) {
> +			/* yes, we can combine them */
> +			int ack = BB_ACK(p[lo]) && BB_ACK(p[hi]);
> +			p[lo] = BB_MAKE(BB_OFFSET(p[lo]), newlen, ack);
> +			memmove(p + hi, p + hi + 1,
> +				(bb->count - hi - 1) * 8);
> +			bb->count--;
> +		}
> +	}
> +	while (sectors) {
> +		/* didn't merge (it all).
> +		 * Need to add a range just before 'hi' */
> +		if (bb->count >= MD_MAX_BADBLOCKS) {
> +			/* No room for more */
> +			rv = 0;
> +			break;
> +		} else {
> +			int this_sectors = sectors;
> +			memmove(p + hi + 1, p + hi,
> +				(bb->count - hi) * 8);
> +			bb->count++;
> +
> +			if (this_sectors > BB_MAX_LEN)
> +				this_sectors = BB_MAX_LEN;
> +			p[hi] = BB_MAKE(s, this_sectors, acknowledged);
> +			sectors -= this_sectors;
> +			s += this_sectors;
> +		}
> +	}
> +
> +	bb->changed = 1;
> +	rcu_assign_pointer(bb->active_page, bb->page);
> +	spin_unlock_irq(&bb->lock);

        md_write_unlock_bb(bb);


> +
> +	return rv;
> +}
> +
> +int rdev_set_badblocks(mdk_rdev_t *rdev, sector_t s, int sectors,
> +		       int acknowledged)
> +{
> +	int rv = md_set_badblocks(&rdev->badblocks,
> +				  s + rdev->data_offset, sectors, acknowledged);
> +	if (rv) {
> +		/* Make sure they get written out promptly */
> +		set_bit(MD_CHANGE_CLEAN, &rdev->mddev->flags);
> +		md_wakeup_thread(rdev->mddev->thread);
> +	}
> +	return rv;
> +}
> +EXPORT_SYMBOL_GPL(rdev_set_badblocks);

I think it would be better if all exported functions in md.c have
prefixed 'md_'.


> +
> +/*
> + * Remove a range of bad blocks from the table.
> + * This may involve extending the table if we spilt a region,
> + * but it must not fail.  So if the table becomes full, we just
> + * drop the remove request.
> + */
> +static int md_clear_badblocks(struct badblocks *bb, sector_t s, int sectors)
> +{
> +	u64 *p;
> +	int lo, hi;
> +	sector_t target = s + sectors;
> +	int rv = 0;
> +
> +	if (bb->shift > 0) {
> +		/* When clearing we round the start up and the end down.
> +		 * This should not matter as the shift should align with
> +		 * the block size and no rounding should ever be needed.
> +		 * However it is better the think a block is bad when it
> +		 * isn't than to think a block is not bad when it is.
> +		 */
> +		s += (1<<bb->shift) - 1;
> +		s >>= bb->shift;
> +		target >>= bb->shift;
> +		sectors = target - s;
> +	}
> +
> +	while (1) {
> +		rcu_assign_pointer(bb->active_page, NULL);
> +		synchronize_rcu();
> +		spin_lock_irq(&bb->lock);
> +		if (bb->active_page == NULL)
> +			break;
> +		/* someone else just unlocked, better retry */
> +		spin_unlock_irq(&bb->lock);
> +	}
> +	/* now have exclusive access to the page */
> +
> +	p = bb->page;
> +	lo = 0;
> +	hi = bb->count;
> +	/* Find the last range that starts before 'target' */
> +	while (hi - lo > 1) {
> +		int mid = (lo + hi) / 2;
> +		sector_t a = BB_OFFSET(p[mid]);
> +		if (a < target)
> +			lo = mid;
> +		else
> +			hi = mid;
> +	}
> +	if (hi > lo) {
> +		/* p[lo] is the last range that could overlap the
> +		 * current range.  Earlier ranges could also overlap,
> +		 * but only this one can overlap the end of the range.
> +		 */
> +		if (BB_OFFSET(p[lo]) + BB_LEN(p[lo]) > target) {
> +			/* Partial overlap, leave the tail of this range */
> +			int ack = BB_ACK(p[lo]);
> +			sector_t a = BB_OFFSET(p[lo]);
> +			sector_t end = a + BB_LEN(p[lo]);
> +
> +			if (a < s) {
> +				/* we need to split this range */
> +				if (bb->count >= MD_MAX_BADBLOCKS) {
> +					rv = 0;
> +					goto out;
> +				}
> +				memmove(p+lo+1, p+lo, (bb->count - lo) * 8);
> +				bb->count++;
> +				p[lo] = BB_MAKE(a, s-a, ack);
> +				lo++;
> +			}
> +			p[lo] = BB_MAKE(target, end - target, ack);
> +			/* there is no longer an overlap */
> +			hi = lo;
> +			lo--;
> +		}
> +		while (lo >= 0 &&
> +		       BB_OFFSET(p[lo]) + BB_LEN(p[lo]) > s) {
> +			/* This range does overlap */
> +			if (BB_OFFSET(p[lo]) < s) {
> +				/* Keep the early parts of this range. */
> +				int ack = BB_ACK(p[lo]);
> +				sector_t start = BB_OFFSET(p[lo]);
> +				p[lo] = BB_MAKE(start, s - start, ack);
> +				/* now low doesn't overlap, so.. */
> +				break;
> +			}
> +			lo--;
> +		}
> +		/* 'lo' is strictly before, 'hi' is strictly after,
> +		 * anything between needs to be discarded
> +		 */
> +		if (hi - lo > 1) {
> +			memmove(p+lo+1, p+hi, (bb->count - hi) * 8);
> +			bb->count -= (hi - lo - 1);
> +		}
> +	}
> +
> +	bb->changed = 1;
> +out:
> +	rcu_assign_pointer(bb->active_page, bb->page);
> +	spin_unlock_irq(&bb->lock);
> +	return rv;
> +}
> +
> +int rdev_clear_badblocks(mdk_rdev_t *rdev, sector_t s, int sectors)
> +{
> +	return md_clear_badblocks(&rdev->badblocks,
> +				  s + rdev->data_offset,
> +				  sectors);
> +}
> +EXPORT_SYMBOL_GPL(rdev_clear_badblocks);

Same here.

Thanks.


> +
> +/*
> + * Acknowledge all bad blocks in a list.
> + * This only succeeds if ->changed is clear.  It is used by
> + * in-kernel metadata updates
> + */
> +void md_ack_all_badblocks(struct badblocks *bb)
> +{
> +	if (bb->page == NULL || bb->changed)
> +		/* no point even trying */
> +		return;
> +	while (1) {
> +		rcu_assign_pointer(bb->active_page, NULL);
> +		synchronize_rcu();
> +		spin_lock_irq(&bb->lock);
> +		if (bb->active_page == NULL)
> +			break;
> +		/* someone else just unlocked, better retry */
> +		spin_unlock_irq(&bb->lock);
> +	}
> +	/* now have exclusive access to the page */
> +
> +	if (bb->changed == 0) {
> +		u64 *p = bb->page;
> +		int i;
> +		for (i = 0; i < bb->count ; i++) {
> +			if (!BB_ACK(p[i])) {
> +				sector_t start = BB_OFFSET(p[i]);
> +				int len = BB_LEN(p[i]);
> +				p[i] = BB_MAKE(start, len, 1);
> +			}
> +		}
> +	}
> +	rcu_assign_pointer(bb->active_page, bb->page);
> +	spin_unlock_irq(&bb->lock);
> +}
> +EXPORT_SYMBOL_GPL(md_ack_all_badblocks);
> +
>  static int md_notify_reboot(struct notifier_block *this,
>  			    unsigned long code, void *x)
>  {
> diff --git a/drivers/md/md.h b/drivers/md/md.h
> index 7d906a9..d327734 100644
> --- a/drivers/md/md.h
> +++ b/drivers/md/md.h
> @@ -29,6 +29,13 @@
>  typedef struct mddev_s mddev_t;
>  typedef struct mdk_rdev_s mdk_rdev_t;
>  
> +/* Bad block numbers are stored sorted in a single page.
> + * 64bits is used for each block or extent.
> + * 54 bits are sector number, 9 bits are extent size,
> + * 1 bit is an 'acknowledged' flag.
> + */
> +#define MD_MAX_BADBLOCKS	(PAGE_SIZE/8)
> +
>  /*
>   * MD's 'extended' device
>   */
> @@ -111,8 +118,48 @@ struct mdk_rdev_s
>  
>  	struct sysfs_dirent *sysfs_state; /* handle for 'state'
>  					   * sysfs entry */
> +
> +	struct badblocks {
> +		int	count;		/* count of bad blocks */
> +		int	shift;		/* shift from sectors to block size
> +					 * a -ve shift means badblocks are
> +					 * disabled.*/
> +		u64	*page;		/* badblock list */
> +		u64	*active_page;	/* either 'page' or 'NULL' */
> +		int	changed;
> +		spinlock_t lock;
> +	} badblocks;
>  };
>  
> +#define BB_LEN_MASK	(0x00000000000001FFULL)
> +#define BB_OFFSET_MASK	(0x7FFFFFFFFFFFFE00ULL)
> +#define BB_ACK_MASK	(0x8000000000000000ULL)
> +#define BB_MAX_LEN	512
> +#define BB_OFFSET(x)	(((x) & BB_OFFSET_MASK) >> 9)
> +#define BB_LEN(x)	(((x) & BB_LEN_MASK) + 1)
> +#define BB_ACK(x)	(!!((x) & BB_ACK_MASK))
> +#define BB_MAKE(a, l, ack) (((a)<<9) | ((l)-1) | ((u64)(!!(ack)) << 63))
> +
> +extern int md_is_badblock(struct badblocks *bb, sector_t s, int sectors,
> +			  sector_t *first_bad, int *bad_sectors);
> +static inline int is_badblock(mdk_rdev_t *rdev, sector_t s, int sectors,
> +			      sector_t *first_bad, int *bad_sectors)
> +{
> +	if (unlikely(rdev->badblocks.count)) {
> +		int rv = md_is_badblock(&rdev->badblocks, rdev->data_offset + s,
> +					sectors,
> +					first_bad, bad_sectors);
> +		if (rv)
> +			*first_bad -= rdev->data_offset;
> +		return rv;
> +	}
> +	return 0;
> +}
> +extern int rdev_set_badblocks(mdk_rdev_t *rdev, sector_t s, int sectors,
> +			      int acknowledged);
> +extern int rdev_clear_badblocks(mdk_rdev_t *rdev, sector_t s, int sectors);
> +extern void md_ack_all_badblocks(struct badblocks *bb);
> +
>  struct mddev_s
>  {
>  	void				*private;
> @@ -517,7 +564,7 @@ extern void mddev_init(mddev_t *mddev);
>  extern int md_run(mddev_t *mddev);
>  extern void md_stop(mddev_t *mddev);
>  extern void md_stop_writes(mddev_t *mddev);
> -extern void md_rdev_init(mdk_rdev_t *rdev);
> +extern int md_rdev_init(mdk_rdev_t *rdev);
>  
>  extern void mddev_suspend(mddev_t *mddev);
>  extern void mddev_resume(mddev_t *mddev);
>
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

  reply	other threads:[~2011-07-22 15:03 UTC|newest]

Thread overview: 65+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2011-07-21  2:58 [md PATCH 00/36] md patches for 3.1 - part 2: bad block logs NeilBrown
2011-07-21  2:58 ` [md PATCH 01/36] md: beginnings of bad block management NeilBrown
2011-07-22 15:03   ` Namhyung Kim [this message]
2011-07-26  2:26     ` NeilBrown
2011-07-26  5:17       ` Namhyung Kim
2011-07-22 16:52   ` Namhyung Kim
2011-07-26  3:20     ` NeilBrown
2011-07-21  2:58 ` [md PATCH 02/36] md/bad-block-log: add sysfs interface for accessing bad-block-log NeilBrown
2011-07-22 15:43   ` Namhyung Kim
2011-07-26  2:29     ` NeilBrown
2011-07-26  5:17       ` Namhyung Kim
2011-07-26  8:48   ` Namhyung Kim
2011-07-26 15:03     ` [PATCH v2] md: add documentation for bad block log Namhyung Kim
2011-07-27  1:05     ` [md PATCH 02/36] md/bad-block-log: add sysfs interface for accessing bad-block-log NeilBrown
2011-07-21  2:58 ` [md PATCH 04/36] md: load/store badblock list from v1.x metadata NeilBrown
2011-07-22 16:34   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 03/36] md: don't allow arrays to contain devices with bad blocks NeilBrown
2011-07-22 15:47   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 05/36] md: Disable bad blocks and v0.90 metadata NeilBrown
2011-07-22 17:02   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 08/36] md: add 'write_error' flag to component devices NeilBrown
2011-07-26 15:22   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 13/36] md/raid1: Handle write errors by updating badblock log NeilBrown
2011-07-27 15:28   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 14/36] md/raid1: record badblocks found during resync etc NeilBrown
2011-07-27 15:39   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 07/36] md/raid1: avoid reading known bad blocks during resync NeilBrown
2011-07-26 14:25   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 11/36] md/raid1: clear bad-block record when write succeeds NeilBrown
2011-07-27  5:05   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 06/36] md/raid1: avoid reading from known bad blocks NeilBrown
2011-07-26 14:06   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 10/36] md/raid1: avoid writing to known-bad blocks on known-bad drives NeilBrown
2011-07-27  4:09   ` Namhyung Kim
2011-07-27  4:19     ` NeilBrown
2011-07-21  2:58 ` [md PATCH 12/36] md/raid1: store behind-write pages in bi_vecs NeilBrown
2011-07-27 15:16   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 09/36] md: make it easier to wait for bad blocks to be acknowledged NeilBrown
2011-07-26 16:04   ` Namhyung Kim
2011-07-27  1:18     ` NeilBrown
2011-07-21  2:58 ` [md PATCH 19/36] md/raid5: write errors should be recorded as bad blocks if possible NeilBrown
2011-07-21  2:58 ` [md PATCH 23/36] md/raid10: Split handle_read_error out from raid10d NeilBrown
2011-07-21  2:58 ` [md PATCH 16/36] md/raid1: factor several functions out or raid1d() NeilBrown
2011-07-27 15:55   ` Namhyung Kim
2011-07-28  1:39     ` NeilBrown
2011-07-21  2:58 ` [md PATCH 24/36] md/raid10: avoid reading from known bad blocks - part 1 NeilBrown
2011-07-21  2:58 ` [md PATCH 21/36] md/raid5: Clear bad blocks on successful write NeilBrown
2011-07-21  2:58 ` [md PATCH 15/36] md/raid1: improve handling of read failure during recovery NeilBrown
2011-07-27 15:45   ` Namhyung Kim
2011-07-21  2:58 ` [md PATCH 17/36] md/raid5: avoid reading from known bad blocks NeilBrown
2011-07-21  2:58 ` [md PATCH 18/36] md/raid5: use bad-block log to improve handling of uncorrectable read errors NeilBrown
2011-07-21  2:58 ` [md PATCH 22/36] md/raid10: simplify/reindent some loops NeilBrown
2011-07-21  2:58 ` [md PATCH 20/36] md/raid5. Don't write to known bad block on doubtful devices NeilBrown
2011-07-21  2:58 ` [md PATCH 27/36] md/raid10: avoid reading known bad blocks during resync/recovery NeilBrown
2011-07-21  2:58 ` [md PATCH 33/36] md/raid10: record bad blocks due to write errors " NeilBrown
2011-07-21  2:58 ` [md PATCH 29/36] md/raid10: avoid writing to known bad blocks on known bad drives NeilBrown
2011-07-21  2:58 ` [md PATCH 30/36] md/raid10: clear bad-block record when write succeeds NeilBrown
2011-07-21  2:58 ` [md PATCH 32/36] md/raid10: attempt to fix read errors during resync/check NeilBrown
2011-07-21  2:58 ` [md PATCH 26/36] md/raid10 - avoid reading from known bad blocks - part 3 NeilBrown
2011-07-21  2:58 ` [md PATCH 31/36] md/raid10: Handle write errors by updating badblock log NeilBrown
2011-07-21  2:58 ` [md PATCH 34/36] md/raid10: simplify read error handling during recovery NeilBrown
2011-07-21  2:58 ` [md PATCH 25/36] md/raid10: avoid reading from known bad blocks - part 2 NeilBrown
2011-07-21  2:58 ` [md PATCH 28/36] md/raid10 record bad blocks as needed during recovery NeilBrown
2011-07-21  2:58 ` [md PATCH 36/36] md/raid10: handle further errors during fix_read_error better NeilBrown
2011-07-21  2:58 ` [md PATCH 35/36] md/raid10: Handle read errors during recovery better NeilBrown

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=878vrq5pda.fsf@gmail.com \
    --to=namhyung@gmail.com \
    --cc=linux-raid@vger.kernel.org \
    --cc=neilb@suse.de \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).