From: Yu Kuai <yukuai1@huaweicloud.com>
To: Xueshi Hu <xueshi.hu@smartx.com>, song@kernel.org
Cc: linux-raid@vger.kernel.org, yukuai1@huaweicloud.com,
"yukuai (C)" <yukuai3@huawei.com>
Subject: Re: [PATCH v3 2/3] md/raid1: don't allow_barrier() before r1bio got freed
Date: Thu, 20 Jul 2023 09:47:32 +0800 [thread overview]
Message-ID: <f572ebcf-bfc7-739e-68f1-506d16d0ff31@huaweicloud.com> (raw)
In-Reply-To: <20230719070954.3084379-3-xueshi.hu@smartx.com>
Hi,
在 2023/07/19 15:09, Xueshi Hu 写道:
> allow_barrier() make reshape possible. Raid reshape changes the
> r1conf::raid_disks and mempool. Free the r1bio firstly and then call
> allow_barrier()
>
After adding fixes tags, feel free to add:
Reviewed-by: Yu Kuai <yukuai3@huawei.com>
> Signed-off-by: Xueshi Hu <xueshi.hu@smartx.com>
> ---
> drivers/md/raid1.c | 15 +++++++++------
> 1 file changed, 9 insertions(+), 6 deletions(-)
>
> diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c
> index 5605c9680818..62e86b7d1561 100644
> --- a/drivers/md/raid1.c
> +++ b/drivers/md/raid1.c
> @@ -313,6 +313,7 @@ static void raid_end_bio_io(struct r1bio *r1_bio)
> {
> struct bio *bio = r1_bio->master_bio;
> struct r1conf *conf = r1_bio->mddev->private;
> + sector_t sector = r1_bio->sector;
>
> /* if nobody has done the final endio yet, do it now */
> if (!test_and_set_bit(R1BIO_Returned, &r1_bio->state)) {
> @@ -323,13 +324,13 @@ static void raid_end_bio_io(struct r1bio *r1_bio)
>
> call_bio_endio(r1_bio);
> }
> +
> + free_r1bio(r1_bio);
> /*
> * Wake up any possible resync thread that waits for the device
> * to go idle. All I/Os, even write-behind writes, are done.
> */
> - allow_barrier(conf, r1_bio->sector);
> -
> - free_r1bio(r1_bio);
> + allow_barrier(conf, sector);
> }
>
> /*
> @@ -1404,6 +1405,7 @@ static void raid1_write_request(struct mddev *mddev, struct bio *bio,
> return;
> }
>
> + retry_write:
> r1_bio = alloc_r1bio(mddev, bio);
> r1_bio->sectors = max_write_sectors;
>
> @@ -1419,7 +1421,6 @@ static void raid1_write_request(struct mddev *mddev, struct bio *bio,
> */
>
> disks = conf->raid_disks * 2;
> - retry_write:
> blocked_rdev = NULL;
> rcu_read_lock();
> max_sectors = r1_bio->sectors;
> @@ -1499,7 +1500,7 @@ static void raid1_write_request(struct mddev *mddev, struct bio *bio,
> for (j = 0; j < i; j++)
> if (r1_bio->bios[j])
> rdev_dec_pending(conf->mirrors[j].rdev, mddev);
> - r1_bio->state = 0;
> + free_r1bio(r1_bio);
> allow_barrier(conf, bio->bi_iter.bi_sector);
>
> if (bio->bi_opf & REQ_NOWAIT) {
> @@ -2529,6 +2530,7 @@ static void handle_read_error(struct r1conf *conf, struct r1bio *r1_bio)
> struct mddev *mddev = conf->mddev;
> struct bio *bio;
> struct md_rdev *rdev;
> + sector_t sector;
>
> clear_bit(R1BIO_ReadError, &r1_bio->state);
> /* we got a read error. Maybe the drive is bad. Maybe just
> @@ -2558,12 +2560,13 @@ static void handle_read_error(struct r1conf *conf, struct r1bio *r1_bio)
> }
>
> rdev_dec_pending(rdev, conf->mddev);
> - allow_barrier(conf, r1_bio->sector);
> + sector = r1_bio->sector;
> bio = r1_bio->master_bio;
>
> /* Reuse the old r1_bio so that the IO_BLOCKED settings are preserved */
> r1_bio->state = 0;
> raid1_read_request(mddev, bio, r1_bio->sectors, r1_bio);
> + allow_barrier(conf, sector);
> }
>
> static void raid1d(struct md_thread *thread)
>
next prev parent reply other threads:[~2023-07-20 1:47 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-07-19 7:09 [PATCH v3 0/3] don't change mempool if in-flight r1bio exists Xueshi Hu
2023-07-19 7:09 ` [PATCH v3 1/3] md/raid1: freeze array more strictly when reshape Xueshi Hu
2023-07-20 1:36 ` Yu Kuai
2023-07-20 1:37 ` Yu Kuai
2023-07-31 14:02 ` Xueshi Hu
2023-08-01 1:24 ` Yu Kuai
2023-07-19 7:09 ` [PATCH v3 2/3] md/raid1: don't allow_barrier() before r1bio got freed Xueshi Hu
2023-07-20 1:47 ` Yu Kuai [this message]
2023-07-19 7:09 ` [PATCH v3 3/3] md/raid1: check array size before reshape Xueshi Hu
2023-07-19 7:38 ` Paul Menzel
2023-07-19 11:51 ` Xueshi Hu
2023-07-20 1:28 ` Yu Kuai
2023-07-28 14:42 ` Xueshi Hu
2023-07-29 0:58 ` Yu Kuai
2023-07-29 3:29 ` Xueshi Hu
2023-07-29 3:36 ` Yu Kuai
2023-07-29 3:51 ` Yu Kuai
2023-07-29 6:16 ` Xueshi Hu
2023-07-29 7:37 ` Yu Kuai
2023-07-29 12:23 ` Xueshi Hu
2023-07-31 1:03 ` Yu Kuai
2023-07-31 3:48 ` Xueshi Hu
2023-07-31 6:22 ` Yu Kuai
2023-07-31 14:12 ` Xueshi Hu
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=f572ebcf-bfc7-739e-68f1-506d16d0ff31@huaweicloud.com \
--to=yukuai1@huaweicloud.com \
--cc=linux-raid@vger.kernel.org \
--cc=song@kernel.org \
--cc=xueshi.hu@smartx.com \
--cc=yukuai3@huawei.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).