linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Xiao Ni <xni@redhat.com>
To: NeilBrown <neilb@suse.de>
Cc: linux-raid@vger.kernel.org
Subject: Re: raid5 reshape is stuck
Date: Wed, 20 May 2015 23:37:57 -0400 (EDT)	[thread overview]
Message-ID: <1552584910.2141570.1432179477679.JavaMail.zimbra@redhat.com> (raw)
In-Reply-To: <20150521094837.6a2d29c4@notabene.brown>



----- Original Message -----
> From: "NeilBrown" <neilb@suse.de>
> To: "Xiao Ni" <xni@redhat.com>
> Cc: linux-raid@vger.kernel.org
> Sent: Thursday, May 21, 2015 7:48:37 AM
> Subject: Re: raid5 reshape is stuck
> 
> On Fri, 15 May 2015 03:00:24 -0400 (EDT) Xiao Ni <xni@redhat.com> wrote:
> 
> > Hi Neil
> > 
> >    I encounter the problem when I reshape a 4-disks raid5 to raid5. It just
> >    can
> > appear with loop devices.
> > 
> >    The steps are:
> > 
> > [root@dhcp-12-158 mdadm-3.3.2]# mdadm -CR /dev/md0 -l5 -n5 /dev/loop[0-4]
> > --assume-clean
> > mdadm: /dev/loop0 appears to be part of a raid array:
> >        level=raid5 devices=6 ctime=Fri May 15 13:47:17 2015
> > mdadm: /dev/loop1 appears to be part of a raid array:
> >        level=raid5 devices=6 ctime=Fri May 15 13:47:17 2015
> > mdadm: /dev/loop2 appears to be part of a raid array:
> >        level=raid5 devices=6 ctime=Fri May 15 13:47:17 2015
> > mdadm: /dev/loop3 appears to be part of a raid array:
> >        level=raid5 devices=6 ctime=Fri May 15 13:47:17 2015
> > mdadm: /dev/loop4 appears to be part of a raid array:
> >        level=raid5 devices=6 ctime=Fri May 15 13:47:17 2015
> > mdadm: Defaulting to version 1.2 metadata
> > mdadm: array /dev/md0 started.
> > [root@dhcp-12-158 mdadm-3.3.2]# mdadm /dev/md0 -a /dev/loop5
> > mdadm: added /dev/loop5
> > [root@dhcp-12-158 mdadm-3.3.2]# mdadm --grow /dev/md0 --raid-devices 6
> > mdadm: Need to backup 10240K of critical section..
> > [root@dhcp-12-158 mdadm-3.3.2]# cat /proc/mdstat
> > Personalities : [raid6] [raid5] [raid4]
> > md0 : active raid5 loop5[5] loop4[4] loop3[3] loop2[2] loop1[1] loop0[0]
> >       8187904 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6]
> >       [UUUUUU]
> >       [>....................]  reshape =  0.0% (0/2046976) finish=6396.8min
> >       speed=0K/sec
> >       
> > unused devices: <none>
> > 
> >    It because the sync_max is set to 0 when run the command --grow
> > 
> > [root@dhcp-12-158 mdadm-3.3.2]# cd /sys/block/md0/md/
> > [root@dhcp-12-158 md]# cat sync_max
> > 0
> > 
> >    I tried reproduce with normal sata devices. The progress of reshape is
> >    no problem. Then
> > I checked the Grow.c. If I use sata devices, in function reshape_array, the
> > return value
> > of set_new_data_offset is 0. But if I used loop devices, it return 1. Then
> > it call the function
> > start_reshape.
> 
> set_new_data_offset returns '0' if there is room on the devices to reduce the
> data offset so that the reshape starts writing to unused space on the array.
> This removes the need for a backup file, or the use of a spare device to
> store a temporary backup.
> It returns '1' if there was no room for relocating the data_offset.
> 
> So on your sata devices (which are presumably larger than your loop devices)
> there was room.  On your loop devices there was not.
> 
> 
> > 
> >    In the function start_reshape it set the sync_max to reshape_progress.
> >    But in sysfs_read it
> > doesn't read reshape_progress. So it's 0 and the sync_max is set to 0. Why
> > it need to set the
> > sync_max at this? I'm not sure about this.
> 
> sync_max is set to 0 so that the reshape does not start until the backup has
> been taken.
> Once the backup is taken, child_monitor() should set sync_max to "max".
> 
> Can you  check if that is happening?
> 
> Thanks,
> NeilBrown
> 
> 

  Thanks very much for the explaining. The problem maybe is fixed. I tried reproduce this with newest
kernel and newest mdadm. Now the problem don't exist. I'll do more tests and give the answer above later.

Best Regards
Xiao

  reply	other threads:[~2015-05-21  3:37 UTC|newest]

Thread overview: 20+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <1612858661.15347659.1431671671467.JavaMail.zimbra@redhat.com>
2015-05-15  7:00 ` raid5 reshape is stuck Xiao Ni
2015-05-19 11:10   ` Xiao Ni
2015-05-20 23:48   ` NeilBrown
2015-05-21  3:37     ` Xiao Ni [this message]
2015-05-21 12:31       ` Xiao Ni
2015-05-22  8:54         ` Xiao Ni
2015-05-25  3:50         ` NeilBrown
2015-05-26 10:00           ` Xiao Ni
2015-05-26 10:48           ` Xiao Ni
2015-05-27  0:02             ` NeilBrown
2015-05-27  1:10               ` NeilBrown
2015-05-27 11:28                 ` Xiao Ni
2015-05-27 11:34                   ` NeilBrown
2015-05-27 12:04                     ` Xiao Ni
2015-05-27 22:59                       ` NeilBrown
2015-05-28  6:32                         ` Xiao Ni
2015-05-28  6:49                           ` NeilBrown
2015-05-29 11:13                             ` XiaoNi
2015-05-29 11:19                               ` NeilBrown
2015-05-29 12:19                                 ` XiaoNi

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1552584910.2141570.1432179477679.JavaMail.zimbra@redhat.com \
    --to=xni@redhat.com \
    --cc=linux-raid@vger.kernel.org \
    --cc=neilb@suse.de \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).