* mdadm --assemble considers event count for spares @ 2013-05-27 10:05 Alexander Lyakas 2013-05-28 1:16 ` NeilBrown 0 siblings, 1 reply; 7+ messages in thread From: Alexander Lyakas @ 2013-05-27 10:05 UTC (permalink / raw) To: linux-raid; +Cc: NeilBrown Hi Neil, It can happen that a spare has a higher event count than a in-array drive. For exampe: RAID1 with two drives is rebuilding one of the drives. Then the "good" drive fails. As a result, MD stops the rebuild and ejects the rebuilding drive from the array. The failed drive stays in the array, because RAID1 never ejects the last drive. However, the "good" drive fails all IOs, so the ejected drive has a larger event count now. Now if MD is stopped and re-assembled, mdadm considers the spare drive as the chosen one: root@vc:/mnt/work/alex/mdadm-neil# ./mdadm --assemble /dev/md200 --name=alex --config=none --homehost=vc --run --auto=md --metadata=1.2 --verbose --verbose /dev/sdc2 /dev/sdd2 mdadm: looking for devices for /dev/md200 mdadm: /dev/sdc2 is identified as a member of /dev/md200, slot 0. mdadm: /dev/sdd2 is identified as a member of /dev/md200, slot -1. mdadm: added /dev/sdc2 to /dev/md200 as 0 (possibly out of date) mdadm: no uptodate device for slot 2 of /dev/md200 mdadm: added /dev/sdd2 to /dev/md200 as -1 mdadm: failed to RUN_ARRAY /dev/md200: Input/output error mdadm: Not enough devices to start the array. Kernel doesn't accept the non-spare drive considering it as non-fresh: May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.679396] md: md200 stopped. May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.686870] md: bind<sdc2> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687623] md: bind<sdd2> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687675] md: kicking non-fresh sdc2 from array! May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687680] md: unbind<sdc2> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687683] md: export_rdev(sdc2) May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693574] md/raid1:md200: active with 0 out of 2 mirrors May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693583] md200: failed to create bitmap (-5) This happens with the latest mdadm from git, and kernel 3.8.2. Is this the expected behavior? Maybe mdadm should not consider spares at all for its "chosen_drive" logic, and perhaps not try to add them to the kernel? Superblocks of both drives: sdc2 - the "good" drive: /dev/sdc2: Magic : a92b4efc Version : 1.2 Feature Map : 0x1 Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b Name : zadara_vc:alex Creation Time : Mon May 27 11:33:50 2013 Raid Level : raid1 Raid Devices : 2 Avail Dev Size : 975063127 (464.95 GiB 499.23 GB) Array Size : 209715200 (200.00 GiB 214.75 GB) Used Dev Size : 419430400 (200.00 GiB 214.75 GB) Data Offset : 2048 sectors Super Offset : 8 sectors Unused Space : before=1968 sectors, after=555632727 sectors State : clean Device UUID : 1f661ca3:fdc8b887:8d3638ab:f2cc0a40 Internal Bitmap : 8 sectors from superblock Update Time : Mon May 27 11:34:57 2013 Checksum : 72a97357 - correct Events : 9 sdd2 - the "rebuilding" drive: /dev/sdd2: Magic : a92b4efc Version : 1.2 Feature Map : 0x1 Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b Name : zadara_vc:alex Creation Time : Mon May 27 11:33:50 2013 Raid Level : raid1 Raid Devices : 2 Avail Dev Size : 976123417 (465.45 GiB 499.78 GB) Array Size : 209715200 (200.00 GiB 214.75 GB) Used Dev Size : 419430400 (200.00 GiB 214.75 GB) Data Offset : 2048 sectors Super Offset : 8 sectors Unused Space : before=1968 sectors, after=556693017 sectors State : clean Device UUID : 9abc7fa9:6bf95a51:51f2cd65:14232e81 Internal Bitmap : 8 sectors from superblock Update Time : Mon May 27 11:35:56 2013 Checksum : 3e793a34 - correct Events : 26 Device Role : spare Array State : A. ('A' == active, '.' == missing, 'R' == replacing) Thanks, Alex. ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: mdadm --assemble considers event count for spares 2013-05-27 10:05 mdadm --assemble considers event count for spares Alexander Lyakas @ 2013-05-28 1:16 ` NeilBrown 2013-05-28 8:56 ` Alexander Lyakas 0 siblings, 1 reply; 7+ messages in thread From: NeilBrown @ 2013-05-28 1:16 UTC (permalink / raw) To: Alexander Lyakas; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 4274 bytes --] On Mon, 27 May 2013 13:05:34 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> wrote: > Hi Neil, > It can happen that a spare has a higher event count than a in-array drive. > For exampe: RAID1 with two drives is rebuilding one of the drives. > Then the "good" drive fails. As a result, MD stops the rebuild and > ejects the rebuilding drive from the array. The failed drive stays in > the array, because RAID1 never ejects the last drive. However, the > "good" drive fails all IOs, so the ejected drive has a larger event > count now. > Now if MD is stopped and re-assembled, mdadm considers the spare drive > as the chosen one: > > root@vc:/mnt/work/alex/mdadm-neil# ./mdadm --assemble /dev/md200 > --name=alex --config=none --homehost=vc --run --auto=md --metadata=1.2 > --verbose --verbose /dev/sdc2 /dev/sdd2 > mdadm: looking for devices for /dev/md200 > mdadm: /dev/sdc2 is identified as a member of /dev/md200, slot 0. > mdadm: /dev/sdd2 is identified as a member of /dev/md200, slot -1. > mdadm: added /dev/sdc2 to /dev/md200 as 0 (possibly out of date) > mdadm: no uptodate device for slot 2 of /dev/md200 > mdadm: added /dev/sdd2 to /dev/md200 as -1 > mdadm: failed to RUN_ARRAY /dev/md200: Input/output error > mdadm: Not enough devices to start the array. > > Kernel doesn't accept the non-spare drive considering it as non-fresh: > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.679396] md: md200 stopped. > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.686870] md: bind<sdc2> > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687623] md: bind<sdd2> > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687675] md: kicking > non-fresh sdc2 from array! > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687680] md: unbind<sdc2> > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687683] md: export_rdev(sdc2) > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693574] > md/raid1:md200: active with 0 out of 2 mirrors > May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693583] md200: > failed to create bitmap (-5) > > This happens with the latest mdadm from git, and kernel 3.8.2. > > Is this the expected behavior? I hadn't thought about it. > Maybe mdadm should not consider spares at all for its "chosen_drive" > logic, and perhaps not try to add them to the kernel? Probably not, no. NeilBrown > > Superblocks of both drives: > sdc2 - the "good" drive: > /dev/sdc2: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x1 > Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b > Name : zadara_vc:alex > Creation Time : Mon May 27 11:33:50 2013 > Raid Level : raid1 > Raid Devices : 2 > > Avail Dev Size : 975063127 (464.95 GiB 499.23 GB) > Array Size : 209715200 (200.00 GiB 214.75 GB) > Used Dev Size : 419430400 (200.00 GiB 214.75 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > Unused Space : before=1968 sectors, after=555632727 sectors > State : clean > Device UUID : 1f661ca3:fdc8b887:8d3638ab:f2cc0a40 > > Internal Bitmap : 8 sectors from superblock > Update Time : Mon May 27 11:34:57 2013 > Checksum : 72a97357 - correct > Events : 9 > > sdd2 - the "rebuilding" drive: > /dev/sdd2: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x1 > Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b > Name : zadara_vc:alex > Creation Time : Mon May 27 11:33:50 2013 > Raid Level : raid1 > Raid Devices : 2 > > Avail Dev Size : 976123417 (465.45 GiB 499.78 GB) > Array Size : 209715200 (200.00 GiB 214.75 GB) > Used Dev Size : 419430400 (200.00 GiB 214.75 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > Unused Space : before=1968 sectors, after=556693017 sectors > State : clean > Device UUID : 9abc7fa9:6bf95a51:51f2cd65:14232e81 > > Internal Bitmap : 8 sectors from superblock > Update Time : Mon May 27 11:35:56 2013 > Checksum : 3e793a34 - correct > Events : 26 > > > Device Role : spare > Array State : A. ('A' == active, '.' == missing, 'R' == replacing) > > > Thanks, > Alex. [-- Attachment #2: signature.asc --] [-- Type: application/pgp-signature, Size: 828 bytes --] ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: mdadm --assemble considers event count for spares 2013-05-28 1:16 ` NeilBrown @ 2013-05-28 8:56 ` Alexander Lyakas 2013-05-28 9:15 ` NeilBrown 0 siblings, 1 reply; 7+ messages in thread From: Alexander Lyakas @ 2013-05-28 8:56 UTC (permalink / raw) To: NeilBrown; +Cc: linux-raid Hi Neil, can you please let me know what have you decided after/whether you had time to think about this issue. Thanks, Alex. On Tue, May 28, 2013 at 4:16 AM, NeilBrown <neilb@suse.de> wrote: > On Mon, 27 May 2013 13:05:34 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> > wrote: > >> Hi Neil, >> It can happen that a spare has a higher event count than a in-array drive. >> For exampe: RAID1 with two drives is rebuilding one of the drives. >> Then the "good" drive fails. As a result, MD stops the rebuild and >> ejects the rebuilding drive from the array. The failed drive stays in >> the array, because RAID1 never ejects the last drive. However, the >> "good" drive fails all IOs, so the ejected drive has a larger event >> count now. >> Now if MD is stopped and re-assembled, mdadm considers the spare drive >> as the chosen one: >> >> root@vc:/mnt/work/alex/mdadm-neil# ./mdadm --assemble /dev/md200 >> --name=alex --config=none --homehost=vc --run --auto=md --metadata=1.2 >> --verbose --verbose /dev/sdc2 /dev/sdd2 >> mdadm: looking for devices for /dev/md200 >> mdadm: /dev/sdc2 is identified as a member of /dev/md200, slot 0. >> mdadm: /dev/sdd2 is identified as a member of /dev/md200, slot -1. >> mdadm: added /dev/sdc2 to /dev/md200 as 0 (possibly out of date) >> mdadm: no uptodate device for slot 2 of /dev/md200 >> mdadm: added /dev/sdd2 to /dev/md200 as -1 >> mdadm: failed to RUN_ARRAY /dev/md200: Input/output error >> mdadm: Not enough devices to start the array. >> >> Kernel doesn't accept the non-spare drive considering it as non-fresh: >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.679396] md: md200 stopped. >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.686870] md: bind<sdc2> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687623] md: bind<sdd2> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687675] md: kicking >> non-fresh sdc2 from array! >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687680] md: unbind<sdc2> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687683] md: export_rdev(sdc2) >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693574] >> md/raid1:md200: active with 0 out of 2 mirrors >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693583] md200: >> failed to create bitmap (-5) >> >> This happens with the latest mdadm from git, and kernel 3.8.2. >> >> Is this the expected behavior? > > I hadn't thought about it. > >> Maybe mdadm should not consider spares at all for its "chosen_drive" >> logic, and perhaps not try to add them to the kernel? > > Probably not, no. > > NeilBrown > > > >> >> Superblocks of both drives: >> sdc2 - the "good" drive: >> /dev/sdc2: >> Magic : a92b4efc >> Version : 1.2 >> Feature Map : 0x1 >> Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b >> Name : zadara_vc:alex >> Creation Time : Mon May 27 11:33:50 2013 >> Raid Level : raid1 >> Raid Devices : 2 >> >> Avail Dev Size : 975063127 (464.95 GiB 499.23 GB) >> Array Size : 209715200 (200.00 GiB 214.75 GB) >> Used Dev Size : 419430400 (200.00 GiB 214.75 GB) >> Data Offset : 2048 sectors >> Super Offset : 8 sectors >> Unused Space : before=1968 sectors, after=555632727 sectors >> State : clean >> Device UUID : 1f661ca3:fdc8b887:8d3638ab:f2cc0a40 >> >> Internal Bitmap : 8 sectors from superblock >> Update Time : Mon May 27 11:34:57 2013 >> Checksum : 72a97357 - correct >> Events : 9 >> >> sdd2 - the "rebuilding" drive: >> /dev/sdd2: >> Magic : a92b4efc >> Version : 1.2 >> Feature Map : 0x1 >> Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b >> Name : zadara_vc:alex >> Creation Time : Mon May 27 11:33:50 2013 >> Raid Level : raid1 >> Raid Devices : 2 >> >> Avail Dev Size : 976123417 (465.45 GiB 499.78 GB) >> Array Size : 209715200 (200.00 GiB 214.75 GB) >> Used Dev Size : 419430400 (200.00 GiB 214.75 GB) >> Data Offset : 2048 sectors >> Super Offset : 8 sectors >> Unused Space : before=1968 sectors, after=556693017 sectors >> State : clean >> Device UUID : 9abc7fa9:6bf95a51:51f2cd65:14232e81 >> >> Internal Bitmap : 8 sectors from superblock >> Update Time : Mon May 27 11:35:56 2013 >> Checksum : 3e793a34 - correct >> Events : 26 >> >> >> Device Role : spare >> Array State : A. ('A' == active, '.' == missing, 'R' == replacing) >> >> >> Thanks, >> Alex. > ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: mdadm --assemble considers event count for spares 2013-05-28 8:56 ` Alexander Lyakas @ 2013-05-28 9:15 ` NeilBrown 2013-05-28 10:50 ` Alexander Lyakas 0 siblings, 1 reply; 7+ messages in thread From: NeilBrown @ 2013-05-28 9:15 UTC (permalink / raw) To: Alexander Lyakas; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 5161 bytes --] On Tue, 28 May 2013 11:56:26 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> wrote: > Hi Neil, > can you please let me know what have you decided after/whether you had > time to think about this issue. I don't actually plan to think about the issue, at least not in the short term. If you would like to propose a concrete solution, then I would probably be motivated to think about that and give you some feedback. NeilBrown > > Thanks, > Alex. > > > > > On Tue, May 28, 2013 at 4:16 AM, NeilBrown <neilb@suse.de> wrote: > > On Mon, 27 May 2013 13:05:34 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> > > wrote: > > > >> Hi Neil, > >> It can happen that a spare has a higher event count than a in-array drive. > >> For exampe: RAID1 with two drives is rebuilding one of the drives. > >> Then the "good" drive fails. As a result, MD stops the rebuild and > >> ejects the rebuilding drive from the array. The failed drive stays in > >> the array, because RAID1 never ejects the last drive. However, the > >> "good" drive fails all IOs, so the ejected drive has a larger event > >> count now. > >> Now if MD is stopped and re-assembled, mdadm considers the spare drive > >> as the chosen one: > >> > >> root@vc:/mnt/work/alex/mdadm-neil# ./mdadm --assemble /dev/md200 > >> --name=alex --config=none --homehost=vc --run --auto=md --metadata=1.2 > >> --verbose --verbose /dev/sdc2 /dev/sdd2 > >> mdadm: looking for devices for /dev/md200 > >> mdadm: /dev/sdc2 is identified as a member of /dev/md200, slot 0. > >> mdadm: /dev/sdd2 is identified as a member of /dev/md200, slot -1. > >> mdadm: added /dev/sdc2 to /dev/md200 as 0 (possibly out of date) > >> mdadm: no uptodate device for slot 2 of /dev/md200 > >> mdadm: added /dev/sdd2 to /dev/md200 as -1 > >> mdadm: failed to RUN_ARRAY /dev/md200: Input/output error > >> mdadm: Not enough devices to start the array. > >> > >> Kernel doesn't accept the non-spare drive considering it as non-fresh: > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.679396] md: md200 stopped. > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.686870] md: bind<sdc2> > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687623] md: bind<sdd2> > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687675] md: kicking > >> non-fresh sdc2 from array! > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687680] md: unbind<sdc2> > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687683] md: export_rdev(sdc2) > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693574] > >> md/raid1:md200: active with 0 out of 2 mirrors > >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693583] md200: > >> failed to create bitmap (-5) > >> > >> This happens with the latest mdadm from git, and kernel 3.8.2. > >> > >> Is this the expected behavior? > > > > I hadn't thought about it. > > > >> Maybe mdadm should not consider spares at all for its "chosen_drive" > >> logic, and perhaps not try to add them to the kernel? > > > > Probably not, no. > > > > NeilBrown > > > > > > > >> > >> Superblocks of both drives: > >> sdc2 - the "good" drive: > >> /dev/sdc2: > >> Magic : a92b4efc > >> Version : 1.2 > >> Feature Map : 0x1 > >> Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b > >> Name : zadara_vc:alex > >> Creation Time : Mon May 27 11:33:50 2013 > >> Raid Level : raid1 > >> Raid Devices : 2 > >> > >> Avail Dev Size : 975063127 (464.95 GiB 499.23 GB) > >> Array Size : 209715200 (200.00 GiB 214.75 GB) > >> Used Dev Size : 419430400 (200.00 GiB 214.75 GB) > >> Data Offset : 2048 sectors > >> Super Offset : 8 sectors > >> Unused Space : before=1968 sectors, after=555632727 sectors > >> State : clean > >> Device UUID : 1f661ca3:fdc8b887:8d3638ab:f2cc0a40 > >> > >> Internal Bitmap : 8 sectors from superblock > >> Update Time : Mon May 27 11:34:57 2013 > >> Checksum : 72a97357 - correct > >> Events : 9 > >> > >> sdd2 - the "rebuilding" drive: > >> /dev/sdd2: > >> Magic : a92b4efc > >> Version : 1.2 > >> Feature Map : 0x1 > >> Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b > >> Name : zadara_vc:alex > >> Creation Time : Mon May 27 11:33:50 2013 > >> Raid Level : raid1 > >> Raid Devices : 2 > >> > >> Avail Dev Size : 976123417 (465.45 GiB 499.78 GB) > >> Array Size : 209715200 (200.00 GiB 214.75 GB) > >> Used Dev Size : 419430400 (200.00 GiB 214.75 GB) > >> Data Offset : 2048 sectors > >> Super Offset : 8 sectors > >> Unused Space : before=1968 sectors, after=556693017 sectors > >> State : clean > >> Device UUID : 9abc7fa9:6bf95a51:51f2cd65:14232e81 > >> > >> Internal Bitmap : 8 sectors from superblock > >> Update Time : Mon May 27 11:35:56 2013 > >> Checksum : 3e793a34 - correct > >> Events : 26 > >> > >> > >> Device Role : spare > >> Array State : A. ('A' == active, '.' == missing, 'R' == replacing) > >> > >> > >> Thanks, > >> Alex. > > [-- Attachment #2: signature.asc --] [-- Type: application/pgp-signature, Size: 828 bytes --] ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: mdadm --assemble considers event count for spares 2013-05-28 9:15 ` NeilBrown @ 2013-05-28 10:50 ` Alexander Lyakas 2013-06-03 23:51 ` NeilBrown 2013-06-17 6:57 ` NeilBrown 0 siblings, 2 replies; 7+ messages in thread From: Alexander Lyakas @ 2013-05-28 10:50 UTC (permalink / raw) To: NeilBrown; +Cc: linux-raid Neil, In my opinion (I may be wrong), a spare drive (raid_disk==-1) doesn't add any information to array assembly. It doesn't have a valid raid slot, and I don't see how its event count is relevant. I don't think a spare can help us much in figuring out array's latest state, which is what assembly code tries to do. So what I was thinking: mdadm --assemble doesn't consider spare drives (raid_disk=-1) at all. It simply skips over them in the initial loop after reading their superblocks. Perhaps it can keep them in a side list. Then array is assembled with non-spare drives only. After array is assembled, we may choose one of the following: # User has to explicitly add the spare drives after array has been assembled. Assemble can warn that some spares have been left out, and tell the user what they are. # Assemble adds the spare drives (perhaps after zeroing their superblocks even), after it assembled the array with non-spare drives. Alex. On Tue, May 28, 2013 at 12:15 PM, NeilBrown <neilb@suse.de> wrote: > On Tue, 28 May 2013 11:56:26 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> > wrote: > >> Hi Neil, >> can you please let me know what have you decided after/whether you had >> time to think about this issue. > > I don't actually plan to think about the issue, at least not in the short > term. > If you would like to propose a concrete solution, then I would probably be > motivated to think about that and give you some feedback. > > NeilBrown > >> >> Thanks, >> Alex. >> >> >> >> >> On Tue, May 28, 2013 at 4:16 AM, NeilBrown <neilb@suse.de> wrote: >> > On Mon, 27 May 2013 13:05:34 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> >> > wrote: >> > >> >> Hi Neil, >> >> It can happen that a spare has a higher event count than a in-array drive. >> >> For exampe: RAID1 with two drives is rebuilding one of the drives. >> >> Then the "good" drive fails. As a result, MD stops the rebuild and >> >> ejects the rebuilding drive from the array. The failed drive stays in >> >> the array, because RAID1 never ejects the last drive. However, the >> >> "good" drive fails all IOs, so the ejected drive has a larger event >> >> count now. >> >> Now if MD is stopped and re-assembled, mdadm considers the spare drive >> >> as the chosen one: >> >> >> >> root@vc:/mnt/work/alex/mdadm-neil# ./mdadm --assemble /dev/md200 >> >> --name=alex --config=none --homehost=vc --run --auto=md --metadata=1.2 >> >> --verbose --verbose /dev/sdc2 /dev/sdd2 >> >> mdadm: looking for devices for /dev/md200 >> >> mdadm: /dev/sdc2 is identified as a member of /dev/md200, slot 0. >> >> mdadm: /dev/sdd2 is identified as a member of /dev/md200, slot -1. >> >> mdadm: added /dev/sdc2 to /dev/md200 as 0 (possibly out of date) >> >> mdadm: no uptodate device for slot 2 of /dev/md200 >> >> mdadm: added /dev/sdd2 to /dev/md200 as -1 >> >> mdadm: failed to RUN_ARRAY /dev/md200: Input/output error >> >> mdadm: Not enough devices to start the array. >> >> >> >> Kernel doesn't accept the non-spare drive considering it as non-fresh: >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.679396] md: md200 stopped. >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.686870] md: bind<sdc2> >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687623] md: bind<sdd2> >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687675] md: kicking >> >> non-fresh sdc2 from array! >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687680] md: unbind<sdc2> >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.687683] md: export_rdev(sdc2) >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693574] >> >> md/raid1:md200: active with 0 out of 2 mirrors >> >> May 27 12:42:28 vsa-00000505-vc-0 kernel: [343203.693583] md200: >> >> failed to create bitmap (-5) >> >> >> >> This happens with the latest mdadm from git, and kernel 3.8.2. >> >> >> >> Is this the expected behavior? >> > >> > I hadn't thought about it. >> > >> >> Maybe mdadm should not consider spares at all for its "chosen_drive" >> >> logic, and perhaps not try to add them to the kernel? >> > >> > Probably not, no. >> > >> > NeilBrown >> > >> > >> > >> >> >> >> Superblocks of both drives: >> >> sdc2 - the "good" drive: >> >> /dev/sdc2: >> >> Magic : a92b4efc >> >> Version : 1.2 >> >> Feature Map : 0x1 >> >> Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b >> >> Name : zadara_vc:alex >> >> Creation Time : Mon May 27 11:33:50 2013 >> >> Raid Level : raid1 >> >> Raid Devices : 2 >> >> >> >> Avail Dev Size : 975063127 (464.95 GiB 499.23 GB) >> >> Array Size : 209715200 (200.00 GiB 214.75 GB) >> >> Used Dev Size : 419430400 (200.00 GiB 214.75 GB) >> >> Data Offset : 2048 sectors >> >> Super Offset : 8 sectors >> >> Unused Space : before=1968 sectors, after=555632727 sectors >> >> State : clean >> >> Device UUID : 1f661ca3:fdc8b887:8d3638ab:f2cc0a40 >> >> >> >> Internal Bitmap : 8 sectors from superblock >> >> Update Time : Mon May 27 11:34:57 2013 >> >> Checksum : 72a97357 - correct >> >> Events : 9 >> >> >> >> sdd2 - the "rebuilding" drive: >> >> /dev/sdd2: >> >> Magic : a92b4efc >> >> Version : 1.2 >> >> Feature Map : 0x1 >> >> Array UUID : 8e051cc5:c536d16e:72b413fa:e7049d4b >> >> Name : zadara_vc:alex >> >> Creation Time : Mon May 27 11:33:50 2013 >> >> Raid Level : raid1 >> >> Raid Devices : 2 >> >> >> >> Avail Dev Size : 976123417 (465.45 GiB 499.78 GB) >> >> Array Size : 209715200 (200.00 GiB 214.75 GB) >> >> Used Dev Size : 419430400 (200.00 GiB 214.75 GB) >> >> Data Offset : 2048 sectors >> >> Super Offset : 8 sectors >> >> Unused Space : before=1968 sectors, after=556693017 sectors >> >> State : clean >> >> Device UUID : 9abc7fa9:6bf95a51:51f2cd65:14232e81 >> >> >> >> Internal Bitmap : 8 sectors from superblock >> >> Update Time : Mon May 27 11:35:56 2013 >> >> Checksum : 3e793a34 - correct >> >> Events : 26 >> >> >> >> >> >> Device Role : spare >> >> Array State : A. ('A' == active, '.' == missing, 'R' == replacing) >> >> >> >> >> >> Thanks, >> >> Alex. >> > > ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: mdadm --assemble considers event count for spares 2013-05-28 10:50 ` Alexander Lyakas @ 2013-06-03 23:51 ` NeilBrown 2013-06-17 6:57 ` NeilBrown 1 sibling, 0 replies; 7+ messages in thread From: NeilBrown @ 2013-06-03 23:51 UTC (permalink / raw) To: Alexander Lyakas; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 1528 bytes --] On Tue, 28 May 2013 13:50:49 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> wrote: > Neil, > In my opinion (I may be wrong), a spare drive (raid_disk==-1) doesn't > add any information to array assembly. It doesn't have a valid raid > slot, and I don't see how its event count is relevant. I don't think a > spare can help us much in figuring out array's latest state, which is > what assembly code tries to do. > So what I was thinking: mdadm --assemble doesn't consider spare drives > (raid_disk=-1) at all. It simply skips over them in the initial loop > after reading their superblocks. Perhaps it can keep them in a side > list. Then array is assembled with non-spare drives only. Sounds reasonable. I would suggest looking at the place where 'most_recent' is set in Assemble.c, and get it to avoid updating 'most_recent' if the current device is a spare. Something like if (most_recent < devcnt) { if (devices[devcnt].i.events > devices[most_recent].i.events) + if (devices[devcnt].i.disk.state == 6) most_recent = devcnt; } Care to give that a try? NeilBrown > > After array is assembled, we may choose one of the following: > # User has to explicitly add the spare drives after array has been > assembled. Assemble can warn that some spares have been left out, and > tell the user what they are. > # Assemble adds the spare drives (perhaps after zeroing their > superblocks even), after it assembled the array with non-spare drives. > > Alex. [-- Attachment #2: signature.asc --] [-- Type: application/pgp-signature, Size: 828 bytes --] ^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: mdadm --assemble considers event count for spares 2013-05-28 10:50 ` Alexander Lyakas 2013-06-03 23:51 ` NeilBrown @ 2013-06-17 6:57 ` NeilBrown 1 sibling, 0 replies; 7+ messages in thread From: NeilBrown @ 2013-06-17 6:57 UTC (permalink / raw) To: Alexander Lyakas; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 1319 bytes --] On Tue, 28 May 2013 13:50:49 +0300 Alexander Lyakas <alex.bolshoy@gmail.com> wrote: > Neil, > In my opinion (I may be wrong), a spare drive (raid_disk==-1) doesn't > add any information to array assembly. It doesn't have a valid raid > slot, and I don't see how its event count is relevant. I don't think a > spare can help us much in figuring out array's latest state, which is > what assembly code tries to do. > So what I was thinking: mdadm --assemble doesn't consider spare drives > (raid_disk=-1) at all. It simply skips over them in the initial loop > after reading their superblocks. Perhaps it can keep them in a side > list. Then array is assembled with non-spare drives only. > > After array is assembled, we may choose one of the following: > # User has to explicitly add the spare drives after array has been > assembled. Assemble can warn that some spares have been left out, and > tell the user what they are. > # Assemble adds the spare drives (perhaps after zeroing their > superblocks even), after it assembled the array with non-spare drives. Hi, I have just committed http://git.neil.brown.name/?p=mdadm.git;a=commitdiff;h=f80057aec5d314798251e318555cb8ac92e4c06f which I believe fixes this issue. If you can test and confirm I would appreciate it. NeilBrown [-- Attachment #2: signature.asc --] [-- Type: application/pgp-signature, Size: 828 bytes --] ^ permalink raw reply [flat|nested] 7+ messages in thread
end of thread, other threads:[~2013-06-17 6:57 UTC | newest] Thread overview: 7+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2013-05-27 10:05 mdadm --assemble considers event count for spares Alexander Lyakas 2013-05-28 1:16 ` NeilBrown 2013-05-28 8:56 ` Alexander Lyakas 2013-05-28 9:15 ` NeilBrown 2013-05-28 10:50 ` Alexander Lyakas 2013-06-03 23:51 ` NeilBrown 2013-06-17 6:57 ` NeilBrown
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).