From: Daniel Sanabria <sanabria.d@gmail.com>
To: Wols Lists <antlists@youngman.org.uk>
Cc: linux-raid@vger.kernel.org
Subject: Re: Inactive arrays
Date: Mon, 12 Sep 2016 20:41:53 +0100 [thread overview]
Message-ID: <CAHscji35kySP4Q8cUpYCXR2P9DBwPcYeWjr_HX==TNX9CPW9NA@mail.gmail.com> (raw)
In-Reply-To: <CAHscji3HdHaJAvwtYNU=Ykc_qohBkfFfrbP0M=FNhuFMK+d-Jg@mail.gmail.com>
ok, I just adjusted system time so that I can start tracking logs.
what I'm noticing however is that fdisk -l is not giving me the expect
partitions (I was expecting at least 2 partitions in every 2.7 disk
similar to what I have in sdd):
[root@lamachine lamachine_220315]# fdisk -l /dev/{sdc,sdd,sde}
Disk /dev/sdc: 2.7 TiB, 3000591900160 bytes, 5860531055 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 4096 bytes
I/O size (minimum/optimal): 4096 bytes / 4096 bytes
Disklabel type: dos
Disk identifier: 0x00000000
Device Boot Start End Sectors Size Id Type
/dev/sdc1 1 4294967295 4294967295 2T ee GPT
Partition 1 does not start on physical sector boundary.
Disk /dev/sdd: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 4096 bytes
I/O size (minimum/optimal): 4096 bytes / 4096 bytes
Disklabel type: gpt
Disk identifier: D3233810-F552-4126-8281-7F71A4938DF9
Device Start End Sectors Size Type
/dev/sdd1 2048 4294969343 4294967296 2T Linux RAID
/dev/sdd2 4294969344 5343545343 1048576000 500G Linux filesystem
Disk /dev/sde: 2.7 TiB, 3000591900160 bytes, 5860531055 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 4096 bytes
I/O size (minimum/optimal): 4096 bytes / 4096 bytes
Disklabel type: dos
Disk identifier: 0x00000000
Device Boot Start End Sectors Size Id Type
/dev/sde1 1 4294967295 4294967295 2T ee GPT
Partition 1 does not start on physical sector boundary.
[root@lamachine lamachine_220315]#
what could've happened here? any ideas why the partition tables ended
up like that?
From previous information I have an idea of what the md128 and md129
are supposed to looks like (also noticed that the device names
changed):
# md128 and md129 details From an old command output
/dev/md128:
Version : 1.2
Creation Time : Fri Oct 24 15:24:38 2014
Raid Level : raid5
Array Size : 4294705152 (4095.75 GiB 4397.78 GB)
Used Dev Size : 2147352576 (2047.88 GiB 2198.89 GB)
Raid Devices : 3
Total Devices : 3
Persistence : Superblock is persistent
Intent Bitmap : Internal
Update Time : Sun Mar 22 06:20:08 2015
State : clean
Active Devices : 3
Working Devices : 3
Failed Devices : 0
Spare Devices : 0
Layout : left-symmetric
Chunk Size : 512K
Name : lamachine:128 (local to host lamachine)
UUID : f2372cb9:d3816fd6:ce86d826:882ec82e
Events : 4041
Number Major Minor RaidDevice State
0 8 49 0 active sync /dev/sdd1
1 8 65 1 active sync /dev/sde1
3 8 81 2 active sync /dev/sdf1
/dev/md129:
Version : 1.2
Creation Time : Mon Nov 10 16:28:11 2014
Raid Level : raid0
Array Size : 1572470784 (1499.63 GiB 1610.21 GB)
Raid Devices : 3
Total Devices : 3
Persistence : Superblock is persistent
Update Time : Mon Nov 10 16:28:11 2014
State : clean
Active Devices : 3
Working Devices : 3
Failed Devices : 0
Spare Devices : 0
Chunk Size : 512K
Name : lamachine:129 (local to host lamachine)
UUID : 895dae98:d1a496de:4f590b8b:cb8ac12a
Events : 0
Number Major Minor RaidDevice State
0 8 50 0 active sync /dev/sdd2
1 8 66 1 active sync /dev/sde2
2 8 82 2 active sync /dev/sdf2
Is there any way to recover the contents of these two arrays ? :(
On 11 September 2016 at 21:06, Daniel Sanabria <sanabria.d@gmail.com> wrote:
> However I'm noticing that the details with this new MB are somewhat different:
>
> [root@lamachine ~]# cat /etc/mdadm.conf
> # mdadm.conf written out by anaconda
> MAILADDR root
> AUTO +imsm +1.x -all
> ARRAY /dev/md2 level=raid5 num-devices=3
> UUID=2cff15d1:e411447b:fd5d4721:03e44022
> ARRAY /dev/md126 level=raid10 num-devices=2
> UUID=9af006ca:8845bbd3:bfe78010:bc810f04
> ARRAY /dev/md127 level=raid0 num-devices=3
> UUID=acd5374f:72628c93:6a906c4b:5f675ce5
> ARRAY /dev/md128 metadata=1.2 spares=1 name=lamachine:128
> UUID=f2372cb9:d3816fd6:ce86d826:882ec82e
> ARRAY /dev/md129 metadata=1.2 name=lamachine:129
> UUID=895dae98:d1a496de:4f590b8b:cb8ac12a
> [root@lamachine ~]# mdadm --detail /dev/md1*
> /dev/md126:
> Version : 0.90
> Creation Time : Thu Dec 3 22:12:12 2009
> Raid Level : raid10
> Array Size : 30719936 (29.30 GiB 31.46 GB)
> Used Dev Size : 30719936 (29.30 GiB 31.46 GB)
> Raid Devices : 2
> Total Devices : 2
> Preferred Minor : 126
> Persistence : Superblock is persistent
>
> Update Time : Tue Jan 12 04:03:41 2016
> State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Layout : near=2
> Chunk Size : 64K
>
> UUID : 9af006ca:8845bbd3:bfe78010:bc810f04
> Events : 0.264152
>
> Number Major Minor RaidDevice State
> 0 8 82 0 active sync set-A /dev/sdf2
> 1 8 1 1 active sync set-B /dev/sda1
> /dev/md127:
> Version : 1.2
> Creation Time : Tue Jul 26 19:00:28 2011
> Raid Level : raid0
> Array Size : 94367232 (90.00 GiB 96.63 GB)
> Raid Devices : 3
> Total Devices : 3
> Persistence : Superblock is persistent
>
> Update Time : Tue Jul 26 19:00:28 2011
> State : clean
> Active Devices : 3
> Working Devices : 3
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 512K
>
> Name : reading.homeunix.com:3
> UUID : acd5374f:72628c93:6a906c4b:5f675ce5
> Events : 0
>
> Number Major Minor RaidDevice State
> 0 8 85 0 active sync /dev/sdf5
> 1 8 21 1 active sync /dev/sdb5
> 2 8 5 2 active sync /dev/sda5
> /dev/md128:
> Version : 1.2
> Raid Level : raid0
> Total Devices : 1
> Persistence : Superblock is persistent
>
> State : inactive
>
> Name : lamachine:128 (local to host lamachine)
> UUID : f2372cb9:d3816fd6:ce86d826:882ec82e
> Events : 4154
>
> Number Major Minor RaidDevice
>
> - 8 49 - /dev/sdd1
> /dev/md129:
> Version : 1.2
> Raid Level : raid0
> Total Devices : 1
> Persistence : Superblock is persistent
>
> State : inactive
>
> Name : lamachine:129 (local to host lamachine)
> UUID : 895dae98:d1a496de:4f590b8b:cb8ac12a
> Events : 0
>
> Number Major Minor RaidDevice
>
> - 8 50 - /dev/sdd2
> [root@lamachine ~]# mdadm --detail /dev/md2*
> /dev/md2:
> Version : 0.90
> Creation Time : Mon Feb 11 07:54:36 2013
> Raid Level : raid5
> Array Size : 511999872 (488.28 GiB 524.29 GB)
> Used Dev Size : 255999936 (244.14 GiB 262.14 GB)
> Raid Devices : 3
> Total Devices : 3
> Preferred Minor : 2
> Persistence : Superblock is persistent
>
> Update Time : Tue Jan 12 02:31:50 2016
> State : clean
> Active Devices : 3
> Working Devices : 3
> Failed Devices : 0
> Spare Devices : 0
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> UUID : 2cff15d1:e411447b:fd5d4721:03e44022 (local to host lamachine)
> Events : 0.611
>
> Number Major Minor RaidDevice State
> 0 8 83 0 active sync /dev/sdf3
> 1 8 18 1 active sync /dev/sdb2
> 2 8 2 2 active sync /dev/sda2
> [root@lamachine ~]# cat /proc/mdstat
> Personalities : [raid10] [raid0] [raid6] [raid5] [raid4]
> md2 : active raid5 sda2[2] sdf3[0] sdb2[1]
> 511999872 blocks level 5, 64k chunk, algorithm 2 [3/3] [UUU]
>
> md127 : active raid0 sda5[2] sdf5[0] sdb5[1]
> 94367232 blocks super 1.2 512k chunks
>
> md129 : inactive sdd2[2](S)
> 524156928 blocks super 1.2
>
> md128 : inactive sdd1[3](S)
> 2147352576 blocks super 1.2
>
> md126 : active raid10 sdf2[0] sda1[1]
> 30719936 blocks 2 near-copies [2/2] [UU]
>
> unused devices: <none>
> [root@lamachine ~]#
>
> On 11 September 2016 at 19:48, Daniel Sanabria <sanabria.d@gmail.com> wrote:
>> ok, system up and running after MB was replaced however the arrays
>> remain inactive.
>>
>> mdadm version is:
>> mdadm - v3.3.4 - 3rd August 2015
>>
>> Here's the output from Phil's lsdrv:
>>
>> [root@lamachine ~]# ./lsdrv
>> PCI [ahci] 00:1f.2 SATA controller: Intel Corporation C600/X79 series
>> chipset 6-Port SATA AHCI Controller (rev 06)
>> ├scsi 0:0:0:0 ATA WDC WD5000AAKS-0 {WD-WCASZ0505379}
>> │└sda 465.76g [8:0] Partitioned (dos)
>> │ ├sda1 29.30g [8:1] MD raid10,near2 (1/2) (w/ sdf2) in_sync
>> {9af006ca-8845-bbd3-bfe7-8010bc810f04}
>> │ │└md126 29.30g [9:126] MD v0.90 raid10,near2 (2) clean, 64k Chunk
>> {9af006ca:8845bbd3:bfe78010:bc810f04}
>> │ │ │ PV LVM2_member 28.03g used, 1.26g free
>> {cE4ePh-RWO8-Wgdy-YPOY-ehyC-KI6u-io1cyH}
>> │ │ └VG vg_bigblackbox 29.29g 1.26g free
>> {VWfuwI-5v2q-w8qf-FEbc-BdGW-3mKX-pZd7hR}
>> │ │ ├dm-2 7.81g [253:2] LV LogVol_opt ext4
>> {b08d7f5e-f15f-4241-804e-edccecab6003}
>> │ │ │└Mounted as /dev/mapper/vg_bigblackbox-LogVol_opt @ /opt
>> │ │ ├dm-0 9.77g [253:0] LV LogVol_root ext4
>> {4dabd6b0-b1a3-464d-8ed7-0aab93fab6c3}
>> │ │ │└Mounted as /dev/mapper/vg_bigblackbox-LogVol_root @ /
>> │ │ ├dm-3 1.95g [253:3] LV LogVol_tmp ext4
>> {f6b46363-170b-4038-83bd-2c5f9f6a1973}
>> │ │ │└Mounted as /dev/mapper/vg_bigblackbox-LogVol_tmp @ /tmp
>> │ │ └dm-1 8.50g [253:1] LV LogVol_var ext4
>> {ab165c61-3d62-4c55-8639-6c2c2bf4b021}
>> │ │ └Mounted as /dev/mapper/vg_bigblackbox-LogVol_var @ /var
>> │ ├sda2 244.14g [8:2] MD raid5 (2/3) (w/ sdb2,sdf3) in_sync
>> {2cff15d1-e411-447b-fd5d-472103e44022}
>> │ │└md2 488.28g [9:2] MD v0.90 raid5 (3) clean, 64k Chunk
>> {2cff15d1:e411447b:fd5d4721:03e44022}
>> │ │ │ ext4 {e9c1c787-496f-4e8f-b62e-35d5b1ff8311}
>> │ │ └Mounted as /dev/md2 @ /home
>> │ ├sda3 1.00k [8:3] Partitioned (dos)
>> │ ├sda5 30.00g [8:5] MD raid0 (2/3) (w/ sdb5,sdf5) in_sync
>> 'reading.homeunix.com:3' {acd5374f-7262-8c93-6a90-6c4b5f675ce5}
>> │ │└md127 90.00g [9:127] MD v1.2 raid0 (3) clean, 512k Chunk, None
>> (None) None {acd5374f:72628c93:6a906c4b:5f675ce5}
>> │ │ │ PV LVM2_member 86.00g used, 3.99g free
>> {VmsWRd-8qHt-bauf-lvAn-FC97-KyH5-gk89ox}
>> │ │ └VG libvirt_lvm 89.99g 3.99g free {t8GQck-f2Eu-iD2V-fnJQ-kBm6-QyKw-dR31PB}
>> │ │ ├dm-6 8.00g [253:6] LV builder2 Partitioned (dos)
>> │ │ ├dm-7 8.00g [253:7] LV builder3 Partitioned (dos)
>> │ │ ├dm-9 8.00g [253:9] LV builder5.3 Partitioned (dos)
>> │ │ ├dm-8 8.00g [253:8] LV builder5.6 Partitioned (dos)
>> │ │ ├dm-5 8.00g [253:5] LV centos_updt Partitioned (dos)
>> │ │ ├dm-10 16.00g [253:10] LV f22lvm Partitioned (dos)
>> │ │ └dm-4 30.00g [253:4] LV win7 Partitioned (dos)
>> │ └sda6 3.39g [8:6] Empty/Unknown
>> ├scsi 1:0:0:0 ATA WDC WD5000AAKS-0 {WD-WCASY7694185}
>> │└sdb 465.76g [8:16] Partitioned (dos)
>> │ ├sdb2 244.14g [8:18] MD raid5 (1/3) (w/ sda2,sdf3) in_sync
>> {2cff15d1-e411-447b-fd5d-472103e44022}
>> │ │└md2 488.28g [9:2] MD v0.90 raid5 (3) clean, 64k Chunk
>> {2cff15d1:e411447b:fd5d4721:03e44022}
>> │ │ ext4 {e9c1c787-496f-4e8f-b62e-35d5b1ff8311}
>> │ ├sdb3 7.81g [8:19] swap {9194f492-881a-4fc3-ac09-ca4e1cc2985a}
>> │ ├sdb4 1.00k [8:20] Partitioned (dos)
>> │ ├sdb5 30.00g [8:21] MD raid0 (1/3) (w/ sda5,sdf5) in_sync
>> 'reading.homeunix.com:3' {acd5374f-7262-8c93-6a90-6c4b5f675ce5}
>> │ │└md127 90.00g [9:127] MD v1.2 raid0 (3) clean, 512k Chunk, None
>> (None) None {acd5374f:72628c93:6a906c4b:5f675ce5}
>> │ │ PV LVM2_member 86.00g used, 3.99g free
>> {VmsWRd-8qHt-bauf-lvAn-FC97-KyH5-gk89ox}
>> │ └sdb6 3.39g [8:22] Empty/Unknown
>> ├scsi 2:x:x:x [Empty]
>> ├scsi 3:x:x:x [Empty]
>> ├scsi 4:x:x:x [Empty]
>> └scsi 5:x:x:x [Empty]
>> PCI [ahci] 0a:00.0 SATA controller: Marvell Technology Group Ltd.
>> 88SE9230 PCIe SATA 6Gb/s Controller (rev 11)
>> ├scsi 6:0:0:0 ATA WDC WD30EZRX-00D {WD-WCC4NCWT13RF}
>> │└sdc 2.73t [8:32] Partitioned (PMBR)
>> ├scsi 7:0:0:0 ATA WDC WD30EZRX-00D {WD-WCC4NPRDD6D7}
>> │└sdd 2.73t [8:48] Partitioned (gpt)
>> │ ├sdd1 2.00t [8:49] MD (none/) spare 'lamachine:128'
>> {f2372cb9-d381-6fd6-ce86-d826882ec82e}
>> │ │└md128 0.00k [9:128] MD v1.2 () inactive, None (None) None
>> {f2372cb9:d3816fd6:ce86d826:882ec82e}
>> │ │ Empty/Unknown
>> │ └sdd2 500.00g [8:50] MD (none/) spare 'lamachine:129'
>> {895dae98-d1a4-96de-4f59-0b8bcb8ac12a}
>> │ └md129 0.00k [9:129] MD v1.2 () inactive, None (None) None
>> {895dae98:d1a496de:4f590b8b:cb8ac12a}
>> │ Empty/Unknown
>> ├scsi 8:0:0:0 ATA WDC WD30EZRX-00D {WD-WCC4N1294906}
>> │└sde 2.73t [8:64] Partitioned (PMBR)
>> ├scsi 9:0:0:0 ATA WDC WD5000AAKS-0 {WD-WMAWF0085724}
>> │└sdf 465.76g [8:80] Partitioned (dos)
>> │ ├sdf1 199.00m [8:81] ext4 {4e51f903-37ca-4479-9197-fac7b2280557}
>> │ │└Mounted as /dev/sdf1 @ /boot
>> │ ├sdf2 29.30g [8:82] MD raid10,near2 (0/2) (w/ sda1) in_sync
>> {9af006ca-8845-bbd3-bfe7-8010bc810f04}
>> │ │└md126 29.30g [9:126] MD v0.90 raid10,near2 (2) clean, 64k Chunk
>> {9af006ca:8845bbd3:bfe78010:bc810f04}
>> │ │ PV LVM2_member 28.03g used, 1.26g free
>> {cE4ePh-RWO8-Wgdy-YPOY-ehyC-KI6u-io1cyH}
>> │ ├sdf3 244.14g [8:83] MD raid5 (0/3) (w/ sda2,sdb2) in_sync
>> {2cff15d1-e411-447b-fd5d-472103e44022}
>> │ │└md2 488.28g [9:2] MD v0.90 raid5 (3) clean, 64k Chunk
>> {2cff15d1:e411447b:fd5d4721:03e44022}
>> │ │ ext4 {e9c1c787-496f-4e8f-b62e-35d5b1ff8311}
>> │ ├sdf4 1.00k [8:84] Partitioned (dos)
>> │ ├sdf5 30.00g [8:85] MD raid0 (0/3) (w/ sda5,sdb5) in_sync
>> 'reading.homeunix.com:3' {acd5374f-7262-8c93-6a90-6c4b5f675ce5}
>> │ │└md127 90.00g [9:127] MD v1.2 raid0 (3) clean, 512k Chunk, None
>> (None) None {acd5374f:72628c93:6a906c4b:5f675ce5}
>> │ │ PV LVM2_member 86.00g used, 3.99g free
>> {VmsWRd-8qHt-bauf-lvAn-FC97-KyH5-gk89ox}
>> │ └sdf6 3.39g [8:86] Empty/Unknown
>> ├scsi 10:x:x:x [Empty]
>> ├scsi 11:x:x:x [Empty]
>> └scsi 12:x:x:x [Empty]
>> PCI [isci] 05:00.0 Serial Attached SCSI controller: Intel Corporation
>> C602 chipset 4-Port SATA Storage Control Unit (rev 06)
>> └scsi 14:x:x:x [Empty]
>> [root@lamachine ~]#
>>
>> Thanks in advance for any recommendations on what steps to take in
>> order to bring these arrays back online.
>>
>> Regards,
>>
>> Daniel
>>
>>
>> On 2 August 2016 at 11:45, Daniel Sanabria <sanabria.d@gmail.com> wrote:
>>> Thanks very much for the response Wol.
>>>
>>> It looks like the PSU is dead (server automatically powers off a few
>>> seconds after power on).
>>>
>>> I'm planning to order a PSU replacement to resume troubleshooting so
>>> please bear with me; maybe the PSU was degraded and couldn't power
>>> some of drives?
>>>
>>> Cheers,
>>>
>>> Daniel
>>>
>>> On 2 August 2016 at 11:17, Wols Lists <antlists@youngman.org.uk> wrote:
>>>> Just a quick first response. I see md128 and md129 are both down, and
>>>> are both listed as one drive, raid0. Bit odd, that ...
>>>>
>>>> What version of mdadm are you using? One of them had a bug (3.2.3 era?)
>>>> that would split an array in two. Is it possible that you should have
>>>> one raid0 array with sdf1 and sdf2? But that's a bit of a weird setup...
>>>>
>>>> I notice also that md126 is raid10 across two drives. That's odd, too.
>>>>
>>>> How much do you know about what the setup should be, and why it was set
>>>> up that way?
>>>>
>>>> Download lspci by Phil Turmel (it requires python2.7, if your machine is
>>>> python3 a quick fix to the shebang at the start should get it to work).
>>>> Post the output from that here.
>>>>
>>>> Cheers,
>>>> Wol
>>>>
>>>> On 02/08/16 08:36, Daniel Sanabria wrote:
>>>>> Hi All,
>>>>>
>>>>> I have a box that I believe was not powered down correctly and after
>>>>> transporting it to a different location it doesn't boot anymore
>>>>> stopping at BIOS check "Verifying DMI Pool Data".
>>>>>
>>>>> The box have 6 drives and after instructing the BIOS to boot from the
>>>>> first drive I managed to boot the OS (Fedora 23) after commenting out
>>>>> 2 /etc/fstab entries , output for "uname -a; cat /etc/fstab" follows:
>>>>>
>>>>> [root@lamachine ~]# uname -a; cat /etc/fstab
>>>>> Linux lamachine 4.3.3-303.fc23.x86_64 #1 SMP Tue Jan 19 18:31:55 UTC
>>>>> 2016 x86_64 x86_64 x86_64 GNU/Linux
>>>>>
>>>>> #
>>>>> # /etc/fstab
>>>>> # Created by anaconda on Tue Mar 24 19:31:21 2015
>>>>> #
>>>>> # Accessible filesystems, by reference, are maintained under '/dev/disk'
>>>>> # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
>>>>> #
>>>>> /dev/mapper/vg_bigblackbox-LogVol_root / ext4
>>>>> defaults 1 1
>>>>> UUID=4e51f903-37ca-4479-9197-fac7b2280557 /boot ext4
>>>>> defaults 1 2
>>>>> /dev/mapper/vg_bigblackbox-LogVol_opt /opt ext4
>>>>> defaults 1 2
>>>>> /dev/mapper/vg_bigblackbox-LogVol_tmp /tmp ext4
>>>>> defaults 1 2
>>>>> /dev/mapper/vg_bigblackbox-LogVol_var /var ext4
>>>>> defaults 1 2
>>>>> UUID=9194f492-881a-4fc3-ac09-ca4e1cc2985a swap swap
>>>>> defaults 0 0
>>>>> /dev/md2 /home ext4 defaults 1 2
>>>>> #/dev/vg_media/lv_media /mnt/media ext4 defaults 1 2
>>>>> #/dev/vg_virt_dir/lv_virt_dir1 /mnt/guest_images/ ext4 defaults 1 2
>>>>> [root@lamachine ~]#
>>>>>
>>>>> When checking mdstat I can see that 2 of the arrays are showing up as
>>>>> inactive, but not sure how to safely activate these so looking for
>>>>> some knowledgeable advice on how to proceed here.
>>>>>
>>>>> Thanks in advance,
>>>>>
>>>>> Daniel
>>>>>
>>>>> Below some more relevant outputs:
>>>>>
>>>>> [root@lamachine ~]# cat /proc/mdstat
>>>>> Personalities : [raid10] [raid6] [raid5] [raid4] [raid0]
>>>>> md127 : active raid0 sda5[0] sdc5[2] sdb5[1]
>>>>> 94367232 blocks super 1.2 512k chunks
>>>>>
>>>>> md2 : active raid5 sda3[0] sdc2[2] sdb2[1]
>>>>> 511999872 blocks level 5, 64k chunk, algorithm 2 [3/3] [UUU]
>>>>>
>>>>> md128 : inactive sdf1[3](S)
>>>>> 2147352576 blocks super 1.2
>>>>>
>>>>> md129 : inactive sdf2[2](S)
>>>>> 524156928 blocks super 1.2
>>>>>
>>>>> md126 : active raid10 sda2[0] sdc1[1]
>>>>> 30719936 blocks 2 near-copies [2/2] [UU]
>>>>>
>>>>> unused devices: <none>
>>>>> [root@lamachine ~]# cat /etc/mdadm.conf
>>>>> # mdadm.conf written out by anaconda
>>>>> MAILADDR root
>>>>> AUTO +imsm +1.x -all
>>>>> ARRAY /dev/md2 level=raid5 num-devices=3
>>>>> UUID=2cff15d1:e411447b:fd5d4721:03e44022
>>>>> ARRAY /dev/md126 level=raid10 num-devices=2
>>>>> UUID=9af006ca:8845bbd3:bfe78010:bc810f04
>>>>> ARRAY /dev/md127 level=raid0 num-devices=3
>>>>> UUID=acd5374f:72628c93:6a906c4b:5f675ce5
>>>>> ARRAY /dev/md128 metadata=1.2 spares=1 name=lamachine:128
>>>>> UUID=f2372cb9:d3816fd6:ce86d826:882ec82e
>>>>> ARRAY /dev/md129 metadata=1.2 name=lamachine:129
>>>>> UUID=895dae98:d1a496de:4f590b8b:cb8ac12a
>>>>> [root@lamachine ~]# mdadm --detail /dev/md1*
>>>>> /dev/md126:
>>>>> Version : 0.90
>>>>> Creation Time : Thu Dec 3 22:12:12 2009
>>>>> Raid Level : raid10
>>>>> Array Size : 30719936 (29.30 GiB 31.46 GB)
>>>>> Used Dev Size : 30719936 (29.30 GiB 31.46 GB)
>>>>> Raid Devices : 2
>>>>> Total Devices : 2
>>>>> Preferred Minor : 126
>>>>> Persistence : Superblock is persistent
>>>>>
>>>>> Update Time : Tue Aug 2 07:46:39 2016
>>>>> State : clean
>>>>> Active Devices : 2
>>>>> Working Devices : 2
>>>>> Failed Devices : 0
>>>>> Spare Devices : 0
>>>>>
>>>>> Layout : near=2
>>>>> Chunk Size : 64K
>>>>>
>>>>> UUID : 9af006ca:8845bbd3:bfe78010:bc810f04
>>>>> Events : 0.264152
>>>>>
>>>>> Number Major Minor RaidDevice State
>>>>> 0 8 2 0 active sync set-A /dev/sda2
>>>>> 1 8 33 1 active sync set-B /dev/sdc1
>>>>> /dev/md127:
>>>>> Version : 1.2
>>>>> Creation Time : Tue Jul 26 19:00:28 2011
>>>>> Raid Level : raid0
>>>>> Array Size : 94367232 (90.00 GiB 96.63 GB)
>>>>> Raid Devices : 3
>>>>> Total Devices : 3
>>>>> Persistence : Superblock is persistent
>>>>>
>>>>> Update Time : Tue Jul 26 19:00:28 2011
>>>>> State : clean
>>>>> Active Devices : 3
>>>>> Working Devices : 3
>>>>> Failed Devices : 0
>>>>> Spare Devices : 0
>>>>>
>>>>> Chunk Size : 512K
>>>>>
>>>>> Name : reading.homeunix.com:3
>>>>> UUID : acd5374f:72628c93:6a906c4b:5f675ce5
>>>>> Events : 0
>>>>>
>>>>> Number Major Minor RaidDevice State
>>>>> 0 8 5 0 active sync /dev/sda5
>>>>> 1 8 21 1 active sync /dev/sdb5
>>>>> 2 8 37 2 active sync /dev/sdc5
>>>>> /dev/md128:
>>>>> Version : 1.2
>>>>> Raid Level : raid0
>>>>> Total Devices : 1
>>>>> Persistence : Superblock is persistent
>>>>>
>>>>> State : inactive
>>>>>
>>>>> Name : lamachine:128 (local to host lamachine)
>>>>> UUID : f2372cb9:d3816fd6:ce86d826:882ec82e
>>>>> Events : 4154
>>>>>
>>>>> Number Major Minor RaidDevice
>>>>>
>>>>> - 8 81 - /dev/sdf1
>>>>> /dev/md129:
>>>>> Version : 1.2
>>>>> Raid Level : raid0
>>>>> Total Devices : 1
>>>>> Persistence : Superblock is persistent
>>>>>
>>>>> State : inactive
>>>>>
>>>>> Name : lamachine:129 (local to host lamachine)
>>>>> UUID : 895dae98:d1a496de:4f590b8b:cb8ac12a
>>>>> Events : 0
>>>>>
>>>>> Number Major Minor RaidDevice
>>>>>
>>>>> - 8 82 - /dev/sdf2
>>>>> [root@lamachine ~]# mdadm --detail /dev/md2
>>>>> /dev/md2:
>>>>> Version : 0.90
>>>>> Creation Time : Mon Feb 11 07:54:36 2013
>>>>> Raid Level : raid5
>>>>> Array Size : 511999872 (488.28 GiB 524.29 GB)
>>>>> Used Dev Size : 255999936 (244.14 GiB 262.14 GB)
>>>>> Raid Devices : 3
>>>>> Total Devices : 3
>>>>> Preferred Minor : 2
>>>>> Persistence : Superblock is persistent
>>>>>
>>>>> Update Time : Mon Aug 1 20:24:23 2016
>>>>> State : clean
>>>>> Active Devices : 3
>>>>> Working Devices : 3
>>>>> Failed Devices : 0
>>>>> Spare Devices : 0
>>>>>
>>>>> Layout : left-symmetric
>>>>> Chunk Size : 64K
>>>>>
>>>>> UUID : 2cff15d1:e411447b:fd5d4721:03e44022 (local to host lamachine)
>>>>> Events : 0.611
>>>>>
>>>>> Number Major Minor RaidDevice State
>>>>> 0 8 3 0 active sync /dev/sda3
>>>>> 1 8 18 1 active sync /dev/sdb2
>>>>> 2 8 34 2 active sync /dev/sdc2
>>>>> [root@lamachine ~]#
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>
next prev parent reply other threads:[~2016-09-12 19:41 UTC|newest]
Thread overview: 37+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-08-02 7:36 Inactive arrays Daniel Sanabria
2016-08-02 10:17 ` Wols Lists
2016-08-02 10:45 ` Daniel Sanabria
2016-08-03 19:18 ` Daniel Sanabria
2016-08-03 21:31 ` Wols Lists
2016-09-11 18:48 ` Daniel Sanabria
2016-09-11 20:06 ` Daniel Sanabria
2016-09-12 19:41 ` Daniel Sanabria [this message]
2016-09-12 21:13 ` Daniel Sanabria
2016-09-12 21:37 ` Chris Murphy
2016-09-13 6:51 ` Daniel Sanabria
2016-09-13 15:04 ` Chris Murphy
2016-09-12 21:39 ` Wols Lists
2016-09-13 6:56 ` Daniel Sanabria
2016-09-13 7:02 ` Adam Goryachev
2016-09-13 15:20 ` Chris Murphy
2016-09-13 19:43 ` Daniel Sanabria
2016-09-13 19:52 ` Chris Murphy
2016-09-13 20:04 ` Daniel Sanabria
2016-09-13 20:13 ` Chris Murphy
2016-09-13 20:29 ` Daniel Sanabria
2016-09-13 20:36 ` Daniel Sanabria
2016-09-13 21:10 ` Chris Murphy
2016-09-13 21:46 ` Daniel Sanabria
2016-09-13 21:26 ` Wols Lists
2016-09-14 4:33 ` Chris Murphy
2016-09-14 10:36 ` Daniel Sanabria
2016-09-14 14:32 ` Chris Murphy
2016-09-14 14:57 ` Daniel Sanabria
2016-09-14 15:15 ` Chris Murphy
2016-09-14 15:47 ` Daniel Sanabria
2016-09-14 16:10 ` Chris Murphy
2016-09-14 16:13 ` Chris Murphy
2016-09-14 18:16 ` Daniel Sanabria
2016-09-14 18:37 ` Chris Murphy
2016-09-14 18:42 ` Wols Lists
2016-09-15 9:21 ` Brad Campbell
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='CAHscji35kySP4Q8cUpYCXR2P9DBwPcYeWjr_HX==TNX9CPW9NA@mail.gmail.com' \
--to=sanabria.d@gmail.com \
--cc=antlists@youngman.org.uk \
--cc=linux-raid@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).