* mdadm.conf issue after updating system
@ 2015-04-22 9:56 Stefan Lamby
0 siblings, 0 replies; only message in thread
From: Stefan Lamby @ 2015-04-22 9:56 UTC (permalink / raw)
To: linux-raid@vger.kernel.org
Hello list.
I just updated my ubuntu system. The update provided a new kernel image also.
Here is what I got back:
update-initramfs: Generating /boot/initrd.img-3.13.0-49-generic
W: mdadm: the array /dev/md/kvm15:10 with UUID
c4540426:9c668fe2:479513f2:42d233b4
W: mdadm: is currently active, but it is not listed in mdadm.conf. if
W: mdadm: it is needed for boot, then YOUR SYSTEM IS NOW UNBOOTABLE!
W: mdadm: please inspect the output of /usr/share/mdadm/mkconf, compare
W: mdadm: it to /etc/mdadm/mdadm.conf, and make the necessary changes.
Is ist OK to just edit mdadm.conf and replace
ARRAY /dev/md/0 metadata=1.2 UUID=75079a2f:acb8c475:85f8ca43:0ad85c4c
name=kvm15:0
with:
ARRAY /dev/md/127 metadata=1.2 UUID=c4540426:9c668fe2:479513f2:42d233b4
name=kvm15:10
Or is there any other action recommended?
Thank you for your help.
Stefan
Here is some additonal information about the system:
root@kvm15:~# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid10] [raid6] [raid5]
[raid4]
md127 : active raid10 sdc1[1] sdb1[5] sdd1[3] sda1[4]
3808330752 blocks super 1.2 512K chunks 2 near-copies [4/4] [UUUU]
unused devices: <none>
root@kvm15:~# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 1,8T 0 disk
└─sda1 8:1 0 1,8T 0 part
└─md127 9:127 0 3,6T 0 raid10
├─vg_raid10-home (dm-0) 252:0 0 1,2T 0 lvm /home
├─vg_raid10-root (dm-1) 252:1 0 93,1G 0 lvm /
├─vg_raid10-var (dm-2) 252:2 0 393,1G 0 lvm /var
├─vg_raid10-tmp (dm-3) 252:3 0 46,6G 0 lvm /tmp
└─vg_raid10-swap (dm-4) 252:4 0 23,3G 0 lvm [SWAP]
sdb 8:16 0 1,8T 0 disk
└─sdb1 8:17 0 1,8T 0 part
└─md127 9:127 0 3,6T 0 raid10
├─vg_raid10-home (dm-0) 252:0 0 1,2T 0 lvm /home
├─vg_raid10-root (dm-1) 252:1 0 93,1G 0 lvm /
├─vg_raid10-var (dm-2) 252:2 0 393,1G 0 lvm /var
├─vg_raid10-tmp (dm-3) 252:3 0 46,6G 0 lvm /tmp
└─vg_raid10-swap (dm-4) 252:4 0 23,3G 0 lvm [SWAP]
sdc 8:32 0 1,8T 0 disk
└─sdc1 8:33 0 1,8T 0 part
└─md127 9:127 0 3,6T 0 raid10
├─vg_raid10-home (dm-0) 252:0 0 1,2T 0 lvm /home
├─vg_raid10-root (dm-1) 252:1 0 93,1G 0 lvm /
├─vg_raid10-var (dm-2) 252:2 0 393,1G 0 lvm /var
├─vg_raid10-tmp (dm-3) 252:3 0 46,6G 0 lvm /tmp
└─vg_raid10-swap (dm-4) 252:4 0 23,3G 0 lvm [SWAP]
sdd 8:48 0 1,8T 0 disk
└─sdd1 8:49 0 1,8T 0 part
└─md127 9:127 0 3,6T 0 raid10
├─vg_raid10-home (dm-0) 252:0 0 1,2T 0 lvm /home
├─vg_raid10-root (dm-1) 252:1 0 93,1G 0 lvm /
├─vg_raid10-var (dm-2) 252:2 0 393,1G 0 lvm /var
├─vg_raid10-tmp (dm-3) 252:3 0 46,6G 0 lvm /tmp
└─vg_raid10-swap (dm-4) 252:4 0 23,3G 0 lvm [SWAP]
sr0 11:0 1 3,7G 0 rom
Gerät boot. Anfang Ende Blöcke Id System
/dev/sda1 * 98435072 3907028991 1904296960 fd Linux raid autodetect
root@kvm15:~# cat /etc/mdadm/mdadm.conf
# mdadm.conf
#
# Please refer to mdadm.conf(5) for information about this file.
#
# by default (built-in), scan all partitions (/proc/partitions) and all
# containers for MD superblocks. alternatively, specify devices to scan, using
# wildcards if desired.
#DEVICE partitions containers
# auto-create devices with Debian standard permissions
CREATE owner=root group=disk mode=0660 auto=yes
# automatically tag new arrays as belonging to the local system
HOMEHOST <system>
# instruct the monitoring daemon where to send mail alerts
MAILADDR root
# definitions of existing MD arrays
ARRAY /dev/md/0 metadata=1.2 UUID=75079a2f:acb8c475:85f8ca43:0ad85c4c
name=kvm15:0
# This file was auto-generated on Tue, 17 Feb 2015 15:57:16 +0100
# by mkconf $Id$
root@kvm15:~# mdadm --detail /dev/md/kvm15\:10
/dev/md/kvm15:10:
Version : 1.2
Creation Time : Fri Mar 6 10:18:15 2015
Raid Level : raid10
Array Size : 3808330752 (3631.91 GiB 3899.73 GB)
Used Dev Size : 1904165376 (1815.95 GiB 1949.87 GB)
Raid Devices : 4
Total Devices : 4
Persistence : Superblock is persistent
Update Time : Wed Apr 22 11:42:28 2015
State : clean
Active Devices : 4
Working Devices : 4
Failed Devices : 0
Spare Devices : 0
Layout : near=2
Chunk Size : 512K
Name : kvm15:10 (local to host kvm15)
UUID : c4540426:9c668fe2:479513f2:42d233b4
Events : 14870
Number Major Minor RaidDevice State
5 8 17 0 active sync /dev/sdb1
1 8 33 1 active sync /dev/sdc1
4 8 1 2 active sync /dev/sda1
3 8 49 3 active sync /dev/sdd1
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] only message in thread
only message in thread, other threads:[~2015-04-22 9:56 UTC | newest]
Thread overview: (only message) (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-04-22 9:56 mdadm.conf issue after updating system Stefan Lamby
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).