linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* Question regarding mdadm.conf
@ 2005-02-17  6:25 Torsten E.
  2005-02-17  6:48 ` Lajber Zoltan
  0 siblings, 1 reply; 7+ messages in thread
From: Torsten E. @ 2005-02-17  6:25 UTC (permalink / raw)
  To: linux-raid

Good morning! :)

End of january I had some problems with an faulty harddisk, which was 
part of an RAID1 (two mirroed disks containing 4 partitions).
I replaced it, and got some realy good instructions from Gordon Henderson.
Last week the same problem occured again, and I decided to replace the 
controller, too. Now its an PCI 4-port S-ATA controller, and everything 
runs fine till now.

My question now is:
the formerly used /etc/mdadm.conf had some lines like:
   ARRAY /dev/md1 level=raid1 num-devices=2
   UUID=66f80d1a:621442c5:2b6eabe8:3a9f5e54
    devices=/dev/sda6,/dev/sdb6
   ARRAY /dev/md0 level=raid1 num-devices=2
   UUID=06b9cdec:65448b34:3add2ff0:2180f5cd
    devices=/dev/sda5,/dev/sdb5

How does I get those UUID information, to add them to the new 
/etc/mdadm.conf?

Have a nice day
Torsten

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Question regarding mdadm.conf
  2005-02-17  6:25 Question regarding mdadm.conf Torsten E.
@ 2005-02-17  6:48 ` Lajber Zoltan
  2005-02-17  7:14   ` Michael Tokarev
  0 siblings, 1 reply; 7+ messages in thread
From: Lajber Zoltan @ 2005-02-17  6:48 UTC (permalink / raw)
  To: Torsten E.; +Cc: linux-raid

Hi!

On Thu, 17 Feb 2005, Torsten E. wrote:

> How does I get those UUID information, to add them to the new
> /etc/mdadm.conf?

Try this one: mdadm --detail /dev/md1 | grep UUID

Bye,
-=Lajbi=----------------------------------------------------------------
 LAJBER Zoltan               Szent Istvan Egyetem,  Informatika Hivatal
   engineer: a mechanism for converting caffeine into designs.


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Question regarding mdadm.conf
  2005-02-17  6:48 ` Lajber Zoltan
@ 2005-02-17  7:14   ` Michael Tokarev
  2005-02-17  8:23     ` Torsten E.
  2005-02-17  8:30     ` Bad blocks Guy
  0 siblings, 2 replies; 7+ messages in thread
From: Michael Tokarev @ 2005-02-17  7:14 UTC (permalink / raw)
  To: Lajber Zoltan; +Cc: Torsten E., linux-raid

Lajber Zoltan wrote:
> Hi!
> 
> On Thu, 17 Feb 2005, Torsten E. wrote:
> 
> 
>>How does I get those UUID information, to add them to the new
>>/etc/mdadm.conf?
> 
> Try this one: mdadm --detail /dev/md1 | grep UUID

I'd say

   mdadm --detail --brief /dev/md1 | grep -v devices=

-- this will give you all information necessary for mdadm.conf,
you can just redirect output into that file.

Note the grep usage.  Someone will disagree with me here, but
there is a reason why to remove devices= line.  Without it,
output from mdadm looks like (on my system anyway):

ARRAY /dev/md1 level=raid1 num-devices=4 UUID=11e92e45:15fcc4a0:cf62e981:a79de494
    devices=/dev/sda1,/dev/sdb1,/dev/sdc1,/dev/sdd1

Ie, it lists all the devices which are parts of the array.
The problem with this is: if, for any reason (dead drive,
adding/removing drives/controllers etc) the devices will
change, and some /dev/sdXY will point to another device wich
is a part of some other raid array, mdadm will refuse to
assemble this array, saying something in a line of "the
UUIDs does not match, aborting".  Without the "devices="
part but with --scan option, mdadm will search all devices
by its own (based on the DEVICE line in mdadm.conf) - this
is somewhat slower as it will try to open each device in
turn, but safer, as it will find all the present components
no matter what.

Someone correct me if I'm wrong... ;)

/mjt

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Question regarding mdadm.conf
  2005-02-17  7:14   ` Michael Tokarev
@ 2005-02-17  8:23     ` Torsten E.
  2005-02-17  8:44       ` Guy
  2005-02-17  8:30     ` Bad blocks Guy
  1 sibling, 1 reply; 7+ messages in thread
From: Torsten E. @ 2005-02-17  8:23 UTC (permalink / raw)
  To: linux-raid

Hi Michael,
Hi Lajber,

Thanks for your hints!

Michael Tokarev scribbled on 17.02.2005 08:14:
> Lajber Zoltan wrote:
> 
>> Hi!
>>
>> On Thu, 17 Feb 2005, Torsten E. wrote:
>>
>>> How does I get those UUID information, to add them to the new
>>> /etc/mdadm.conf?
>>
>> Try this one: mdadm --detail /dev/md1 | grep UUID
> 
> I'd say
> 
>   mdadm --detail --brief /dev/md1 | grep -v devices=

Didn't work for me ... but as I (hopefully!) understood the basic usage 
I run:
   mdadm --detail --brief /dev/md* >> /tmp/mdtest
After adding some more lines (DEVICE /dev/sda*, DEVICE /dev/sdb*, 
MAILADDR admin) I simply renamed it to /etc/mdadm.conf ... its not a 
good way, but for me its an useable (and so its good again ;)).

> -- this will give you all information necessary for mdadm.conf,
> you can just redirect output into that file.
> 
> Note the grep usage.  Someone will disagree with me here, but
> there is a reason why to remove devices= line.  Without it,
> output from mdadm looks like (on my system anyway):
> 
> ARRAY /dev/md1 level=raid1 num-devices=4 
> UUID=11e92e45:15fcc4a0:cf62e981:a79de494
>    devices=/dev/sda1,/dev/sdb1,/dev/sdc1,/dev/sdd1
> 
> Ie, it lists all the devices which are parts of the array.
> The problem with this is: if, for any reason (dead drive,
> adding/removing drives/controllers etc) the devices will
> change, and some /dev/sdXY will point to another device wich
> is a part of some other raid array, mdadm will refuse to
> assemble this array, saying something in a line of "the
> UUIDs does not match, aborting".  Without the "devices="
> part but with --scan option, mdadm will search all devices
> by its own (based on the DEVICE line in mdadm.conf) - this
> is somewhat slower as it will try to open each device in
> turn, but safer, as it will find all the present components
> no matter what.
> 
> Someone correct me if I'm wrong... ;)
> 
> /mjt

Have a nice day!! :)
Torsten

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Bad blocks
  2005-02-17  7:14   ` Michael Tokarev
  2005-02-17  8:23     ` Torsten E.
@ 2005-02-17  8:30     ` Guy
  1 sibling, 0 replies; 7+ messages in thread
From: Guy @ 2005-02-17  8:30 UTC (permalink / raw)
  To: linux-raid

About 1 month ago the topic was bad blocks.  I have been monitoring the bad
blocks on my disks and I find I have had 3 new bad blocks since Jan 18.
Each on a different disk.  I have 17 disks, SEAGATE ST118202LC.

These bad blocks did not cause any problems with md.  I believe they were
readable or write errors, but re-mapped since I have AWRE and ARRE turned
on.

I don't know what is a normal rate, but based on the last month, I would
expect about 2.29 defects per disk per year.  I have had the disks in use
for about 2.5 years.  I have 81 defects.  That comes to 1.905 per disk per
year.  Not to far off the mark!  And I have replaced disks 2-3 times in 2.5
years.  The failed disks failed due to a cable problem, the disks may have
been just fine, but I took them apart for the magnets before I realized the
power cable was at fault.  The cable is now repaired.

Anyway, if these were read errors, md would have caused me lots of problems.
So, we need md to deal with bad blocks without kicking out the disk.

I have been using this command to monitor the disks:
sginfo -G /dev/sda | grep "in grown table"

My current bad block status (Defect list) for 17 disks:
0 entries (0 bytes) in grown table.
0 entries (0 bytes) in grown table.
8 entries (64 bytes) in grown table.
5 entries (40 bytes) in grown table.
0 entries (0 bytes) in grown table.
12 entries (96 bytes) in grown table.
20 entries (160 bytes) in grown table.
0 entries (0 bytes) in grown table.
6 entries (48 bytes) in grown table.
0 entries (0 bytes) in grown table.
6 entries (48 bytes) in grown table.
28 entries (224 bytes) in grown table.
0 entries (0 bytes) in grown table.
2 entries (16 bytes) in grown table.
3 entries (24 bytes) in grown table.
4 entries (32 bytes) in grown table.
0 entries (0 bytes) in grown table.

Does anyone know how many defects is considered too many?

Guy


^ permalink raw reply	[flat|nested] 7+ messages in thread

* RE: Question regarding mdadm.conf
  2005-02-17  8:23     ` Torsten E.
@ 2005-02-17  8:44       ` Guy
  2005-02-17 11:11         ` GrantC
  0 siblings, 1 reply; 7+ messages in thread
From: Guy @ 2005-02-17  8:44 UTC (permalink / raw)
  To: 'Torsten E.', linux-raid

In the past, Neil has recommended using a device line like this:
DEVICE partitions

From "man mdadm.conf":
Alternatively, a device line can contain  the  word  partitions.
    This  will  cause mdadm to read /proc/partitions and include all
    devices and partitions found there-in.  mdadm does not  use  the
    names  from /proc/partitions but only the major and minor device
    numbers.  It scans /dev to find the name that matches  the  num-
    bers.

Guy

-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Torsten E.
Sent: Thursday, February 17, 2005 3:24 AM
To: linux-raid@vger.kernel.org
Subject: Re: Question regarding mdadm.conf

Hi Michael,
Hi Lajber,

Thanks for your hints!

Michael Tokarev scribbled on 17.02.2005 08:14:
> Lajber Zoltan wrote:
> 
>> Hi!
>>
>> On Thu, 17 Feb 2005, Torsten E. wrote:
>>
>>> How does I get those UUID information, to add them to the new
>>> /etc/mdadm.conf?
>>
>> Try this one: mdadm --detail /dev/md1 | grep UUID
> 
> I'd say
> 
>   mdadm --detail --brief /dev/md1 | grep -v devices=

Didn't work for me ... but as I (hopefully!) understood the basic usage 
I run:
   mdadm --detail --brief /dev/md* >> /tmp/mdtest
After adding some more lines (DEVICE /dev/sda*, DEVICE /dev/sdb*, 
MAILADDR admin) I simply renamed it to /etc/mdadm.conf ... its not a 
good way, but for me its an useable (and so its good again ;)).

> -- this will give you all information necessary for mdadm.conf,
> you can just redirect output into that file.
> 
> Note the grep usage.  Someone will disagree with me here, but
> there is a reason why to remove devices= line.  Without it,
> output from mdadm looks like (on my system anyway):
> 
> ARRAY /dev/md1 level=raid1 num-devices=4 
> UUID=11e92e45:15fcc4a0:cf62e981:a79de494
>    devices=/dev/sda1,/dev/sdb1,/dev/sdc1,/dev/sdd1
> 
> Ie, it lists all the devices which are parts of the array.
> The problem with this is: if, for any reason (dead drive,
> adding/removing drives/controllers etc) the devices will
> change, and some /dev/sdXY will point to another device wich
> is a part of some other raid array, mdadm will refuse to
> assemble this array, saying something in a line of "the
> UUIDs does not match, aborting".  Without the "devices="
> part but with --scan option, mdadm will search all devices
> by its own (based on the DEVICE line in mdadm.conf) - this
> is somewhat slower as it will try to open each device in
> turn, but safer, as it will find all the present components
> no matter what.
> 
> Someone correct me if I'm wrong... ;)
> 
> /mjt

Have a nice day!! :)
Torsten
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: Question regarding mdadm.conf
  2005-02-17  8:44       ` Guy
@ 2005-02-17 11:11         ` GrantC
  0 siblings, 0 replies; 7+ messages in thread
From: GrantC @ 2005-02-17 11:11 UTC (permalink / raw)
  To: Guy; +Cc: 'Torsten E.', linux-raid

On Thu, 17 Feb 2005 03:44:40 -0500, you wrote:

>In the past, Neil has recommended using a device line like this:
>DEVICE partitions
>
. . .
>>   mdadm --detail --brief /dev/md1 | grep -v devices=

peetoo:~$  mdadm --detail --brief /dev/md1
ARRAY /dev/md1 level=raid0 num-devices=2 UUID=732ae5f8:e5d69879:30435be8:aafb7e94
   devices=/dev/hda6,/dev/hdc6

peetoo:~$  mdadm --detail --brief /dev/md1 | grep -v " devices="
ARRAY /dev/md1 level=raid0 num-devices=2 UUID=732ae5f8:e5d69879:30435be8:aafb7e94

Is that what was intended?

Cheers
Grant.




^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2005-02-17 11:11 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2005-02-17  6:25 Question regarding mdadm.conf Torsten E.
2005-02-17  6:48 ` Lajber Zoltan
2005-02-17  7:14   ` Michael Tokarev
2005-02-17  8:23     ` Torsten E.
2005-02-17  8:44       ` Guy
2005-02-17 11:11         ` GrantC
2005-02-17  8:30     ` Bad blocks Guy

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).