* Can't resolve mismatch_count > 0 for a raid 1 array
@ 2009-04-08 10:03 Steven Ellis
2009-04-08 21:49 ` Goswin von Brederlow
2009-04-08 21:50 ` Bill Davidsen
0 siblings, 2 replies; 6+ messages in thread
From: Steven Ellis @ 2009-04-08 10:03 UTC (permalink / raw)
To: Linux RAID
I've resolved most of my raid issues by re-housing the affected system
and replacing the motherboard, but across the 3 boards I've tried I
always have an issue with my /dev/md1 array producing mismatch_count
of 128 or 256.
System is running Centos 5.2 with a Xen Dom0 kernel
This md1 volume is a pair of 40GB HDs raid1 on an IDE controller which
I them have a bunch of LVM's that are my Xen guests.
Is there any chance that these mismatch_count values are due to swap
partitions for the Xen guests?
Steve
Steven Ellis - Technical Director
OpenMedia Limited
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: Can't resolve mismatch_count > 0 for a raid 1 array
2009-04-08 10:03 Can't resolve mismatch_count > 0 for a raid 1 array Steven Ellis
@ 2009-04-08 21:49 ` Goswin von Brederlow
2009-04-09 0:07 ` Steven Ellis
2009-04-08 21:50 ` Bill Davidsen
1 sibling, 1 reply; 6+ messages in thread
From: Goswin von Brederlow @ 2009-04-08 21:49 UTC (permalink / raw)
To: Linux RAID
Steven Ellis <steven@openmedia.co.nz> writes:
> I've resolved most of my raid issues by re-housing the affected system
> and replacing the motherboard, but across the 3 boards I've tried I
> always have an issue with my /dev/md1 array producing mismatch_count
> of 128 or 256.
>
> System is running Centos 5.2 with a Xen Dom0 kernel
>
> This md1 volume is a pair of 40GB HDs raid1 on an IDE controller which
> I them have a bunch of LVM's that are my Xen guests.
>
> Is there any chance that these mismatch_count values are due to swap
> partitions for the Xen guests?
>
> Steve
Have you repaired the raid or do you just check and check and check?
MfG
Goswin
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: Can't resolve mismatch_count > 0 for a raid 1 array
2009-04-08 10:03 Can't resolve mismatch_count > 0 for a raid 1 array Steven Ellis
2009-04-08 21:49 ` Goswin von Brederlow
@ 2009-04-08 21:50 ` Bill Davidsen
2009-04-08 22:00 ` Iustin Pop
1 sibling, 1 reply; 6+ messages in thread
From: Bill Davidsen @ 2009-04-08 21:50 UTC (permalink / raw)
To: Steven Ellis; +Cc: Linux RAID
Steven Ellis wrote:
> I've resolved most of my raid issues by re-housing the affected system
> and replacing the motherboard, but across the 3 boards I've tried I
> always have an issue with my /dev/md1 array producing mismatch_count
> of 128 or 256.
>
> System is running Centos 5.2 with a Xen Dom0 kernel
>
> This md1 volume is a pair of 40GB HDs raid1 on an IDE controller which
> I them have a bunch of LVM's that are my Xen guests.
>
> Is there any chance that these mismatch_count values are due to swap
> partitions for the Xen guests?
That's the cause, and since md code doesn't currently have a clue which
copy is "right" it's always a problem if you do something like suspend
to disk. You probably don't do that with xen images, but swap and raid1
almost always have a mismatch.
--
bill davidsen <davidsen@tmr.com>
CTO TMR Associates, Inc
"You are disgraced professional losers. And by the way, give us our money back."
- Representative Earl Pomeroy, Democrat of North Dakota
on the A.I.G. executives who were paid bonuses after a federal bailout.
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: Can't resolve mismatch_count > 0 for a raid 1 array
2009-04-08 21:50 ` Bill Davidsen
@ 2009-04-08 22:00 ` Iustin Pop
2009-04-09 0:13 ` Steven Ellis
0 siblings, 1 reply; 6+ messages in thread
From: Iustin Pop @ 2009-04-08 22:00 UTC (permalink / raw)
To: Bill Davidsen; +Cc: Steven Ellis, Linux RAID
On Wed, Apr 08, 2009 at 05:50:46PM -0400, Bill Davidsen wrote:
> Steven Ellis wrote:
>> I've resolved most of my raid issues by re-housing the affected system
>> and replacing the motherboard, but across the 3 boards I've tried I
>> always have an issue with my /dev/md1 array producing mismatch_count
>> of 128 or 256.
>>
>> System is running Centos 5.2 with a Xen Dom0 kernel
>>
>> This md1 volume is a pair of 40GB HDs raid1 on an IDE controller which
>> I them have a bunch of LVM's that are my Xen guests.
>>
>> Is there any chance that these mismatch_count values are due to swap
>> partitions for the Xen guests?
>
> That's the cause, and since md code doesn't currently have a clue which
> copy is "right" it's always a problem if you do something like suspend
> to disk. You probably don't do that with xen images, but swap and raid1
> almost always have a mismatch.
But only because (for non-xen guests) the raid1 code and the swap code /
data live in the same address space, and could be changed in between the
two writes.
I would be surprised if this happens for xen guests, where the address
space is not shared; once the xen guest initiates a write, dom0 gets the
data and writes it from its internal buffer, not the domU's one which
could be modified.
At least, that's how I think things happen.
regards,
iustin
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: Can't resolve mismatch_count > 0 for a raid 1 array
2009-04-08 21:49 ` Goswin von Brederlow
@ 2009-04-09 0:07 ` Steven Ellis
0 siblings, 0 replies; 6+ messages in thread
From: Steven Ellis @ 2009-04-09 0:07 UTC (permalink / raw)
To: Linux RAID
On Thu, April 9, 2009 9:49 am, Goswin von Brederlow wrote:
> Steven Ellis <steven@openmedia.co.nz> writes:
>
>> I've resolved most of my raid issues by re-housing the affected system
>> and replacing the motherboard, but across the 3 boards I've tried I
>> always have an issue with my /dev/md1 array producing mismatch_count
>> of 128 or 256.
>>
>> System is running Centos 5.2 with a Xen Dom0 kernel
>>
>> This md1 volume is a pair of 40GB HDs raid1 on an IDE controller which
>> I them have a bunch of LVM's that are my Xen guests.
>>
>> Is there any chance that these mismatch_count values are due to swap
>> partitions for the Xen guests?
>>
>> Steve
>
> Have you repaired the raid or do you just check and check and check?
>
Yes I've repaired the raid several times. Then when I put the Xen guests
under load and re-run check the errors re-appear.
--------------------------------------------
Steven Ellis - Technical Director
OpenMedia Limited - The Home of myPVR
email - steven@openmedia.co.nz
website - http://www.openmedia.co.nz
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: Can't resolve mismatch_count > 0 for a raid 1 array
2009-04-08 22:00 ` Iustin Pop
@ 2009-04-09 0:13 ` Steven Ellis
0 siblings, 0 replies; 6+ messages in thread
From: Steven Ellis @ 2009-04-09 0:13 UTC (permalink / raw)
To: Linux RAID
On Thu, April 9, 2009 10:00 am, Iustin Pop wrote:
> On Wed, Apr 08, 2009 at 05:50:46PM -0400, Bill Davidsen wrote:
>> Steven Ellis wrote:
>>> I've resolved most of my raid issues by re-housing the affected system
>>> and replacing the motherboard, but across the 3 boards I've tried I
>>> always have an issue with my /dev/md1 array producing mismatch_count
>>> of 128 or 256.
>>>
>>> System is running Centos 5.2 with a Xen Dom0 kernel
>>>
>>> This md1 volume is a pair of 40GB HDs raid1 on an IDE controller which
>>> I them have a bunch of LVM's that are my Xen guests.
>>>
>>> Is there any chance that these mismatch_count values are due to swap
>>> partitions for the Xen guests?
>>
>> That's the cause, and since md code doesn't currently have a clue which
>> copy is "right" it's always a problem if you do something like suspend
>> to disk. You probably don't do that with xen images, but swap and raid1
>> almost always have a mismatch.
>
> But only because (for non-xen guests) the raid1 code and the swap code /
> data live in the same address space, and could be changed in between the
> two writes.
>
> I would be surprised if this happens for xen guests, where the address
> space is not shared; once the xen guest initiates a write, dom0 gets the
> data and writes it from its internal buffer, not the domU's one which
> could be modified.
>
> At least, that's how I think things happen.
The box has 5 Raid arrays
# cat /proc/mdstat
Personalities : [raid1] [raid6] [raid5] [raid4]
md3 : active raid1 sdb1[0] sda1[1]
976759936 blocks [2/2] [UU]
md0 : active raid1 hdb1[0] hda1[1]
128384 blocks [2/2] [UU]
md4 : active raid1 hdb2[1] hda2[0]
522048 blocks [2/2] [UU]
md2 : active raid5 hdh1[3] hdg1[2] hdf1[1] hde1[0]
732587712 blocks level 5, 64k chunk, algorithm 2 [4/4] [UUUU]
md1 : active raid1 hdb3[0] hda3[1]
38427392 blocks [2/2] [UU]
unused devices: <none>
md4 is the swap partition for the Xen Server, and md0 is the boot
partition. Neither of these have any issues.
md2 is a raid5 set that also hasn't been reporting any issues.
md3 is a new SATA based raid1 set that I had some issues with when using a
different motherboard, but doesn't produce any errors no even under
serious load.
md1 contains the root file system for my Xen server, plus the root + swap
partitions for my various Xen guests. This is the volume that is
generating the mismatch_count errors.
Now most of my Xen guests are presented with two LVM allocated partitions
out of md1, eg guest_root and guest_swap.
I do have an exception to this for one guest where I present a single LVM
partition as a virtual HD to the guest which then manages the
swap/root/home partitions itself.
I'm wondering if this presentation of a partition as a disk is the issue.
Steve
--------------------------------------------
Steven Ellis - Technical Director
OpenMedia Limited - The Home of myPVR
email - steven@openmedia.co.nz
website - http://www.openmedia.co.nz
^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2009-04-09 0:13 UTC | newest]
Thread overview: 6+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-04-08 10:03 Can't resolve mismatch_count > 0 for a raid 1 array Steven Ellis
2009-04-08 21:49 ` Goswin von Brederlow
2009-04-09 0:07 ` Steven Ellis
2009-04-08 21:50 ` Bill Davidsen
2009-04-08 22:00 ` Iustin Pop
2009-04-09 0:13 ` Steven Ellis
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).