* Cluster Aware MD Driver
[not found] ` <20030104164314.26621.qmail@escalade.vistahp.com>
@ 2003-01-04 19:13 ` Steven Dake
2003-01-04 21:06 ` Brian Jackson
0 siblings, 1 reply; 5+ messages in thread
From: Steven Dake @ 2003-01-04 19:13 UTC (permalink / raw)
To: Brian Jackson; +Cc: opengfs-users, linux-raid
Brian,
The way to fix this problem is to enable the MD layer and MD tools to be
cluster aware. I hope sometime before March 03 to have this
functionality implemented in 2.4 and mdadm 1.0.1.
While I have not spent a ton of time thinking about it, I'll share what
I'm thinking in case someone else wants to get a jump start since I wont
have time until Feb to start working on it.
What I am thinking of doing is making one node in a cluster the master,
and all other nodes nonmasters. The master is responsible for commiting
superblock and resync changes to disk. The other nodes are responsible
for updating their status if any one node changes state.
During startup, a userland daemon runs which opens a TCP port. Then the
daemon attempts to connect to a list of other servers in a configuration
file (/etc/clusterips). On each connection, a master of the cluster is
elected by an election algortihm. Once the daemon makes connections, it
culls its file descriptor list such that it only has one connection per
server to each other server. Whenever a state change occurs (raid set
faulty, raid array failed, raid hot added, raid hot removed, raid start,
raid stop, etc) the state change is transmitted via the server to all
other server nodes. When the server receives a state change message, it
will handle appropriately by using the md ioctls to update its internal
state. There are more complications, such as heartbeating to detect
dead nodes, etc. All of this keeps the arrays in sync across the entire
cluster. The big changes here are to the mdadm package to use the local
server to execute operations and to create a server which processes the
state changes and executes the appropriate MD layer ioctls.
The question you might be asking is, how do you protect from each server
overwritting similiar data such as the superblock or resync data. The
trick is to add an ioctl to the MD layer that turns writes on or off.
During master election above, the master will turn its writes on. All
non-masters will turn their writes off before any RAID arrays are
started. Also, resyncs have to be communicated across the cluster such
that /proc/mdstat displays correct information so I believe an ioctl
will have to be added to indicate this and allow resync from a specific
spot in case the current master dies in a resync. The changes to the MD
layer should be fairly minimal and noninvasive and also work well for
non-cluster configurations. The writes will default to on to ensure
that non-clusters work properly even with autostart.
The only downside of this approach is that RAID autostart can no longer
be used. The only solution to supporting that feature for clusters is
to write all of the above userspace code into the kernel which would be
a big pain and likely not accepted into the mainline.
Thanks
-steve
Brian Jackson wrote:
> At the moment if one node goes down, the rest of the nodes will
> continue to run as expected. The only single point of failure from a
> node point of view is the lock server. Dominik is working on
> integrating the OpenDLM to get rid of that single point of failure. If
> one disk in the shared storage dies, that will bring you down. I have
> tried to stack pool on top of software raid, but the MD driver doesn't
> play well with clusters so that won't work. We are trying to figure
> out a way to fix this, but when GFS was originally designed the pool
> driver(that does GFS's volume management) only had stripping(raid0). I
> hope this answers your questions.
> --Brian Jackson
> P.S. I am not sure, but it sounds like you have some misconception
> about how OpenGFS works. It does not use the disks on your nodes. It
> uses some kind of shared storage.
> Vittorio Ballestra writes:
>
>> Escuse me if this is a stupid or repetead question, I've searched
>> into the docs (?) and the mailinglists but I'm not able to understand
>> if openGFS has some sort of fault-tollerance.
>> This are my doubts :
>> What happens if one host is down or one disk on one host is down ?
>> Will the entire openGFS filesystem be down ? If one disk is broken
>> and its content corrupted, will the whole openGFS be corrupted ?
>> If openGFS is not supporting any kind of fault-tollerance, can one
>> use raid disks on each node ?
>> Thanks,
>> V
>>
>> -------------------------------------------------------
>> This sf.net email is sponsored by:ThinkGeek
>> Welcome to geek heaven.
>> http://thinkgeek.com/sf
>> _______________________________________________
>> Opengfs-users mailing list
>> Opengfs-users@lists.sourceforge.net
>> https://lists.sourceforge.net/lists/listinfo/opengfs-users
>
>
>
>
> -------------------------------------------------------
> This sf.net email is sponsored by:ThinkGeek
> Welcome to geek heaven.
> http://thinkgeek.com/sf
> _______________________________________________
> Opengfs-users mailing list
> Opengfs-users@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/opengfs-users
>
>
>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: Cluster Aware MD Driver
2003-01-04 19:13 ` Steven Dake
@ 2003-01-04 21:06 ` Brian Jackson
0 siblings, 0 replies; 5+ messages in thread
From: Brian Jackson @ 2003-01-04 21:06 UTC (permalink / raw)
To: Steven Dake; +Cc: opengfs-users, linux-raid
Steven Dake writes:
> Brian,
>
> The way to fix this problem is to enable the MD layer and MD tools to be
> cluster aware. I hope sometime before March 03 to have this functionality
> implemented in 2.4 and mdadm 1.0.1.
>
> While I have not spent a ton of time thinking about it, I'll share what
> I'm thinking in case someone else wants to get a jump start since I wont
> have time until Feb to start working on it.
I haven't looked at their implementation in depth, but I know EVMS has the
beginnings of this. Might be worth a peek.
>
> What I am thinking of doing is making one node in a cluster the master,
> and all other nodes nonmasters. The master is responsible for commiting
> superblock and resync changes to disk. The other nodes are responsible
> for updating their status if any one node changes state.
>
> During startup, a userland daemon runs which opens a TCP port. Then the
> daemon attempts to connect to a list of other servers in a configuration
> file (/etc/clusterips). On each connection, a master of the cluster is
> elected by an election algortihm. Once the daemon makes connections, it
> culls its file descriptor list such that it only has one connection per
> server to each other server. Whenever a state change occurs (raid set
> faulty, raid array failed, raid hot added, raid hot removed, raid start,
> raid stop, etc) the state change is transmitted via the server to all
> other server nodes. When the server receives a state change message, it
> will handle appropriately by using the md ioctls to update its internal
> state. There are more complications, such as heartbeating to detect dead
> nodes, etc. All of this keeps the arrays in sync across the entire
> cluster. The big changes here are to the mdadm package to use the local
> server to execute operations and to create a server which processes the
> state changes and executes the appropriate MD layer ioctls.
One thing you might want to think about is that most people who are looking
at a cluster capable raid are already going to have some sort of cluster
management software. It might be useful to use the transports available.
Maybe a plugin system that uses what you were saying as the default method,
but can also use a plugin written to take advantage of an existing cluster
management system. Just an idea.
>
> The question you might be asking is, how do you protect from each server
> overwritting similiar data such as the superblock or resync data. The
> trick is to add an ioctl to the MD layer that turns writes on or off.
> During master election above, the master will turn its writes on. All
> non-masters will turn their writes off before any RAID arrays are started.
> Also, resyncs have to be communicated across the cluster such that
> /proc/mdstat displays correct information so I believe an ioctl will have
> to be added to indicate this and allow resync from a specific spot in case
> the current master dies in a resync. The changes to the MD layer should
> be fairly minimal and noninvasive and also work well for non-cluster
> configurations. The writes will default to on to ensure that non-clusters
> work properly even with autostart.
Maybe the ability to write or not could be a mdadm switch. Something like:
mdadm -A --non-master
that would keep the changes to the MD drivers to a minimum(I think, but I
may be thinking the wrong way), but require manual intervention if the
master were to die(or at least some sort of outside intervention).
>
> The only downside of this approach is that RAID autostart can no longer be
> used. The only solution to supporting that feature for clusters is to
> write all of the above userspace code into the kernel which would be a big
> pain and likely not accepted into the mainline.
That won't be a big deal for most HA/LB type clusters. I think it will
matter more to the HPC crowd though(they might be using GFS as /). Even then
they can probably get most of the functionality they need with early
userspace.
>
> Thanks
> -steve
--Brian Jackson
P.S. I have some other ideas about this I might send later on. Also LMB from
the Linux-ha list might be interested too. I think he wrote some patches
that sort of made MD work in a cluster.
>
> Brian Jackson wrote:
>
>> At the moment if one node goes down, the rest of the nodes will continue
>> to run as expected. The only single point of failure from a node point of
>> view is the lock server. Dominik is working on integrating the OpenDLM to
>> get rid of that single point of failure. If one disk in the shared
>> storage dies, that will bring you down. I have tried to stack pool on top
>> of software raid, but the MD driver doesn't play well with clusters so
>> that won't work. We are trying to figure out a way to fix this, but when
>> GFS was originally designed the pool driver(that does GFS's volume
>> management) only had stripping(raid0). I hope this answers your
>> questions.
>> --Brian Jackson
>> P.S. I am not sure, but it sounds like you have some misconception about
>> how OpenGFS works. It does not use the disks on your nodes. It uses some
>> kind of shared storage.
>> Vittorio Ballestra writes:
>>
>>> Escuse me if this is a stupid or repetead question, I've searched into
>>> the docs (?) and the mailinglists but I'm not able to understand if
>>> openGFS has some sort of fault-tollerance.
>>> This are my doubts :
>>> What happens if one host is down or one disk on one host is down ? Will
>>> the entire openGFS filesystem be down ? If one disk is broken and its
>>> content corrupted, will the whole openGFS be corrupted ?
>>> If openGFS is not supporting any kind of fault-tollerance, can one use
>>> raid disks on each node ?
>>> Thanks,
>>> V
>>>
>>> -------------------------------------------------------
>>> This sf.net email is sponsored by:ThinkGeek
>>> Welcome to geek heaven.
>>> http://thinkgeek.com/sf
>>> _______________________________________________
>>> Opengfs-users mailing list
>>> Opengfs-users@lists.sourceforge.net
>>> https://lists.sourceforge.net/lists/listinfo/opengfs-users
>>
>>
>>
>>
>> -------------------------------------------------------
>> This sf.net email is sponsored by:ThinkGeek
>> Welcome to geek heaven.
>> http://thinkgeek.com/sf
>> _______________________________________________
>> Opengfs-users mailing list
>> Opengfs-users@lists.sourceforge.net
>> https://lists.sourceforge.net/lists/listinfo/opengfs-users
>>
>>
>>
>
>
>
> -------------------------------------------------------
> This sf.net email is sponsored by:ThinkGeek
> Welcome to geek heaven.
> http://thinkgeek.com/sf
> _______________________________________________
> Opengfs-users mailing list
> Opengfs-users@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/opengfs-users
^ permalink raw reply [flat|nested] 5+ messages in thread
* RE: Cluster Aware MD Driver
@ 2003-01-07 14:54 Cress, Andrew R
0 siblings, 0 replies; 5+ messages in thread
From: Cress, Andrew R @ 2003-01-07 14:54 UTC (permalink / raw)
To: 'Brian Jackson', Steven Dake; +Cc: opengfs-users, linux-raid
I also had a couple of comments below. My .02.
Andy
-----Original Message-----
From: Brian Jackson
Sent: Saturday, January 04, 2003 4:06 PM
To: Steven Dake
Cc: opengfs-users@lists.sourceforge.net; linux-raid@vger.kernel.org
Subject: Re: Cluster Aware MD Driver
[...]
One thing you might want to think about is that most people who are looking
at a cluster capable raid are already going to have some sort of cluster
management software. It might be useful to use the transports available.
Maybe a plugin system that uses what you were saying as the default method,
but can also use a plugin written to take advantage of an existing cluster
management system. Just an idea.
[Andy] I would agree, but state this more strongly, in that utilizing the
existing
cluster management software would be a requirement for most of these
customers.
Things like the heartbeat, network transport, election process, and
master/slave
relationships would be things that would come with that software, and need
to be
administered from a common interface in the cluster, so using a plugin or
dynamic
library for these functions sounds like a good approach.
>
> The question you might be asking is, how do you protect from each server
> overwritting similiar data such as the superblock or resync data.
[...]
> The writes will default to on to ensure that non-clusters
> work properly even with autostart.
Maybe the ability to write or not could be a mdadm switch. Something like:
mdadm -A --non-master
that would keep the changes to the MD drivers to a minimum(I think, but I
may be thinking the wrong way), but require manual intervention if the
master were to die(or at least some sort of outside intervention).
[Andy] The outside intervention would be from the cluster management
software, since
I don't think manual intervention would do. If there were an API to toggle
this, the
cluster management software would be able to change who the master was if
the master
node went down.
[...]
^ permalink raw reply [flat|nested] 5+ messages in thread
* Cluster Aware MD Driver
@ 2007-06-13 19:16 Xinwei Hu
2007-06-13 19:50 ` Mike Snitzer
0 siblings, 1 reply; 5+ messages in thread
From: Xinwei Hu @ 2007-06-13 19:16 UTC (permalink / raw)
To: linux-raid
Hi all,
Steven Dake proposed a solution* to make MD layer and tools to be cluster
aware in early 2003. But it seems that no progressing is made since then. I'd
like to pick this one up again. :)
So far as I understand, Steven's proposal still applies to currently MD
implementation mostly, except we have bitmap now. And bitmap can be
workarounded via set_bitmap_file.
The problem is that it seems we need a kernel<->userspace interface to sync
the mddev struct across all nodes, but I don't find out how.
I'm new to the MD driver, so correct me if I'm wrong. And you suggestions
are really appreciated.
Thanks.
* http://osdir.com/ml/raid/2003-01/msg00013.html
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: Cluster Aware MD Driver
2007-06-13 19:16 Cluster Aware MD Driver Xinwei Hu
@ 2007-06-13 19:50 ` Mike Snitzer
0 siblings, 0 replies; 5+ messages in thread
From: Mike Snitzer @ 2007-06-13 19:50 UTC (permalink / raw)
To: Xinwei Hu; +Cc: linux-raid
Is the goal to have the MD device be directly accessible from all
nodes? This strategy seems flawed in that it speaks to updating MD
superblocks then in-memory Linux data structures across a cluster.
The reality is if we're talking about shared storage the MD management
only needs to happen in one node. Others can weigh in on this but the
current MD really doesn't want to be cluster-aware.
IMHO, this cluster awareness really doesn't belong in MD/mdadm. A
high-level cluster management tool should be doing this MD
ownership/coordination work. The MD ownership can be transferred
accordingly if/when the current owner fails, etc. But this implies
that the MD is only ever active on one node at any given point in
time.
Mike
On 6/13/07, Xinwei Hu <hxinwei@gmail.com> wrote:
> Hi all,
>
> Steven Dake proposed a solution* to make MD layer and tools to be cluster
> aware in early 2003. But it seems that no progressing is made since then. I'd
> like to pick this one up again. :)
>
> So far as I understand, Steven's proposal still applies to currently MD
> implementation mostly, except we have bitmap now. And bitmap can be
> workarounded via set_bitmap_file.
>
> The problem is that it seems we need a kernel<->userspace interface to sync
> the mddev struct across all nodes, but I don't find out how.
>
> I'm new to the MD driver, so correct me if I'm wrong. And you suggestions
> are really appreciated.
>
> Thanks.
>
> * http://osdir.com/ml/raid/2003-01/msg00013.html
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2007-06-13 19:50 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2007-06-13 19:16 Cluster Aware MD Driver Xinwei Hu
2007-06-13 19:50 ` Mike Snitzer
-- strict thread matches above, loose matches on Subject: below --
2003-01-07 14:54 Cress, Andrew R
[not found] <200301041016.10224.vittorio.ballestra@infogestnet.it>
[not found] ` <20030104164314.26621.qmail@escalade.vistahp.com>
2003-01-04 19:13 ` Steven Dake
2003-01-04 21:06 ` Brian Jackson
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).