All of lore.kernel.org
 help / color / mirror / Atom feed
* [RFC] ct_sync 0.15
@ 2004-08-13 13:35 KOVACS Krisztian
  2004-08-13 14:01 ` [nf-failover] " Jonathan G - Mailing Lists
  0 siblings, 1 reply; 4+ messages in thread
From: KOVACS Krisztian @ 2004-08-13 13:35 UTC (permalink / raw)
  To: Netfilter-failover list; +Cc: netfilter-devel


  Hi,

  I think finally I managed to get ct_sync in a state where some more
public review/testing would certainly do good for the project. This week
I've been successfully running overnight stress tests on our two-node
cluster, and both of the nodes survived all of the tests without locking
up, panicking, etc. OK, I know this is not too much, but somehow I felt
this would be a bare minimum we should reach before asking anyone to try
out ct_sync. Of course, there is still a lot to do. Just to mention two
things which certainly need a lot of work: protocol implementation is
really the bare minimum which is capable of doing anything, and
expectation support is missing completely. The current implementation
supports replicating conntrack entries, and supports NAT. So basically
all the simple things (not using expectations) should work reasonably
well. This is why I've bumped the version number to 0.15.

  The aim of this e-mail is twofold:

      * First, I just thought that ct_sync is now ready for some
        testing. Unfortunately our development and testing environment
        is very limited, so anyone being able to help us out and do some
        testing would be of a big help.
      * After doing some work to stabilize the current code, there is
        now need to discuss a few things before going on with
        implementation. So, I'd be happy if there was some discussion on
        these things before starting to implement anything.

  I'll try to summarize the most important open problems I've come
through. This list does not contain things I consider trivial, for
example exporting important internal constants as tunable parameters
through sysctl.

     1. There should some facility by which one can select which
        connections have to be replicated. This way it would be possible
        to limit replication traffic to the bare minimum. For example,
        there is no point in replicating conntrack entries for
        connections whose endpoint is one of the nodes (administrative
        SSH traffic, for example). A per-conntrack flag would be needed,
        just like CONNMARK, which could be set for conntracks needing
        replication with a simple iptables rule. Actually, CONNMARK is
        enough, if we choose a given bit of the mark as the SYNC bit.
        Besides this, we should decide if we needed a SYNC or a NOSYNC
        bit, that is, if the default mode of operation should be "sync
        or not to sync".
     2. The error recovery functions in the protocol layer should be
        revamped. The protocol is a plain sequence numbered, NACK-based
        one using multicast UDP. Lost packets are detected by the
        receiver when receiving the next packet with the inappropriate
        sequence number (not 'seq of the last packet + 1') is received.
        When this occurs, the node sends a recovery request containing
        the last successfully received sequence number. When the 

-- 
 Regards,
   Krisztian KOVACS

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [nf-failover] [RFC] ct_sync 0.15
  2004-08-13 13:35 [RFC] ct_sync 0.15 KOVACS Krisztian
@ 2004-08-13 14:01 ` Jonathan G - Mailing Lists
  2004-08-13 17:35   ` KOVACS Krisztian
  0 siblings, 1 reply; 4+ messages in thread
From: Jonathan G - Mailing Lists @ 2004-08-13 14:01 UTC (permalink / raw)
  To: KOVACS Krisztian; +Cc: Netfilter-failover list, netfilter-devel


Hi Krisztian,

    my partner and me are preparing for the next week a firewall/High 
Availability/Load Balancer cluster made of the following software:

- Security: IPTables 1.2.8
- HA: KeepAlived (only the VRRPv2 part)
- LB: Pen (siag.nu/pen) for tcp connections


We would be so pleased and interested in test you package and discuss a 
real wold implementation (the installation of this cluster (2 nodes) is 
for an ISP).

Drop me an email to tell how start and we will help all of us get more 
real results with our system.

;-)

BR,

jonathan


KOVACS Krisztian wrote:

>   Hi,
> 
>   I think finally I managed to get ct_sync in a state where some more
> public review/testing would certainly do good for the project. This week
> I've been successfully running overnight stress tests on our two-node
> cluster, and both of the nodes survived all of the tests without locking
> up, panicking, etc. OK, I know this is not too much, but somehow I felt
> this would be a bare minimum we should reach before asking anyone to try
> out ct_sync. Of course, there is still a lot to do. Just to mention two
> things which certainly need a lot of work: protocol implementation is
> really the bare minimum which is capable of doing anything, and
> expectation support is missing completely. The current implementation
> supports replicating conntrack entries, and supports NAT. So basically
> all the simple things (not using expectations) should work reasonably
> well. This is why I've bumped the version number to 0.15.
> 
>   The aim of this e-mail is twofold:
> 
>       * First, I just thought that ct_sync is now ready for some
>         testing. Unfortunately our development and testing environment
>         is very limited, so anyone being able to help us out and do some
>         testing would be of a big help.
>       * After doing some work to stabilize the current code, there is
>         now need to discuss a few things before going on with
>         implementation. So, I'd be happy if there was some discussion on
>         these things before starting to implement anything.
> 
>   I'll try to summarize the most important open problems I've come
> through. This list does not contain things I consider trivial, for
> example exporting important internal constants as tunable parameters
> through sysctl.
> 
>      1. There should some facility by which one can select which
>         connections have to be replicated. This way it would be possible
>         to limit replication traffic to the bare minimum. For example,
>         there is no point in replicating conntrack entries for
>         connections whose endpoint is one of the nodes (administrative
>         SSH traffic, for example). A per-conntrack flag would be needed,
>         just like CONNMARK, which could be set for conntracks needing
>         replication with a simple iptables rule. Actually, CONNMARK is
>         enough, if we choose a given bit of the mark as the SYNC bit.
>         Besides this, we should decide if we needed a SYNC or a NOSYNC
>         bit, that is, if the default mode of operation should be "sync
>         or not to sync".
>      2. The error recovery functions in the protocol layer should be
>         revamped. The protocol is a plain sequence numbered, NACK-based
>         one using multicast UDP. Lost packets are detected by the
>         receiver when receiving the next packet with the inappropriate
>         sequence number (not 'seq of the last packet + 1') is received.
>         When this occurs, the node sends a recovery request containing
>         the last successfully received sequence number. When the 
> 

-- 
   :::: Jonathan Gonzalez Fernandez ::::

    (o>  mail  : jonathan@surestorm.com
    //\  jabber: surestorm@jabber.org
    V_/  site  : www.surestorm.com

   ::: Registered Linux User #333386 :::

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [nf-failover] [RFC] ct_sync 0.15
  2004-08-13 14:01 ` [nf-failover] " Jonathan G - Mailing Lists
@ 2004-08-13 17:35   ` KOVACS Krisztian
  2004-08-13 18:07     ` Jonathan G - Mailing Lists
  0 siblings, 1 reply; 4+ messages in thread
From: KOVACS Krisztian @ 2004-08-13 17:35 UTC (permalink / raw)
  To: Jonathan G - Mailing Lists; +Cc: Netfilter-failover list, netfilter-devel


  Hi,

On Fri, Aug 13, 2004 at 04:01:00PM +0200, Jonathan G - Mailing Lists wrote:
>    my partner and me are preparing for the next week a firewall/High 
> Availability/Load Balancer cluster made of the following software:
> 
> - Security: IPTables 1.2.8
> - HA: KeepAlived (only the VRRPv2 part)
> - LB: Pen (siag.nu/pen) for tcp connections
> 
> 
> We would be so pleased and interested in test you package and discuss a 
> real wold implementation (the installation of this cluster (2 nodes) is 
> for an ISP).

  Wow, great. Primarily I would like if you could do some testing. I
don't think ct_sync is production-ready, but definitely needs testing in
as many different environments as possible. So, if you have some extra
time and could give ct_sync a test run, I'd be really happy.

> Drop me an email to tell how start and we will help all of us get more 
> real results with our system.

  Hmmm, in theory all you have to do is check out the netfilter-ha
repository from CVS, and all instructions should be given in the README
file there. (Probably the most problematic part is getting the patches
applied to your kernel. If you have problems doing so, I'll happily help
with creating an aggregated patch.)

  To check out the tree, use:
$ cvs -d :pserver:cvs:cvs@pserver.netfilter.org:/cvspublic co netfilter-ha

-- 
 KOVACS Krisztian

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [nf-failover] [RFC] ct_sync 0.15
  2004-08-13 17:35   ` KOVACS Krisztian
@ 2004-08-13 18:07     ` Jonathan G - Mailing Lists
  0 siblings, 0 replies; 4+ messages in thread
From: Jonathan G - Mailing Lists @ 2004-08-13 18:07 UTC (permalink / raw)
  To: KOVACS Krisztian; +Cc: Netfilter-failover list, netfilter-devel

OK Krisztian, as i said on Tuesday i will start patching the kernel and 
i will be in contact.

;-)
BR,

jonathan


KOVACS Krisztian wrote:
>   Hi,
> 
> On Fri, Aug 13, 2004 at 04:01:00PM +0200, Jonathan G - Mailing Lists wrote:
> 
>>   my partner and me are preparing for the next week a firewall/High 
>>Availability/Load Balancer cluster made of the following software:
>>
>>- Security: IPTables 1.2.8
>>- HA: KeepAlived (only the VRRPv2 part)
>>- LB: Pen (siag.nu/pen) for tcp connections
>>
>>
>>We would be so pleased and interested in test you package and discuss a 
>>real wold implementation (the installation of this cluster (2 nodes) is 
>>for an ISP).
> 
> 
>   Wow, great. Primarily I would like if you could do some testing. I
> don't think ct_sync is production-ready, but definitely needs testing in
> as many different environments as possible. So, if you have some extra
> time and could give ct_sync a test run, I'd be really happy.
> 
> 
>>Drop me an email to tell how start and we will help all of us get more 
>>real results with our system.
> 
> 
>   Hmmm, in theory all you have to do is check out the netfilter-ha
> repository from CVS, and all instructions should be given in the README
> file there. (Probably the most problematic part is getting the patches
> applied to your kernel. If you have problems doing so, I'll happily help
> with creating an aggregated patch.)
> 
>   To check out the tree, use:
> $ cvs -d :pserver:cvs:cvs@pserver.netfilter.org:/cvspublic co netfilter-ha
> 

-- 
   :::: Jonathan Gonzalez Fernandez ::::

    (o>  mail  : jonathan@surestorm.com
    //\  jabber: surestorm@jabber.org
    V_/  site  : www.surestorm.com

   ::: Registered Linux User #333386 :::

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2004-08-13 18:07 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2004-08-13 13:35 [RFC] ct_sync 0.15 KOVACS Krisztian
2004-08-13 14:01 ` [nf-failover] " Jonathan G - Mailing Lists
2004-08-13 17:35   ` KOVACS Krisztian
2004-08-13 18:07     ` Jonathan G - Mailing Lists

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.