On Thu, Dec 10, 2009 at 01:23:52PM +0100, Beck Ingmar wrote:
> 
> 
> Hi Florian,
> 
> why improves bonding of the replication-connection not the performance.
> 
> We currently have a direct connection for the 1 Gb DRBD replication and
> I
> wanted to activate an active-active bonding (with a further LAN-port). I
> would expect nearly double of replication-performance. Andreas Kurz had
> spoken at a workshop for that.
> 
> Or depends it on whether the connection between the nodes directly or
> via 1-n switches.

Quoting linux/Documentation/networking/bonding.txt:

...

balance-rr: This mode is the only mode that will permit a single
        TCP/IP connection to stripe traffic across multiple
        interfaces. It is therefore the only mode that will allow a
        single TCP/IP stream to utilize more than one interface's
        worth of throughput.  This comes at a cost, however: the
        striping generally results in peer systems receiving packets out
        of order, causing TCP/IP's congestion control system to kick
        in, often by retransmitting segments.

        It is possible to adjust TCP/IP's congestion limits by
        altering the net.ipv4.tcp_reordering sysctl parameter.  The
        usual default value is 3, and the maximum useful value is 127.
        For a four interface balance-rr bond, expect that a single
        TCP/IP stream will utilize no more than approximately 2.3
        interface's worth of throughput, even after adjusting
        tcp_reordering.

...

This likely has been correct at some point in time, or for 100 MBit/s.

In my experience, bonding two 1GBit links can yield about 1.7 to 1.8,
adding the third drops back to about 1.0 but adds latency,
and adding a forth will deteriorate to less then a single link.

This is for a _sinlge_ tcp bulk transfer.  Of course, the sum total over
all transfers on those bonded links will be higher.

But for the single tcp bulk transfer, in my experience, more than two
1GBit links in bonding-rr are counterproductive.

May be highly variable with different NICs, boards,
processors and bus and processor speeds.


-- 
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com

DRBD® and LINBIT® are registered trademarks of LINBIT, Austria.
__
please don't Cc me, but send to list   --   I'm subscribed
_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to