Hi Tim, Thanks for your reply. I actually came across your stack overflow post and I tried it out but it didn’t work for me.
I ended up using multiple bridges between datacenter. This allows us to have more than one tcp session to get more acks on the wire. Basically I needed to adjust the number of bridges based on the amount of latency I was experiencing. It was trial and error approach to determine the correct number of bridges. Brian > On Jan 12, 2021, at 10:20 PM, Tim Bain <tb...@alumni.duke.edu> wrote: > > Early in my time as an ActiveMQ user, I ran into unexpectedly poor > performance between a network of ActiveMQ 5.x brokers across a high-latency > network link. The particular situation I had involved links that were idle > long enough for the TCP congestion window to contract back to its min value > between bursts of traffic, resulting in overall low throughput. The > solution ended up being tweaks to the TCP configuration on the hosts in > question, though it also turned out that the problem would only occur in > certain somewhat unlikely scenarios and was mainly an artifact of my > simplistic test case. Full details are at > https://stackoverflow.com/questions/25494929/3x-latency-observed-in-activemq-only-for-higher-wan-latencies > . > > This may or may not be relevant to your situation (though it's worth > investigating), but there's a relevant general conclusion to be drawn: > sometimes the root cause of a problem exists outside of the broker itself, > and so your attempts to characterize the problem need to consider not only > the broker but also other elements such as the network stack. (Your > environment won't give you any control over the network itself from the > sound of it, otherwise I'd include that in the list of things to > instrument.) > > Tim > > On Tue, Jan 12, 2021, 8:23 AM Brian Ramprasad <bri...@cs.toronto.edu> wrote: > >> Hi Everyone, >> >> >> I am trying setup a network of 3 brokers, one in each Amazon datacenter. >> The latency between the datacenters is quite high (25-150ms) and it is >> severely impacting the throughput as compared to running all 3 brokers in >> the same datacenter. The 3 brokers are connected in a pipeline (1->2->3) >> using Core bridges. >> >> I have tried increasing the tcpSendBufferSize and the tcpReceiveBufferSize >> on the acceptors in the broker.xml but it does not seem to improve the >> throughput. When I use a really small buffer, the throughput goes down even >> further so I know that changing the buffer size is having an effect. >> >> Does anyone have an idea of what other settings I can try to improve the >> throughput? >> >> Thanks! >> >> Brian R >> >> >>