http://activemq.apache.org/producer-flow-control.html

On Sep 11, 2011, at 6:08 PM, bbansal wrote:

> Hello folks, 
> 
> I am evaluating ActiveMQ for some simple scenarios. The web-server will push
> notifications to the queue/topic to be consumed by one or many consumers.
> The one requirement is web-server should not get impacted or should be able
> to write at their speed even if consumers goes down etc. 
> 
> ActiveMQ is performing very well with about 1500 QPS (8 producer thread,
> persistence, kaha-db) Kahadb parameters being used are 
> 
> enableJournalDiskSyncs="false" indexWriteBatchSize="1000"
> enableIndexWriteAsync="true
> 
> The system works great if consumers are all caught up, the issue is when I
> am trying to test scenarios with backlogged data (keep running producer for
> 30 mins or so) and then start consumers. Consumer show good consumption rate
> but the producers (8 threads same as before) cannot do more than 120 QPS.
> This is a drop of more than 90% degradation. 
> 
> I ran profiler on the code (Jprofiler) and looks like the writers are
> getting stuck for write locks while competing with the removeAsyncMessages()
> or call to clear messages which got acknowledged from clients etc. 
> 
> I saw similar complaints for some other folks, Is there some settings we can
> use to fix the problem ? I dont want to degrade any guarantee level (eg.
> disable acks etc). 
> 
> Would be more than happy to run experiments with different settings if folks
> have some suggestions. 
> 
> 
> 
> --
> View this message in context: 
> http://activemq.2283324.n4.nabble.com/Backlog-data-causes-producers-to-slow-down-tp3806018p3806018.html
> Sent from the ActiveMQ - User mailing list archive at Nabble.com.

Reply via email to