The publisher tried to publish 10G data (msg size 10k) to 3 brokers. This error occurred so frequently and caused big delay.
[2013-06-06 11:13:29,491] INFO Shutting down producer (kafka.producer.Producer) [2013-06-06 11:13:29,492] INFO Beging shutting down ProducerSendThread (kafka.producer.async.ProducerSendThread) [2013-06-06 11:13:32,804] INFO Shutting down producer (kafka.producer.Producer) [2013-06-06 11:13:32,804] INFO Beging shutting down ProducerSendThread (kafka.producer.async.ProducerSendThread) [2013-06-06 11:13:33,716] WARN Failed to send producer request with correlation id 334159 to broker 1 with data for partitions [libomirror8,9],[libomirror8,6],[libomirror8,0],[libomirror8,3] (kafka.producer.async.DefaultEventHandler) kafka.common.KafkaException: Incomplete response (ProducerResponse(334159,Map([libomirror8,9] -> ProducerResponseStatus(0,96820), [libomirror8,6] -> ProducerResponseStatus(2,0), [libomirror8,0] -> ProducerResponseStatus(0,0)))) for producer request (Name: ProducerRequest; Version: 0; CorrelationId: 334159; ClientId: ProducerPerformance; RequiredAcks: -1; AckTimeoutMs: 3000 ms; TopicAndPartition: [libomirror8,9] -> 102660,[libomirror8,6] -> 102660,[libomirror8,3] -> 102660,[libomirror8,0] -> 102660) at kafka.producer.async.DefaultEventHandler.kafka$producer$async$DefaultEventHandler$$send(DefaultEventHandler.scala:250) at kafka.producer.async.DefaultEventHandler$$anonfun$dispatchSerializedData$1.apply(DefaultEventHandler.scala:107) at kafka.producer.async.DefaultEventHandler$$anonfun$dispatchSerializedData$1.apply(DefaultEventHandler.scala:101) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:80) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:80) at scala.collection.Iterator$class.foreach(Iterator.scala:631) at scala.collection.mutable.HashTable$$anon$1.foreach(HashTable.scala:161) at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:194) at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) at scala.collection.mutable.HashMap.foreach(HashMap.scala:80) at kafka.producer.async.DefaultEventHandler.dispatchSerializedData(DefaultEventHandler.scala:101) at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:73) at kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104) at kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87) at kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:67) at scala.collection.immutable.Stream.foreach(Stream.scala:254) at kafka.producer.async.ProducerSendThread.processEvents(ProducerSendThread.scala:66) at kafka.producer.async.ProducerSendThread.run(ProducerSendThread.scala:44) [2013-06-06 11:13:33,728] INFO Back off for 100 ms before retrying send. Remaining retries = 3 (kafka.producer.async.DefaultEventHandler) Regards, Libo