Hi Jun,

Thanks for clarification, I'll give a try with ack=-1 (in producer).

However, i did a fallback to older version of kafka (*kafka_2.10-0.8.2.1*),
and i don't see this issue (loss of messages).

looks like kafka_2.11-0.9.0.1 has issues(BUG) during replication.

Thanks,

Regards,
Mazhar Shaikh.



On Thu, Aug 18, 2016 at 10:30 PM, Jun Rao <j...@confluent.io> wrote:

> Mazhar,
>
> There is probably a mis-understanding. Ack=-1 (or all) doesn't mean waiting
> for all replicas. It means waiting for all replicas that are in sync. So,
> if a replica is down, it will be removed from the in-sync replicas, which
> allows the producer to continue with fewer replicas.
>
> For the connection issue that you saw in the log, this could happen when a
> connection is idle for some time. It won't break the replication logic
> since a new connection will be created automatically. You can increase the
> socket idle time on the broker if you want to turn off this behavior.
>
> Thanks,
>
> Jun
>
> On Thu, Aug 18, 2016 at 12:07 AM, Mazhar Shaikh <
> mazhar.shaikh...@gmail.com>
> wrote:
>
> > Hi Jun,
> >
> > Setting to -1, may solve this issue.
> > But it will cause producer buffer full in load test resulting to failures
> > and drop of messages from client(producer side)
> > Hence, this will not actually solve the problem.
> >
> > I need to fix this from kafka broker side, so that there is no impact on
> > producer or consumer.
> >
> > From the logs looks like there is connection problem during between
> brokers
> > and kafka broker is loosing records during this process.
> >
> > But why is kafka broker loosing records,
> >
> > I feel this is a BUG in kafka.
> >
> > [2016-08-17 12:54:50,293] TRACE [Controller 2]: checking need to trigger
> > partition rebalance (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,294] DEBUG [Controller 2]: preferred replicas by
> > broker Map(0 -> Map([topic1,45] -> List(0, 1), [topic1,17] -> List(0, 1),
> > [topic1,19] -> List(0, 1), [topic1,42] -> List(0, 1), [topic1,43] ->
> > List(0, 1), [topic1,44] -> List(0, 1), [topic1,16] -> List(0, 1),
> > [topic1,46] -> List(0, 1), [topic1,20] -> List(0, 1), [topic1,41] ->
> > List(0, 1), [topic1,18] -> List(0, 1), [topic1,22] -> List(0, 1),
> > [topic1,40] -> List(0, 1), [topic1,47] -> List(0, 1), [topic1,23] ->
> > List(0, 1), [topic1,21] -> List(0, 1)), 5 -> Map([topic1,78] -> List(5,
> 3),
> > [topic1,84] -> List(5, 3), [topic1,87] -> List(5, 3), [topic1,74] ->
> > List(5, 3), [topic1,81] -> List(5, 3), [topic1,73] -> List(5, 3),
> > [topic1,80] -> List(5, 3), [topic1,77] -> List(5, 3), [topic1,75] ->
> > List(5, 3), [topic1,85] -> List(5, 3), [topic1,76] -> List(5, 3),
> > [topic1,83] -> List(5, 3), [topic1,86] -> List(5, 3), [topic1,72] ->
> > List(5, 3), [topic1,79] -> List(5, 3), [topic1,82] -> List(5, 3)), 1 ->
> > Map([topic1,92] -> List(1, 0), [topic1,95] -> List(1, 0), [topic1,69] ->
> > List(1, 0), [topic1,93] -> List(1, 0), [topic1,70] -> List(1, 0),
> > [topic1,67] -> List(1, 0), [topic1,65] -> List(1, 0), [topic1,88] ->
> > List(1, 0), [topic1,90] -> List(1, 0), [topic1,66] -> List(1, 0),
> > [topic1,94] -> List(1, 0), [topic1,64] -> List(1, 0), [topic1,89] ->
> > List(1, 0), [topic1,68] -> List(1, 0), [topic1,71] -> List(1, 0),
> > [topic1,91] -> List(1, 0)), 2 -> Map([topic1,8] -> List(2, 4), [topic1,3]
> > -> List(2, 4), [topic1,15] -> List(2, 4), [topic1,2] -> List(2, 4),
> > [topic1,1] -> List(2, 4), [topic1,6] -> List(2, 4), [topic1,9] -> List(2,
> > 4), [topic1,12] -> List(2, 4), [topic1,14] -> List(2, 4), [topic1,11] ->
> > List(2, 4), [topic1,13] -> List(2, 4), [topic1,0] -> List(2, 4),
> [topic1,4]
> > -> List(2, 4), [topic1,5] -> List(2, 4), [topic1,10] -> List(2, 4),
> > [topic1,7] -> List(2, 4)), 3 -> Map([topic1,33] -> List(3, 5),
> [topic1,30]
> > -> List(3, 5), [topic1,24] -> List(3, 5), [topic1,36] -> List(3, 5),
> > [topic1,38] -> List(3, 5), [topic1,26] -> List(3, 5), [topic1,27] ->
> > List(3, 5), [topic1,39] -> List(3, 5), [topic1,29] -> List(3, 5),
> > [topic1,34] -> List(3, 5), [topic1,28] -> List(3, 5), [topic1,32] ->
> > List(3, 5), [topic1,35] -> List(3, 5), [topic1,25] -> List(3, 5),
> > [topic1,31] -> List(3, 5), [topic1,37] -> List(3, 5)), 4 ->
> Map([topic1,53]
> > -> List(4, 2), [topic1,56] -> List(4, 2), [topic1,49] -> List(4, 2),
> > [topic1,50] -> List(4, 2), [topic1,51] -> List(4, 2), [topic1,58] ->
> > List(4, 2), [topic1,63] -> List(4, 2), [topic1,54] -> List(4, 2),
> > [topic1,48] -> List(4, 2), [topic1,61] -> List(4, 2), [topic1,62] ->
> > List(4, 2), [topic1,57] -> List(4, 2), [topic1,60] -> List(4, 2),
> > [topic1,52] -> List(4, 2), [topic1,55] -> List(4, 2), [topic1,59] ->
> > List(4, 2))) (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,294] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,294] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 0 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,294] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,294] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 5 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,294] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,294] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 1 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,295] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 2 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,295] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 3 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:54:50,295] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 4 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:55:32,783] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:55:32,894] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,36],
> > [topic1,30], [topic1,31], [topic1,86], [topic1,78], [topic1,74],
> > [topic1,82], [topic1,33]) (kafka.controller.
> IsrChangeNotificationListener)
> > [2016-08-17 12:55:32,896] WARN [Controller-2-to-broker-2-send-thread],
> > Controller 2 epoch 2 fails to send request {controller_id=2,controller_
> > epoch=2,partition_states=[{topic=topic1,partition=82,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=30,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=78,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=12,replicas=[5,3]},{topic=topic1,partition=86,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=31,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=20,replicas=[3,5]},{topic=topic1,partition=36,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=74,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=33,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]}],live_brokers=[{id=5,end_
> > points=[{port=9092,host=b5.kafka,security_protocol_type=
> > 0}]},{id=3,end_points=[{port=9092,host=b3.kafka,security_
> > protocol_type=0}]},{id=2,end_points=[{port=9092,host=b2.
> > kafka,security_protocol_type=0}]},{id=1,end_points=[{port=
> > 9092,host=b1.kafka,security_protocol_type=0}]},{id=4,end_
> > points=[{port=9092,host=b4.kafka,security_protocol_type=
> > 0}]},{id=0,end_points=[{port=9092,host=b0.kafka,security_
> > protocol_type=0}]}]}
> > to broker Node(2, b2.kafka, 9092). Reconnecting to broker.
> > (kafka.controller.RequestSendThread)
> > java.io.IOException: Connection to 2 was disconnected before the response
> > was read
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:87)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:84)
> >         at scala.Option.foreach(Option.scala:257)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:84)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:80)
> >         at kafka.utils.NetworkClientBlockingOps$.recurse$1(
> > NetworkClientBlockingOps.scala:129)
> >         at kafka.utils.NetworkClientBlockingOps$.kafka$utils$
> > NetworkClientBlockingOps$$pollUntilFound$extension(
> > NetworkClientBlockingOps.
> > scala:139)
> >         at kafka.utils.NetworkClientBlockingOps$.blockingSendAndReceive$
> > extension(NetworkClientBlockingOps.scala:80)
> >         at kafka.controller.RequestSendThread.liftedTree1$
> > 1(ControllerChannelManager.scala:180)
> >         at kafka.controller.RequestSendThread.doWork(
> > ControllerChannelManager.scala:171)
> >         at kafka.utils.ShutdownableThread.run(
> ShutdownableThread.scala:63)
> > [2016-08-17 12:55:32,897] WARN [Controller-2-to-broker-5-send-thread],
> > Controller 2 epoch 2 fails to send request {controller_id=2,controller_
> > epoch=2,partition_states=[{topic=topic1,partition=82,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=30,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=78,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=12,replicas=[5,3]},{topic=topic1,partition=86,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=31,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=20,replicas=[3,5]},{topic=topic1,partition=36,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=74,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=33,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]}],live_brokers=[{id=0,end_
> > points=[{port=9092,host=b0.kafka,security_protocol_type=
> > 0}]},{id=5,end_points=[{port=9092,host=b5.kafka,security_
> > protocol_type=0}]},{id=3,end_points=[{port=9092,host=b3.
> > kafka,security_protocol_type=0}]},{id=1,end_points=[{port=
> > 9092,host=b1.kafka,security_protocol_type=0}]},{id=2,end_
> > points=[{port=9092,host=b2.kafka,security_protocol_type=
> > 0}]},{id=4,end_points=[{port=9092,host=b4.kafka,security_
> > protocol_type=0}]}]}
> > to broker Node(5, b5.kafka, 9092). Reconnecting to broker.
> > (kafka.controller.RequestSendThread)
> > java.io.IOException: Connection to 5 was disconnected before the response
> > was read
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:87)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:84)
> >         at scala.Option.foreach(Option.scala:257)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:84)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:80)
> >         at kafka.utils.NetworkClientBlockingOps$.recurse$1(
> > NetworkClientBlockingOps.scala:129)
> >         at kafka.utils.NetworkClientBlockingOps$.kafka$utils$
> > NetworkClientBlockingOps$$pollUntilFound$extension(
> > NetworkClientBlockingOps.
> > scala:139)
> >         at kafka.utils.NetworkClientBlockingOps$.blockingSendAndReceive$
> > extension(NetworkClientBlockingOps.scala:80)
> >         at kafka.controller.RequestSendThread.liftedTree1$
> > 1(ControllerChannelManager.scala:180)
> >         at kafka.controller.RequestSendThread.doWork(
> > ControllerChannelManager.scala:171)
> >         at kafka.utils.ShutdownableThread.run(
> ShutdownableThread.scala:63)
> > [2016-08-17 12:55:32,898] WARN [Controller-2-to-broker-4-send-thread],
> > Controller 2 epoch 2 fails to send request {controller_id=2,controller_
> > epoch=2,partition_states=[{topic=topic1,partition=82,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=30,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=78,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=12,replicas=[5,3]},{topic=topic1,partition=86,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=31,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=20,replicas=[3,5]},{topic=topic1,partition=36,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=74,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=33,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]}],live_brokers=[{id=3,end_
> > points=[{port=9092,host=b3.kafka,security_protocol_type=
> > 0}]},{id=1,end_points=[{port=9092,host=b1.kafka,security_
> > protocol_type=0}]},{id=4,end_points=[{port=9092,host=b4.
> > kafka,security_protocol_type=0}]},{id=2,end_points=[{port=
> > 9092,host=b2.kafka,security_protocol_type=0}]},{id=0,end_
> > points=[{port=9092,host=b0.kafka,security_protocol_type=
> > 0}]},{id=5,end_points=[{port=9092,host=b5.kafka,security_
> > protocol_type=0}]}]}
> > to broker Node(4, b4.kafka, 9092). Reconnecting to broker.
> > (kafka.controller.RequestSendThread)
> > java.io.IOException: Connection to 4 was disconnected before the response
> > was read
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:87)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:84)
> >         at scala.Option.foreach(Option.scala:257)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:84)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:80)
> >         at kafka.utils.NetworkClientBlockingOps$.recurse$1(
> > NetworkClientBlockingOps.scala:129)
> >         at kafka.utils.NetworkClientBlockingOps$.kafka$utils$
> > NetworkClientBlockingOps$$pollUntilFound$extension(
> > NetworkClientBlockingOps.
> > scala:139)
> >         at kafka.utils.NetworkClientBlockingOps$.blockingSendAndReceive$
> > extension(NetworkClientBlockingOps.scala:80)
> >         at kafka.controller.RequestSendThread.liftedTree1$
> > 1(ControllerChannelManager.scala:180)
> >         at kafka.controller.RequestSendThread.doWork(
> > ControllerChannelManager.scala:171)
> >         at kafka.utils.ShutdownableThread.run(
> ShutdownableThread.scala:63)
> > [2016-08-17 12:55:32,900] WARN [Controller-2-to-broker-1-send-thread],
> > Controller 2 epoch 2 fails to send request {controller_id=2,controller_
> > epoch=2,partition_states=[{topic=topic1,partition=82,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=30,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=78,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=12,replicas=[5,3]},{topic=topic1,partition=86,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=31,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=20,replicas=[3,5]},{topic=topic1,partition=36,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=74,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=33,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]}],live_brokers=[{id=5,end_
> > points=[{port=9092,host=b5.kafka,security_protocol_type=
> > 0}]},{id=0,end_points=[{port=9092,host=b0.kafka,security_
> > protocol_type=0}]},{id=3,end_points=[{port=9092,host=b3.
> > kafka,security_protocol_type=0}]},{id=1,end_points=[{port=
> > 9092,host=b1.kafka,security_protocol_type=0}]},{id=4,end_
> > points=[{port=9092,host=b4.kafka,security_protocol_type=
> > 0}]},{id=2,end_points=[{port=9092,host=b2.kafka,security_
> > protocol_type=0}]}]}
> > to broker Node(1, b1.kafka, 9092). Reconnecting to broker.
> > (kafka.controller.RequestSendThread)
> > java.io.IOException: Connection to 1 was disconnected before the response
> > was read
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:87)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:84)
> >         at scala.Option.foreach(Option.scala:257)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:84)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:80)
> >         at kafka.utils.NetworkClientBlockingOps$.recurse$1(
> > NetworkClientBlockingOps.scala:129)
> >         at kafka.utils.NetworkClientBlockingOps$.kafka$utils$
> > NetworkClientBlockingOps$$pollUntilFound$extension(
> > NetworkClientBlockingOps.
> > scala:139)
> >         at kafka.utils.NetworkClientBlockingOps$.blockingSendAndReceive$
> > extension(NetworkClientBlockingOps.scala:80)
> >         at kafka.controller.RequestSendThread.liftedTree1$
> > 1(ControllerChannelManager.scala:180)
> >         at kafka.controller.RequestSendThread.doWork(
> > ControllerChannelManager.scala:171)
> >         at kafka.utils.ShutdownableThread.run(
> ShutdownableThread.scala:63)
> > [2016-08-17 12:55:32,902] WARN [Controller-2-to-broker-3-send-thread],
> > Controller 2 epoch 2 fails to send request {controller_id=2,controller_
> > epoch=2,partition_states=[{topic=topic1,partition=82,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=30,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=78,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=12,replicas=[5,3]},{topic=topic1,partition=86,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=31,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=20,replicas=[3,5]},{topic=topic1,partition=36,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=74,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=33,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]}],live_brokers=[{id=3,end_
> > points=[{port=9092,host=b3.kafka,security_protocol_type=
> > 0}]},{id=0,end_points=[{port=9092,host=b0.kafka,security_
> > protocol_type=0}]},{id=5,end_points=[{port=9092,host=b5.
> > kafka,security_protocol_type=0}]},{id=2,end_points=[{port=
> > 9092,host=b2.kafka,security_protocol_type=0}]},{id=1,end_
> > points=[{port=9092,host=b1.kafka,security_protocol_type=
> > 0}]},{id=4,end_points=[{port=9092,host=b4.kafka,security_
> > protocol_type=0}]}]}
> > to broker Node(3, b3.kafka, 9092). Reconnecting to broker.
> > (kafka.controller.RequestSendThread)
> > java.io.IOException: Connection to 3 was disconnected before the response
> > was read
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:87)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:84)
> >         at scala.Option.foreach(Option.scala:257)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:84)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:80)
> >         at kafka.utils.NetworkClientBlockingOps$.recurse$1(
> > NetworkClientBlockingOps.scala:129)
> >         at kafka.utils.NetworkClientBlockingOps$.kafka$utils$
> > NetworkClientBlockingOps$$pollUntilFound$extension(
> > NetworkClientBlockingOps.
> > scala:139)
> >         at kafka.utils.NetworkClientBlockingOps$.blockingSendAndReceive$
> > extension(NetworkClientBlockingOps.scala:80)
> >         at kafka.controller.RequestSendThread.liftedTree1$
> > 1(ControllerChannelManager.scala:180)
> >         at kafka.controller.RequestSendThread.doWork(
> > ControllerChannelManager.scala:171)
> >         at kafka.utils.ShutdownableThread.run(
> ShutdownableThread.scala:63)
> > [2016-08-17 12:55:32,903] WARN [Controller-2-to-broker-0-send-thread],
> > Controller 2 epoch 2 fails to send request {controller_id=2,controller_
> > epoch=2,partition_states=[{topic=topic1,partition=82,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=30,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=78,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=12,replicas=[5,3]},{topic=topic1,partition=86,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=31,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=20,replicas=[3,5]},{topic=topic1,partition=36,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]},{topic=topic1,partition=74,
> > controller_epoch=2,leader=5,leader_epoch=4,isr=[5,3],zk_
> > version=14,replicas=[5,3]},{topic=topic1,partition=33,
> > controller_epoch=2,leader=3,leader_epoch=2,isr=[3,5],zk_
> > version=18,replicas=[3,5]}],live_brokers=[{id=4,end_
> > points=[{port=9092,host=b4.kafka,security_protocol_type=
> > 0}]},{id=3,end_points=[{port=9092,host=b3.kafka,security_
> > protocol_type=0}]},{id=1,end_points=[{port=9092,host=b1.
> > kafka,security_protocol_type=0}]},{id=2,end_points=[{port=
> > 9092,host=b2.kafka,security_protocol_type=0}]},{id=5,end_
> > points=[{port=9092,host=b5.kafka,security_protocol_type=
> > 0}]},{id=0,end_points=[{port=9092,host=b0.kafka,security_
> > protocol_type=0}]}]}
> > to broker Node(0, b0.kafka, 9092). Reconnecting to broker.
> > (kafka.controller.RequestSendThread)
> > java.io.IOException: Connection to 0 was disconnected before the response
> > was read
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:87)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1$$anonfun$apply$1.apply(
> > NetworkClientBlockingOps.scala:84)
> >         at scala.Option.foreach(Option.scala:257)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:84)
> >         at kafka.utils.NetworkClientBlockingOps$$anonfun$
> > blockingSendAndReceive$extension$1.apply(NetworkClientBlockingOps.
> > scala:80)
> >         at kafka.utils.NetworkClientBlockingOps$.recurse$1(
> > NetworkClientBlockingOps.scala:129)
> >         at kafka.utils.NetworkClientBlockingOps$.kafka$utils$
> > NetworkClientBlockingOps$$pollUntilFound$extension(
> > NetworkClientBlockingOps.
> > scala:139)
> >         at kafka.utils.NetworkClientBlockingOps$.blockingSendAndReceive$
> > extension(NetworkClientBlockingOps.scala:80)
> >         at kafka.controller.RequestSendThread.liftedTree1$
> > 1(ControllerChannelManager.scala:180)
> >         at kafka.controller.RequestSendThread.doWork(
> > ControllerChannelManager.scala:171)
> >         at kafka.utils.ShutdownableThread.run(
> ShutdownableThread.scala:63)
> > [2016-08-17 12:55:32,927] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:55:33,162] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:55:33,169] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,50])
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:55:33,194] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:55:33,198] INFO [Controller-2-to-broker-2-send-thread],
> > Controller 2 connected to Node(2, b2.kafka, 9092) for sending state
> change
> > requests (kafka.controller.RequestSendThread)
> > [2016-08-17 12:55:33,199] INFO [Controller-2-to-broker-5-send-thread],
> > Controller 2 connected to Node(5, b5.kafka, 9092) for sending state
> change
> > requests (kafka.controller.RequestSendThread)
> > [2016-08-17 12:55:33,200] INFO [Controller-2-to-broker-4-send-thread],
> > Controller 2 connected to Node(4, b4.kafka, 9092) for sending state
> change
> > requests (kafka.controller.RequestSendThread)
> > [2016-08-17 12:55:33,202] INFO [Controller-2-to-broker-1-send-thread],
> > Controller 2 connected to Node(1, b1.kafka, 9092) for sending state
> change
> > requests (kafka.controller.RequestSendThread)
> > [2016-08-17 12:55:33,204] INFO [Controller-2-to-broker-0-send-thread],
> > Controller 2 connected to Node(0, b0.kafka, 9092) for sending state
> change
> > requests (kafka.controller.RequestSendThread)
> > [2016-08-17 12:55:33,207] INFO [Controller-2-to-broker-3-send-thread],
> > Controller 2 connected to Node(3, b3.kafka, 9092) for sending state
> change
> > requests (kafka.controller.RequestSendThread)
> > [2016-08-17 12:55:39,981] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:55:40,018] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,34],
> > [topic1,30], [topic1,31], [topic1,25], [topic1,29], [topic1,38],
> > [topic1,35], [topic1,33]) (kafka.controller.
> IsrChangeNotificationListener)
> > [2016-08-17 12:55:40,377] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:55:40,388] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,86],
> > [topic1,78], [topic1,82]) (kafka.controller.
> IsrChangeNotificationListener)
> > [2016-08-17 12:55:40,409] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 12:59:50,293] TRACE [Controller 2]: checking need to trigger
> > partition rebalance (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] DEBUG [Controller 2]: preferred replicas by
> > broker Map(0 -> Map([topic1,45] -> List(0, 1), [topic1,17] -> List(0, 1),
> > [topic1,19] -> List(0, 1), [topic1,42] -> List(0, 1), [topic1,43] ->
> > List(0, 1), [topic1,44] -> List(0, 1), [topic1,16] -> List(0, 1),
> > [topic1,46] -> List(0, 1), [topic1,20] -> List(0, 1), [topic1,41] ->
> > List(0, 1), [topic1,18] -> List(0, 1), [topic1,22] -> List(0, 1),
> > [topic1,40] -> List(0, 1), [topic1,47] -> List(0, 1), [topic1,23] ->
> > List(0, 1), [topic1,21] -> List(0, 1)), 5 -> Map([topic1,78] -> List(5,
> 3),
> > [topic1,84] -> List(5, 3), [topic1,87] -> List(5, 3), [topic1,74] ->
> > List(5, 3), [topic1,81] -> List(5, 3), [topic1,73] -> List(5, 3),
> > [topic1,80] -> List(5, 3), [topic1,77] -> List(5, 3), [topic1,75] ->
> > List(5, 3), [topic1,85] -> List(5, 3), [topic1,76] -> List(5, 3),
> > [topic1,83] -> List(5, 3), [topic1,86] -> List(5, 3), [topic1,72] ->
> > List(5, 3), [topic1,79] -> List(5, 3), [topic1,82] -> List(5, 3)), 1 ->
> > Map([topic1,92] -> List(1, 0), [topic1,95] -> List(1, 0), [topic1,69] ->
> > List(1, 0), [topic1,93] -> List(1, 0), [topic1,70] -> List(1, 0),
> > [topic1,67] -> List(1, 0), [topic1,65] -> List(1, 0), [topic1,88] ->
> > List(1, 0), [topic1,90] -> List(1, 0), [topic1,66] -> List(1, 0),
> > [topic1,94] -> List(1, 0), [topic1,64] -> List(1, 0), [topic1,89] ->
> > List(1, 0), [topic1,68] -> List(1, 0), [topic1,71] -> List(1, 0),
> > [topic1,91] -> List(1, 0)), 2 -> Map([topic1,8] -> List(2, 4), [topic1,3]
> > -> List(2, 4), [topic1,15] -> List(2, 4), [topic1,2] -> List(2, 4),
> > [topic1,1] -> List(2, 4), [topic1,6] -> List(2, 4), [topic1,9] -> List(2,
> > 4), [topic1,12] -> List(2, 4), [topic1,14] -> List(2, 4), [topic1,11] ->
> > List(2, 4), [topic1,13] -> List(2, 4), [topic1,0] -> List(2, 4),
> [topic1,4]
> > -> List(2, 4), [topic1,5] -> List(2, 4), [topic1,10] -> List(2, 4),
> > [topic1,7] -> List(2, 4)), 3 -> Map([topic1,33] -> List(3, 5),
> [topic1,30]
> > -> List(3, 5), [topic1,24] -> List(3, 5), [topic1,36] -> List(3, 5),
> > [topic1,38] -> List(3, 5), [topic1,26] -> List(3, 5), [topic1,27] ->
> > List(3, 5), [topic1,39] -> List(3, 5), [topic1,29] -> List(3, 5),
> > [topic1,34] -> List(3, 5), [topic1,28] -> List(3, 5), [topic1,32] ->
> > List(3, 5), [topic1,35] -> List(3, 5), [topic1,25] -> List(3, 5),
> > [topic1,31] -> List(3, 5), [topic1,37] -> List(3, 5)), 4 ->
> Map([topic1,53]
> > -> List(4, 2), [topic1,56] -> List(4, 2), [topic1,49] -> List(4, 2),
> > [topic1,50] -> List(4, 2), [topic1,51] -> List(4, 2), [topic1,58] ->
> > List(4, 2), [topic1,63] -> List(4, 2), [topic1,54] -> List(4, 2),
> > [topic1,48] -> List(4, 2), [topic1,61] -> List(4, 2), [topic1,62] ->
> > List(4, 2), [topic1,57] -> List(4, 2), [topic1,60] -> List(4, 2),
> > [topic1,52] -> List(4, 2), [topic1,55] -> List(4, 2), [topic1,59] ->
> > List(4, 2))) (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 0 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 5 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 1 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,294] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 2 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,295] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 3 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 12:59:50,295] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 4 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:00:39,546] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:39,604] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,5])
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:39,649] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:39,888] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,071] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,37],
> > [topic1,27], [topic1,34], [topic1,32], [topic1,24], [topic1,39],
> > [topic1,36], [topic1,30], [topic1,31], [topic1,25], [topic1,29],
> > [topic1,38], [topic1,26], [topic1,35], [topic1,33], [topic1,28])
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,103] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,261] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,283] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,72],
> > [topic1,80]) (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,296] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,656] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,662] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,55])
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:40,934] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:47,335] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:47,393] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,37],
> > [topic1,27], [topic1,34], [topic1,32], [topic1,24], [topic1,39],
> > [topic1,36], [topic1,30], [topic1,31], [topic1,25], [topic1,29],
> > [topic1,38], [topic1,26], [topic1,35], [topic1,33], [topic1,28])
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:47,423] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:47,897] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:00:47,944] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,5],
> > [topic1,3], [topic1,7], [topic1,11], [topic1,2], [topic1,6], [topic1,1],
> > [topic1,10], [topic1,14], [topic1,9], [topic1,15]) (kafka.controller.
> > IsrChangeNotificationListener)
> > [2016-08-17 13:00:48,020] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:04:50,293] TRACE [Controller 2]: checking need to trigger
> > partition rebalance (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,295] DEBUG [Controller 2]: preferred replicas by
> > broker Map(0 -> Map([topic1,45] -> List(0, 1), [topic1,17] -> List(0, 1),
> > [topic1,19] -> List(0, 1), [topic1,42] -> List(0, 1), [topic1,43] ->
> > List(0, 1), [topic1,44] -> List(0, 1), [topic1,16] -> List(0, 1),
> > [topic1,46] -> List(0, 1), [topic1,20] -> List(0, 1), [topic1,41] ->
> > List(0, 1), [topic1,18] -> List(0, 1), [topic1,22] -> List(0, 1),
> > [topic1,40] -> List(0, 1), [topic1,47] -> List(0, 1), [topic1,23] ->
> > List(0, 1), [topic1,21] -> List(0, 1)), 5 -> Map([topic1,78] -> List(5,
> 3),
> > [topic1,84] -> List(5, 3), [topic1,87] -> List(5, 3), [topic1,74] ->
> > List(5, 3), [topic1,81] -> List(5, 3), [topic1,73] -> List(5, 3),
> > [topic1,80] -> List(5, 3), [topic1,77] -> List(5, 3), [topic1,75] ->
> > List(5, 3), [topic1,85] -> List(5, 3), [topic1,76] -> List(5, 3),
> > [topic1,83] -> List(5, 3), [topic1,86] -> List(5, 3), [topic1,72] ->
> > List(5, 3), [topic1,79] -> List(5, 3), [topic1,82] -> List(5, 3)), 1 ->
> > Map([topic1,92] -> List(1, 0), [topic1,95] -> List(1, 0), [topic1,69] ->
> > List(1, 0), [topic1,93] -> List(1, 0), [topic1,70] -> List(1, 0),
> > [topic1,67] -> List(1, 0), [topic1,65] -> List(1, 0), [topic1,88] ->
> > List(1, 0), [topic1,90] -> List(1, 0), [topic1,66] -> List(1, 0),
> > [topic1,94] -> List(1, 0), [topic1,64] -> List(1, 0), [topic1,89] ->
> > List(1, 0), [topic1,68] -> List(1, 0), [topic1,71] -> List(1, 0),
> > [topic1,91] -> List(1, 0)), 2 -> Map([topic1,8] -> List(2, 4), [topic1,3]
> > -> List(2, 4), [topic1,15] -> List(2, 4), [topic1,2] -> List(2, 4),
> > [topic1,1] -> List(2, 4), [topic1,6] -> List(2, 4), [topic1,9] -> List(2,
> > 4), [topic1,12] -> List(2, 4), [topic1,14] -> List(2, 4), [topic1,11] ->
> > List(2, 4), [topic1,13] -> List(2, 4), [topic1,0] -> List(2, 4),
> [topic1,4]
> > -> List(2, 4), [topic1,5] -> List(2, 4), [topic1,10] -> List(2, 4),
> > [topic1,7] -> List(2, 4)), 3 -> Map([topic1,33] -> List(3, 5),
> [topic1,30]
> > -> List(3, 5), [topic1,24] -> List(3, 5), [topic1,36] -> List(3, 5),
> > [topic1,38] -> List(3, 5), [topic1,26] -> List(3, 5), [topic1,27] ->
> > List(3, 5), [topic1,39] -> List(3, 5), [topic1,29] -> List(3, 5),
> > [topic1,34] -> List(3, 5), [topic1,28] -> List(3, 5), [topic1,32] ->
> > List(3, 5), [topic1,35] -> List(3, 5), [topic1,25] -> List(3, 5),
> > [topic1,31] -> List(3, 5), [topic1,37] -> List(3, 5)), 4 ->
> Map([topic1,53]
> > -> List(4, 2), [topic1,56] -> List(4, 2), [topic1,49] -> List(4, 2),
> > [topic1,50] -> List(4, 2), [topic1,51] -> List(4, 2), [topic1,58] ->
> > List(4, 2), [topic1,63] -> List(4, 2), [topic1,54] -> List(4, 2),
> > [topic1,48] -> List(4, 2), [topic1,61] -> List(4, 2), [topic1,62] ->
> > List(4, 2), [topic1,57] -> List(4, 2), [topic1,60] -> List(4, 2),
> > [topic1,52] -> List(4, 2), [topic1,55] -> List(4, 2), [topic1,59] ->
> > List(4, 2))) (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,295] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 0 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 5 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 1 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 2 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 3 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:04:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 4 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:05:34,317] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:34,365] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,80],
> > [topic1,40], [topic1,21], [topic1,31], [topic1,84], [topic1,33])
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:34,388] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:36,426] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:36,437] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,92])
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:36,699] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:40,225] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:40,239] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,80],
> > [topic1,84]) (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:40,246] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:40,958] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:41,006] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,22],
> > [topic1,16], [topic1,20], [topic1,19], [topic1,40], [topic1,21],
> > [topic1,18], [topic1,47], [topic1,44], [topic1,45], [topic1,42],
> > [topic1,46], [topic1,43], [topic1,23]) (kafka.controller.
> > IsrChangeNotificationListener)
> > [2016-08-17 13:05:41,067] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:42,517] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:05:42,622] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,37],
> > [topic1,27], [topic1,34], [topic1,32], [topic1,24], [topic1,39],
> > [topic1,30], [topic1,31], [topic1,25], [topic1,29], [topic1,38],
> > [topic1,26], [topic1,35], [topic1,33], [topic1,28]) (kafka.controller.
> > IsrChangeNotificationListener)
> > [2016-08-17 13:05:42,690] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:09:50,293] TRACE [Controller 2]: checking need to trigger
> > partition rebalance (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,295] DEBUG [Controller 2]: preferred replicas by
> > broker Map(0 -> Map([topic1,45] -> List(0, 1), [topic1,17] -> List(0, 1),
> > [topic1,19] -> List(0, 1), [topic1,42] -> List(0, 1), [topic1,43] ->
> > List(0, 1), [topic1,44] -> List(0, 1), [topic1,16] -> List(0, 1),
> > [topic1,46] -> List(0, 1), [topic1,20] -> List(0, 1), [topic1,41] ->
> > List(0, 1), [topic1,18] -> List(0, 1), [topic1,22] -> List(0, 1),
> > [topic1,40] -> List(0, 1), [topic1,47] -> List(0, 1), [topic1,23] ->
> > List(0, 1), [topic1,21] -> List(0, 1)), 5 -> Map([topic1,78] -> List(5,
> 3),
> > [topic1,84] -> List(5, 3), [topic1,87] -> List(5, 3), [topic1,74] ->
> > List(5, 3), [topic1,81] -> List(5, 3), [topic1,73] -> List(5, 3),
> > [topic1,80] -> List(5, 3), [topic1,77] -> List(5, 3), [topic1,75] ->
> > List(5, 3), [topic1,85] -> List(5, 3), [topic1,76] -> List(5, 3),
> > [topic1,83] -> List(5, 3), [topic1,86] -> List(5, 3), [topic1,72] ->
> > List(5, 3), [topic1,79] -> List(5, 3), [topic1,82] -> List(5, 3)), 1 ->
> > Map([topic1,92] -> List(1, 0), [topic1,95] -> List(1, 0), [topic1,69] ->
> > List(1, 0), [topic1,93] -> List(1, 0), [topic1,70] -> List(1, 0),
> > [topic1,67] -> List(1, 0), [topic1,65] -> List(1, 0), [topic1,88] ->
> > List(1, 0), [topic1,90] -> List(1, 0), [topic1,66] -> List(1, 0),
> > [topic1,94] -> List(1, 0), [topic1,64] -> List(1, 0), [topic1,89] ->
> > List(1, 0), [topic1,68] -> List(1, 0), [topic1,71] -> List(1, 0),
> > [topic1,91] -> List(1, 0)), 2 -> Map([topic1,8] -> List(2, 4), [topic1,3]
> > -> List(2, 4), [topic1,15] -> List(2, 4), [topic1,2] -> List(2, 4),
> > [topic1,1] -> List(2, 4), [topic1,6] -> List(2, 4), [topic1,9] -> List(2,
> > 4), [topic1,12] -> List(2, 4), [topic1,14] -> List(2, 4), [topic1,11] ->
> > List(2, 4), [topic1,13] -> List(2, 4), [topic1,0] -> List(2, 4),
> [topic1,4]
> > -> List(2, 4), [topic1,5] -> List(2, 4), [topic1,10] -> List(2, 4),
> > [topic1,7] -> List(2, 4)), 3 -> Map([topic1,33] -> List(3, 5),
> [topic1,30]
> > -> List(3, 5), [topic1,24] -> List(3, 5), [topic1,36] -> List(3, 5),
> > [topic1,38] -> List(3, 5), [topic1,26] -> List(3, 5), [topic1,27] ->
> > List(3, 5), [topic1,39] -> List(3, 5), [topic1,29] -> List(3, 5),
> > [topic1,34] -> List(3, 5), [topic1,28] -> List(3, 5), [topic1,32] ->
> > List(3, 5), [topic1,35] -> List(3, 5), [topic1,25] -> List(3, 5),
> > [topic1,31] -> List(3, 5), [topic1,37] -> List(3, 5)), 4 ->
> Map([topic1,53]
> > -> List(4, 2), [topic1,56] -> List(4, 2), [topic1,49] -> List(4, 2),
> > [topic1,50] -> List(4, 2), [topic1,51] -> List(4, 2), [topic1,58] ->
> > List(4, 2), [topic1,63] -> List(4, 2), [topic1,54] -> List(4, 2),
> > [topic1,48] -> List(4, 2), [topic1,61] -> List(4, 2), [topic1,62] ->
> > List(4, 2), [topic1,57] -> List(4, 2), [topic1,60] -> List(4, 2),
> > [topic1,52] -> List(4, 2), [topic1,55] -> List(4, 2), [topic1,59] ->
> > List(4, 2))) (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,295] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 0 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,295] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 5 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 1 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 2 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 3 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,296] DEBUG [Controller 2]: topics not in preferred
> > replica Map() (kafka.controller.KafkaController)
> > [2016-08-17 13:09:50,297] TRACE [Controller 2]: leader imbalance ratio
> for
> > broker 4 is 0.000000 (kafka.controller.KafkaController)
> > [2016-08-17 13:10:37,278] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:10:37,292] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,67],
> > [topic1,95]) (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:10:37,304] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:10:43,375] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:10:43,383] DEBUG Sending MetadataRequest to
> > Brokers:ArrayBuffer(0, 5, 1, 2, 3, 4) for TopicAndPartitions:Set([
> > topic1,67],
> > [topic1,95]) (kafka.controller.IsrChangeNotificationListener)
> > [2016-08-17 13:10:43,394] DEBUG [IsrChangeNotificationListener] Fired!!!
> > (kafka.controller.IsrChangeNotificationListener)
> >
> > Thanks
> >
> > Regards,
> > Mazhar Shaikh.
> >
> >
> >
> > On Wed, Aug 17, 2016 at 9:50 PM, Jun Rao <j...@confluent.io> wrote:
> >
> > > Yes, you can try setting it to -1 in 0.8.1, which is the equivalent of
> > > "all" in 0.9 and above.
> > >
> > > Thanks,
> > >
> > > Jun
> > >
> > > On Wed, Aug 17, 2016 at 8:32 AM, Mazhar Shaikh <
> > mazhar.shaikh...@gmail.com
> > > >
> > > wrote:
> > >
> > > > Hi Jun,
> > > >
> > > > I'm using default configuration (ack=1),
> > > > changing it t0 all or 2 will not help, as the producer queue will be
> > > > exhausted is any kafka broker goes down for long time.
> > > >
> > > >
> > > > Thanks.
> > > >
> > > > Regards,
> > > > Mazhar Shaikh.
> > > >
> > > >
> > > > On Wed, Aug 17, 2016 at 8:11 PM, Jun Rao <j...@confluent.io> wrote:
> > > >
> > > > > Are you using acks=1 or acks=all in the producer? Only the latter
> > > > > guarantees acked messages won't be lost after leader failure.
> > > > >
> > > > > Thanks,
> > > > >
> > > > > Jun
> > > > >
> > > > > On Wed, Aug 10, 2016 at 11:41 PM, Mazhar Shaikh <
> > > > > mazhar.shaikh...@gmail.com>
> > > > > wrote:
> > > > >
> > > > > > Hi Kafka Team,
> > > > > >
> > > > > > I'm using kafka (kafka_2.11-0.9.0.1) with librdkafka (0.8.1) API
> > for
> > > > > > producer
> > > > > > During a run of 2hrs, I notice the total number of messaged ack'd
> > by
> > > > > > librdkafka delivery report is greater than the maxoffset of a
> > > partition
> > > > > in
> > > > > > kafka broker.
> > > > > > I'm running kafka broker with replication factor of 2.
> > > > > >
> > > > > > Here, message has been lost between librdkafka - kafka broker.
> > > > > >
> > > > > > As librdkafka is providing success delivery report for all the
> > > > messages.
> > > > > >
> > > > > > Looks like kafka broker is dropping the messages after
> > acknowledging
> > > > > > librdkafka.
> > > > > >
> > > > > > Requesting you help in solving this issue.
> > > > > >
> > > > > > Thank you.
> > > > > >
> > > > > >
> > > > > > Regards
> > > > > > Mazhar Shaikh
> > > > > >
> > > > >
> > > >
> > >
> >
>

Reply via email to