Hi, Kafka version 1.0.0 (can't upgrade to another version yet due to legacy dependency)
The stream application uses low level processor API and maintains state. A topic is setup with 30 partitions and I had split to 2 stream application instances consuming the same topic, each with 15 threads. The application starts fine and moves well until REBALANCING occur. When it does, the application takes long time to move to RUNNING status by itself. During this time no exception, no additional logging occurs in the application. 1) Could this behavior be due to an issue on Kafka broker OR is this related to the stream application? 2) What logging can we increase to get additional insight as to what cause this pause state for a significant period of time (this is impacting the throughput)? FYI, we have set the REQUEST TIMEOUT to max integer value to avoid timeout. In the event we have a single application with 30 threads, I don't see this long pause, but that means we need to increase the number of threads and memory, which is vertical scaling and not feasible for handling a topic with significant volume. *Instance 1:* 2018-07-29 01:45:43 INFO StreamStateListener22 - Stream application moved from RUNNING to REBALANCING 2018-07-29 02:15:59 INFO StreamStateListener22 - Stream application moved from REBALANCING to RUNNING 2018-07-29 05:19:18 INFO StreamStateListener22 - Stream application moved from RUNNING to REBALANCING 2018-07-29 05:54:00 INFO StreamStateListener22 - Stream application moved from REBALANCING to RUNNING *Instance 2:* 2018-07-29 01:45:58 INFO StreamStateListener22 - Stream application moved from RUNNING to REBALANCING 2018-07-29 02:41:22 INFO StreamStateListener22 - Stream application moved from REBALANCING to RUNNING 2018-07-29 05:19:33 INFO StreamStateListener22 - Stream application moved from RUNNING to REBALANCING 2018-07-29 05:54:14 INFO StreamStateListener22 - Stream application moved from REBALANCING to RUNNING Thanks, Siva