Tom Schamberger created FLINK-11186: ---------------------------------------
Summary: Support for event-time balancing for multiple Kafka comsumer partitions Key: FLINK-11186 URL: https://issues.apache.org/jira/browse/FLINK-11186 Project: Flink Issue Type: New Feature Components: DataStream API, Kafka Connector Reporter: Tom Schamberger Currently, it is not possible with Flink to back-pressure individual Kafka partitions, which are faster in terms of event-time. This leads to unnecessary memory consumption and can lead to deadlocks in the case of back-pressure. When multiple Kafka topics are consumed, succeeding event-time window operators have to wait until the last Kafka partition has produced a sufficient watermark to be triggered. If individual Kafka partitions differ in read performance or the event-time of messages within partitions is not monotonically distributed, this can lead to a situation, where 'fast' partitions (event-time makes fast progress) outperform slower partitions until back-pressuring prevents all partitions from being further consumed. This leads to a deadlock of the application. I suggest, that windows should be able to back-pressure individual partitions, which progress faster in terms of event-time, so that slow partitions can keep up. -- This message was sent by Atlassian JIRA (v7.6.3#76005)