Hi all, I have read the below in the documentation :
"To maximize throughput, set setBufferTimeout(-1) which will remove the timeout and buffers will only be flushed when they are full. To minimize latency, set the timeout to a value close to 0 (for example 5 or 10 ms). A buffer timeout of 0 should be avoided, because it can cause severe performance degradation." why a 0 BufferTimeout cause severe performance degradation, shouldnt it provide min latency, what is meant by perf. degradation there. On the otherhand, can we say that min latency is always > BufferTimeout. Best, -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/