Whoops, I said I'd put the specific exception at the bottom of the
e-mail.  It probably isn't the important part of this thread, but
might suggest when this situation can occur.  Also of note, this is
occurring on Kafka Streams 0.10.2.1.


20:56:07.061 [StreamThread-3] ERROR o.a.k.s.p.internals.StreamThread -
stream-thread [StreamThread-3] Failed to commit StreamTask 0_4 state:
org.apache.kafka.streams.errors.StreamsException: task [0_4] exception
caught when producing
        at 
org.apache.kafka.streams.processor.internals.RecordCollectorImpl.checkForException(RecordCollectorImpl.java:121)
        at 
org.apache.kafka.streams.processor.internals.RecordCollectorImpl.flush(RecordCollectorImpl.java:129)
        at 
org.apache.kafka.streams.processor.internals.StreamTask$1.run(StreamTask.java:76)
        at 
org.apache.kafka.streams.processor.internals.StreamsMetricsImpl.measureLatencyNs(StreamsMetricsImpl.java:188)
        at 
org.apache.kafka.streams.processor.internals.StreamTask.commit(StreamTask.java:280)
        at 
org.apache.kafka.streams.processor.internals.StreamThread.commitOne(StreamThread.java:807)
        at 
org.apache.kafka.streams.processor.internals.StreamThread.commitAll(StreamThread.java:794)
        at 
org.apache.kafka.streams.processor.internals.StreamThread.maybeCommit(StreamThread.java:769)
        at 
org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:647)
        at 
org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:361)
Caused by: org.apache.kafka.common.errors.TimeoutException: Failed to
update metadata after 60000 ms.

On Fri, May 19, 2017 at 9:47 AM, Mathieu Fenniak
<mathieu.fenn...@replicon.com> wrote:
> Hi Kafka devs,
>
> This morning I observed a specific Kafka Streams aggregation that
> ended up with an incorrect computed output after a Kafka Streams
> thread crashed with an unhandled exception.
>
> The topology is pretty simple -- a single KTable source, group by a
> field in the value, aggregate that adds up another field, output to a
> topic.
>
> Here's the sequence of events that appears to have occurred:
>
> 1. A new record (record A) is received by the source KTable, and put
> in the KTable RocksDB state store.
>
> 2. While processing record A, an exception happens preventing
> producing to Kafka. (specific exception at end of e-mail).
>
> 3. The stream thread throws an unhandled exception and stops.
>
> 4. The state stores are closed and flushed.  Record A is now in the
> local state store.
>
> 5. The consumer group rebalances.
>
> 6. A different thread, in the same process, on the same host, picks up the 
> task.
>
> 7. New thread initializes its state store for the KTable, but it's on
> the same host as the original thread, so it still contains the k/v for
> record A.
>
> 8. New thread resumes consuming at the last committed offset, which is
> before record A.
>
> 9. When processing record A, the new thread reads the value that was
> written to the state store in step #1 by record A's key.
>
> 10. The repartition map receives a Change with both an oldValue and a
> newValue, and forwards a Change(null, v) and Change(v, null)
>
> 11. The aggregation ends up both subtracting and adding the value of
> record A, resulting in an incorrect output.
>
> As a result of this sequence, my aggregate output went from a value of
> 0, to negative (subtracting record A), to 0.  And stayed there.
>
> Does this seem like a feasible series of events?  Is this a bug in KS,
> or, is it behavior that maybe can't be improved without exactly-once?
> I'd think the best behavior would be for writes to the RocksDB state
> store to be transactional and only commit when the producer commits,
> but, there's a lot of overhead involved in that.
>
> Mathieu

Reply via email to