Michael Hornung created KAFKA-13683: ---------------------------------------
Summary: Streams - Transactional Producer - Transaction with key xyz went wrong with exception: Timeout expired after 60000milliseconds while awaiting InitProducerId Key: KAFKA-13683 URL: https://issues.apache.org/jira/browse/KAFKA-13683 Project: Kafka Issue Type: Bug Components: streams Affects Versions: 3.0.0, 2.7.0, 2.6.0 Reporter: Michael Hornung Fix For: 3.0.0, 2.7.0, 2.6.0 Attachments: AkkaHttpRestServer.scala, timeoutException.png We have an urgent issue with our customer using kafka transactional producer with kafka cluster with 3 or more nodes. We are using confluent on azure. We this exception regularly: "Transaction with key XYZ went wrong with exception: Timeout expired after 60000milliseconds while awaiting InitProducerId" (see attachment) We assume that the cause is a node which is down and the producer still sends messages to the “down node”. We are using kafa streams 3.0. *We expect that if a node is down kafka producer is intelligent enough to not send messages to this node any more.* *What’s the solution of this issue? Is there any config we have to set?* *This request is urgent because our costumer will soon have production issues.* *Additional information* * send record --> see attachment “AkkaHttpRestServer.scala” – line 100 * producer config --> see attachment “AkkaHttpRestServer.scala” – line 126 -- This message was sent by Atlassian Jira (v8.20.1#820001)