Thanks. I like that you're moving Kafka toward supporting this dual-write pattern. Each use case needs to consider the tradeoffs. You already summarized the pros very well in the KIP. I would summarize the cons as follows:
- you sacrifice availability - each write requires both DB and Kafka to be available so I think your overall application availability is 1 - p(DB is unavailable)*p(Kafka is unavailable). - latency will be higher and throughput lower - each write requires both writes to DB and Kafka while holding an exclusive lock in DB. - you need to create a producer per unit of concurrency in your app which has some overhead in the app and Kafka side (number of connections, poor batching). I assume the producers would need to be configured for low latency (linger.ms=0) - there's some complexity in managing stable transactional ids for each producer/concurrency unit in your application. With k8s deployment, you may need to switch to something like a StatefulSet that gives each pod a stable identity across restarts. On top of that pod identity which you can use as a prefix, you then assign unique transactional ids to each concurrency unit (thread/goroutine). On Wed, Aug 23, 2023 at 12:53 PM Artem Livshits <alivsh...@confluent.io.invalid> wrote: > Hi Roger, > > Thank you for the feedback. You make a very good point that we also > discussed internally. Adding support for multiple concurrent > transactions in one producer could be valuable but it seems to be a fairly > large and independent change that would deserve a separate KIP. If such > support is added we could modify 2PC functionality to incorporate that. > > > Maybe not too bad but a bit of pain to manage these ids inside each > process and across all application processes. > > I'm not sure if supporting multiple transactions in one producer would make > id management simpler: we'd need to store a piece of data per transaction, > so whether it's N producers with a single transaction or N transactions > with a single producer, it's still roughly the same amount of data to > manage. In fact, managing transactional ids (current proposal) might be > easier, because the id is controlled by the application and it knows how to > complete the transaction after crash / restart; while a TID would be > generated by Kafka and that would create a question of starting Kafka > transaction, but not saving its TID and then crashing, then figuring out > which transactions to abort and etc. > > > 2) creating a separate producer for each concurrency slot in the > application > > This is a very valid concern. Maybe we'd need to have some multiplexing of > transactional logical "streams" over the same connection. Seems like a > separate KIP, though. > > > Otherwise, it seems you're left with single-threaded model per > application process? > > That's a fair assessment. Not necessarily exactly single-threaded per > application, but a single producer per thread model (i.e. an application > could have a pool of threads + producers to increase concurrency). > > -Artem > > On Tue, Aug 22, 2023 at 7:22 PM Roger Hoover <roger.hoo...@gmail.com> > wrote: > > > Artem, > > > > Thanks for the reply. > > > > If I understand correctly, Kafka does not support concurrent transactions > > from the same producer (transactional id). I think this means that > > applications that want to support in-process concurrency (say > thread-level > > concurrency with row-level DB locking) would need to manage separate > > transactional ids and producers per thread and then store txn state > > accordingly. The potential usability downsides I see are > > 1) managing a set of transactional ids for each application process that > > scales up to it's max concurrency. Maybe not too bad but a bit of pain > to > > manage these ids inside each process and across all application > processes. > > 2) creating a separate producer for each concurrency slot in the > > application - this could create a lot more producers and resultant > > connections to Kafka than the typical model of a single producer per > > process. > > > > Otherwise, it seems you're left with single-threaded model per > application > > process? > > > > Thanks, > > > > Roger > > > > On Tue, Aug 22, 2023 at 5:11 PM Artem Livshits > > <alivsh...@confluent.io.invalid> wrote: > > > > > Hi Roger, Arjun, > > > > > > Thank you for the questions. > > > > It looks like the application must have stable transactional ids over > > > time? > > > > > > The transactional id should uniquely identify a producer instance and > > needs > > > to be stable across the restarts. If the transactional id is not > stable > > > across restarts, then zombie messages from a previous incarnation of > the > > > producer may violate atomicity. If there are 2 producer instances > > > concurrently producing data with the same transactional id, they are > > going > > > to constantly fence each other and most likely make little or no > > progress. > > > > > > The name might be a little bit confusing as it may be mistaken for a > > > transaction id / TID that uniquely identifies every transaction. The > > name > > > and the semantics were defined in the original exactly-once-semantics > > (EoS) > > > proposal ( > > > > > > > > > https://cwiki.apache.org/confluence/display/KAFKA/KIP-98+-+Exactly+Once+Delivery+and+Transactional+Messaging > > > ) > > > and KIP-939 just build on top of that. > > > > > > > I'm curious to understand what happens if the producer dies, and does > > not > > > come up and recover the pending transaction within the transaction > > timeout > > > interval. > > > > > > If the producer / application never comes back, the transaction will > > remain > > > in prepared (a.k.a. "in-doubt") state until an operator forcefully > > > terminates the transaction. That's why there is a new ACL is defined > in > > > this proposal -- this functionality should only provided to > applications > > > that implement proper recovery logic. > > > > > > -Artem > > > > > > On Tue, Aug 22, 2023 at 12:52 AM Arjun Satish <arjun.sat...@gmail.com> > > > wrote: > > > > > > > Hello Artem, > > > > > > > > Thanks for the KIP. > > > > > > > > I have the same question as Roger on concurrent writes, and an > > additional > > > > one on consumer behavior. Typically, transactions will timeout if not > > > > committed within some time interval. With the proposed changes in > this > > > KIP, > > > > consumers cannot consume past the ongoing transaction. I'm curious to > > > > understand what happens if the producer dies, and does not come up > and > > > > recover the pending transaction within the transaction timeout > > interval. > > > Or > > > > are we saying that when used in this 2PC context, we should configure > > > these > > > > transaction timeouts to very large durations? > > > > > > > > Thanks in advance! > > > > > > > > Best, > > > > Arjun > > > > > > > > > > > > On Mon, Aug 21, 2023 at 1:06 PM Roger Hoover <roger.hoo...@gmail.com > > > > > > wrote: > > > > > > > > > Hi Artem, > > > > > > > > > > Thanks for writing this KIP. Can you clarify the requirements a > bit > > > more > > > > > for managing transaction state? It looks like the application must > > > have > > > > > stable transactional ids over time? What is the granularity of > > those > > > > ids > > > > > and producers? Say the application is a multi-threaded Java web > > > server, > > > > > can/should all the concurrent threads share a transactional id and > > > > > producer? That doesn't seem right to me unless the application is > > > using > > > > > global DB locks that serialize all requests. Instead, if the > > > application > > > > > uses row-level DB locks, there could be multiple, concurrent, > > > independent > > > > > txns happening in the same JVM so it seems like the granularity > > > managing > > > > > transactional ids and txn state needs to line up with granularity > of > > > the > > > > DB > > > > > locking. > > > > > > > > > > Does that make sense or am I misunderstanding? > > > > > > > > > > Thanks, > > > > > > > > > > Roger > > > > > > > > > > On Wed, Aug 16, 2023 at 11:40 PM Artem Livshits > > > > > <alivsh...@confluent.io.invalid> wrote: > > > > > > > > > > > Hello, > > > > > > > > > > > > This is a discussion thread for > > > > > > > > > > > > > > > > > > > > > > > > > > > https://cwiki.apache.org/confluence/display/KAFKA/KIP-939%3A+Support+Participation+in+2PC > > > > > > . > > > > > > > > > > > > The KIP proposes extending Kafka transaction support (that > already > > > uses > > > > > 2PC > > > > > > under the hood) to enable atomicity of dual writes to Kafka and > an > > > > > external > > > > > > database, and helps to fix a long standing Flink issue. > > > > > > > > > > > > An example of code that uses the dual write recipe with JDBC and > > > should > > > > > > work for most SQL databases is here > > > > > > https://github.com/apache/kafka/pull/14231. > > > > > > > > > > > > The FLIP for the sister fix in Flink is here > > > > > > > > > > > > > > > > > > > > > https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=255071710 > > > > > > > > > > > > -Artem > > > > > > > > > > > > > > > > > > > > >