Hi,

I want to dump events in a kafka topic into datalake as a type 1 snapshot
in iceberg. Type 1 means a record having a key will overwrite the previous
record having the same key. Each key will have only one record in the
snapshot.

Note that I want to simplify the long path:
kafka -> (streaming job) -> type 2 historical -> (spark job) -> type 1
snapshot
TO the short path:
kafka -> (streaming job) -> type 1 snapshot.

The short path requires the streaming job directly update the type 1
snapshot using row level update (e.g. merge into). The streaming job can be
either spark structured streaming (SSS) or flink.

According to https://iceberg.apache.org/spark-structured-streaming/, SSS
does not support "merge into".

According to https://iceberg.apache.org/flink/#insert-into, Flink streaming
job does not support "merge into" or "insert into".

Is this a dead end? Appreciate any hints.

Reply via email to