I'm not aware of any way to control the sink order, afaik each Table#executeInsert will generate a separate job on its own. You may be able to hack it around by having a custom DynamicTableSink that for each record sends it to tidb and then to kafka.
May I ask why you need that? If the notification system after the Kafka sink depends on tidb, perhaps you need a retry system there that can wait for tidb to ingest and process those data? On Thu, Oct 14, 2021 at 10:40 AM WuKong <wukon...@foxmail.com> wrote: > Hi all: > I have two Flink SQL , the same source from Kafka, and one SQL > sink data into Tidb ,another one SQL sink Kafka to notify downstream > system, how can I control the sink order , I wish If source Kafka data > come, first sink Tidb and after that sink Kafka . > > ------------------------------ > --- > Best, > WuKong >