Re: Flink-1.12 Sql on Job two SQL sink control order

2021-10-18 Thread Arvid Heise
Are you running in Batch? Then you probably need to write 2 SQL jobs (or statements). In streaming, the notion of order doesn't make much sense. But maybe I misunderstood your use case. On Thu, Oct 14, 2021 at 11:37 AM Francesco Guardiani < france...@ververica.com> wrote: > I'm not aware of any

Re: Flink-1.12 Sql on Job two SQL sink control order

2021-10-14 Thread Francesco Guardiani
I'm not aware of any way to control the sink order, afaik each Table#executeInsert will generate a separate job on its own. You may be able to hack it around by having a custom DynamicTableSink that for each record sends it to tidb and then to kafka. May I ask why you need that? If the notificatio

Flink-1.12 Sql on Job two SQL sink control order

2021-10-14 Thread WuKong
Hi all: I have two Flink SQL , the same source from Kafka, and one SQL sink data into Tidb ,another one SQL sink Kafka to notify downstream system, how can I control the sink order , I wish If source Kafka data come, first sink Tidb and after that sink Kafka . --- Best, WuKong