Hey everyone,

I'm trying to use Flink SQL to construct a set of transformations for my
application. Let's say the topology just has three steps:

- SQL Source
- SQL SELECT statement
- SQL Sink (via INSERT)

The sink I'm using (JDBC) would really benefit from data partitioning (by
PK ID) to avoid conflicting transactions and deadlocks. I can force Flink
to partition the data by the PK ID before the INSERT by resorting to
DataStream API and leveraging the keyBy method, then transforming
DataStream back to the Table again...

Is there a simpler way to do this? I understand that, for example, a GROUP
BY statement will probably perform similar data shuffling, but what if I
have a simple SELECT followed by INSERT?

Thank you!

Reply via email to