Hi community,

I have a Hive table that stores tens of millions rows of data. In my Flink
job, I want to process the data in batch manner:

- Split the data into batches, each batch has (maybe) 10,000 rows.
- For each batch, call a batchPut() API on my redis client to dump in Redis.

Doing so in a streaming manner is not expected, as that will cause too many
round trips between Flink workers and Redis.

Is there a way to do that? I find little clue in Flink docs, since almost
all APIs feel better suited for streaming processing by default.

Thank you!

Best,
Yik San

Reply via email to