Hello team,
I'm currently in the process of configuring a Flink job. This job entails
reading files from a specified directory and then transmitting the data to
a Kafka sink. I've already successfully designed a Flink job that reads the
file contents in a streaming manner and effectively sends them to Kafka.
However, my specific requirement is a bit more intricate. I need the job to
not only read these files and push the data to Kafka but also relocate the
processed file to a different directory once all of its contents have been
processed. Following this, the job should seamlessly transition to
processing the next file in the source directory. Additionally, I have some
concerns regarding how the job will behave if it encounters a restart.
Could you please advise if this is achievable, and if so, provide guidance
or code to implement it?

I'm also quite interested in how the job will handle situations where the
source has a parallelism greater than 2 or 3, and how it can accurately
monitor the completion of reading all contents in each file.

Thanks and Regards,
Arjun

Reply via email to