Hi Yuxia, Thank you so much for your response. Much appreciated. Here, by CDC I meant the incremental changes that have to be pushed from Kafka to my processing layer which is Flink.
Let me go through the links shared by you. Sid. On Mon, Jun 27, 2022 at 6:39 AM yuxia <luoyu...@alumni.sjtu.edu.cn> wrote: > > I mean CDC should be handled on the Kafka side. > What do you mean about that? Do you mean the the Kafka should store the > message with the cdc format like debezium[1], Canal[2], MaxWell[3], OGG[4]? > > > Or should I need to use Table API > > I'm afraid not. Seems you can still use Flink Datastream API as Table > API makes no difference for your case. > > BTW, you can try flink cdc [5] > > [1] > https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/formats/debezium/ > [2] > https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/formats/canal/ > [3] > https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/formats/maxwell/ > [4] > https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/formats/ogg/ > [5] https://ververica.github.io/flink-cdc-connectors/ > > > Best regards, > Yuxia > > ------------------------------ > *发件人: *"Sid" <flinkbyhe...@gmail.com> > *收件人: *"User" <user@flink.apache.org> > *发送时间: *星期六, 2022年 6 月 25日 下午 6:32:22 > *主题: *How to make current application cdc > > Hello, > > I have a current flow where the data from the Flink-Kafka connector is > captured and processed using Flink Datastream API and stored in Kafka > topics. However, I would like to make it CDC enabled. I went through an > article where it was mentioned that it should be handled on the Kafka side > while capturing the data. I mean CDC should be handled on the Kafka side. > Or should I need to use Table API? > So, any ideas/links are much appreciated as I am trying to understand > these concepts. > > TIA, > Sid > >