Hi Jiabao, My usecase is that when I start my flink job it should load and process all the existing data in a collection and also wait and process any new data that comes along the way. As I notice that flink-connector-mongodb would process all the existing data, so do I still need this connector or I can use flink-connector-mongodb-cdc to process both existing and new data ?
Thanks Sachin On Fri, Aug 16, 2024 at 3:46 PM Jiabao Sun <jiabao...@apache.org> wrote: > Hi Sachin, > > flink-connector-mongodb supports batch reading and writing to MongoDB, > similar to flink-connector-jdbc, while flink-connector-mongodb-cdc supports > streaming MongoDB changes. > > If you need to stream MongoDB changes, you should use > flink-connector-mongodb-cdc. > You can refer to the following documentation about mongodb cdc. > > > https://nightlies.apache.org/flink/flink-cdc-docs-release-3.1/docs/connectors/flink-sources/mongodb-cdc/ > > Best, > Jiabao > > On 2024/08/16 09:46:47 Sachin Mittal wrote: > > Hi, > > I have a scenario where I load a collection from MongoDB inside Flink > using > > flink-connector-mongodb. > > What I additionally want is any future changes (insert/updates) to that > > collection is also streamed inside my Flink Job. > > > > What I was thinking of is to use a CDC connector to stream data to my > Flink > > job. > > > > When researching this I found Flink CDC and they have a CDC connector for > > MongoDB - flink-connector-mongodb-cdc > > > > > > However I am not able to figure out how to stream those changes also to > my > > Job which is also reading from the same collection. > > > > Thanks > > Sachin > > >