Hi Leonard,
     Yes i did see Xianqian’s reply however i thought my email did not go
through as the community is often very active but I did not receive a
response until Xianqian’s reply.

Thank you Xianqian we are currently trying out the suggestions for altering
schema via the catalog.

Thank you Andrew Otto for the docs as i will I will take a look at this.




On Wed, 5 Mar 2025 at 5:47 AM, Leonard Xu <xbjt...@gmail.com> wrote:

> Hey Taher,
>
> Xianqian has replied your email, did you subscribe Flink dev mailing-list?
>
> Best,
> Leonard
>
> > 2025年2月11日 16:52,Xiqian YU <kono....@outlook.com> 写道:
> >
> > Hi Taher,
> >
> > Since we’re creating a DataStream-based pipeline job with SQL Server
> CDC, schema change events must be handled manually. A possible approach
> would be:
> >
> > 1) Enable schema change events with `.includeSchemaChanges(true)`
> option, so DDL events will be parsed and encoded in `SourceRecord`s.
> >
> > 2) Write a customized `DebeziumDeserializationSchema` class and parse
> schema change events.
> `MySqlEventDeserializer#deserializeSchemaChangeRecord` could be used as a
> reference [1].
> >
> > 3) Evolve sink schema of Paimon tables with `PaimonCatalog` manually.
> `PaimonMetadataApplier` [2] is an existing schema evolving implementation
> supporting a few frequently used schema change events.
> >
> > Also, CDC Pipeline framework [3] has provided a fully-automatic schema
> sensing and evolving solution, but unfortunately Microsoft SQL Server
> source is not supported yet until we close #3445 [4] or #3507 [5].
> >
> > [1]
> https://github.com/apache/flink-cdc/blob/master/flink-cdc-connect/flink-cdc-pipeline-connectors/flink-cdc-pipeline-connector-mysql/src/main/java/org/apache/flink/cdc/connectors/mysql/source/MySqlEventDeserializer.java
> > [2]
> https://github.com/apache/flink-cdc/blob/master/flink-cdc-connect/flink-cdc-pipeline-connectors/flink-cdc-pipeline-connector-paimon/src/main/java/org/apache/flink/cdc/connectors/paimon/sink/PaimonMetadataApplier.java
> > [3]
> https://nightlies.apache.org/flink/flink-cdc-docs-release-3.3/docs/core-concept/data-pipeline/
> > [4] https://github.com/apache/flink-cdc/pull/3445
> > [5] https://github.com/apache/flink-cdc/pull/3507
> >
> > Best Regards,
> > Xiqian
> >
> > Taher Koitawala <taher...@gmail.com> 於 2025年2月11日 15:59 寫道:
> >
> > Hi Devs,
> >      As a POC we are trying to create a steaming pipeline from MSSQL cdc
> > to Paimon:
> >
> > To do this we are doing
> > 1. msSql server cdc operator
> > 2. Transform operator
> > 3. paimon sink
> >
> > We have written the cdc connector with is a
> JsonDebeziumDeserialisedSchema
> > String
> >
> > I wish to write this paimon in a table format with same columns as
> source.
> >
> > As far as i know paimon automatically handles schema updates like new
> field
> > additions.
> >
> > Please can someone point me on how to write this stream efficiently to
> > paimon table with schema updates?
> >
> > For now i have SouceFunction<String>
> >
> > Which is the record mentioned above!
> >
> > Regards,
> > Taher Koitawala
> >
>
>

Reply via email to