Hi!

This isn't exactly what you are looking for, but we have been looking into
some of these things at WMF too.
See John Mwangi's experiment project and writeup:
https://github.com/john-mwangi/mariadb-iceberg-pipeline/blob/main/WRITEUP.md

And the attached task with details in comments.
T373144 [SPIKE] Learn and document how to use Flink-CDC from MediaWiki
MariaDB locally <https://phabricator.wikimedia.org/T373144>

Maybe there will be some insights you can use there?

Good luck!

On Tue, Mar 4, 2025 at 9:44 AM Taher Koitawala <taher...@gmail.com> wrote:

> Hi users, please help me here
> ---------- Forwarded message ---------
> From: Taher Koitawala <taher...@gmail.com>
> Date: Tue, 11 Feb 2025 at 11:59 AM
> Subject: Flink CDC to Paimon
> To: Selvam Ragav <selvam.ra...@gmail.com>, <d...@flink.apache.org>
>
>
> Hi Devs,
>        As a POC we are trying to create a steaming pipeline from MSSQL cdc
> to Paimon:
>
> To do this we are doing
> 1. msSql server cdc operator
> 2. Transform operator
> 3. paimon sink
>
> We have written the cdc connector with is a JsonDebeziumDeserialisedSchema
> String
>
> I wish to write this paimon in a table format with same columns as source.
>
> As far as i know paimon automatically handles schema updates like new
> field additions.
>
> Please can someone point me on how to write this stream efficiently to
> paimon table with schema updates?
>
> For now i have SouceFunction<String>
>
> Which is the record mentioned above!
>
> Regards,
> Taher Koitawala
>
>
>
>
>

Reply via email to