Hi,

The readme on the Flink CDC connectors [1] say that Oracle Databases
version 11, 12, 19 are supported with Oracle Driver 19.3.0.0.

Best regards,

Martijn

[1] https://github.com/ververica/flink-cdc-connectors/blob/master/README.md

On Fri, 11 Feb 2022 at 08:37, mohan radhakrishnan <
radhakrishnan.mo...@gmail.com> wrote:

> Thanks. I looked at it. Our primary DB is Oracle and MySql. Flink CDC
> Connector uses Debezium. I think. So ververica doesn't have a Flink CDC
> Connector for Oracle ?
>
> On Mon, Feb 7, 2022 at 3:03 PM Leonard Xu <xbjt...@gmail.com> wrote:
>
>> Hello, mohan
>>
>> 1. Does flink have any support to track any missed source Jdbc CDC
>> records ?
>>
>>
>> Flink CDC Connector provides Exactly once semantics which means they
>> won’t miss records. Tips: The Flink JDBC Connector only
>> Scan the database once which can not continuously read CDC stream.
>>
>> 2. What is the equivalent of Kafka consumer groups ?
>>
>>
>> Different database has different CDC mechanism, it’s serverId which used
>> to mark a slave for MySQL/MariaDB, it’s slot name for PostgresSQL.
>>
>>
>> 3. Delivering to kafka from flink is not exactly once. Is that right ?
>>
>>
>> No, both Flink CDC Connector and Flink Kafka Connector provide exactly
>> once implementation.
>>
>> BTW, if your destination is Elasticsearch, the quick start demo[1] may
>> help you.
>>
>> Best,
>> Leonard
>>
>> [1]
>> https://ververica.github.io/flink-cdc-connectors/master/content/quickstart/mysql-postgres-tutorial.html
>>
>>
>>
>> Thanks
>>
>> On Friday, February 4, 2022, mohan radhakrishnan <
>> radhakrishnan.mo...@gmail.com> wrote:
>>
>>> Hello,
>>>                So the jdbc source connector is  kafka and transformation
>>> is done by flink (flink sql) ? But that connector can miss records. I
>>> thought. Started looking at flink for this and other use cases.
>>> Can I see the alternative to spring cloudstreams( kafka streams )? Since
>>> I am learning flink, kafka streams' changelog topics and exactly-once
>>> delivery and dlqs seemed good for our cŕitical push notifications.
>>>
>>> We also needed a  elastic  sink.
>>>
>>> Thanks
>>>
>>> On Friday, February 4, 2022, Dawid Wysakowicz <dwysakow...@apache.org>
>>> wrote:
>>>
>>>> Hi Mohan,
>>>>
>>>> I don't know much about Kafka Connect, so I will not talk about its
>>>> features and differences to Flink. Flink on its own does not have a
>>>> capability to read a CDC stream directly from a DB. However there is the
>>>> flink-cdc-connectors[1] projects which embeds the standalone Debezium
>>>> engine inside of Flink's source and can process DB changelog with all
>>>> processing guarantees that Flink provides.
>>>>
>>>> As for the idea of processing further with Kafka Streams. Why not
>>>> process data with Flink? What do you miss in Flink?
>>>>
>>>> Best,
>>>>
>>>> Dawid
>>>>
>>>> [1] https://github.com/ververica/flink-cdc-connectors
>>>>
>>>> On 04/02/2022 13:55, mohan radhakrishnan wrote:
>>>>
>>>>> Hi,
>>>>>      When I was looking for CDC I realized Flink uses Kafka Connector
>>>>> to stream to Flink. The idea is to send it forward to Kafka and consume it
>>>>> using Kafka Streams.
>>>>>
>>>>> Are there source DLQs or additional mechanisms to detect failures to
>>>>> read from the DB ?
>>>>>
>>>>> We don't want to use Debezium and our CDC is based on queries.
>>>>>
>>>>> What mechanisms does Flink have that a Kafka Connect worker does not ?
>>>>> Kafka Connect workers can go down and source data can be lost.
>>>>>
>>>>> Does the idea  to send it forward to Kafka and consume it using Kafka
>>>>> Streams make sense ? The checkpointing feature of Flink can help ? I plan
>>>>> to use Kafka Streams for 'Exactly-once Delivery' and changelog topics.
>>>>>
>>>>> Could you point out relevant material to read ?
>>>>>
>>>>> Thanks,
>>>>> Mohan
>>>>>
>>>>
>>

Reply via email to