H Márton,

Thank you very much for your answer.

The point with Kafka makes sense. It offers huge bag of potential connectors 
that could be used.
But … not everybody wants or needs Kafka. This brings additional architectural 
complication and delays, which might not be acceptable by everybody.
That’s why you do have your own connectors anyway.

The Flink connector which reads from Oracle utilizes the LogMiner technology, 
which 
Is not acceptable for every user. It has big limitation regarding speed.
You can overcome that only with a binary reader of the database redo log (like 
10 times 
faster and delay even up to 50-100ms).

The reason I am asking is not just to create some additional connector just for 
fun.
My main concern is if there is actual demand from users for bigger speed of 
getting changes from the source database or having lower delay.
You can find a lot of information in the net about differences between a 
log-based and 
one which is using logminer technology.

I think, that would be enough for a start. Please tell me what you think about 
it.
Would anyone consider using such connector?

Regards,
Adam Leszczyński


> On 4 Jan 2023, at 12:07, Márton Balassi <balassi.mar...@gmail.com> wrote:
> 
> (cc Leonard)
> 
> Hi Adam,
> 
> From an architectural perspective if you land the records to Kafka or other
> message broker Flink will be able to process them, at this point I do not
> see much merit trying to circumvent this step.
> There is a related project in the Flink space called CDC connectors [1], I
> highly encourage you to check that out for context and ccd Leonard one of
> its primary maintainers.
> 
> [1] https://github.com/ververica/flink-cdc-connectors/
> 
> On Tue, Jan 3, 2023 at 8:40 PM Adam Leszczyński <aleszczyn...@bersler.com>
> wrote:
> 
>> Hi Flink Team,
>> 
>> I’m the author of OpenLogReplictor - open source parser of Oracle redo
>> logs which allows to send transactions
>> to some message bus. Currently the sink that is implemented is just text
>> file or Kafka topic.
>> Also transactions can be sent using plain tcp connection or some message
>> queue like ZeroMQ.
>> Code is GPL and all versions from 11.2 are supported. No LogMiner needed.
>> 
>> Transactions can be sent using json or protobuf format. Currently the code
>> has reached GA and is actually used in production.
>> 
>> The architecture is modular and allows very easily to add other sinks like
>> for example Apache Flink.
>> Actually I’m going towards approach that OpenLogReplicator could used
>> Kubernetes and work in HA.
>> 
>> Well… that is the general direction. Do you think there could some
>> application of this soft with Apache Flink?
>> For example very easily there could be some client which could connect to
>> OpenLogReplicator using tcp connection
>> and get transactions and just send them to Apache Flink. An example of
>> such client is also present in GitHub repo.
>> https://github.com/bersler/OpenLogReplicator
>> 
>> Is there any rational for such integration? Or just a waste of time cause
>> nobody would use it anyway?
>> 
>> Kind regards,
>> Adam Leszczyński
>> 
>> 

Reply via email to