Dear all,

I'm working for a big project and one of the challenge is to read Kafka
topics and copy them via Hive command into Hive managed tables in order to
enable ACID HIVE properties.

I try it but I have a issue with back pressure:
- The first window read 20.000 events and wrote them in Hive tables
- The second, third, ... send only 100 events because the write in Hive
take more time than the read of a Kafka topic. But writing 100 events or
50.000 events takes +/- the same time for Hive.

Someone have already do this source and sink? Could you help on this?
Or have you some tips?
It seems that defining a size window on number of event instead time is not
possible. Is it true?

Thank you for your help

Youssef

-- 
♻ Be green, keep it on the screen

Reply via email to