Thank you very much for the reply, Neha, I have a question about consumer,
I consume the data from kafka and write into DB, of course I have to create
a hash map in memory, load data into memory and bulk copy to DB instead of
insert into DB line by line. Does it mean I need to ack each message while
load to memory?

thanks



On Thu, Dec 4, 2014 at 1:21 PM, Neha Narkhede <n...@confluent.io> wrote:

> This is specific for pentaho but may be useful -
> https://github.com/RuckusWirelessIL/pentaho-kafka-consumer
>
> On Thu, Dec 4, 2014 at 12:58 PM, Sa Li <sal...@gmail.com> wrote:
>
> > Hello, all
> >
> > I never developed a kafka consumer, I want to be able to make an advanced
> > kafka consumer in java to consume the data and continuously write the
> data
> > into postgresql DB. I am thinking to create a map in memory and getting a
> > predefined number of messages in memory then write into DB in batch, is
> > there a API or sample code to allow me to do this?
> >
> >
> > thanks
> >
> >
> > --
> >
> > Alec Li
> >
>
>
>
> --
> Thanks,
> Neha
>



-- 

Alec Li

Reply via email to