if you are using scala/akka this will handle the batching and acks for you.
https://github.com/sclasen/akka-kafka#akkabatchconsumer On Fri, Dec 5, 2014 at 9:21 AM, Sa Li <sal...@gmail.com> wrote: > Thank you very much for the reply, Neha, I have a question about consumer, > I consume the data from kafka and write into DB, of course I have to create > a hash map in memory, load data into memory and bulk copy to DB instead of > insert into DB line by line. Does it mean I need to ack each message while > load to memory? > > thanks > > > > On Thu, Dec 4, 2014 at 1:21 PM, Neha Narkhede <n...@confluent.io> wrote: > > > This is specific for pentaho but may be useful - > > https://github.com/RuckusWirelessIL/pentaho-kafka-consumer > > > > On Thu, Dec 4, 2014 at 12:58 PM, Sa Li <sal...@gmail.com> wrote: > > > > > Hello, all > > > > > > I never developed a kafka consumer, I want to be able to make an > advanced > > > kafka consumer in java to consume the data and continuously write the > > data > > > into postgresql DB. I am thinking to create a map in memory and > getting a > > > predefined number of messages in memory then write into DB in batch, is > > > there a API or sample code to allow me to do this? > > > > > > > > > thanks > > > > > > > > > -- > > > > > > Alec Li > > > > > > > > > > > -- > > Thanks, > > Neha > > > > > > -- > > Alec Li >