Sorry, I guess I missed that. The followup discussion was around the simple consumer :)
I'm not sure why the OP didn't find this solution acceptable. On Wed, Sep 3, 2014 at 8:29 AM, Philip O'Toole <philip.oto...@yahoo.com> wrote: > That's what I said in my first reply. :-) > > ----------------------------------------- > http://www.philipotoole.com > > > On Tuesday, September 2, 2014 10:37 PM, Gwen Shapira <gshap...@cloudera.com> > wrote: > > > I believe a simpler solution would be to create multiple > ConsumerConnector, each with 1 thread (single ConsumerStream) and use > commitOffset API to commit all partitions managed by each > ConsumerConnector after the thread finished processing the messages. > > Does that solve the problem, Bhavesh? > > Gwen > > On Tue, Sep 2, 2014 at 5:47 PM, Philip O'Toole > <philip.oto...@yahoo.com.invalid> wrote: >> Yeah, from reading that I suspect you need the SimpleConsumer. Try it out >> and see. >> >> Philip >> >> >> ----------------------------------------- >> http://www.philipotoole.com >> >> >> On Tuesday, September 2, 2014 5:43 PM, Bhavesh Mistry >> <mistry.p.bhav...@gmail.com> wrote: >> >> >> >> Hi Philip, >> >> Yes, We have disabled auto commit but, we need to be able to read from >> particular offset if we manage the offset ourself in some storage(DB). >> High Level consumer does not allow per partition management plug-ability. >> >> I like to have the High Level consumers Failover and auto rebalancing. We >> just need plug ability of offset management. >> >> Thanks, >> >> Bhavesh >> >> >> >> On Tue, Sep 2, 2014 at 5:20 PM, Philip O'Toole < >> philip.oto...@yahoo.com.invalid> wrote: >> >>> No, you'll need to write your own failover. >>> >>> I'm not sure I follow your second question, but the high-level Consumer >>> should be able to do what you want if you disable auto-commit. I'm not >>> sure >>> what else you're asking. >>> >>> >>> Philip >>> >>> >>> ----------------------------------------- >>> http://www.philipotoole.com >>> >>> >>> On Tuesday, September 2, 2014 5:15 PM, Bhavesh Mistry < >>> mistry.p.bhav...@gmail.com> wrote: >>> >>> >>> >>> Hi Philip, >>> >>> Thanks for the update. With Simple Consumer I will not get failover and >>> rebalance that is provided out of box. what is other option not to block >>> reading and keep processing and commit only when batch is done. >>> >>> Thanks, >>> >>> Bhavesh >>> >>> >>> >>> On Tue, Sep 2, 2014 at 4:43 PM, Philip O'Toole < >>> philip.oto...@yahoo.com.invalid> wrote: >>> >>> > Either use the SimpleConsumer which gives you much finer-grained >>> > control, >>> > or (this worked with 0.7) spin up a ConsumerConnection (this is a >>> HighLevel >>> > consumer concept) per partition, turn off auto-commit. >>> > >>> > Philip >>> > >>> > >>> > ----------------------------------------- >>> > http://www.philipotoole.com >>> > >>> > >>> > On Tuesday, September 2, 2014 4:38 PM, Bhavesh Mistry < >>> > mistry.p.bhav...@gmail.com> wrote: >>> > >>> > >>> > >>> > Hi Kafka Group, >>> > >>> > I have to pull the data from the Topic and index into Elastic Search >>> > with >>> > Bulk API and wanted to commit only batch that has been committed and >>> still >>> > continue to read from topic further on same topic. I have auto commit >>> > to >>> > be off. >>> > >>> > >>> > List<Message> batch ..... >>> > >>> > while (iterator.hasNext()) { >>> > batch.add(iterator.next().message()); >>> > if(batch size is 50 ){ >>> > //===>>>> Once the bulk API is successful it will commit the >>> offset >>> > to zookeeper... >>> > executor.submit(new Thread() process batch and commit batch, >>> > cconsumerConnector) >>> > batch = new batch buffer.... >>> > } >>> > } >>> > >>> > This commitOffset API commits all messages that have been read so far. >>> > What is best way to continue reading and only commit another thread >>> finish >>> > batch process is successful. This will lead to fragmentation of the >>> > Consumer offset so what is best way to implement continuous reading >>> stream >>> > and commit the rage offset. >>> > >>> > Is Simple Consumer a better approach for this. >>> > >>> > >>> > Thanks, >>> > >>> > Bhavesh >>> > >>> > >>> > >>> > >>> > >>> > >>> > >>> > Thanks, >>> > Bhavesh >>> > >>> > >