Fredo,

Yes, setting "auto.offset.reset=largest" should allow the consumer to reset
the fetching offset to log end when this happens. And if your consumer keep
lagging frequently you probably should consider either increase parallelism
on consumer clients or increase log.retention for the specific topic.

Guozhang

On Thu, Dec 24, 2015 at 12:51 AM, Fredo Lee <buaatianwa...@gmail.com> wrote:

> This problem is caused by that the consumer  offset lag producer offset,
> and the messages are deleted, then i use the lagged offset of fetch
> messages.But those messages have been deleted.
>
> Can this be resoled with using `auto.offset.reset=largest` ?
>
> 2015-12-24 16:32 GMT+08:00 Fredo Lee <buaatianwa...@gmail.com>:
>
> > When using kafka-0.8.2.0, Some questions happaned to me.
> >
> > I created one topic called `test` with partitions 60, replicaiton-factors
> > 2 and set log.retention.hours to 24 , Then i send some messages to
> `test`.
> > some days later, i create a consumer for this topic. but i got `out of
> > range` (i store my offset on kafka)
> >
> > So my questions is:
> >
> > 1. the `__consumer_offset` topic is just normal topic like other? Is
> > its log.retention.hours same to other topics?
> > 2. If I set log.retention.hours to 24 hours and store some  some offsets
> > in  `_consumer_offset_32/00000000000000.log`, which offset would i get
> > after 24 hours.
> >
> > for the reason of topic deletion, will i get 0 offset?
> > I doubt that my `out of range exception` is caused by this i described
> > above.
> >
> > My english is so pool. hope people can understand.
> >
> >
> > WanliTian
> >
>



-- 
-- Guozhang

Reply via email to