Yes, you can try setting it to -1 in 0.8.1, which is the equivalent of "all" in 0.9 and above.
Thanks, Jun On Wed, Aug 17, 2016 at 8:32 AM, Mazhar Shaikh <mazhar.shaikh...@gmail.com> wrote: > Hi Jun, > > I'm using default configuration (ack=1), > changing it t0 all or 2 will not help, as the producer queue will be > exhausted is any kafka broker goes down for long time. > > > Thanks. > > Regards, > Mazhar Shaikh. > > > On Wed, Aug 17, 2016 at 8:11 PM, Jun Rao <j...@confluent.io> wrote: > > > Are you using acks=1 or acks=all in the producer? Only the latter > > guarantees acked messages won't be lost after leader failure. > > > > Thanks, > > > > Jun > > > > On Wed, Aug 10, 2016 at 11:41 PM, Mazhar Shaikh < > > mazhar.shaikh...@gmail.com> > > wrote: > > > > > Hi Kafka Team, > > > > > > I'm using kafka (kafka_2.11-0.9.0.1) with librdkafka (0.8.1) API for > > > producer > > > During a run of 2hrs, I notice the total number of messaged ack'd by > > > librdkafka delivery report is greater than the maxoffset of a partition > > in > > > kafka broker. > > > I'm running kafka broker with replication factor of 2. > > > > > > Here, message has been lost between librdkafka - kafka broker. > > > > > > As librdkafka is providing success delivery report for all the > messages. > > > > > > Looks like kafka broker is dropping the messages after acknowledging > > > librdkafka. > > > > > > Requesting you help in solving this issue. > > > > > > Thank you. > > > > > > > > > Regards > > > Mazhar Shaikh > > > > > >