You will need to configure request.required.acks properly. See http://kafka.apache.org/documentation.html#producerconfigs for details.
Thanks, Jun On Tue, Dec 10, 2013 at 1:55 AM, Nishant Kumar <nish.a...@gmail.com> wrote: > Hi All, > > I am using kafka 0.8. > > > My producers configurations are as follows > > kafka8.bytearray.producer.type=sync > > kafka8.producer.batch.num.messages=100 > > kafka8.producer.topic.metadata.refresh.interval.ms=600000 > > kafka8.producer.retry.backoff.ms=100 > > kafka8.producer.message.send.max.retries=3 > > My Kafaka server. properties are as > > # The number of messages to accept before forcing a flush of data to > disk > log.flush.interval.messages=500 > > # The maximum amount of time a message can sit in a log before we > force a flush > log.flush.interval.ms=100 > > # Per-topic overrides for log.flush.interval.ms > #log.flush.intervals.ms.per.topic=topic1:1000, topic2:3000 > > > Specified Sync property in producer. properties file > > # specifies whether the messages are sent asynchronously (async) > or synchronously (sync) > producer.type=sync > > > > My consumer is running in a separate jar. Consumer config are > > zookeeper.connect=IP > group.id=consumerGroup > fetch.message.max.bytes=1000000000 > zookeeper.session.timeout.ms=60000 > auto.offset.reset=smallest > zookeeper.sync.time.ms=200 > auto.commit.enable=false > > If my data generator and consumer are running parallel and suddenly > kafka is restarted, less > records are consumed then expected. > > e.g. If i set the number of records to be produce are 3000 after that > it throws an exception. My consumer runs > in parallel to that, mean while if i restart my kafka ,my consumer is > only able to > get 2400 approx records. approximately 600 records are missing even if > i am running kafaka in synchronized mode. > > I am not able to know why this data lose is happening. If you have > any idea regarding this. > Please help me to know what i am missing here in this case. > > Regards, > > Nishant Kumar >