Hi, experts

Again, we still having the issues of losing data, see we see data 5000
records, but only find 4500 records on brokers, we did set required.acks -1
to make sure all brokers ack, but that only cause the long latency, but not
cure the data lost.


thanks


On Mon, Jan 5, 2015 at 9:55 AM, Xiaoyu Wang <xw...@rocketfuel.com> wrote:

> @Sa,
>
> the required.acks is producer side configuration. Set to -1 means requiring
> ack from all brokers.
>
> On Fri, Jan 2, 2015 at 1:51 PM, Sa Li <sal...@gmail.com> wrote:
>
> > Thanks a lot, Tim, this is the config of brokers
> >
> > ----------
> > broker.id=1
> > port=9092
> > host.name=10.100.70.128
> > num.network.threads=4
> > num.io.threads=8
> > socket.send.buffer.bytes=1048576
> > socket.receive.buffer.bytes=1048576
> > socket.request.max.bytes=104857600
> > auto.leader.rebalance.enable=true
> > auto.create.topics.enable=true
> > default.replication.factor=3
> >
> > log.dirs=/tmp/kafka-logs-1
> > num.partitions=8
> >
> > log.flush.interval.messages=10000
> > log.flush.interval.ms=1000
> > log.retention.hours=168
> > log.segment.bytes=536870912
> > log.cleanup.interval.mins=1
> >
> > zookeeper.connect=10.100.70.128:2181,10.100.70.28:2181,10.100.70.29:2181
> > zookeeper.connection.timeout.ms=1000000
> >
> > -----------------------
> >
> >
> > We actually play around request.required.acks in producer config, -1
> cause
> > long latency, 1 is the parameter to cause messages lost. But I am not
> sure,
> > if this is the reason to lose the records.
> >
> >
> > thanks
> >
> > AL
> >
> >
> >
> >
> >
> >
> >
> > On Fri, Jan 2, 2015 at 9:59 AM, Timothy Chen <tnac...@gmail.com> wrote:
> >
> > > What's your configured required.acks? And also are you waiting for all
> > > your messages to be acknowledged as well?
> > >
> > > The new producer returns futures back, but you still need to wait for
> > > the futures to complete.
> > >
> > > Tim
> > >
> > > On Fri, Jan 2, 2015 at 9:54 AM, Sa Li <sal...@gmail.com> wrote:
> > > > Hi, all
> > > >
> > > > We are sending the message from a producer, we send 100000 records,
> but
> > > we
> > > > see only 99573 records for that topics, we confirm this by consume
> this
> > > > topic and check the log size in kafka web console.
> > > >
> > > > Any ideas for the message lost, what is the reason to cause this?
> > > >
> > > > thanks
> > > >
> > > > --
> > > >
> > > > Alec Li
> > >
> >
> >
> >
> > --
> >
> > Alec Li
> >
>



-- 

Alec Li

Reply via email to