Is there anything on the broker log? Is it possible that your client and broker are not running on the same version?
Jiangjie (Becket) Qin On 7/15/15, 11:40 AM, "Vadim Bobrov" <vadimbob...@gmail.com> wrote: >caught it, thanks for help! >any ideas what to do? > >TRACE 2015-07-15 18:37:58,070 [chaos-akka.actor.jms-dispatcher-1019 ] >kafka.network.BoundedByteBufferSend - 113 bytes written. >ERROR 2015-07-15 18:37:58,078 [chaos-akka.actor.jms-dispatcher-1019 ] >kafka.consumer.ZookeeperConsumerConnector - >[chaos-test_ds2.outsmartinc.com-1436984542831-4d399e71], Error while >committing offsets. >java.io.EOFException: Received -1 when reading from channel, socket has >likely been closed. > at kafka.utils.Utils$.read(Utils.scala:381) > at >kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.s >cala:54) > at >kafka.network.Receive$class.readCompletely(Transmission.scala:56) > at >kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferRec >eive.scala:29) > at >kafka.network.BlockingChannel.receive(BlockingChannel.scala:111) > at >kafka.consumer.ZookeeperConsumerConnector.liftedTree2$1(ZookeeperConsumerC >onnector.scala:313) > at >kafka.consumer.ZookeeperConsumerConnector.commitOffsets(ZookeeperConsumerC >onnector.scala:310) > at >kafka.javaapi.consumer.ZookeeperConsumerConnector.commitOffsets(ZookeeperC >onsumerConnector.scala:111) > at >com.os.messaging.kafka.KafkaMessageService$class.acknowledge(KafkaMessageS >ervice.scala:55) > at com.os.actor.Main$$anonfun$2$$anon$2.acknowledge(Main.scala:84) > at >com.os.actor.acquisition.MessageListenerActor$$anonfun$8.applyOrElse(Messa >geListenerActor.scala:180) > at >com.os.actor.acquisition.MessageListenerActor$$anonfun$8.applyOrElse(Messa >geListenerActor.scala:164) > at scala.PartialFunction$OrElse.apply(PartialFunction.scala:162) > at akka.actor.FSM$class.processEvent(FSM.scala:607) > at >com.os.actor.acquisition.MessageListenerActor.processEvent(MessageListener >Actor.scala:32) > at akka.actor.FSM$class.akka$actor$FSM$$processMsg(FSM.scala:598) > at akka.actor.FSM$$anonfun$receive$1.applyOrElse(FSM.scala:592) > at akka.actor.Actor$class.aroundReceive(Actor.scala:465) > at >com.os.actor.acquisition.MessageListenerActor.aroundReceive(MessageListene >rActor.scala:32) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) > at akka.dispatch.Mailbox.run(Mailbox.scala:221) > at akka.dispatch.Mailbox.exec(Mailbox.scala:231) > at >scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at >scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java >:1339) > at >scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at >scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.ja >va:107) > > >On Wed, Jul 15, 2015 at 1:36 PM, Vadim Bobrov <vadimbob...@gmail.com> >wrote: > >> thanks Joel and Jiangjie, >> I have figured it out. In addition to my log4j 2 config file I also >>needed >> a log4j 1 config file, then it works. Let me trace what happens when the >> offsets are not committed and report back >> >> On Wed, Jul 15, 2015 at 1:33 PM, Joel Koshy <jjkosh...@gmail.com> wrote: >> >>> - You can also change the log4j level dynamically via the >>> kafka.Log4jController mbean. >>> - You can also look at offset commit request metrics (mbeans) on the >>> broker (just to check if _any_ offset commits are coming through >>> during the period you see no moving offsets). >>> - The alternative is to just consume the offsets topic. >>> >>> On Wed, Jul 15, 2015 at 05:30:17PM +0000, Jiangjie Qin wrote: >>> > I am not sure how your project was setup. But I think it depends on >>>what >>> > log4j property file you specified when you started your application. >>>Can >>> > you check if you have log4j appender defined and the loggers are >>> directed >>> > to the correct appender? >>> > >>> > Thanks, >>> > >>> > Jiangjie (Becket) Qin >>> > >>> > On 7/15/15, 8:10 AM, "Vadim Bobrov" <vadimbob...@gmail.com> wrote: >>> > >>> > >Thanks Jiangjie, >>> > > >>> > >unfortunately turning trace level on does not seem to work (any log >>> level >>> > >actually) I am using log4j2 (through slf4j) and despite including >>> log4j1 >>> > >bridge and these lines: >>> > > >>> > ><Logger name="org.apache.kafka" level="trace"/> >>> > ><Logger name="kafka" level="trace"/> >>> > > >>> > >in my conf file I could not squeeze out any logging from kafka. >>>Logging >>> > >for >>> > >all other libs (like zookeeper e.g.) work perfectly. Am I doing >>> something >>> > >wrong? >>> > > >>> > > >>> > >On Tue, Jul 14, 2015 at 6:55 PM, Jiangjie Qin >>> <j...@linkedin.com.invalid> >>> > >wrote: >>> > > >>> > >> Hi Vadim, >>> > >> >>> > >> Can you turn on trace level logging on your consumer and search >>>for >>> > >> "offset commit response² in the log? >>> > >> Also maybe take a look at the log to see if there is any exception >>> > >>thrown. >>> > >> >>> > >> Thanks, >>> > >> >>> > >> Jiangjie (Becket) Qin >>> > >> >>> > >> On 7/14/15, 11:06 AM, "Vadim Bobrov" <vadimbob...@gmail.com> >>>wrote: >>> > >> >>> > >> >just caught this error again. I issue commitOffsets - no error >>>but >>> no >>> > >> >committng offsets either. __consumer_offsets watching shows no >>>new >>> > >> >messages >>> > >> >either. Then in a few minutes I issue commitOffsets again - all >>> > >>committed. >>> > >> >Unless I am doing something terribly wrong this is very >>>unreliable >>> > >> > >>> > >> >On Tue, Jul 14, 2015 at 1:49 PM, Joel Koshy <jjkosh...@gmail.com> >>> > >>wrote: >>> > >> > >>> > >> >> Actually, how are you committing offsets? Are you using the old >>> > >> >> (zookeeperconsumerconnector) or new KafkaConsumer? >>> > >> >> >>> > >> >> It is true that the current APIs don't return any result, but >>>it >>> > >>would >>> > >> >> help to check if anything is getting into the offsets topic - >>> unless >>> > >> >> you are seeing errors in the logs, the offset commit should >>> succeed >>> > >> >> (if you are indeed explicitly committing offsets). >>> > >> >> >>> > >> >> Thanks, >>> > >> >> >>> > >> >> Joel >>> > >> >> >>> > >> >> On Tue, Jul 14, 2015 at 12:19:01PM -0400, Vadim Bobrov wrote: >>> > >> >> > Thanks, Joel, I will but regardless of my findings the basic >>> > >>problem >>> > >> >>will >>> > >> >> > still be there: there is no guarantee that the offsets will >>>be >>> > >> >>committed >>> > >> >> > after commitOffsets. Because commitOffsets does not return >>>its >>> exit >>> > >> >> status, >>> > >> >> > nor does it block as I understand until offsets are >>>committed. >>> In >>> > >> >>other >>> > >> >> > words, there is no way to know that it has, in fact, commited >>> the >>> > >> >>offsets >>> > >> >> > >>> > >> >> > or am I missing something? And then another question - why >>>does >>> it >>> > >> >>seem >>> > >> >> to >>> > >> >> > depend on the number of consumed messages? >>> > >> >> > >>> > >> >> > On Tue, Jul 14, 2015 at 11:36 AM, Joel Koshy < >>> jjkosh...@gmail.com> >>> > >> >> wrote: >>> > >> >> > >>> > >> >> > > Can you take a look at the kafka commit rate mbean on your >>> > >>consumer? >>> > >> >> > > Also, can you consume the offsets topic while you are >>> committing >>> > >> >> > > offsets and see if/what offsets are getting committed? >>> > >> >> > > ( >>> http://www.slideshare.net/jjkoshy/offset-management-in-kafka/32) >>> > >> >> > > >>> > >> >> > > Thanks, >>> > >> >> > > >>> > >> >> > > Joel >>> > >> >> > > >>> > >> >> > > On Tue, Jul 14, 2015 at 11:12:03AM -0400, Vadim Bobrov >>>wrote: >>> > >> >> > > > I am trying to replace ActiveMQ with Kafka in our >>> environment >>> > >> >> however I >>> > >> >> > > > have encountered a strange problem that basically >>>prevents >>> from >>> > >> >>using >>> > >> >> > > Kafka >>> > >> >> > > > in production. The problem is that sometimes the offsets >>>are >>> > >>not >>> > >> >> > > committed. >>> > >> >> > > > >>> > >> >> > > > I am using Kafka 0.8.2.1, offset storage = kafka, high >>>level >>> > >> >> consumer, >>> > >> >> > > > auto-commit = off. Every N messages I issue >>>commitOffsets(). >>> > >>Now >>> > >> >> here is >>> > >> >> > > > the problem - if N is below a certain number (180 000 for >>> me) >>> > >>it >>> > >> >> works >>> > >> >> > > and >>> > >> >> > > > the offset is moving. If N is 180 000 or more the offset >>>is >>> not >>> > >> >> updated >>> > >> >> > > > after commitOffsets >>> > >> >> > > > >>> > >> >> > > > I am looking at offsets using kafka-run-class.sh >>> > >> >> > > > kafka.tools.ConsumerOffsetChecker >>> > >> >> > > > Any help? >>> > >> >> > > >>> > >> >> > > >>> > >> >> >>> > >> >> >>> > >> >>> > >> >>> > >>> >>> -- >>> Joel >>> >> >>