Arjun, Have you looked at https://cwiki.apache.org/confluence/display/KAFKA/FAQ#FAQ-Myconsumerseemstohavestopped,why ?
Thanks, Neha On Tue, Feb 25, 2014 at 5:04 AM, Arjun <ar...@socialtwist.com> wrote: > The thing i found is my ConsumerFetcherThreads are not going beyond > BoundedByteBufferReceive.readFrom. When i added a few more traces in that > function i found that the call is stalling after exceptIncomplete function. > I guess Utils.read is stalling for more than 30 sec, which is the socket > time out time. > Is there anyway i can increase the socket timeout over there. I am not > getting why the consumers are getting stuck over there. > > There are no error on the brokers. > > Thanks > Arjun Narasimha Kota > > > > On Tuesday 25 February 2014 05:45 PM, Arjun wrote: > >> Adding to this, i have started my logs in trace mode. I fount that the >> Consumer fetcher threads are sending the meta data but are not receiving >> any. >> I see all the >> "TRACE >> [ConsumerFetcherThread-group1_www.taf-dev.com-1393329622308-6e15dd12-0-1] >> [kafka.network.BoundedByteBufferSend] 205 bytes written" >> but no reading is taking place. I may be looking at some thing wrong. >> >> Can some one please help me out in this. >> >> Thanks >> Arjun Narasimha kota >> >> >> On Tuesday 25 February 2014 03:46 PM, Arjun wrote: >> >>> Apart from that i get this stack trace >>> >>> 25 Feb 2014 15:45:22,636 WARN [ConsumerFetcherThread-group1_ >>> www.taf-dev.com-1393322165136-8318b07d-0-0] >>> [kafka.consumer.ConsumerFetcherThread] >>> [ConsumerFetcherThread-group1_www.taf-dev.com-1393322165136-8318b07d-0-0], >>> Error in fetch Name: FetchRequest; Version: 0; CorrelationId: 32; ClientId: >>> group1-ConsumerFetcherThread-group1_www.taf-dev.com-1393322165136-8318b07d-0-0; >>> ReplicaId: -1; MaxWait: 10000000 ms; MinBytes: 1 bytes; RequestInfo: >>> [taf.referral.emails.service,8] -> PartitionFetchInfo(1702,1048576) >>> java.net.SocketTimeoutException >>> at sun.nio.ch.SocketAdaptor$SocketInputStream.read( >>> SocketAdaptor.java:201) >>> at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:86) >>> at java.nio.channels.Channels$ReadableByteChannelImpl.read( >>> Channels.java:221) >>> at kafka.utils.Utils$.read(Utils.scala:395) >>> at kafka.network.BoundedByteBufferReceive.readFrom( >>> BoundedByteBufferReceive.scala:54) >>> at kafka.network.Receive$class.readCompletely(Transmission.scala:56) >>> at kafka.network.BoundedByteBufferReceive.readCompletely( >>> BoundedByteBufferReceive.scala:29) >>> at kafka.network.BlockingChannel.receive(BlockingChannel.scala:100) >>> at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer. >>> scala:81) >>> at kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$ >>> $sendRequest(SimpleConsumer.scala:71) >>> at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$ >>> apply$mcV$sp$1.apply$mcV$sp(SimpleConsumer.scala:110) >>> at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$ >>> apply$mcV$sp$1.apply(SimpleConsumer.scala:110) >>> at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$ >>> apply$mcV$sp$1.apply(SimpleConsumer.scala:110) >>> at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) >>> at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply$mcV$sp( >>> SimpleConsumer.scala:109) >>> at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply( >>> SimpleConsumer.scala:109) >>> at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply( >>> SimpleConsumer.scala:109) >>> at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) >>> at kafka.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:108) >>> at kafka.server.AbstractFetcherThread.processFetchRequest( >>> AbstractFetcherThread.scala:94) >>> at kafka.server.AbstractFetcherThread.doWork( >>> AbstractFetcherThread.scala:86) >>> at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:51) >>> >>> >>> Does it effect anything. I havent looked at it as it was just a warning. >>> Should i be worried about this? >>> >>> >>> Thanks >>> Arjun Narasimha kota >>> >>> On Tuesday 25 February 2014 03:45 PM, Arjun wrote: >>> >>>> Hi, >>>> >>>> I am using kafka 0.8. I have 3 brokers on three systems and 3 >>>> zookeepers running. >>>> I am using the high level consumer which is in examples folder of >>>> kafka. I am able to push the messages into the queue, but retriving the >>>> messages is taking some time. Is there any way i can tune this. >>>> >>>> I get this info in the kafka console, does the consumer slowness >>>> because of this?? >>>> Reconnect due to socket error: null >>>> >>>> The producer is pushing the messages as i can see that using Consumer >>>> offset checker tool.I can also see there is a lag in the consumer messages >>>> in this. >>>> >>>> >>>> Thanks >>>> Arjun Narasimha Kota >>>> >>> >>> >> >