Hi Grant, Yes, I saw exception in Spark and Kafka. In Kafka server logs I get this exception: kafka.common.OffsetOutOfRangeException: Request for offset 2823 but we only have log segments in the range 2824 to 2824. at kafka.log.Log.read(Log.scala:380) at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:530) at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:476) at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:471) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) at scala.collection.immutable.Map$Map1.foreach(Map.scala:105) at scala.collection.TraversableLike$class.map(TraversableLike.scala:206) at scala.collection.immutable.Map$Map1.map(Map.scala:93) at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:471) at kafka.server.KafkaApis$FetchRequestPurgatory.expire(KafkaApis.scala:783) at kafka.server.KafkaApis$FetchRequestPurgatory.expire(KafkaApis.scala:765) at kafka.server.RequestPurgatory$ExpiredRequestReaper.run(RequestPurgatory.sca
Similar kind of exception comes to Spark Job. Here are my versions : Spark - 1.4.1 Kafka - 0.8.1 I changed retention on config using this command : ./kafka-topics.sh --alter --zookeeper XXX:2181 --topic MyTopic --config retention.ms=1440 (I believe this is in minutes) I am also noticing something in Kafka. When I run below command on broker: ./kafka-run-class.sh kafka.tools.GetOffsetShell --broker-list vdc-vm8.apple.com:9092 --topic MyTopic --time -2 Earliest offset is being set to latest just in few seconds. Am I co-relating this issue correctly? Here is my example on a new Topic. Initial output of this command is ./kafka-run-class.sh kafka.tools.GetOffsetShell --broker-list vdc-vm8.apple.com:9092 --topic MyTopic --time -2 SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. MyTopic:0:60 I published 4 messages to Kafka. Immediately after few seconds, command output is: MyTopic:0:64 Isn't this supposed to stay at 60 for longer time based on retention policy? Thanks, Leena On Thu, Aug 6, 2015 at 12:09 PM, Grant Henke <ghe...@cloudera.com> wrote: > Does this Spark Jira match up with what you are seeing or sound related? > https://issues.apache.org/jira/browse/SPARK-8474 > > What versions of Spark and Kafka are you using? Can you include more of the > spark log? Any errors shown in the Kafka log? > > Thanks, > Grant > > On Thu, Aug 6, 2015 at 1:17 PM, Cassa L <lcas...@gmail.com> wrote: > > > Hi, > > Has anyone tried streaming API of Spark with Kafka? I am experimenting > new > > Spark API to read from Kafka. > > KafkaUtils.createDirectStream(...) > > > > Every now and then, I get following error "spark > > kafka.common.OffsetOutOfRangeException" and my spark script stops > working. > > I have simple topic with just one partition. > > > > I would appreciate any clues on how to debug this issue. > > > > Thanks, > > LCassa > > > > > > -- > Grant Henke > Software Engineer | Cloudera > gr...@cloudera.com | twitter.com/gchenke | linkedin.com/in/granthenke >