Dmitri,

Could you reproduce this easily? Are you using a load balancer? Earlier,
another user had the same issue and eventually figured out that the problem
is in the network router.

Thanks,

Jun

On Wed, Nov 28, 2012 at 11:34 AM, Dmitri Priimak <
prii...@highwire.stanford.edu> wrote:

> Hi.
>
> In the kafka broker (version 0.7.0) log I see occasionally following error
> message
>
>  FATAL Halting due to unrecoverable I/O error while handling producer
> request: Unexpected end of
> ZLIB input stream (kafka.server.KafkaRequestHandlers)
> java.io.EOFException: Unexpected end of ZLIB input stream
>         at
> java.util.zip.InflaterInputStream.fill(InflaterInputStream.java:223)
>         at
> java.util.zip.InflaterInputStream.read(InflaterInputStream.java:141)
>         at java.util.zip.GZIPInputStream.read(GZIPInputStream.java:90)
>         at java.io.FilterInputStream.read(FilterInputStream.java:90)
>         at
> kafka.message.CompressionUtils$$anonfun$decompress$4.apply$mcI$sp(CompressionUtils.scala:123)
>         at
> kafka.message.CompressionUtils$$anonfun$decompress$4.apply(CompressionUtils.scala:123)
>         at
> kafka.message.CompressionUtils$$anonfun$decompress$4.apply(CompressionUtils.scala:123)
>         at
> scala.collection.immutable.Stream$$anonfun$continually$1.apply(Stream.scala:598)
>         at
> scala.collection.immutable.Stream$$anonfun$continually$1.apply(Stream.scala:598)
>         at scala.collection.immutable.Stream$Cons.tail(Stream.scala:555)
>         at scala.collection.immutable.Stream$Cons.tail(Stream.scala:549)
>         at
> scala.collection.immutable.Stream$$anonfun$takeWhile$1.apply(Stream.scala:394)
>         at
> scala.collection.immutable.Stream$$anonfun$takeWhile$1.apply(Stream.scala:394)
>         at scala.collection.immutable.Stream$Cons.tail(Stream.scala:555)
>         at scala.collection.immutable.Stream$Cons.tail(Stream.scala:549)
>         at scala.collection.immutable.Stream.foreach(Stream.scala:255)
>         at
> kafka.message.CompressionUtils$.decompress(CompressionUtils.scala:123)
>         at
> kafka.message.ByteBufferMessageSet$$anon$1.makeNextOuter(ByteBufferMessageSet.scala:124)
>         at
> kafka.message.ByteBufferMessageSet$$anon$1.makeNext(ByteBufferMessageSet.scala:138)
>         at
> kafka.message.ByteBufferMessageSet$$anon$1.makeNext(ByteBufferMessageSet.scala:82)
>         at
> kafka.utils.IteratorTemplate.maybeComputeNext(IteratorTemplate.scala:59)
>         at kafka.utils.IteratorTemplate.hasNext(IteratorTemplate.scala:51)
>         at scala.collection.Iterator$class.foreach(Iterator.scala:631)
>         at kafka.utils.IteratorTemplate.foreach(IteratorTemplate.scala:30)
>         at
> scala.collection.IterableLike$class.foreach(IterableLike.scala:79)
>         at kafka.message.MessageSet.foreach(MessageSet.scala:87)
>         at kafka.log.Log.append(Log.scala:202)
>         at
>
> kafka.server.KafkaRequestHandlers.kafka$server$KafkaRequestHandlers$$handleProducerRequest(KafkaRequestHandlers.scala:75)
>         at
>
> kafka.server.KafkaRequestHandlers$$anonfun$handleMultiProducerRequest$1.apply(KafkaRequestHandlers.scala:68)
>         at
>
> kafka.server.KafkaRequestHandlers$$anonfun$handleMultiProducerRequest$1.apply(KafkaRequestHandlers.scala:68)
>         at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
>         at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
>         at
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34)
>         at scala.collection.mutable.ArrayOps.foreach(ArrayOps.scala:34)
>         at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
>         at scala.collection.mutable.ArrayOps.map(ArrayOps.scala:34)
>         at
> kafka.server.KafkaRequestHandlers.handleMultiProducerRequest(KafkaRequestHandlers.scala:68)
>         at
> kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$4.apply(KafkaRequestHandlers.scala:46)
>         at
> kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$4.apply(KafkaRequestHandlers.scala:46)
>         at kafka.network.Processor.handle(SocketServer.scala:289)
>         at kafka.network.Processor.read(SocketServer.scala:312)
>         at kafka.network.Processor.run(SocketServer.scala:207)
>         at java.lang.Thread.run(Thread.java:662)
>
> At which point broker actually dies. Shouldn't it keep working even even
> if there is some such error?
> Also, does anyone else saw this error? And is it fixed in the newer
> versions?
>
> --
> Dmitri Priimak
>
>

Reply via email to