There is a setting that controls the maximum message size. This is to
ensure the messages can be read on the server and by all consumers without
running out of memory or exceeding the consumer fetch size. In 0.7.x this
setting is controlled by the broker configuration max.message.size.

-Jay


On Tue, Jan 29, 2013 at 12:18 AM, Bo Sun <docsun...@gmail.com> wrote:

> hi all
> i've got a exception .
> kafka.common.MessageSizeTooLargeException: payload size of 1772597 larger
> than 1000000
>         at
>
> kafka.message.ByteBufferMessageSet.verifyMessageSize(ByteBufferMessageSet.scala:93)
>         at kafka.producer.SyncProducer.send(SyncProducer.scala:122)
>         at
>
> kafka.producer.ProducerPool$$anonfun$send$1.apply$mcVI$sp(ProducerPool.scala:114)
>         at
> kafka.producer.ProducerPool$$anonfun$send$1.apply(ProducerPool.scala:100)
>         at
> kafka.producer.ProducerPool$$anonfun$send$1.apply(ProducerPool.scala:100)
>         at
>
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:57)
>         at
> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:43)
>         at kafka.producer.ProducerPool.send(ProducerPool.scala:100)
>         at kafka.producer.Producer.zkSend(Producer.scala:140)
>         at kafka.producer.Producer.send(Producer.scala:99)
>         at kafka.javaapi.producer.Producer.send(Producer.scala:103)
>
> i dont know why. plz help me .thanks
>

Reply via email to