Hello

I finally upgraded kafka 0.7 to kafka 0.8 and a few kafka 0.8 clusters are
being tested now.

Today, I got alerted with the following messages:

 "data": {
    "exceptionMessage": "Found a message larger than the maximum fetch size
of this consumer on topic nf_errors_log partition 0 at fetch offset
76736251. Increase the fetch size, or decrease the maximum message size the
broker will allow.",
    "exceptionStackTrace": "kafka.common.MessageSizeTooLargeException:
Found a message larger than the maximum fetch size of this consumer on
topic nf_errors_log partition 0 at fetch offset 76736251. Increase the
fetch size, or decrease the maximum message size the broker will allow.
    "exceptionType": "kafka.common.MessageSizeTooLargeException"
  },
  "description": "RuntimeException aborted realtime
processing[nf_errors_log]"

What I don't understand is, I am using all default properties, which means

broker's message.max.bytes is 1000000
consumer's fetch.message.max.bytes is 1024 * 1024 greater than broker's
message.max.bytes

How could this happen? I am using snappy compression.

Thank you
Best, Jae

Reply via email to