Thanks for the details Luke. At what point would you consider a message too big?
Are you using compression? Bert On Thursday, June 26, 2014, Luke Forehand < luke.foreh...@networkedinsights.com> wrote: > I have used 50MB message size and it is not a great idea. First of all > you need to make sure you have these settings in sync: > message.max.bytes > replica.fetch.max.bytes > fetch.message.max.bytes > > I had not set the replica fetch setting and didn't realize one of my > partitions was not replicating after a large message was produced. I also > ran into heap issues with having to fetch such a large message, lots of > unnecessary garbage collection. I suggest breaking down your message into > smaller chunks. In my case, I decided to break an XML input stream (which > had a root element wrapping a ridiculously large number of children) into > smaller messages, having to parse the large xml root document and re-wrap > each child element with a shallow clone of its parents as I iterated the > stream. > > -Luke > > ________________________________________ > From: Denny Lee <denny.g....@gmail.com <javascript:;>> > Sent: Tuesday, June 24, 2014 10:35 AM > To: users@kafka.apache.org <javascript:;> > Subject: Experiences with larger message sizes > > By any chance has anyone worked with using Kafka with message sizes that > are approximately 50MB in size? Based on from some of the previous threads > there are probably some concerns on memory pressure due to the compression > on the broker and decompression on the consumer and a best practices on > ensuring batch size (to ultimately not have the compressed message exceed > message size limit). > > Any other best practices or thoughts concerning this scenario? > > Thanks! > Denny >