Re: Question on MessageSizeTooLargeException

2014-01-21 Thread Jun Rao
Great. Please open a jira and attach your patch there. Thanks, Jun On Mon, Jan 20, 2014 at 10:37 PM, Bae, Jae Hyeon wrote: > Nope, just packaging for Netflix cloud environment. > > The first one is, producer discovery(metadata.broker.list) is integrated > with Netflix Eureka. > The second one

Re: Question on MessageSizeTooLargeException

2014-01-20 Thread Bae, Jae Hyeon
Nope, just packaging for Netflix cloud environment. The first one is, producer discovery(metadata.broker.list) is integrated with Netflix Eureka. The second one is, yammer metric library is connected with Netflix Servo. Except these two big things, I fixed a few lines to fit into our monitoring en

Re: Question on MessageSizeTooLargeException

2014-01-20 Thread Jun Rao
What kind of customization are you performing? Are you changing the wire and on-disk protocols? Thanks, Jun On Mon, Jan 20, 2014 at 10:02 AM, Bae, Jae Hyeon wrote: > Due to short retention period, I don't have that log segment now. > > How I am developing kafka is, > > I forked apache/kafka i

Re: Question on MessageSizeTooLargeException

2014-01-20 Thread Bae, Jae Hyeon
Due to short retention period, I don't have that log segment now. How I am developing kafka is, I forked apache/kafka into my personal repo and customized a little bit. I kept tracking 0.8 branch but you seems moved to trunk branch. I will update it to trunk branch or 0.8.0 tag. Thank you Best,

Re: Question on MessageSizeTooLargeException

2014-01-20 Thread Jun Rao
Could you use our DumpLogSegment tool on the relevant log segment and see if the log is corrupted? Also, are you using the 0.8.0 release? Thanks, Jun On Sun, Jan 19, 2014 at 10:09 PM, Bae, Jae Hyeon wrote: > Hello > > I finally upgraded kafka 0.7 to kafka 0.8 and a few kafka 0.8 clusters are

Question on MessageSizeTooLargeException

2014-01-19 Thread Bae, Jae Hyeon
Hello I finally upgraded kafka 0.7 to kafka 0.8 and a few kafka 0.8 clusters are being tested now. Today, I got alerted with the following messages: "data": { "exceptionMessage": "Found a message larger than the maximum fetch size of this consumer on topic nf_errors_log partition 0 at fetch