The RC candidate build is here: http://home.apache.org/~gwenshap/kafka-0.10.2.1-rc1/ <http://home.apache.org/~gwenshap/kafka-0.10.2.1-rc1/>
Eno > On 17 Apr 2017, at 17:20, Mahendra Kariya <mahendra.kar...@go-jek.com> wrote: > > Thanks! > > In the meantime, is the jar published somewhere on github or as a part of > build pipeline? > > On Mon, Apr 17, 2017 at 9:18 PM, Eno Thereska <eno.there...@gmail.com> > wrote: > >> Not yet, but as soon as 0.10.2 is voted it should be. Hopefully this week. >> >> Eno >>> On 17 Apr 2017, at 13:25, Mahendra Kariya <mahendra.kar...@go-jek.com> >> wrote: >>> >>> Are the bug fix releases published to Maven central repo? >>> >>> On Sat, Apr 1, 2017 at 12:26 PM, Eno Thereska <eno.there...@gmail.com> >>> wrote: >>> >>>> Hi Sachin, >>>> >>>> In the bug fix release for 0.10.2 (and in trunk) we have now set >>>> max.poll.interval to infinite since from our experience with streams >> this >>>> should not be something that users set: https://github.com/apache/ >>>> kafka/pull/2770/files <https://github.com/apache/kafka/pull/2770/files >>> . >>>> >>>> We're in the process of documenting that change. For now you can >> increase >>>> the request timeout without worrying about max.poll.interval anymore. In >>>> fact I'd suggest you also increase max.poll.interval as we've done it >> above. >>>> >>>> Thanks >>>> Eno >>>> >>>>> On 1 Apr 2017, at 03:28, Sachin Mittal <sjmit...@gmail.com> wrote: >>>>> >>>>> Should this timeout be less than max poll interval value? if yes than >>>>> generally speaking what should be the ratio between two or range for >> this >>>>> timeout value . >>>>> >>>>> Thanks >>>>> Sachin >>>>> >>>>> On 1 Apr 2017 04:57, "Matthias J. Sax" <matth...@confluent.io> wrote: >>>>> >>>>> Yes, you can increase ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG >>>>> >>>>> >>>>> -Matthias >>>>> >>>>> >>>>> On 3/31/17 11:32 AM, Sachin Mittal wrote: >>>>>> Hi, >>>>>> So I have added the config ProducerConfig.RETRIES_CONFIG, >>>>> Integer.MAX_VALUE >>>>>> and the NotLeaderForPartitionException is gone. >>>>>> >>>>>> However we see a new exception especially under heavy load: >>>>>> org.apache.kafka.streams.errors.StreamsException: task [0_1] >> exception >>>>>> caught when producing >>>>>> at >>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl. >>>>> checkForException(RecordCollectorImpl.java:119) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>> at >>>>>> org.apache.kafka.streams.processor.internals. >> RecordCollectorImpl.flush( >>>>> RecordCollectorImpl.java:127) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] at >>>>>> org.apache.kafka.streams.processor.internals. >>>> StreamTask$1.run(StreamTask. >>>>> java:76) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>> at >>>>>> org.apache.kafka.streams.processor.internals.StreamsMetricsImpl. >>>>> measureLatencyNs(StreamsMetricsImpl.java:188) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>> at >>>>>> org.apache.kafka.streams.processor.internals. >>>> StreamTask.commit(StreamTask. >>>>> java:280) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] at >>>>>> org.apache.kafka.streams.processor.internals.StreamThread.commitOne( >>>>> StreamThread.java:787) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>> at >>>>>> org.apache.kafka.streams.processor.internals.StreamThread.commitAll( >>>>> StreamThread.java:774) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] at >>>>>> org.apache.kafka.streams.processor.internals. >> StreamThread.maybeCommit( >>>>> StreamThread.java:749) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>> at >>>>>> org.apache.kafka.streams.processor.internals.StreamThread.runLoop( >>>>> StreamThread.java:671) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] at >>>>>> org.apache.kafka.streams.processor.internals. >>>>> StreamThread.run(StreamThread.java:378) >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>> org.apache.kafka.common.errors.TimeoutException: Expiring 1 record(s) >>>> for >>>>>> new-part-advice-key-table-changelog-1: 30001 ms has passed since last >>>>> append >>>>>> >>>>>> So any idea as why TimeoutException is happening. >>>>>> Is this controlled by >>>>>> ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG >>>>>> >>>>>> If yes >>>>>> What should the value be set in this given that out consumer >>>>>> max.poll.interval.ms is defaul 5 minutes. >>>>>> >>>>>> Is there any other setting that we should try to avoid such errors >> which >>>>>> causes stream thread to die. >>>>>> >>>>>> Thanks >>>>>> Sachin >>>>>> >>>>>> >>>>>> On Sun, Mar 26, 2017 at 1:39 AM, Eno Thereska <eno.there...@gmail.com >>> >>>>>> wrote: >>>>>> >>>>>>> Hi Sachin, >>>>>>> >>>>>>> Not in this case. >>>>>>> >>>>>>> Thanks >>>>>>> Eno >>>>>>> >>>>>>>> On Mar 25, 2017, at 6:19 PM, Sachin Mittal <sjmit...@gmail.com> >>>> wrote: >>>>>>>> >>>>>>>> OK. >>>>>>>> I will try this out. >>>>>>>> >>>>>>>> Do I need to change anything for >>>>>>>> max.in.flight.requests.per.connection >>>>>>>> >>>>>>>> Thanks >>>>>>>> Sachin >>>>>>>> >>>>>>>> >>>>>>>> On Sat, Mar 25, 2017 at 10:59 PM, Eno Thereska < >>>> eno.there...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Sachin, >>>>>>>>> >>>>>>>>> For this particular error, “org.apache.kafka.common.errors. >>>>>>>>> NotLeaderForPartitionException: This server is not the leader for >>>> that >>>>>>>>> topic-partition.”, could you try setting the number of retries to >>>>>>> something >>>>>>>>> large like this: >>>>>>>>> >>>>>>>>> Properties props = new Properties(); >>>>>>>>> props.put(StreamsConfig.APPLICATION_ID_CONFIG, applicationId); >>>>>>>>> ... >>>>>>>>> props.put(ProducerConfig.RETRIES_CONFIG, Integer.MAX_VALUE); >>>>>>>>> >>>>>>>>> This will retry the produce requests and should hopefully solve >> your >>>>>>>>> immediate problem. >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> Eno >>>>>>>>> >>>>>>>>> >>>>>>>>> On 25/03/2017, 08:35, "Sachin Mittal" <sjmit...@gmail.com> wrote: >>>>>>>>> >>>>>>>>> Hi, >>>>>>>>> We have encountered another case of series of errors which I would >>>>>>> need >>>>>>>>> more help in understanding. >>>>>>>>> >>>>>>>>> In logs we see message like this: >>>>>>>>> ERROR 2017-03-25 03:41:40,001 [kafka-producer-network-thread | >>>>>>>>> 85-StreamThread-3-producer]: >>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl >> - >>>>>>>>> task >>>>>>>>> [0_1] Error sending record to topic new-part-advice-key-table- >>>>>>> changelog. >>>>>>>>> No >>>>>>>>> more offsets will be recorded for this task and the exception will >>>>>>>>> eventually be thrown >>>>>>>>> >>>>>>>>> then some millisecond later >>>>>>>>> ERROR 2017-03-25 03:41:40,149 [StreamThread-3]: >>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread - >>>>>>>>> stream-thread >>>>>>>>> [StreamThread-3] Failed while executing StreamTask 0_1 due to >> flush >>>>>>>>> state: >>>>>>>>> org.apache.kafka.streams.errors.StreamsException: task [0_1] >>>>>>> exception >>>>>>>>> caught when producing >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl. >>>>>>>>> checkForException(RecordCollectorImpl.java:119) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>>>>>>> RecordCollectorImpl.flush(RecordCollectorImpl.java:127) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>> StreamTask.flushState( >>>>>>>>> StreamTask.java:422) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >> StreamThread$4.apply( >>>>>>>>> StreamThread.java:555) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread. >>>>>>>>> performOnAllTasks(StreamThread.java:513) >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>>>>>>> StreamThread.flushAllState(StreamThread.java:551) >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread. >>>>>>>>> shutdownTasksAndState(StreamThread.java:463) >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>> StreamThread.shutdown( >>>>>>>>> StreamThread.java:408) >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>>>>>>> StreamThread.run(StreamThread.java:389) >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> org.apache.kafka.common.errors.NotLeaderForPartitionException: >>>> This >>>>>>>>> server >>>>>>>>> is not the leader for that topic-partition. >>>>>>>>> >>>>>>>>> finally we get this >>>>>>>>> ERROR 2017-03-25 03:41:45,724 [StreamThread-3]: >>>>>>>>> com.advice.TestKafkaAdvice >>>>>>>>> - Uncaught exception: >>>>>>>>> org.apache.kafka.streams.errors.StreamsException: Exception >> caught >>>>>>> in >>>>>>>>> process. taskId=0_1, processor=KSTREAM-SOURCE-0000000000, >>>>>>>>> topic=advice-stream, partition=1, offset=48062286 >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>>>>>>> StreamTask.process(StreamTask.java:216) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >> StreamThread.runLoop( >>>>>>>>> StreamThread.java:651) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>>>>>>> StreamThread.run(StreamThread.java:378) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> Caused by: org.apache.kafka.streams.errors.StreamsException: task >>>>>>>>> [0_1] >>>>>>>>> exception caught when producing >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl. >>>>>>>>> checkForException(RecordCollectorImpl.java:119) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>>>>> RecordCollectorImpl.send( >>>>>>>>> RecordCollectorImpl.java:76) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> at >>>>>>>>> org.apache.kafka.streams.processor.internals. >>>>>>> RecordCollectorImpl.send( >>>>>>>>> RecordCollectorImpl.java:64) >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] >>>>>>>>> >>>>>>>>> >>>>>>>>> Again it is not clear why in this case we need to shut down the >>>>>>> steams >>>>>>>>> thread and eventually the application. Shouldn't we capture this >>>>>>> error >>>>>>>>> too? >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> Sachin >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>> >>>> >> >>