Thanks!

In the meantime, is the jar published somewhere on github or as a part of
build pipeline?

On Mon, Apr 17, 2017 at 9:18 PM, Eno Thereska <eno.there...@gmail.com>
wrote:

> Not yet, but as soon as 0.10.2 is voted it should be. Hopefully this week.
>
> Eno
> > On 17 Apr 2017, at 13:25, Mahendra Kariya <mahendra.kar...@go-jek.com>
> wrote:
> >
> > Are the bug fix releases published to Maven central repo?
> >
> > On Sat, Apr 1, 2017 at 12:26 PM, Eno Thereska <eno.there...@gmail.com>
> > wrote:
> >
> >> Hi Sachin,
> >>
> >> In the bug fix release for 0.10.2 (and in trunk) we have now set
> >> max.poll.interval to infinite since from our experience with streams
> this
> >> should not be something that users set: https://github.com/apache/
> >> kafka/pull/2770/files <https://github.com/apache/kafka/pull/2770/files
> >.
> >>
> >> We're in the process of documenting that change. For now you can
> increase
> >> the request timeout without worrying about max.poll.interval anymore. In
> >> fact I'd suggest you also increase max.poll.interval as we've done it
> above.
> >>
> >> Thanks
> >> Eno
> >>
> >>> On 1 Apr 2017, at 03:28, Sachin Mittal <sjmit...@gmail.com> wrote:
> >>>
> >>> Should this timeout be less than max poll interval value? if yes than
> >>> generally speaking what should be the ratio between two or range for
> this
> >>> timeout value .
> >>>
> >>> Thanks
> >>> Sachin
> >>>
> >>> On 1 Apr 2017 04:57, "Matthias J. Sax" <matth...@confluent.io> wrote:
> >>>
> >>> Yes, you can increase ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
> >>>
> >>>
> >>> -Matthias
> >>>
> >>>
> >>> On 3/31/17 11:32 AM, Sachin Mittal wrote:
> >>>> Hi,
> >>>> So I have added the config ProducerConfig.RETRIES_CONFIG,
> >>> Integer.MAX_VALUE
> >>>> and the NotLeaderForPartitionException is gone.
> >>>>
> >>>> However we see a new exception especially under heavy load:
> >>>> org.apache.kafka.streams.errors.StreamsException: task [0_1]
> exception
> >>>> caught when producing
> >>>> at
> >>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>> checkForException(RecordCollectorImpl.java:119)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>> at
> >>>> org.apache.kafka.streams.processor.internals.
> RecordCollectorImpl.flush(
> >>> RecordCollectorImpl.java:127)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>> org.apache.kafka.streams.processor.internals.
> >> StreamTask$1.run(StreamTask.
> >>> java:76)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>> at
> >>>> org.apache.kafka.streams.processor.internals.StreamsMetricsImpl.
> >>> measureLatencyNs(StreamsMetricsImpl.java:188)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>> at
> >>>> org.apache.kafka.streams.processor.internals.
> >> StreamTask.commit(StreamTask.
> >>> java:280)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>> org.apache.kafka.streams.processor.internals.StreamThread.commitOne(
> >>> StreamThread.java:787)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>> at
> >>>> org.apache.kafka.streams.processor.internals.StreamThread.commitAll(
> >>> StreamThread.java:774)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>> org.apache.kafka.streams.processor.internals.
> StreamThread.maybeCommit(
> >>> StreamThread.java:749)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>> at
> >>>> org.apache.kafka.streams.processor.internals.StreamThread.runLoop(
> >>> StreamThread.java:671)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>> org.apache.kafka.streams.processor.internals.
> >>> StreamThread.run(StreamThread.java:378)
> >>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>> org.apache.kafka.common.errors.TimeoutException: Expiring 1 record(s)
> >> for
> >>>> new-part-advice-key-table-changelog-1: 30001 ms has passed since last
> >>> append
> >>>>
> >>>> So any idea as why TimeoutException is happening.
> >>>> Is this controlled by
> >>>> ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
> >>>>
> >>>> If yes
> >>>> What should the value be set in this given that out consumer
> >>>> max.poll.interval.ms is defaul 5 minutes.
> >>>>
> >>>> Is there any other setting that we should try to avoid such errors
> which
> >>>> causes stream thread to die.
> >>>>
> >>>> Thanks
> >>>> Sachin
> >>>>
> >>>>
> >>>> On Sun, Mar 26, 2017 at 1:39 AM, Eno Thereska <eno.there...@gmail.com
> >
> >>>> wrote:
> >>>>
> >>>>> Hi Sachin,
> >>>>>
> >>>>> Not in this case.
> >>>>>
> >>>>> Thanks
> >>>>> Eno
> >>>>>
> >>>>>> On Mar 25, 2017, at 6:19 PM, Sachin Mittal <sjmit...@gmail.com>
> >> wrote:
> >>>>>>
> >>>>>> OK.
> >>>>>> I will try this out.
> >>>>>>
> >>>>>> Do I need to change anything for
> >>>>>> max.in.flight.requests.per.connection
> >>>>>>
> >>>>>> Thanks
> >>>>>> Sachin
> >>>>>>
> >>>>>>
> >>>>>> On Sat, Mar 25, 2017 at 10:59 PM, Eno Thereska <
> >> eno.there...@gmail.com>
> >>>>>> wrote:
> >>>>>>
> >>>>>>> Hi Sachin,
> >>>>>>>
> >>>>>>> For this particular error, “org.apache.kafka.common.errors.
> >>>>>>> NotLeaderForPartitionException: This server is not the leader for
> >> that
> >>>>>>> topic-partition.”, could you try setting the number of retries to
> >>>>> something
> >>>>>>> large like this:
> >>>>>>>
> >>>>>>> Properties props = new Properties();
> >>>>>>> props.put(StreamsConfig.APPLICATION_ID_CONFIG, applicationId);
> >>>>>>> ...
> >>>>>>> props.put(ProducerConfig.RETRIES_CONFIG, Integer.MAX_VALUE);
> >>>>>>>
> >>>>>>> This will retry the produce requests and should hopefully solve
> your
> >>>>>>> immediate problem.
> >>>>>>>
> >>>>>>> Thanks
> >>>>>>> Eno
> >>>>>>>
> >>>>>>>
> >>>>>>> On 25/03/2017, 08:35, "Sachin Mittal" <sjmit...@gmail.com> wrote:
> >>>>>>>
> >>>>>>>  Hi,
> >>>>>>>  We have encountered another case of series of errors which I would
> >>>>> need
> >>>>>>>  more help in understanding.
> >>>>>>>
> >>>>>>>  In logs we see message like this:
> >>>>>>>  ERROR 2017-03-25 03:41:40,001 [kafka-producer-network-thread |
> >>>>>>>  85-StreamThread-3-producer]:
> >>>>>>>  org.apache.kafka.streams.processor.internals.RecordCollectorImpl
> -
> >>>>>>> task
> >>>>>>>  [0_1] Error sending record to topic new-part-advice-key-table-
> >>>>> changelog.
> >>>>>>> No
> >>>>>>>  more offsets will be recorded for this task and the exception will
> >>>>>>>  eventually be thrown
> >>>>>>>
> >>>>>>>  then some millisecond later
> >>>>>>>  ERROR 2017-03-25 03:41:40,149 [StreamThread-3]:
> >>>>>>>  org.apache.kafka.streams.processor.internals.StreamThread -
> >>>>>>> stream-thread
> >>>>>>>  [StreamThread-3] Failed while executing StreamTask 0_1 due to
> flush
> >>>>>>> state:
> >>>>>>>  org.apache.kafka.streams.errors.StreamsException: task [0_1]
> >>>>> exception
> >>>>>>>  caught when producing
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>>>>>> checkForException(RecordCollectorImpl.java:119)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >>>>>>> RecordCollectorImpl.flush(RecordCollectorImpl.java:127)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >> StreamTask.flushState(
> >>>>>>> StreamTask.java:422)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> StreamThread$4.apply(
> >>>>>>> StreamThread.java:555)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.StreamThread.
> >>>>>>> performOnAllTasks(StreamThread.java:513)
> >>>>>>>  [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >>>>>>> StreamThread.flushAllState(StreamThread.java:551)
> >>>>>>>  [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.StreamThread.
> >>>>>>> shutdownTasksAndState(StreamThread.java:463)
> >>>>>>>  [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >> StreamThread.shutdown(
> >>>>>>> StreamThread.java:408)
> >>>>>>>  [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >>>>>>> StreamThread.run(StreamThread.java:389)
> >>>>>>>  [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>  org.apache.kafka.common.errors.NotLeaderForPartitionException:
> >> This
> >>>>>>> server
> >>>>>>>  is not the leader for that topic-partition.
> >>>>>>>
> >>>>>>>  finally we get this
> >>>>>>>  ERROR 2017-03-25 03:41:45,724 [StreamThread-3]:
> >>>>>>> com.advice.TestKafkaAdvice
> >>>>>>>  - Uncaught exception:
> >>>>>>>  org.apache.kafka.streams.errors.StreamsException: Exception
> caught
> >>>>> in
> >>>>>>>  process. taskId=0_1, processor=KSTREAM-SOURCE-0000000000,
> >>>>>>>  topic=advice-stream, partition=1, offset=48062286
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >>>>>>> StreamTask.process(StreamTask.java:216)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> StreamThread.runLoop(
> >>>>>>> StreamThread.java:651)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >>>>>>> StreamThread.run(StreamThread.java:378)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>  Caused by: org.apache.kafka.streams.errors.StreamsException: task
> >>>>>>> [0_1]
> >>>>>>>  exception caught when producing
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>>>>>> checkForException(RecordCollectorImpl.java:119)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >>>>> RecordCollectorImpl.send(
> >>>>>>> RecordCollectorImpl.java:76)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>      at
> >>>>>>>  org.apache.kafka.streams.processor.internals.
> >>>>> RecordCollectorImpl.send(
> >>>>>>> RecordCollectorImpl.java:64)
> >>>>>>>  ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>
> >>>>>>>
> >>>>>>>  Again it is not clear why in this case we need to shut down the
> >>>>> steams
> >>>>>>>  thread and eventually the application. Shouldn't we capture this
> >>>>> error
> >>>>>>> too?
> >>>>>>>
> >>>>>>>  Thanks
> >>>>>>>  Sachin
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>
> >>>>>
> >>>>
> >>
> >>
>
>

Reply via email to