sure

On Tue, Nov 17, 2015 at 11:30 AM, Gyula Fóra <gyula.f...@gmail.com> wrote:

> Should I open a JIRA for this?
>
> Gyula Fóra <gyula.f...@gmail.com> ezt írta (időpont: 2015. nov. 17., K,
> 11:30):
>
> > Thanks for the quick response and thorough explanation :)
> >
> > Gyula
> >
> > Robert Metzger <rmetz...@apache.org> ezt írta (időpont: 2015. nov. 17.,
> > K, 11:27):
> >
> >> I would try that approach first
> >>
> >> On Tue, Nov 17, 2015 at 11:26 AM, Stephan Ewen <se...@apache.org>
> wrote:
> >>
> >> > Another idea: The bug is fixed in the Zookeeper Client that Kafka
> uses,
> >> so
> >> > if we can bump the transitive dependency, that might fix it...
> >> >
> >> > On Tue, Nov 17, 2015 at 11:19 AM, Stephan Ewen <se...@apache.org>
> >> wrote:
> >> >
> >> > > Hey!
> >> > >
> >> > > The problem here is that there is no such thing as proper thread
> >> killing
> >> > > in Java (at least it makes everything unstable if you do). Threads
> >> need
> >> > to
> >> > > exit cooperatively.
> >> > >
> >> > > The Kafka Function calls simply are uninterruptibly stuck and never
> >> > return
> >> > > (pretty bad bug in their Zookeeper Client). As far as I know one
> >> cannot
> >> > > clean this up properly unless one kills the process.
> >> > >
> >> > > We could try and work around this by running the Zookeeper commit
> in a
> >> > > dedicated lightweight thread that shares no resources and thus does
> >> not
> >> > > make the system unstable if stopped (against better advise ;-) )
> >> > >
> >> > > Stephan
> >> > >
> >> > >
> >> > >
> >> > >
> >> > >
> >> > > On Tue, Nov 17, 2015 at 10:43 AM, Ufuk Celebi <u...@apache.org>
> wrote:
> >> > >
> >> > >> https://issues.apache.org/jira/browse/KAFKA-824
> >> > >>
> >> > >> This has been fixed for Kafka’s 0.9.0 version.
> >> > >>
> >> > >> We should investigate why the job gets stuck though. Do you have a
> >> stack
> >> > >> trace or any logs available?
> >> > >>
> >> > >> – Ufuk
> >> > >>
> >> > >> > On 17 Nov 2015, at 09:24, Gyula Fóra <gyf...@apache.org> wrote:
> >> > >> >
> >> > >> > Hey guys,
> >> > >> >
> >> > >> > I ran into some issue with the kafka consumers.
> >> > >> >
> >> > >> > I am reading from more than 50 topics with parallelism 1, and
> while
> >> > >> running
> >> > >> > the job I got the following exception during the checkpoint
> >> > notification
> >> > >> > (offset committing):
> >> > >> >
> >> > >> > java.lang.RuntimeException: Error while confirming checkpoint
> >> > >> > at org.apache.flink.runtime.taskmanager.Task$2.run(Task.java:935)
> >> > >> > at
> >> > >>
> >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> >> > >> > at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> >> > >> > at java.lang.Thread.run(Thread.java:745)
> >> > >> > Caused by: java.lang.NullPointerException
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> org.I0Itec.zkclient.ZkConnection.writeDataReturnStat(ZkConnection.java:115)
> >> > >> > at org.I0Itec.zkclient.ZkClient$10.call(ZkClient.java:817)
> >> > >> > at
> >> org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:675)
> >> > >> > at
> >> org.I0Itec.zkclient.ZkClient.writeDataReturnStat(ZkClient.java:813)
> >> > >> > at org.I0Itec.zkclient.ZkClient.writeData(ZkClient.java:808)
> >> > >> > at org.I0Itec.zkclient.ZkClient.writeData(ZkClient.java:777)
> >> > >> > at kafka.utils.ZkUtils$.updatePersistentPath(ZkUtils.scala:332)
> >> > >> > at kafka.utils.ZkUtils.updatePersistentPath(ZkUtils.scala)
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> org.apache.flink.streaming.connectors.kafka.internals.ZookeeperOffsetHandler.setOffsetInZooKeeper(ZookeeperOffsetHandler.java:112)
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> org.apache.flink.streaming.connectors.kafka.internals.ZookeeperOffsetHandler.commit(ZookeeperOffsetHandler.java:80)
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer.notifyCheckpointComplete(FlinkKafkaConsumer.java:542)
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.notifyOfCompletedCheckpoint(AbstractUdfStreamOperator.java:176)
> >> > >> > at
> >> > >> >
> >> > >>
> >> >
> >>
> org.apache.flink.streaming.runtime.tasks.StreamTask.notifyCheckpointComplete(StreamTask.java:478)
> >> > >> > at org.apache.flink.runtime.taskmanager.Task$2.run(Task.java:931)
> >> > >> > ... 5 more
> >> > >> >
> >> > >> > This happened at the same time on multiple kafka consumers. Could
> >> this
> >> > >> be
> >> > >> > some Zookeeper related issue? Maybe we should be aware of this.
> >> > >> >
> >> > >> > Another problem is that subsequently the whole pipeline got stuck
> >> at
> >> > the
> >> > >> > sources while canceling so the job could never restart. Maybe it
> >> would
> >> > >> be
> >> > >> > worth killing the whole thing and restart in these situations?
> >> > >> >
> >> > >> > Thanks,
> >> > >> > Gyula
> >> > >>
> >> > >>
> >> > >
> >> >
> >>
> >
>

Reply via email to