I think that actually there's an Exception thrown within the code that I
suspect it's not reported anywhere..could it be?

On Mon, Jun 29, 2015 at 3:28 PM, Flavio Pompermaier <pomperma...@okkam.it>
wrote:

> Which file and which JVM options do I have to modify to try options 1 and
> 3..?
>
>    1. Don't fill the JVMs up to the limit with objects. Give more memory
>    to the JVM, or give less memory to Flink managed memory
>    2. Use more JVMs, i.e., a higher parallelism
>    3. Use a concurrent garbage collector, like G1
>
> Actually, when I run the code from Eclipse I see an exception do to an
> error in the data (because I try to read a URI that contains illegal
> characters) but I don't think the program reach that point, I don't see
> anywhere an exception and the error occur later on in the code..
>
> However, all of your options seems related to a scalability problem, where
> I should add more resources to complete the work...while it works locally
> in the IDE where I have less resources (except the gc that I use default
> settings while I don't know if the cluster has some default ones)..isn't it
> strange?
>
> On Mon, Jun 29, 2015 at 2:29 PM, Stephan Ewen <se...@apache.org> wrote:
>
>> Hi Flavio!
>>
>> I had a look at the logs. There seems nothing suspicious - at some point,
>> the TaskManager and JobManager declare each other unreachable.
>>
>> A pretty common cause for that is that the JVMs stall for a long time due
>> to garbage collection. The JobManager cannot see the difference between a
>> JVM that is irresponsive (due to garbage collection) and a JVM that is dead.
>>
>> Here is what you can do to prevent long garbage collection stalls:
>>
>>  - Don't fill the JVMs up to the limit with objects. Give more memory to
>> the JVM, or give less memory to Flink managed memory.
>>  - Use more JVMs, i.e., a higher parallelism.
>>  - Use a concurrent garbage collector, like G1.
>>
>>
>> Greetings,
>> Stephan
>>
>>
>> On Mon, Jun 29, 2015 at 12:39 PM, Stephan Ewen <se...@apache.org> wrote:
>>
>>> Hi Flavio!
>>>
>>> Can you post the JobManager's log here? It should have the message about
>>> what is going wrong...
>>>
>>> Stephan
>>>
>>>
>>> On Mon, Jun 29, 2015 at 11:43 AM, Flavio Pompermaier <
>>> pomperma...@okkam.it> wrote:
>>>
>>>> Hi to all,
>>>>
>>>> I'm restarting the discussion about a problem I alredy dicussed on this
>>>> mailing list (but that started with a different subject).
>>>> I'm running Flink 0.9.0 on CDH 5.1.3 so I compiled the sources as:
>>>>
>>>> mvn clean  install -Dhadoop.version=2.3.0-cdh5.1.3
>>>> -Dhbase.version=0.98.1-cdh5.1.3 -Dhadoop.core.version=2.3.0-mr1-cdh5.1.3
>>>> -DskipTests -Pvendor-repos
>>>>
>>>> The problem I'm facing is that the cluster start successfully but when
>>>> I run my job (from the web-client) I get, after some time, this exception:
>>>>
>>>> 16:35:41,636 WARN  akka.remote.RemoteWatcher
>>>>           - Detected unreachable: [akka.tcp://flink@192.168.234.83:6123
>>>> ]
>>>> 16:35:46,605 INFO  org.apache.flink.runtime.taskmanager.TaskManager   -
>>>> Disconnecting from JobManager: JobManager is no longer reachable
>>>> 16:35:46,614 INFO  org.apache.flink.runtime.taskmanager.TaskManager   -
>>>> Cancelling all computations and discarding all cached data.
>>>> 16:35:46,644 INFO  org.apache.flink.runtime.taskmanager.Task
>>>>       - Attempting to fail task externally CHAIN GroupReduce (GroupReduce
>>>> at compactDataSources(MyClass.java:213)) -> Combine(Distinct at
>>>> compactDataSources(MyClass.java:213)) (8/36)
>>>> 16:35:46,669 INFO  org.apache.flink.runtime.taskmanager.Task
>>>>       - CHAIN GroupReduce (GroupReduce at
>>>> compactDataSources(MyClass.java:213)) -> Combine(Distinct at
>>>> compactDataSources(MyClass.java:213)) (8/36) switched to FAILED with
>>>> exception.
>>>> java.lang.Exception: Disconnecting from JobManager: JobManager is no
>>>> longer reachable
>>>>         at org.apache.flink.runtime.taskmanager.TaskManager.org
>>>> $apache$flink$runtime$taskmanager$TaskManager$$handleJobManagerDisconnect(TaskManager.scala:741)
>>>>         at
>>>> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:267)
>>>>         at
>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>>>>         at
>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>>>>         at
>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>>>>         at
>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36)
>>>>         at
>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)
>>>>         at
>>>> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>>>>         at
>>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)
>>>>         at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>>>>         at
>>>> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:114)
>>>>         at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>>>>         at
>>>> akka.actor.dungeon.DeathWatch$class.receivedTerminated(DeathWatch.scala:46)
>>>>         at akka.actor.ActorCell.receivedTerminated(ActorCell.scala:369)
>>>>         at akka.actor.ActorCell.autoReceiveMessage(ActorCell.scala:501)
>>>>         at akka.actor.ActorCell.invoke(ActorCell.scala:486)
>>>>         at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>>>>         at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>>>>         at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>>>>         at
>>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>>>         at
>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)
>>>>         at
>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)
>>>>         at
>>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>>>         at
>>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>>> 16:35:46,767 INFO  org.apache.flink.runtime.taskmanager.Task
>>>>           - Triggering cancellation of task code CHAIN GroupReduce
>>>> (GroupReduce at compactDataSources(MyClass.java:213)) -> Combine(Distinct
>>>> at compactDataSources(MyClass.java:213)) (8/36)
>>>> (57a0ad78726d5ba7255aa87038250c51).
>>>>
>>>> The job instead runs correctly from the IDE (Eclipse). How can I
>>>> understand/debug what's wrong?
>>>>
>>>> Best,
>>>> Flavio
>>>>
>>>>
>>>
>>
>
>

Reply via email to