I think that actually there's an Exception thrown within the code that I suspect it's not reported anywhere..could it be?
On Mon, Jun 29, 2015 at 3:28 PM, Flavio Pompermaier <pomperma...@okkam.it> wrote: > Which file and which JVM options do I have to modify to try options 1 and > 3..? > > 1. Don't fill the JVMs up to the limit with objects. Give more memory > to the JVM, or give less memory to Flink managed memory > 2. Use more JVMs, i.e., a higher parallelism > 3. Use a concurrent garbage collector, like G1 > > Actually, when I run the code from Eclipse I see an exception do to an > error in the data (because I try to read a URI that contains illegal > characters) but I don't think the program reach that point, I don't see > anywhere an exception and the error occur later on in the code.. > > However, all of your options seems related to a scalability problem, where > I should add more resources to complete the work...while it works locally > in the IDE where I have less resources (except the gc that I use default > settings while I don't know if the cluster has some default ones)..isn't it > strange? > > On Mon, Jun 29, 2015 at 2:29 PM, Stephan Ewen <se...@apache.org> wrote: > >> Hi Flavio! >> >> I had a look at the logs. There seems nothing suspicious - at some point, >> the TaskManager and JobManager declare each other unreachable. >> >> A pretty common cause for that is that the JVMs stall for a long time due >> to garbage collection. The JobManager cannot see the difference between a >> JVM that is irresponsive (due to garbage collection) and a JVM that is dead. >> >> Here is what you can do to prevent long garbage collection stalls: >> >> - Don't fill the JVMs up to the limit with objects. Give more memory to >> the JVM, or give less memory to Flink managed memory. >> - Use more JVMs, i.e., a higher parallelism. >> - Use a concurrent garbage collector, like G1. >> >> >> Greetings, >> Stephan >> >> >> On Mon, Jun 29, 2015 at 12:39 PM, Stephan Ewen <se...@apache.org> wrote: >> >>> Hi Flavio! >>> >>> Can you post the JobManager's log here? It should have the message about >>> what is going wrong... >>> >>> Stephan >>> >>> >>> On Mon, Jun 29, 2015 at 11:43 AM, Flavio Pompermaier < >>> pomperma...@okkam.it> wrote: >>> >>>> Hi to all, >>>> >>>> I'm restarting the discussion about a problem I alredy dicussed on this >>>> mailing list (but that started with a different subject). >>>> I'm running Flink 0.9.0 on CDH 5.1.3 so I compiled the sources as: >>>> >>>> mvn clean install -Dhadoop.version=2.3.0-cdh5.1.3 >>>> -Dhbase.version=0.98.1-cdh5.1.3 -Dhadoop.core.version=2.3.0-mr1-cdh5.1.3 >>>> -DskipTests -Pvendor-repos >>>> >>>> The problem I'm facing is that the cluster start successfully but when >>>> I run my job (from the web-client) I get, after some time, this exception: >>>> >>>> 16:35:41,636 WARN akka.remote.RemoteWatcher >>>> - Detected unreachable: [akka.tcp://flink@192.168.234.83:6123 >>>> ] >>>> 16:35:46,605 INFO org.apache.flink.runtime.taskmanager.TaskManager - >>>> Disconnecting from JobManager: JobManager is no longer reachable >>>> 16:35:46,614 INFO org.apache.flink.runtime.taskmanager.TaskManager - >>>> Cancelling all computations and discarding all cached data. >>>> 16:35:46,644 INFO org.apache.flink.runtime.taskmanager.Task >>>> - Attempting to fail task externally CHAIN GroupReduce (GroupReduce >>>> at compactDataSources(MyClass.java:213)) -> Combine(Distinct at >>>> compactDataSources(MyClass.java:213)) (8/36) >>>> 16:35:46,669 INFO org.apache.flink.runtime.taskmanager.Task >>>> - CHAIN GroupReduce (GroupReduce at >>>> compactDataSources(MyClass.java:213)) -> Combine(Distinct at >>>> compactDataSources(MyClass.java:213)) (8/36) switched to FAILED with >>>> exception. >>>> java.lang.Exception: Disconnecting from JobManager: JobManager is no >>>> longer reachable >>>> at org.apache.flink.runtime.taskmanager.TaskManager.org >>>> $apache$flink$runtime$taskmanager$TaskManager$$handleJobManagerDisconnect(TaskManager.scala:741) >>>> at >>>> org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1.applyOrElse(TaskManager.scala:267) >>>> at >>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) >>>> at >>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) >>>> at >>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) >>>> at >>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36) >>>> at >>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29) >>>> at >>>> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) >>>> at >>>> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29) >>>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465) >>>> at >>>> org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:114) >>>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) >>>> at >>>> akka.actor.dungeon.DeathWatch$class.receivedTerminated(DeathWatch.scala:46) >>>> at akka.actor.ActorCell.receivedTerminated(ActorCell.scala:369) >>>> at akka.actor.ActorCell.autoReceiveMessage(ActorCell.scala:501) >>>> at akka.actor.ActorCell.invoke(ActorCell.scala:486) >>>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) >>>> at akka.dispatch.Mailbox.run(Mailbox.scala:221) >>>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231) >>>> at >>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>>> at >>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253) >>>> at >>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346) >>>> at >>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>>> at >>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>>> 16:35:46,767 INFO org.apache.flink.runtime.taskmanager.Task >>>> - Triggering cancellation of task code CHAIN GroupReduce >>>> (GroupReduce at compactDataSources(MyClass.java:213)) -> Combine(Distinct >>>> at compactDataSources(MyClass.java:213)) (8/36) >>>> (57a0ad78726d5ba7255aa87038250c51). >>>> >>>> The job instead runs correctly from the IDE (Eclipse). How can I >>>> understand/debug what's wrong? >>>> >>>> Best, >>>> Flavio >>>> >>>> >>> >> > >