It seems Flink used a different exception graph outside of my IDE (intellij)
The job anatomy is: load data from csv and build an initial graph => reduce that graph (remove loops and combine multi edges) => extend the modified graph by a new vertex value => run a gather-scatter iteration I have to extend the vertex value, because each vertex need its incident edges inside the iteration. My CustomVertexValue is able to hold all incident edges. Vertex<Double, CustomVertexValue> vertex Flink is try to optimize the execution graph, but that’s the issue. Maybe Flink provides an influence by the programmer? Best and thanks Marc Am 10.06.2017 um 00:49 schrieb Greg Hogan <c...@greghogan.com<mailto:c...@greghogan.com>>: Have you looked at org.apache.flink.gelly.GraphExtension.CustomVertexValue.createInitSemiCluster(CustomVertexValue.java:51)? On Jun 9, 2017, at 4:53 PM, Kaepke, Marc <marc.kae...@haw-hamburg.de<mailto:marc.kae...@haw-hamburg.de>> wrote: Hi everyone, I don’t have any exceptions if I execute my Gelly job in my IDE (local) directly. The next step is an execution with a real kubernetes cluster (1 JobManager and 3 TaskManager on dedicated machines). The word count example is running without exceptions. My Gelly job throws following exception and I don’t know why. org.apache.flink.client.program.ProgramInvocationException: The program execution failed: Job execution failed. at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:478) at org.apache.flink.client.program.StandaloneClusterClient.submitJob(StandaloneClusterClient.java:105) at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:442) at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:429) at org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:62) at org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:926) at org.apache.flink.api.java.DataSet.collect(DataSet.java:410) at org.apache.flink.gelly.Algorithm.SemiClusteringPregel.run(SemiClusteringPregel.java:84) at org.apache.flink.gelly.Algorithm.SemiClusteringPregel.run(SemiClusteringPregel.java:34) at org.apache.flink.graph.Graph.run(Graph.java:1850) at org.apache.flink.gelly.job.GellyMain.main(GellyMain.java:128) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:528) at org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:419) at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:381) at org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:838) at org.apache.flink.client.CliFrontend.run(CliFrontend.java:259) at org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1086) at org.apache.flink.client.CliFrontend$2.call(CliFrontend.java:1133) at org.apache.flink.client.CliFrontend$2.call(CliFrontend.java:1130) at org.apache.flink.runtime.security.HadoopSecurityContext$1.run(HadoopSecurityContext.java:43) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) at org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:40) at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1129) Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution failed. at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$6.apply$mcV$sp(JobManager.scala:933) at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$6.apply(JobManager.scala:876) at org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$6.apply(JobManager.scala:876) at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:397) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Caused by: java.lang.NullPointerException at org.apache.flink.gelly.GraphExtension.CustomVertexValue.createInitSemiCluster(CustomVertexValue.java:51) at org.apache.flink.gelly.PreModification.IncidentEdgesCollector.iterateEdges(IncidentEdgesCollector.java:37) at org.apache.flink.graph.Graph$ApplyCoGroupFunctionOnAllEdges.coGroup(Graph.java:1252) at org.apache.flink.runtime.operators.CoGroupDriver.run(CoGroupDriver.java:159) at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:490) at org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:355) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:702) at java.lang.Thread.run(Thread.java:748) I guess the trigger is the coGroup function, but I’m not sure and need your help. Best, Marc