Hi Mihail, could you share your code or at least the implementations of getVerticesDataSet() and InitVerticesMapper so I can take a look? Where is InitVerticesMapper called above?
Cheers, Vasia. On 26 June 2015 at 10:51, Mihail Vieru <vi...@informatik.hu-berlin.de> wrote: > Hi Robert, > > I'm using the same input data, as well as the same parameters I use in the > IDE's run configuration. > I don't run the job on the cluster (yet), but locally, by starting Flink > with the start-local.sh script. > > > I will try to explain my code a bit. The *Integer[] *array is initialized > in the *getVerticesDataSet()* method. > > * DataSet<Vertex<Integer, Tuple2<Integer[],String> >> vertices = > getVerticesDataSet(env);* > * ...* > * Graph<Integer, Tuple2<Integer[],String>, NullValue> graph = > Graph.fromDataSet(vertices, edges, env);* > * ...* > * Graph<Integer, Tuple2<Integer[],String>, NullValue> > intermediateGraph = * > * graph.run(new APSP<Integer>(srcVertexId, maxIterations));* > > > In APSP I'm addressing it in the *InitVerticesMapper*, but is now > suddenly empty. > > Best, > Mihail > > > On 26.06.2015 10:00, Robert Metzger wrote: > > Hi Mihail, > > the NPE has been thrown from > *graphdistance.APSP$InitVerticesMapper.map(APSP.java:74)*. I guess that > is code written by you or a library you are using. > Maybe the data you are using on the cluster is different from your local > test data? > > Best, > Robert > > > On Thu, Jun 25, 2015 at 7:41 PM, Mihail Vieru < > vi...@informatik.hu-berlin.de> wrote: > >> Hi, >> >> I get an ArrayIndexOutOfBoundsException when I run my job from a JAR in >> the CLI. >> This doesn't occur in the IDE. >> >> I've build the JAR using the "maven-shade-plugin" and the pom.xml >> configuration Robert has provided here: >> >> https://stackoverflow.com/questions/30102523/linkage-failure-when-running-apache-flink-jobs >> I specify the entry point using the "-c" option. >> >> The array the Exception refers to is actually initialized when a vertices >> dataset is read from the file system. >> >> Any ideas on what could cause this issue? >> >> Best, >> Mihail >> >> P.S.: the stack trace: >> >> *org.apache.flink.client.program.ProgramInvocationException: The program >> execution failed: Job execution failed.* >> * at org.apache.flink.client.program.Client.run(Client.java:413)* >> * at org.apache.flink.client.program.Client.run(Client.java:356)* >> * at org.apache.flink.client.program.Client.run(Client.java:349)* >> * at >> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)* >> * at graphdistance.KAPSPNaiveJob.main(KAPSPNaiveJob.java:56)* >> * at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)* >> * at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)* >> * at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)* >> * at java.lang.reflect.Method.invoke(Method.java:606)* >> * at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)* >> * at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)* >> * at org.apache.flink.client.program.Client.run(Client.java:315)* >> * at >> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)* >> * at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)* >> * at >> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)* >> * at org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)* >> *Caused by: org.apache.flink.runtime.client.JobExecutionException: Job >> execution failed.* >> * at >> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:314)* >> * at >> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)* >> * at >> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)* >> * at >> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)* >> * at >> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36)* >> * at >> org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)* >> * at >> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)* >> * at >> org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)* >> * at akka.actor.Actor$class.aroundReceive(Actor.scala:465)* >> * at >> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:92)* >> * at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)* >> * at akka.actor.ActorCell.invoke(ActorCell.scala:487)* >> * at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)* >> * at akka.dispatch.Mailbox.run(Mailbox.scala:221)* >> * at akka.dispatch.Mailbox.exec(Mailbox.scala:231)* >> * at >> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)* >> * at >> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)* >> * at >> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)* >> * at >> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)* >> *Caused by: java.lang.ArrayIndexOutOfBoundsException: 0* >> * at graphdistance.APSP$InitVerticesMapper.map(APSP.java:74)* >> * at graphdistance.APSP$InitVerticesMapper.map(APSP.java:48)* >> * at org.apache.flink.graph.Graph$2.map(Graph.java:389)* >> * at org.apache.flink.graph.Graph$2.map(Graph.java:387)* >> * at >> org.apache.flink.runtime.operators.MapDriver.run(MapDriver.java:97)* >> * at >> org.apache.flink.runtime.operators.RegularPactTask.run(RegularPactTask.java:496)* >> * at >> org.apache.flink.runtime.operators.RegularPactTask.invoke(RegularPactTask.java:362)* >> * at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)* >> * at java.lang.Thread.run(Thread.java:745)* >> > > >