Hi Robert,

I'm using the same input data, as well as the same parameters I use in the IDE's run configuration. I don't run the job on the cluster (yet), but locally, by starting Flink with the start-local.sh script.


I will try to explain my code a bit. The /Integer[] /array is initialized in the /getVerticesDataSet()/ method.

/ DataSet<Vertex<Integer, Tuple2<Integer[],String> >> vertices = getVerticesDataSet(env);//
//        ...//
// Graph<Integer, Tuple2<Integer[],String>, NullValue> graph = Graph.fromDataSet(vertices, edges, env);//
//        ...//
// Graph<Integer, Tuple2<Integer[],String>, NullValue> intermediateGraph = //
//                graph.run(new APSP<Integer>(srcVertexId, maxIterations));/


In APSP I'm addressing it in the /InitVerticesMapper/, but is now suddenly empty.

Best,
Mihail

On 26.06.2015 10:00, Robert Metzger wrote:
Hi Mihail,

the NPE has been thrown from /graphdistance.APSP$InitVerticesMapper.map(APSP.java:74)/. I guess that is code written by you or a library you are using. Maybe the data you are using on the cluster is different from your local test data?

Best,
Robert


On Thu, Jun 25, 2015 at 7:41 PM, Mihail Vieru <vi...@informatik.hu-berlin.de <mailto:vi...@informatik.hu-berlin.de>> wrote:

    Hi,

    I get an ArrayIndexOutOfBoundsException when I run my job from a
    JAR in the CLI.
    This doesn't occur in the IDE.

    I've build the JAR using the "maven-shade-plugin" and the pom.xml
    configuration Robert has provided here:
    
https://stackoverflow.com/questions/30102523/linkage-failure-when-running-apache-flink-jobs
    I specify the entry point using the "-c" option.

    The array the Exception refers to is actually initialized when a
    vertices dataset is read from the file system.

    Any ideas on what could cause this issue?

    Best,
    Mihail

    P.S.: the stack trace:
    /
    //org.apache.flink.client.program.ProgramInvocationException: The
    program execution failed: Job execution failed.//
    //    at org.apache.flink.client.program.Client.run(Client.java:413)//
    //    at org.apache.flink.client.program.Client.run(Client.java:356)//
    //    at org.apache.flink.client.program.Client.run(Client.java:349)//
    //    at
    
org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)//
    //    at graphdistance.KAPSPNaiveJob.main(KAPSPNaiveJob.java:56)//
    //    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)//
    //    at
    
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)//
    //    at
    
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)//
    //    at java.lang.reflect.Method.invoke(Method.java:606)//
    //    at
    
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)//
    //    at
    
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)//
    //    at org.apache.flink.client.program.Client.run(Client.java:315)//
    //    at
    org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)//
    //    at
    org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)//
    //    at
    org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:880)//
    //    at
    org.apache.flink.client.CliFrontend.main(CliFrontend.java:922)//
    //Caused by:
    org.apache.flink.runtime.client.JobExecutionException: Job
    execution failed.//
    //    at
    
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:314)//
    //    at
    
scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)//
    //    at
    
scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)//
    //    at
    
scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)//
    //    at
    
org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:36)//
    //    at
    
org.apache.flink.runtime.ActorLogMessages$$anon$1.apply(ActorLogMessages.scala:29)//
    //    at
    scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)//
    //    at
    
org.apache.flink.runtime.ActorLogMessages$$anon$1.applyOrElse(ActorLogMessages.scala:29)//
    //    at akka.actor.Actor$class.aroundReceive(Actor.scala:465)//
    //    at
    
org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:92)//
    //    at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)//
    //    at akka.actor.ActorCell.invoke(ActorCell.scala:487)//
    //    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)//
    //    at akka.dispatch.Mailbox.run(Mailbox.scala:221)//
    //    at akka.dispatch.Mailbox.exec(Mailbox.scala:231)//
    //    at
    scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)//
    //    at
    
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)//
    //    at
    scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)//
    //    at
    
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)//
    //Caused by: java.lang.*ArrayIndexOutOfBoundsException*: 0//
    //    at graphdistance.APSP$InitVerticesMapper.map(APSP.java:74)//
    //    at graphdistance.APSP$InitVerticesMapper.map(APSP.java:48)//
    //    at org.apache.flink.graph.Graph$2.map(Graph.java:389)//
    //    at org.apache.flink.graph.Graph$2.map(Graph.java:387)//
    //    at
    org.apache.flink.runtime.operators.MapDriver.run(MapDriver.java:97)//
    //    at
    
org.apache.flink.runtime.operators.RegularPactTask.run(RegularPactTask.java:496)//
    //    at
    
org.apache.flink.runtime.operators.RegularPactTask.invoke(RegularPactTask.java:362)//
    //    at
    org.apache.flink.runtime.taskmanager.Task.run(Task.java:559)//
    //    at java.lang.Thread.run(Thread.java:745)/



Reply via email to