Actually, the ProgramAbortException is something that the web interface
does in order to run the program to a part where the plan is created (for
visualization), but the program is not executed.

Apparently this logic is a bit broken in the latest version, or you are
catching and re-throwing the exception.

On Tue, Aug 11, 2015 at 3:44 PM, Michael Huelfenhaus <
m.huelfenh...@davengo.com> wrote:

> I can now reproduce this error and the successful run.
>
> When load the jar in the web interface and start the plan the
>
> org.apache.flink.client.program.Client$ProgramAbortException
>
>
> happens when I reload the page with the error the plan is shown correct
> and the execution succeeds.
>
> - Michael
>
>
> Am 11.08.2015 um 15:29 schrieb Michael Huelfenhaus <
> m.huelfenh...@davengo.com>:
>
> Hey Stephan
>
> the error disappeared after restarting the computer.
>
> now I got another one but this is also strange because the program ran
> once successful and is now failing again without me knowingly changing
> something.
>
> - Michael
>
> An error occurred while invoking the program:
>
> The program caused an error:
>
>
> org.apache.flink.client.program.Client$ProgramAbortException
>       at 
> org.apache.flink.streaming.api.environment.StreamPlanEnvironment.execute(StreamPlanEnvironment.java:68)
>       at com.davengo.rfidcloud.flink.DaoJoin.main(DaoJoin.java:68)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:483)
>       at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
>       at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
>       at 
> org.apache.flink.client.program.Client.getOptimizedPlan(Client.java:241)
>       at org.apache.flink.client.CliFrontend.info(CliFrontend.java:377)
>       at 
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:933)
>       at 
> org.apache.flink.client.web.JobSubmissionServlet.doGet(JobSubmissionServlet.java:173)
>       at javax.servlet.http.HttpServlet.service(HttpServlet.java:734)
>       at javax.servlet.http.HttpServlet.service(HttpServlet.java:847)
>       at 
> org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:532)
>       at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:453)
>       at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:227)
>       at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:965)
>       at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:388)
>       at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:187)
>       at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:901)
>       at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
>       at 
> org.eclipse.jetty.server.handler.HandlerList.handle(HandlerList.java:47)
>       at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:113)
>       at org.eclipse.jetty.server.Server.handle(Server.java:352)
>       at 
> org.eclipse.jetty.server.HttpConnection.handleRequest(HttpConnection.java:596)
>       at 
> org.eclipse.jetty.server.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:1048)
>       at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:549)
>       at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:211)
>       at 
> org.eclipse.jetty.server.HttpConnection.handle(HttpConnection.java:425)
>       at 
> org.eclipse.jetty.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:489)
>       at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:436)
>       at java.lang.Thread.run(Thread.java:745)
>
>
>
> Am 11.08.2015 um 13:22 schrieb Stephan Ewen <se...@apache.org>:
>
> Hi Michael!
>
> Can you try the following:
>
> 1) Check whether the class is actually in the jar file (unzip / grep)
>
> 2) Can you check in the logs what is the path of the BLOB server temp
> directory (JobManager and TaskManager). Are there any strange characters /
> spaces in the path?
>
> 3) Can you check in the TaskManager log whether you see an entry that it
> downloads or pulls the BLOB (here the jar file) from the JobManager?
>
> Thanks,
> Stephan
>
>
>
>
> On Tue, Aug 11, 2015 at 1:04 PM, Michael Huelfenhaus <
> m.huelfenh...@davengo.com> wrote:
>
>> I tried this before without success.
>>
>> Same exception the named class was not found.
>>
>> -Michael
>>
>> Am 06.08.2015 um 14:07 schrieb Matthias J. Sax <
>> mj...@informatik.hu-berlin.de>:
>>
>> > Never mind. Just saw that this in not the problem...
>> >
>> > Sounds weird to me. Maybe you can try to name the class. Anonymous
>> > classes should not be a problem, but it should be worth a try.
>> >
>> > -Matthias
>> >
>> >
>> >
>> > On 08/06/2015 01:51 PM, Matthias J. Sax wrote:
>> >> If I see it correctly your jar contains
>> >>
>> >>> com/davengo/rfidcloud/flink/DaoJoin$1.class
>> >>
>> >> But your error message says
>> >>
>> >>> ClassNotFoundException: com.otter.ist.flink.DaoJoin$1
>> >>
>> >> Both are different packages. Your jar seems not be correctly packaged.
>> >>
>> >>
>> >> -Matthias
>> >>
>> >> On 08/06/2015 12:46 PM, Michael Huelfenhaus wrote:
>> >>> I am back at work next Tuesday, any further ideas would be great until
>> >>> then, for now I am continuing inside ecplise.
>> >>>
>> >>> Am 06.08.2015 um 11:27 schrieb Michael Huelfenhaus
>> >>> <m.huelfenh...@davengo.com <mailto:m.huelfenh...@davengo.com>>:
>> >>>
>> >>>> hi,
>> >>>>
>> >>>>> how did you build the jar file?
>> >>>>
>> >>>> mvn clean install -Pbuild-jar
>> >>>>
>> >>>>> Have you checked whether your classes are in the jar file?
>> >>>>
>> >>>> yes, this seems alright for me
>> >>>>
>> >>>>> jar tf target/flink-test-0.1.jar
>> >>>> META-INF/MANIFEST.MF
>> >>>> META-INF/
>> >>>> com/
>> >>>> com/davengo/
>> >>>> com/davengo/rfidcloud/
>> >>>> com/davengo/rfidcloud/flink/
>> >>>> com/davengo/rfidcloud/flink/DaoJoin$1.class
>> >>>> com/davengo/rfidcloud/flink/DaoJoin.class
>> >>>> com/davengo/rfidcloud/flink/streampojos/
>> >>>> com/davengo/rfidcloud/flink/streampojos/EpcTuple.class
>> >>>> log4j.properties
>> >>>> META-INF/maven/
>> >>>> META-INF/maven/com.davengo.rfidcloud.flink/
>> >>>> META-INF/maven/com.davengo.rfidcloud.flink/flink-test/
>> >>>> META-INF/maven/com.davengo.rfidcloud.flink/flink-test/pom.xml
>> >>>> META-INF/maven/com.davengo.rfidcloud.flink/flink-test/pom.properties
>> >>>>
>> >>>> Am 06.08.2015 um 11:21 schrieb Robert Metzger <rmetz...@apache.org
>> >>>> <mailto:rmetz...@apache.org>>:
>> >>>>
>> >>>>> Hi,
>> >>>>>
>> >>>>> how did you build the jar file?
>> >>>>> Have you checked whether your classes are in the jar file?
>> >>>>>
>> >>>>> On Thu, Aug 6, 2015 at 11:08 AM, Michael Huelfenhaus
>> >>>>> <m.huelfenh...@davengo.com <mailto:m.huelfenh...@davengo.com>>
>> wrote:
>> >>>>>
>> >>>>>    Hello everybody
>> >>>>>
>> >>>>>    I am truing to build a very simple streaming application with the
>> >>>>>    nightly build of flink 0.10, my code runs fine in eclipse.
>> >>>>>
>> >>>>>    But when I build and deploy the jar locally I always get
>> >>>>>    java.lang.ClassNotFoundException: com.otter.ist.flink.DaoJoin$1
>> >>>>>
>> >>>>>    There is also no plan visible in the web interface.
>> >>>>>
>> >>>>>    I start the local flink 0.10 with start-local-streaming.sh  after
>> >>>>>    building it from the git code
>> >>>>>
>> >>>>>    Below you find the complete error, my code and the pom.xml any
>> >>>>>    help is appreciated.
>> >>>>>
>> >>>>>    Cheers Michael
>> >>>>>
>> >>>>>
>> >>>>>    error log from web interface:
>> >>>>>    An error occurred while invoking the program:
>> >>>>>
>> >>>>>    The main method caused an error.
>> >>>>>
>> >>>>>
>> >>>>>    org.apache.flink.runtime.client.JobExecutionException: Job
>> >>>>>    execution failed.
>> >>>>>            at
>> >>>>>
>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1.applyOrElse(JobManager.scala:364)
>> >>>>>            at
>> >>>>>
>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>> >>>>>            at
>> >>>>>
>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>> >>>>>            at
>> >>>>>
>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.runtime.LeaderSessionMessages$$anonfun$receive$1.applyOrElse(LeaderSessionMessages.scala:40)
>> >>>>>            at
>> >>>>>
>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
>> >>>>>            at
>> >>>>>
>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
>> >>>>>            at
>> >>>>>
>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28)
>> >>>>>            at
>> >>>>>
>> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28)
>> >>>>>            at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:101)
>> >>>>>            at
>> akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
>> >>>>>            at akka.actor.ActorCell.invoke(ActorCell.scala:487)
>> >>>>>            at
>> akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254)
>> >>>>>            at akka.dispatch.Mailbox.run(Mailbox.scala:221)
>> >>>>>            at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
>> >>>>>            at
>> >>>>>
>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>> >>>>>            at
>> >>>>>
>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>> >>>>>            at
>> >>>>>
>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>> >>>>>            at
>> >>>>>
>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>> >>>>>    Caused by: java.lang.Exception: Call to registerInputOutput() of
>> >>>>>    invokable failed
>> >>>>>            at
>> >>>>>    org.apache.flink.runtime.taskmanager.Task.run(Task.java:526)
>> >>>>>            at java.lang.Thread.run(Thread.java:745)
>> >>>>>    Caused by:
>> >>>>>    org.apache.flink.streaming.runtime.tasks.StreamTaskException:
>> >>>>>    Cannot instantiate user function.
>> >>>>>            at
>> >>>>>
>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperator(StreamConfig.java:207)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.streaming.runtime.tasks.OutputHandler.createChainedCollector(OutputHandler.java:173)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.streaming.runtime.tasks.OutputHandler.createChainedCollector(OutputHandler.java:159)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.streaming.runtime.tasks.OutputHandler.<init>(OutputHandler.java:107)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.streaming.runtime.tasks.StreamTask.registerInputOutput(StreamTask.java:99)
>> >>>>>            at
>> >>>>>    org.apache.flink.runtime.taskmanager.Task.run(Task.java:523)
>> >>>>>            ... 1 more
>> >>>>>    Caused by: java.lang.ClassNotFoundException:
>> >>>>>    com.otter.ist.flink.DaoJoin$1
>> >>>>>            at java.net.URLClassLoader$1.run(URLClassLoader.java:372)
>> >>>>>            at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
>> >>>>>            at java.security.AccessController.doPrivileged(Native
>> Method)
>> >>>>>            at
>> java.net.URLClassLoader.findClass(URLClassLoader.java:360)
>> >>>>>            at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>> >>>>>            at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>> >>>>>            at java.lang.Class.forName0(Native Method)
>> >>>>>            at java.lang.Class.forName(Class.java:344)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.util.InstantiationUtil$ClassLoaderObjectInputStream.resolveClass(InstantiationUtil.java:71)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1613)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1518)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1774)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1993)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1918)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801)
>> >>>>>            at
>> >>>>>
>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351)
>> >>>>>            at
>> >>>>>    java.io.ObjectInputStream.readObject(ObjectInputStream.java:371)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:302)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:264)
>> >>>>>            at
>> >>>>>
>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperator(StreamConfig.java:205)
>> >>>>>            ... 6 more
>> >>>>>
>> >>>>>    my code:
>> >>>>>
>> >>>>>    package com.otter.ist.flink;
>> >>>>>
>> >>>>>    import org.apache.flink.api.common.functions.MapFunction;
>> >>>>>    import org.apache.flink.api.java.tuple.Tuple3;
>> >>>>>    import org.apache.flink.streaming.api.datastream.DataStream;
>> >>>>>    import
>> >>>>>
>> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
>> >>>>>
>> >>>>>    public class DaoJoin {
>> >>>>>
>> >>>>>            public static void main(String[] args) throws Exception {
>> >>>>>                    //
>> >>>>>
>> *************************************************************************
>> >>>>>                    // PROGRAM
>> >>>>>                    //
>> >>>>>
>> *************************************************************************
>> >>>>>
>> >>>>>                    if (!parseParameters(args)) {
>> >>>>>                            return;
>> >>>>>                    }
>> >>>>>
>> >>>>>                    // set up the execution environment
>> >>>>>                    final StreamExecutionEnvironment env =
>> >>>>>    StreamExecutionEnvironment.createLocalEnvironment();
>> >>>>>    //              final StreamExecutionEnvironment env =
>> >>>>>    StreamExecutionEnvironment.getExecutionEnvironment();
>> >>>>>
>> >>>>>
>> >>>>>                    // get input data
>> >>>>>                    DataStream<String> text = getTextDataStream(env);
>> >>>>>
>> >>>>>                    DataStream<Tuple3<String, String, Integer>> epcs
>> >>>>>    = text.map(new MapFunction<String, Tuple3<String, String,
>> >>>>>    Integer>>(){
>> >>>>>
>> >>>>>                            private static final long
>> >>>>>    serialVersionUID = -7889264579632622427L;
>> >>>>>
>> >>>>>                            @Override
>> >>>>>                            public Tuple3<String, String, Integer>
>> >>>>>    map(String line) throws Exception {
>> >>>>>                                    String[] fields = line.split("
>> ");
>> >>>>>
>> >>>>>                                    return new Tuple3<String, String,
>> >>>>>    Integer>(fields[0], fields[1], Integer.parseInt(fields[2]));
>> >>>>>                            }
>> >>>>>
>> >>>>>                    });
>> >>>>>
>> >>>>>                    // emit result
>> >>>>>                    if (fileOutput) {
>> >>>>>                            epcs.writeAsText(outputPath);
>> >>>>>                    } else {
>> >>>>>                            epcs.print();
>> >>>>>                    }
>> >>>>>                    System.out.println(env.getExecutionPlan());
>> >>>>>
>> >>>>>                    // execute program
>> >>>>>                    env.execute("DaoJoin");
>> >>>>>            }
>> >>>>>
>> >>>>>            //
>> >>>>>
>> *************************************************************************
>> >>>>>            // UTIL METHODS
>> >>>>>            //
>> >>>>>
>> *************************************************************************
>> >>>>>
>> >>>>>            private static boolean fileOutput = false;
>> >>>>>            private static String textPath;
>> >>>>>            private static String outputPath;
>> >>>>>
>> >>>>>            private static boolean parseParameters(String[] args) {
>> >>>>>
>> >>>>>                    if (args.length > 0) {
>> >>>>>                            // parse input arguments
>> >>>>>                            fileOutput = true;
>> >>>>>                            if (args.length == 2) {
>> >>>>>                                    textPath = args[0];
>> >>>>>                                    outputPath = args[1];
>> >>>>>                            } else {
>> >>>>>                                    System.err.println("Usage:
>> >>>>>    DaoJoin <text path> <result path>");
>> >>>>>                                    return false;
>> >>>>>                            }
>> >>>>>                            System.out.println("fileout: " +
>> fileOutput);
>> >>>>>                    } else {
>> >>>>>                            System.out.println("Executing WordCount
>> >>>>>    example with built-in default data.");
>> >>>>>                            System.out.println("  Provide parameters
>> >>>>>    to read input data from a file.");
>> >>>>>                            System.out.println("  Usage: WordCount
>> >>>>>    <text path> <result path>");
>> >>>>>                    }
>> >>>>>                    return true;
>> >>>>>            }
>> >>>>>
>> >>>>>            private static DataStream<String>
>> >>>>>    getTextDataStream(StreamExecutionEnvironment env) {
>> >>>>>                            // read the text file from given input
>> path
>> >>>>>                            return env.readTextFile(textPath);
>> >>>>>            }
>> >>>>>    }
>> >>>>>
>> >>>>>
>> >>>>>    the pom.xml
>> >>>>>    <project xmlns="http://maven.apache.org/POM/4.0.0";
>> >>>>>    xmlns:xsi="http://xwww.w3.org/2001/XMLSchema-instance";
>> >>>>>            xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
>> >>>>>    http://maven.apache.org/xsd/maven-4.0.0.xsd";>
>> >>>>>            <modelVersion>4.0.0</modelVersion>
>> >>>>>
>> >>>>>            <groupId>com.otter.ist.flink</groupId>
>> >>>>>            <artifactId>flink-test</artifactId>
>> >>>>>            <version>0.1</version>
>> >>>>>            <packaging>jar</packaging>
>> >>>>>
>> >>>>>            <name>DaoJoin</name>
>> >>>>>            <url>http://www.otter.com <http://www.otter.com/></url>
>> >>>>>
>> >>>>>            <properties>
>> >>>>>
>> >>>>>
>> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
>> >>>>>            </properties>
>> >>>>>
>> >>>>>            <repositories>
>> >>>>>                    <repository>
>> >>>>>                            <id>apache.snapshots</id>
>> >>>>>                            <name>Apache Development Snapshot
>> >>>>>    Repository</name>
>> >>>>>
>> >>>>>    <url>
>> https://repository.apache.org/content/repositories/snapshots/</url>
>> >>>>>                            <releases>
>> >>>>>                                    <enabled>false</enabled>
>> >>>>>                            </releases>
>> >>>>>                            <snapshots>
>> >>>>>                                    <enabled>true</enabled>
>> >>>>>                            </snapshots>
>> >>>>>                    </repository>
>> >>>>>            </repositories>
>> >>>>>
>> >>>>>            <!--
>> >>>>>
>> >>>>>                    Execute "mvn clean package -Pbuild-jar"
>> >>>>>                    to build a jar file out of this project!
>> >>>>>
>> >>>>>                    How to use the Flink Quickstart pom:
>> >>>>>
>> >>>>>                    a) Adding new dependencies:
>> >>>>>                            You can add dependencies to the list
>> below.
>> >>>>>                            Please check if the maven-shade-plugin
>> >>>>>    below is filtering out your dependency
>> >>>>>                            and remove the exclude from there.
>> >>>>>
>> >>>>>                    b) Build a jar for running on the cluster:
>> >>>>>                            There are two options for creating a jar
>> >>>>>    from this project
>> >>>>>
>> >>>>>                            b.1) "mvn clean package" -> this will
>> >>>>>    create a fat jar which contains all
>> >>>>>                                            dependencies necessary
>> >>>>>    for running the jar created by this pom in a cluster.
>> >>>>>                                            The "maven-shade-plugin"
>> >>>>>    excludes everything that is provided on a running Flink cluster.
>> >>>>>
>> >>>>>                            b.2) "mvn clean package -Pbuild-jar" ->
>> >>>>>    This will also create a fat-jar, but with much
>> >>>>>                                            nicer dependency
>> >>>>>    exclusion handling. This approach is preferred and leads to
>> >>>>>                                            much cleaner jar files.
>> >>>>>            -->
>> >>>>>
>> >>>>>            <dependencies>
>> >>>>>                    <dependency>
>> >>>>>                            <groupId>org.apache.flink</groupId>
>> >>>>>                            <artifactId>flink-java</artifactId>
>> >>>>>                            <version>0.10-SNAPSHOT</version>
>> >>>>>                    </dependency>
>> >>>>>                    <dependency>
>> >>>>>                            <groupId>org.apache.flink</groupId>
>> >>>>>
>> <artifactId>flink-streaming-core</artifactId>
>> >>>>>                            <version>0.10-SNAPSHOT</version>
>> >>>>>                    </dependency>
>> >>>>>                    <dependency>
>> >>>>>                            <groupId>org.apache.flink</groupId>
>> >>>>>                            <artifactId>flink-clients</artifactId>
>> >>>>>                            <version>0.10-SNAPSHOT</version>
>> >>>>>                    </dependency>
>> >>>>>            </dependencies>
>> >>>>>
>> >>>>>            <build>
>> >>>>>                    <plugins>
>> >>>>>                            <!-- We use the maven-shade plugin to
>> >>>>>    create a fat jar that contains all dependencies
>> >>>>>                            except flink and it's transitive
>> >>>>>    dependencies. The resulting fat-jar can be executed
>> >>>>>                            on a cluster. Change the value of
>> >>>>>    Program-Class if your program entry point changes. -->
>> >>>>>                            <plugin>
>> >>>>>
>> >>>>>    <groupId>org.apache.maven.plugins</groupId>
>> >>>>>
>> >>>>>    <artifactId>maven-shade-plugin</artifactId>
>> >>>>>                                    <version>2.3</version>
>> >>>>>                                    <executions>
>> >>>>>                                            <!-- Run shade goal on
>> >>>>>    package phase -->
>> >>>>>                                            <execution>
>> >>>>>
>> >>>>>    <phase>package</phase>
>> >>>>>                                                    <goals>
>> >>>>>
>> >>>>>    <goal>shade</goal>
>> >>>>>                                                    </goals>
>> >>>>>                                                    <configuration>
>> >>>>>
>> <artifactSet>
>> >>>>>
>> >>>>>    <excludes>
>> >>>>>
>> >>>>>          <!-- This list contains all dependencies of flink-dist
>> >>>>>
>> >>>>>          Everything else will be packaged into the fat-jar
>> >>>>>
>> >>>>>          -->
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-shaded-*</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-core</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-java</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-scala</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-runtime</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-optimizer</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-clients</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-spargel</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-avro</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-java-examples</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-scala-examples</exclude>
>> >>>>>
>> >>>>>
>> <exclude>org.apache.flink:flink-streaming-examples</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.flink:flink-streaming-core</exclude>
>> >>>>>
>> >>>>>
>> >>>>>          <!-- Also exclude very big transitive dependencies of Flink
>> >>>>>
>> >>>>>
>> >>>>>          WARNING: You have to remove these excludes if your code
>> >>>>>    relies on other
>> >>>>>
>> >>>>>          versions of these dependencies.
>> >>>>>
>> >>>>>
>> >>>>>          -->
>> >>>>>
>> >>>>>          <exclude>org.scala-lang:scala-library</exclude>
>> >>>>>
>> >>>>>          <exclude>org.scala-lang:scala-compiler</exclude>
>> >>>>>
>> >>>>>          <exclude>org.scala-lang:scala-reflect</exclude>
>> >>>>>
>> >>>>>          <exclude>com.amazonaws:aws-java-sdk</exclude>
>> >>>>>
>> >>>>>          <exclude>com.typesafe.akka:akka-actor_*</exclude>
>> >>>>>
>> >>>>>          <exclude>com.typesafe.akka:akka-remote_*</exclude>
>> >>>>>
>> >>>>>          <exclude>com.typesafe.akka:akka-slf4j_*</exclude>
>> >>>>>
>> >>>>>          <exclude>io.netty:netty-all</exclude>
>> >>>>>
>> >>>>>          <exclude>io.netty:netty</exclude>
>> >>>>>
>> >>>>>          <exclude>org.eclipse.jetty:jetty-server</exclude>
>> >>>>>
>> >>>>>          <exclude>org.eclipse.jetty:jetty-continuation</exclude>
>> >>>>>
>> >>>>>          <exclude>org.eclipse.jetty:jetty-http</exclude>
>> >>>>>
>> >>>>>          <exclude>org.eclipse.jetty:jetty-io</exclude>
>> >>>>>
>> >>>>>          <exclude>org.eclipse.jetty:jetty-util</exclude>
>> >>>>>
>> >>>>>          <exclude>org.eclipse.jetty:jetty-security</exclude>
>> >>>>>
>> >>>>>          <exclude>org.eclipse.jetty:jetty-servlet</exclude>
>> >>>>>
>> >>>>>          <exclude>commons-fileupload:commons-fileupload</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.avro:avro</exclude>
>> >>>>>
>> >>>>>          <exclude>commons-collections:commons-collections</exclude>
>> >>>>>
>> >>>>>          <exclude>org.codehaus.jackson:jackson-core-asl</exclude>
>> >>>>>
>> >>>>>          <exclude>org.codehaus.jackson:jackson-mapper-asl</exclude>
>> >>>>>
>> >>>>>          <exclude>com.thoughtworks.paranamer:paranamer</exclude>
>> >>>>>
>> >>>>>          <exclude>org.xerial.snappy:snappy-java</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.commons:commons-compress</exclude>
>> >>>>>
>> >>>>>          <exclude>org.tukaani:xz</exclude>
>> >>>>>
>> >>>>>          <exclude>com.esotericsoftware.kryo:kryo</exclude>
>> >>>>>
>> >>>>>          <exclude>com.esotericsoftware.minlog:minlog</exclude>
>> >>>>>
>> >>>>>          <exclude>org.objenesis:objenesis</exclude>
>> >>>>>
>> >>>>>          <exclude>com.twitter:chill_*</exclude>
>> >>>>>
>> >>>>>          <exclude>com.twitter:chill-java</exclude>
>> >>>>>
>> >>>>>          <exclude>com.twitter:chill-avro_*</exclude>
>> >>>>>
>> >>>>>          <exclude>com.twitter:chill-bijection_*</exclude>
>> >>>>>
>> >>>>>          <exclude>com.twitter:bijection-core_*</exclude>
>> >>>>>
>> >>>>>          <exclude>com.twitter:bijection-avro_*</exclude>
>> >>>>>
>> >>>>>          <exclude>commons-lang:commons-lang</exclude>
>> >>>>>
>> >>>>>          <exclude>junit:junit</exclude>
>> >>>>>
>> >>>>>          <exclude>de.javakaffee:kryo-serializers</exclude>
>> >>>>>
>> >>>>>          <exclude>joda-time:joda-time</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.commons:commons-lang3</exclude>
>> >>>>>
>> >>>>>          <exclude>org.slf4j:slf4j-api</exclude>
>> >>>>>
>> >>>>>          <exclude>org.slf4j:slf4j-log4j12</exclude>
>> >>>>>
>> >>>>>          <exclude>log4j:log4j</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.commons:commons-math</exclude>
>> >>>>>
>> >>>>>
>> >>>>>    <exclude>org.apache.sling:org.apache.sling.commons.json</exclude>
>> >>>>>
>> >>>>>          <exclude>commons-logging:commons-logging</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.httpcomponents:httpclient</exclude>
>> >>>>>
>> >>>>>          <exclude>org.apache.httpcomponents:httpcore</exclude>
>> >>>>>
>> >>>>>          <exclude>commons-codec:commons-codec</exclude>
>> >>>>>
>> >>>>>          <exclude>com.fasterxml.jackson.core:jackson-core</exclude>
>> >>>>>
>> >>>>>
>> <exclude>com.fasterxml.jackson.core:jackson-databind</exclude>
>> >>>>>
>> >>>>>
>> >>>>>    <exclude>com.fasterxml.jackson.core:jackson-annotations</exclude>
>> >>>>>
>> >>>>>          <exclude>org.codehaus.jettison:jettison</exclude>
>> >>>>>
>> >>>>>          <exclude>stax:stax-api</exclude>
>> >>>>>
>> >>>>>          <exclude>com.typesafe:config</exclude>
>> >>>>>
>> >>>>>          <exclude>org.uncommons.maths:uncommons-maths</exclude>
>> >>>>>
>> >>>>>          <exclude>com.github.scopt:scopt_*</exclude>
>> >>>>>
>> >>>>>          <exclude>org.mortbay.jetty:servlet-api</exclude>
>> >>>>>
>> >>>>>          <exclude>commons-io:commons-io</exclude>
>> >>>>>
>> >>>>>          <exclude>commons-cli:commons-cli</exclude>
>> >>>>>
>> >>>>>    </excludes>
>> >>>>>
>> >>>>>    </artifactSet>
>> >>>>>                                                            <filters>
>> >>>>>
>> >>>>>    <filter>
>> >>>>>
>> >>>>>          <artifact>org.apache.flink:*</artifact>
>> >>>>>
>> >>>>>          <excludes>
>> >>>>>
>> >>>>>                  <exclude>org/apache/flink/shaded/**</exclude>
>> >>>>>
>> >>>>>                  <exclude>web-docs/**</exclude>
>> >>>>>
>> >>>>>          </excludes>
>> >>>>>
>> >>>>>    </filter>
>> >>>>>
>> </filters>
>> >>>>>
>> >>>>>    <transformers>
>> >>>>>
>> >>>>>    <!-- add Main-Class to manifest file -->
>> >>>>>
>> >>>>>    <transformer
>> >>>>>
>> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
>> >>>>>
>> >>>>>          <mainClass>com.otter.ist.flink.DaoJoin</mainClass>
>> >>>>>
>> >>>>>    </transformer>
>> >>>>>
>> >>>>>    </transformers>
>> >>>>>
>> >>>>>    <createDependencyReducedPom>false</createDependencyReducedPom>
>> >>>>>                                                    </configuration>
>> >>>>>                                            </execution>
>> >>>>>                                    </executions>
>> >>>>>                            </plugin>
>> >>>>>
>> >>>>>                            <!-- Configure the jar plugin to add the
>> >>>>>    main class as a manifest entry -->
>> >>>>>                            <plugin>
>> >>>>>
>> >>>>>    <groupId>org.apache.maven.plugins</groupId>
>> >>>>>
>> >>>>>    <artifactId>maven-jar-plugin</artifactId>
>> >>>>>                                    <version>2.5</version>
>> >>>>>                                    <configuration>
>> >>>>>                                            <archive>
>> >>>>>                                                    <manifestEntries>
>> >>>>>
>> >>>>>    <Main-Class>com.otter.ist.flink.DaoJoin</Main-Class>
>> >>>>>
>> </manifestEntries>
>> >>>>>                                            </archive>
>> >>>>>                                    </configuration>
>> >>>>>                            </plugin>
>> >>>>>
>> >>>>>                            <plugin>
>> >>>>>
>> >>>>>    <groupId>org.apache.maven.plugins</groupId>
>> >>>>>
>> >>>>>    <artifactId>maven-compiler-plugin</artifactId>
>> >>>>>                                    <version>3.1</version>
>> >>>>>                                    <configuration>
>> >>>>>                                            <source>1.8</source> <!--
>> >>>>>    If you want to use Java 8, change this to "1.8" -->
>> >>>>>                                            <target>1.8</target> <!--
>> >>>>>    If you want to use Java 8, change this to "1.8" -->
>> >>>>>                                    </configuration>
>> >>>>>                            </plugin>
>> >>>>>                    </plugins>
>> >>>>>
>> >>>>>
>> >>>>>                    <!-- If you want to use Java 8 Lambda Expressions
>> >>>>>    uncomment the following lines -->
>> >>>>>                    <!--
>> >>>>>                    <pluginManagement>
>> >>>>>                            <plugins>
>> >>>>>                                    <plugin>
>> >>>>>
>> >>>>>    <artifactId>maven-compiler-plugin</artifactId>
>> >>>>>                                            <configuration>
>> >>>>>
>> <source>1.8</source>
>> >>>>>
>> <target>1.8</target>
>> >>>>>
>> >>>>>    <compilerId>jdt</compilerId>
>> >>>>>                                            </configuration>
>> >>>>>                                            <dependencies>
>> >>>>>                                                    <dependency>
>> >>>>>
>> >>>>>    <groupId>org.eclipse.tycho</groupId>
>> >>>>>
>> >>>>>    <artifactId>tycho-compiler-jdt</artifactId>
>> >>>>>
>> >>>>>    <version>0.21.0</version>
>> >>>>>                                                    </dependency>
>> >>>>>                                            </dependencies>
>> >>>>>                                    </plugin>
>> >>>>>
>> >>>>>                                    <plugin>
>> >>>>>
>> >>>>>    <groupId>org.eclipse.m2e</groupId>
>> >>>>>
>> >>>>>    <artifactId>lifecycle-mapping</artifactId>
>> >>>>>                                            <version>1.0.0</version>
>> >>>>>                                            <configuration>
>> >>>>>
>> >>>>>    <lifecycleMappingMetadata>
>> >>>>>
>> >>>>>    <pluginExecutions>
>> >>>>>
>> >>>>>    <pluginExecution>
>> >>>>>
>> >>>>>          <pluginExecutionFilter>
>> >>>>>
>> >>>>>                  <groupId>org.apache.maven.plugins</groupId>
>> >>>>>
>> >>>>>                  <artifactId>maven-assembly-plugin</artifactId>
>> >>>>>
>> >>>>>                  <versionRange>[2.4,)</versionRange>
>> >>>>>
>> >>>>>                  <goals>
>> >>>>>
>> >>>>>                          <goal>single</goal>
>> >>>>>
>> >>>>>                  </goals>
>> >>>>>
>> >>>>>          </pluginExecutionFilter>
>> >>>>>
>> >>>>>          <action>
>> >>>>>
>> >>>>>                  <ignore/>
>> >>>>>
>> >>>>>          </action>
>> >>>>>
>> >>>>>    </pluginExecution>
>> >>>>>
>> >>>>>    <pluginExecution>
>> >>>>>
>> >>>>>          <pluginExecutionFilter>
>> >>>>>
>> >>>>>                  <groupId>org.apache.maven.plugins</groupId>
>> >>>>>
>> >>>>>                  <artifactId>maven-compiler-plugin</artifactId>
>> >>>>>
>> >>>>>                  <versionRange>[3.1,)</versionRange>
>> >>>>>
>> >>>>>                  <goals>
>> >>>>>
>> >>>>>                          <goal>testCompile</goal>
>> >>>>>
>> >>>>>                          <goal>compile</goal>
>> >>>>>
>> >>>>>                  </goals>
>> >>>>>
>> >>>>>          </pluginExecutionFilter>
>> >>>>>
>> >>>>>          <action>
>> >>>>>
>> >>>>>                  <ignore/>
>> >>>>>
>> >>>>>          </action>
>> >>>>>
>> >>>>>    </pluginExecution>
>> >>>>>
>> >>>>>    </pluginExecutions>
>> >>>>>
>> >>>>>    </lifecycleMappingMetadata>
>> >>>>>                                            </configuration>
>> >>>>>                                    </plugin>
>> >>>>>                            </plugins>
>> >>>>>                    </pluginManagement>
>> >>>>>                    -->
>> >>>>>
>> >>>>>            </build>
>> >>>>>            <profiles>
>> >>>>>                    <profile>
>> >>>>>                            <!-- A profile that does everyting
>> correctly:
>> >>>>>                            We set the Flink dependencies to
>> provided -->
>> >>>>>                            <id>build-jar</id>
>> >>>>>                            <activation>
>> >>>>>
>> >>>>>    <activeByDefault>false</activeByDefault>
>> >>>>>                            </activation>
>> >>>>>                            <dependencies>
>> >>>>>                                    <dependency>
>> >>>>>
>> >>>>>    <groupId>org.apache.flink</groupId>
>> >>>>>
>> >>>>>    <artifactId>flink-java</artifactId>
>> >>>>>
>> >>>>>    <version>0.10-SNAPSHOT</version>
>> >>>>>                                            <scope>provided</scope>
>> >>>>>                                    </dependency>
>> >>>>>                                    <dependency>
>> >>>>>
>> >>>>>    <groupId>org.apache.flink</groupId>
>> >>>>>
>> >>>>>    <artifactId>flink-streaming-core</artifactId>
>> >>>>>
>> >>>>>    <version>0.10-SNAPSHOT</version>
>> >>>>>                                            <scope>provided</scope>
>> >>>>>                                    </dependency>
>> >>>>>                                    <dependency>
>> >>>>>
>> >>>>>    <groupId>org.apache.flink</groupId>
>> >>>>>
>> >>>>>    <artifactId>flink-clients</artifactId>
>> >>>>>
>> >>>>>    <version>0.10-SNAPSHOT</version>
>> >>>>>                                            <scope>provided</scope>
>> >>>>>                                    </dependency>
>> >>>>>                            </dependencies>
>> >>>>>                    </profile>
>> >>>>>            </profiles>
>> >>>>>    </project>
>> >>>>>
>> >>>>>
>> >>>>
>> >>>
>> >>
>> >
>>
>>
>
>
>

Reply via email to