I tried this before without success. Same exception the named class was not found.
-Michael Am 06.08.2015 um 14:07 schrieb Matthias J. Sax <mj...@informatik.hu-berlin.de>: > Never mind. Just saw that this in not the problem... > > Sounds weird to me. Maybe you can try to name the class. Anonymous > classes should not be a problem, but it should be worth a try. > > -Matthias > > > > On 08/06/2015 01:51 PM, Matthias J. Sax wrote: >> If I see it correctly your jar contains >> >>> com/davengo/rfidcloud/flink/DaoJoin$1.class >> >> But your error message says >> >>> ClassNotFoundException: com.otter.ist.flink.DaoJoin$1 >> >> Both are different packages. Your jar seems not be correctly packaged. >> >> >> -Matthias >> >> On 08/06/2015 12:46 PM, Michael Huelfenhaus wrote: >>> I am back at work next Tuesday, any further ideas would be great until >>> then, for now I am continuing inside ecplise. >>> >>> Am 06.08.2015 um 11:27 schrieb Michael Huelfenhaus >>> <m.huelfenh...@davengo.com <mailto:m.huelfenh...@davengo.com>>: >>> >>>> hi, >>>> >>>>> how did you build the jar file? >>>> >>>> mvn clean install -Pbuild-jar >>>> >>>>> Have you checked whether your classes are in the jar file? >>>> >>>> yes, this seems alright for me >>>> >>>>> jar tf target/flink-test-0.1.jar >>>> META-INF/MANIFEST.MF >>>> META-INF/ >>>> com/ >>>> com/davengo/ >>>> com/davengo/rfidcloud/ >>>> com/davengo/rfidcloud/flink/ >>>> com/davengo/rfidcloud/flink/DaoJoin$1.class >>>> com/davengo/rfidcloud/flink/DaoJoin.class >>>> com/davengo/rfidcloud/flink/streampojos/ >>>> com/davengo/rfidcloud/flink/streampojos/EpcTuple.class >>>> log4j.properties >>>> META-INF/maven/ >>>> META-INF/maven/com.davengo.rfidcloud.flink/ >>>> META-INF/maven/com.davengo.rfidcloud.flink/flink-test/ >>>> META-INF/maven/com.davengo.rfidcloud.flink/flink-test/pom.xml >>>> META-INF/maven/com.davengo.rfidcloud.flink/flink-test/pom.properties >>>> >>>> Am 06.08.2015 um 11:21 schrieb Robert Metzger <rmetz...@apache.org >>>> <mailto:rmetz...@apache.org>>: >>>> >>>>> Hi, >>>>> >>>>> how did you build the jar file? >>>>> Have you checked whether your classes are in the jar file? >>>>> >>>>> On Thu, Aug 6, 2015 at 11:08 AM, Michael Huelfenhaus >>>>> <m.huelfenh...@davengo.com <mailto:m.huelfenh...@davengo.com>> wrote: >>>>> >>>>> Hello everybody >>>>> >>>>> I am truing to build a very simple streaming application with the >>>>> nightly build of flink 0.10, my code runs fine in eclipse. >>>>> >>>>> But when I build and deploy the jar locally I always get >>>>> java.lang.ClassNotFoundException: com.otter.ist.flink.DaoJoin$1 >>>>> >>>>> There is also no plan visible in the web interface. >>>>> >>>>> I start the local flink 0.10 with start-local-streaming.sh after >>>>> building it from the git code >>>>> >>>>> Below you find the complete error, my code and the pom.xml any >>>>> help is appreciated. >>>>> >>>>> Cheers Michael >>>>> >>>>> >>>>> error log from web interface: >>>>> An error occurred while invoking the program: >>>>> >>>>> The main method caused an error. >>>>> >>>>> >>>>> org.apache.flink.runtime.client.JobExecutionException: Job >>>>> execution failed. >>>>> at >>>>> >>>>> org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1.applyOrElse(JobManager.scala:364) >>>>> at >>>>> >>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) >>>>> at >>>>> >>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) >>>>> at >>>>> >>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) >>>>> at >>>>> >>>>> org.apache.flink.runtime.LeaderSessionMessages$$anonfun$receive$1.applyOrElse(LeaderSessionMessages.scala:40) >>>>> at >>>>> >>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) >>>>> at >>>>> >>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) >>>>> at >>>>> >>>>> scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) >>>>> at >>>>> >>>>> org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33) >>>>> at >>>>> >>>>> org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28) >>>>> at >>>>> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) >>>>> at >>>>> >>>>> org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28) >>>>> at akka.actor.Actor$class.aroundReceive(Actor.scala:465) >>>>> at >>>>> >>>>> org.apache.flink.runtime.jobmanager.JobManager.aroundReceive(JobManager.scala:101) >>>>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) >>>>> at akka.actor.ActorCell.invoke(ActorCell.scala:487) >>>>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) >>>>> at akka.dispatch.Mailbox.run(Mailbox.scala:221) >>>>> at akka.dispatch.Mailbox.exec(Mailbox.scala:231) >>>>> at >>>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>>>> at >>>>> >>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >>>>> at >>>>> >>>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>>>> at >>>>> >>>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>>>> Caused by: java.lang.Exception: Call to registerInputOutput() of >>>>> invokable failed >>>>> at >>>>> org.apache.flink.runtime.taskmanager.Task.run(Task.java:526) >>>>> at java.lang.Thread.run(Thread.java:745) >>>>> Caused by: >>>>> org.apache.flink.streaming.runtime.tasks.StreamTaskException: >>>>> Cannot instantiate user function. >>>>> at >>>>> >>>>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperator(StreamConfig.java:207) >>>>> at >>>>> >>>>> org.apache.flink.streaming.runtime.tasks.OutputHandler.createChainedCollector(OutputHandler.java:173) >>>>> at >>>>> >>>>> org.apache.flink.streaming.runtime.tasks.OutputHandler.createChainedCollector(OutputHandler.java:159) >>>>> at >>>>> >>>>> org.apache.flink.streaming.runtime.tasks.OutputHandler.<init>(OutputHandler.java:107) >>>>> at >>>>> >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.registerInputOutput(StreamTask.java:99) >>>>> at >>>>> org.apache.flink.runtime.taskmanager.Task.run(Task.java:523) >>>>> ... 1 more >>>>> Caused by: java.lang.ClassNotFoundException: >>>>> com.otter.ist.flink.DaoJoin$1 >>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:372) >>>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:360) >>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>>> at java.lang.Class.forName0(Native Method) >>>>> at java.lang.Class.forName(Class.java:344) >>>>> at >>>>> >>>>> org.apache.flink.util.InstantiationUtil$ClassLoaderObjectInputStream.resolveClass(InstantiationUtil.java:71) >>>>> at >>>>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1613) >>>>> at >>>>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1518) >>>>> at >>>>> >>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1774) >>>>> at >>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>>>> at >>>>> >>>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1993) >>>>> at >>>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1918) >>>>> at >>>>> >>>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) >>>>> at >>>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) >>>>> at >>>>> java.io.ObjectInputStream.readObject(ObjectInputStream.java:371) >>>>> at >>>>> >>>>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:302) >>>>> at >>>>> >>>>> org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:264) >>>>> at >>>>> >>>>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperator(StreamConfig.java:205) >>>>> ... 6 more >>>>> >>>>> my code: >>>>> >>>>> package com.otter.ist.flink; >>>>> >>>>> import org.apache.flink.api.common.functions.MapFunction; >>>>> import org.apache.flink.api.java.tuple.Tuple3; >>>>> import org.apache.flink.streaming.api.datastream.DataStream; >>>>> import >>>>> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; >>>>> >>>>> public class DaoJoin { >>>>> >>>>> public static void main(String[] args) throws Exception { >>>>> // >>>>> >>>>> ************************************************************************* >>>>> // PROGRAM >>>>> // >>>>> >>>>> ************************************************************************* >>>>> >>>>> if (!parseParameters(args)) { >>>>> return; >>>>> } >>>>> >>>>> // set up the execution environment >>>>> final StreamExecutionEnvironment env = >>>>> StreamExecutionEnvironment.createLocalEnvironment(); >>>>> // final StreamExecutionEnvironment env = >>>>> StreamExecutionEnvironment.getExecutionEnvironment(); >>>>> >>>>> >>>>> // get input data >>>>> DataStream<String> text = getTextDataStream(env); >>>>> >>>>> DataStream<Tuple3<String, String, Integer>> epcs >>>>> = text.map(new MapFunction<String, Tuple3<String, String, >>>>> Integer>>(){ >>>>> >>>>> private static final long >>>>> serialVersionUID = -7889264579632622427L; >>>>> >>>>> @Override >>>>> public Tuple3<String, String, Integer> >>>>> map(String line) throws Exception { >>>>> String[] fields = line.split(" "); >>>>> >>>>> return new Tuple3<String, String, >>>>> Integer>(fields[0], fields[1], Integer.parseInt(fields[2])); >>>>> } >>>>> >>>>> }); >>>>> >>>>> // emit result >>>>> if (fileOutput) { >>>>> epcs.writeAsText(outputPath); >>>>> } else { >>>>> epcs.print(); >>>>> } >>>>> System.out.println(env.getExecutionPlan()); >>>>> >>>>> // execute program >>>>> env.execute("DaoJoin"); >>>>> } >>>>> >>>>> // >>>>> >>>>> ************************************************************************* >>>>> // UTIL METHODS >>>>> // >>>>> >>>>> ************************************************************************* >>>>> >>>>> private static boolean fileOutput = false; >>>>> private static String textPath; >>>>> private static String outputPath; >>>>> >>>>> private static boolean parseParameters(String[] args) { >>>>> >>>>> if (args.length > 0) { >>>>> // parse input arguments >>>>> fileOutput = true; >>>>> if (args.length == 2) { >>>>> textPath = args[0]; >>>>> outputPath = args[1]; >>>>> } else { >>>>> System.err.println("Usage: >>>>> DaoJoin <text path> <result path>"); >>>>> return false; >>>>> } >>>>> System.out.println("fileout: " + fileOutput); >>>>> } else { >>>>> System.out.println("Executing WordCount >>>>> example with built-in default data."); >>>>> System.out.println(" Provide parameters >>>>> to read input data from a file."); >>>>> System.out.println(" Usage: WordCount >>>>> <text path> <result path>"); >>>>> } >>>>> return true; >>>>> } >>>>> >>>>> private static DataStream<String> >>>>> getTextDataStream(StreamExecutionEnvironment env) { >>>>> // read the text file from given input path >>>>> return env.readTextFile(textPath); >>>>> } >>>>> } >>>>> >>>>> >>>>> the pom.xml >>>>> <project xmlns="http://maven.apache.org/POM/4.0.0" >>>>> xmlns:xsi="http://xwww.w3.org/2001/XMLSchema-instance" >>>>> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 >>>>> http://maven.apache.org/xsd/maven-4.0.0.xsd"> >>>>> <modelVersion>4.0.0</modelVersion> >>>>> >>>>> <groupId>com.otter.ist.flink</groupId> >>>>> <artifactId>flink-test</artifactId> >>>>> <version>0.1</version> >>>>> <packaging>jar</packaging> >>>>> >>>>> <name>DaoJoin</name> >>>>> <url>http://www.otter.com <http://www.otter.com/></url> >>>>> >>>>> <properties> >>>>> >>>>> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> >>>>> </properties> >>>>> >>>>> <repositories> >>>>> <repository> >>>>> <id>apache.snapshots</id> >>>>> <name>Apache Development Snapshot >>>>> Repository</name> >>>>> >>>>> >>>>> <url>https://repository.apache.org/content/repositories/snapshots/</url> >>>>> <releases> >>>>> <enabled>false</enabled> >>>>> </releases> >>>>> <snapshots> >>>>> <enabled>true</enabled> >>>>> </snapshots> >>>>> </repository> >>>>> </repositories> >>>>> >>>>> <!-- >>>>> >>>>> Execute "mvn clean package -Pbuild-jar" >>>>> to build a jar file out of this project! >>>>> >>>>> How to use the Flink Quickstart pom: >>>>> >>>>> a) Adding new dependencies: >>>>> You can add dependencies to the list below. >>>>> Please check if the maven-shade-plugin >>>>> below is filtering out your dependency >>>>> and remove the exclude from there. >>>>> >>>>> b) Build a jar for running on the cluster: >>>>> There are two options for creating a jar >>>>> from this project >>>>> >>>>> b.1) "mvn clean package" -> this will >>>>> create a fat jar which contains all >>>>> dependencies necessary >>>>> for running the jar created by this pom in a cluster. >>>>> The "maven-shade-plugin" >>>>> excludes everything that is provided on a running Flink cluster. >>>>> >>>>> b.2) "mvn clean package -Pbuild-jar" -> >>>>> This will also create a fat-jar, but with much >>>>> nicer dependency >>>>> exclusion handling. This approach is preferred and leads to >>>>> much cleaner jar files. >>>>> --> >>>>> >>>>> <dependencies> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-java</artifactId> >>>>> <version>0.10-SNAPSHOT</version> >>>>> </dependency> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-streaming-core</artifactId> >>>>> <version>0.10-SNAPSHOT</version> >>>>> </dependency> >>>>> <dependency> >>>>> <groupId>org.apache.flink</groupId> >>>>> <artifactId>flink-clients</artifactId> >>>>> <version>0.10-SNAPSHOT</version> >>>>> </dependency> >>>>> </dependencies> >>>>> >>>>> <build> >>>>> <plugins> >>>>> <!-- We use the maven-shade plugin to >>>>> create a fat jar that contains all dependencies >>>>> except flink and it's transitive >>>>> dependencies. The resulting fat-jar can be executed >>>>> on a cluster. Change the value of >>>>> Program-Class if your program entry point changes. --> >>>>> <plugin> >>>>> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> >>>>> <artifactId>maven-shade-plugin</artifactId> >>>>> <version>2.3</version> >>>>> <executions> >>>>> <!-- Run shade goal on >>>>> package phase --> >>>>> <execution> >>>>> >>>>> <phase>package</phase> >>>>> <goals> >>>>> >>>>> <goal>shade</goal> >>>>> </goals> >>>>> <configuration> >>>>> <artifactSet> >>>>> >>>>> <excludes> >>>>> >>>>> <!-- This list contains all dependencies of flink-dist >>>>> >>>>> Everything else will be packaged into the fat-jar >>>>> >>>>> --> >>>>> >>>>> <exclude>org.apache.flink:flink-shaded-*</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-core</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-java</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-scala</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-runtime</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-optimizer</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-clients</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-spargel</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-avro</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-java-examples</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-scala-examples</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-streaming-examples</exclude> >>>>> >>>>> <exclude>org.apache.flink:flink-streaming-core</exclude> >>>>> >>>>> >>>>> <!-- Also exclude very big transitive dependencies of Flink >>>>> >>>>> >>>>> WARNING: You have to remove these excludes if your code >>>>> relies on other >>>>> >>>>> versions of these dependencies. >>>>> >>>>> >>>>> --> >>>>> >>>>> <exclude>org.scala-lang:scala-library</exclude> >>>>> >>>>> <exclude>org.scala-lang:scala-compiler</exclude> >>>>> >>>>> <exclude>org.scala-lang:scala-reflect</exclude> >>>>> >>>>> <exclude>com.amazonaws:aws-java-sdk</exclude> >>>>> >>>>> <exclude>com.typesafe.akka:akka-actor_*</exclude> >>>>> >>>>> <exclude>com.typesafe.akka:akka-remote_*</exclude> >>>>> >>>>> <exclude>com.typesafe.akka:akka-slf4j_*</exclude> >>>>> >>>>> <exclude>io.netty:netty-all</exclude> >>>>> >>>>> <exclude>io.netty:netty</exclude> >>>>> >>>>> <exclude>org.eclipse.jetty:jetty-server</exclude> >>>>> >>>>> <exclude>org.eclipse.jetty:jetty-continuation</exclude> >>>>> >>>>> <exclude>org.eclipse.jetty:jetty-http</exclude> >>>>> >>>>> <exclude>org.eclipse.jetty:jetty-io</exclude> >>>>> >>>>> <exclude>org.eclipse.jetty:jetty-util</exclude> >>>>> >>>>> <exclude>org.eclipse.jetty:jetty-security</exclude> >>>>> >>>>> <exclude>org.eclipse.jetty:jetty-servlet</exclude> >>>>> >>>>> <exclude>commons-fileupload:commons-fileupload</exclude> >>>>> >>>>> <exclude>org.apache.avro:avro</exclude> >>>>> >>>>> <exclude>commons-collections:commons-collections</exclude> >>>>> >>>>> <exclude>org.codehaus.jackson:jackson-core-asl</exclude> >>>>> >>>>> <exclude>org.codehaus.jackson:jackson-mapper-asl</exclude> >>>>> >>>>> <exclude>com.thoughtworks.paranamer:paranamer</exclude> >>>>> >>>>> <exclude>org.xerial.snappy:snappy-java</exclude> >>>>> >>>>> <exclude>org.apache.commons:commons-compress</exclude> >>>>> >>>>> <exclude>org.tukaani:xz</exclude> >>>>> >>>>> <exclude>com.esotericsoftware.kryo:kryo</exclude> >>>>> >>>>> <exclude>com.esotericsoftware.minlog:minlog</exclude> >>>>> >>>>> <exclude>org.objenesis:objenesis</exclude> >>>>> >>>>> <exclude>com.twitter:chill_*</exclude> >>>>> >>>>> <exclude>com.twitter:chill-java</exclude> >>>>> >>>>> <exclude>com.twitter:chill-avro_*</exclude> >>>>> >>>>> <exclude>com.twitter:chill-bijection_*</exclude> >>>>> >>>>> <exclude>com.twitter:bijection-core_*</exclude> >>>>> >>>>> <exclude>com.twitter:bijection-avro_*</exclude> >>>>> >>>>> <exclude>commons-lang:commons-lang</exclude> >>>>> >>>>> <exclude>junit:junit</exclude> >>>>> >>>>> <exclude>de.javakaffee:kryo-serializers</exclude> >>>>> >>>>> <exclude>joda-time:joda-time</exclude> >>>>> >>>>> <exclude>org.apache.commons:commons-lang3</exclude> >>>>> >>>>> <exclude>org.slf4j:slf4j-api</exclude> >>>>> >>>>> <exclude>org.slf4j:slf4j-log4j12</exclude> >>>>> >>>>> <exclude>log4j:log4j</exclude> >>>>> >>>>> <exclude>org.apache.commons:commons-math</exclude> >>>>> >>>>> >>>>> <exclude>org.apache.sling:org.apache.sling.commons.json</exclude> >>>>> >>>>> <exclude>commons-logging:commons-logging</exclude> >>>>> >>>>> <exclude>org.apache.httpcomponents:httpclient</exclude> >>>>> >>>>> <exclude>org.apache.httpcomponents:httpcore</exclude> >>>>> >>>>> <exclude>commons-codec:commons-codec</exclude> >>>>> >>>>> <exclude>com.fasterxml.jackson.core:jackson-core</exclude> >>>>> >>>>> <exclude>com.fasterxml.jackson.core:jackson-databind</exclude> >>>>> >>>>> >>>>> <exclude>com.fasterxml.jackson.core:jackson-annotations</exclude> >>>>> >>>>> <exclude>org.codehaus.jettison:jettison</exclude> >>>>> >>>>> <exclude>stax:stax-api</exclude> >>>>> >>>>> <exclude>com.typesafe:config</exclude> >>>>> >>>>> <exclude>org.uncommons.maths:uncommons-maths</exclude> >>>>> >>>>> <exclude>com.github.scopt:scopt_*</exclude> >>>>> >>>>> <exclude>org.mortbay.jetty:servlet-api</exclude> >>>>> >>>>> <exclude>commons-io:commons-io</exclude> >>>>> >>>>> <exclude>commons-cli:commons-cli</exclude> >>>>> >>>>> </excludes> >>>>> >>>>> </artifactSet> >>>>> <filters> >>>>> >>>>> <filter> >>>>> >>>>> <artifact>org.apache.flink:*</artifact> >>>>> >>>>> <excludes> >>>>> >>>>> <exclude>org/apache/flink/shaded/**</exclude> >>>>> >>>>> <exclude>web-docs/**</exclude> >>>>> >>>>> </excludes> >>>>> >>>>> </filter> >>>>> </filters> >>>>> >>>>> <transformers> >>>>> >>>>> <!-- add Main-Class to manifest file --> >>>>> >>>>> <transformer >>>>> >>>>> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer"> >>>>> >>>>> <mainClass>com.otter.ist.flink.DaoJoin</mainClass> >>>>> >>>>> </transformer> >>>>> >>>>> </transformers> >>>>> >>>>> <createDependencyReducedPom>false</createDependencyReducedPom> >>>>> </configuration> >>>>> </execution> >>>>> </executions> >>>>> </plugin> >>>>> >>>>> <!-- Configure the jar plugin to add the >>>>> main class as a manifest entry --> >>>>> <plugin> >>>>> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> >>>>> <artifactId>maven-jar-plugin</artifactId> >>>>> <version>2.5</version> >>>>> <configuration> >>>>> <archive> >>>>> <manifestEntries> >>>>> >>>>> <Main-Class>com.otter.ist.flink.DaoJoin</Main-Class> >>>>> </manifestEntries> >>>>> </archive> >>>>> </configuration> >>>>> </plugin> >>>>> >>>>> <plugin> >>>>> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> >>>>> <artifactId>maven-compiler-plugin</artifactId> >>>>> <version>3.1</version> >>>>> <configuration> >>>>> <source>1.8</source> <!-- >>>>> If you want to use Java 8, change this to "1.8" --> >>>>> <target>1.8</target> <!-- >>>>> If you want to use Java 8, change this to "1.8" --> >>>>> </configuration> >>>>> </plugin> >>>>> </plugins> >>>>> >>>>> >>>>> <!-- If you want to use Java 8 Lambda Expressions >>>>> uncomment the following lines --> >>>>> <!-- >>>>> <pluginManagement> >>>>> <plugins> >>>>> <plugin> >>>>> >>>>> <artifactId>maven-compiler-plugin</artifactId> >>>>> <configuration> >>>>> <source>1.8</source> >>>>> <target>1.8</target> >>>>> >>>>> <compilerId>jdt</compilerId> >>>>> </configuration> >>>>> <dependencies> >>>>> <dependency> >>>>> >>>>> <groupId>org.eclipse.tycho</groupId> >>>>> >>>>> <artifactId>tycho-compiler-jdt</artifactId> >>>>> >>>>> <version>0.21.0</version> >>>>> </dependency> >>>>> </dependencies> >>>>> </plugin> >>>>> >>>>> <plugin> >>>>> >>>>> <groupId>org.eclipse.m2e</groupId> >>>>> >>>>> <artifactId>lifecycle-mapping</artifactId> >>>>> <version>1.0.0</version> >>>>> <configuration> >>>>> >>>>> <lifecycleMappingMetadata> >>>>> >>>>> <pluginExecutions> >>>>> >>>>> <pluginExecution> >>>>> >>>>> <pluginExecutionFilter> >>>>> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> >>>>> <artifactId>maven-assembly-plugin</artifactId> >>>>> >>>>> <versionRange>[2.4,)</versionRange> >>>>> >>>>> <goals> >>>>> >>>>> <goal>single</goal> >>>>> >>>>> </goals> >>>>> >>>>> </pluginExecutionFilter> >>>>> >>>>> <action> >>>>> >>>>> <ignore/> >>>>> >>>>> </action> >>>>> >>>>> </pluginExecution> >>>>> >>>>> <pluginExecution> >>>>> >>>>> <pluginExecutionFilter> >>>>> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> >>>>> <artifactId>maven-compiler-plugin</artifactId> >>>>> >>>>> <versionRange>[3.1,)</versionRange> >>>>> >>>>> <goals> >>>>> >>>>> <goal>testCompile</goal> >>>>> >>>>> <goal>compile</goal> >>>>> >>>>> </goals> >>>>> >>>>> </pluginExecutionFilter> >>>>> >>>>> <action> >>>>> >>>>> <ignore/> >>>>> >>>>> </action> >>>>> >>>>> </pluginExecution> >>>>> >>>>> </pluginExecutions> >>>>> >>>>> </lifecycleMappingMetadata> >>>>> </configuration> >>>>> </plugin> >>>>> </plugins> >>>>> </pluginManagement> >>>>> --> >>>>> >>>>> </build> >>>>> <profiles> >>>>> <profile> >>>>> <!-- A profile that does everyting correctly: >>>>> We set the Flink dependencies to provided --> >>>>> <id>build-jar</id> >>>>> <activation> >>>>> >>>>> <activeByDefault>false</activeByDefault> >>>>> </activation> >>>>> <dependencies> >>>>> <dependency> >>>>> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-java</artifactId> >>>>> >>>>> <version>0.10-SNAPSHOT</version> >>>>> <scope>provided</scope> >>>>> </dependency> >>>>> <dependency> >>>>> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-streaming-core</artifactId> >>>>> >>>>> <version>0.10-SNAPSHOT</version> >>>>> <scope>provided</scope> >>>>> </dependency> >>>>> <dependency> >>>>> >>>>> <groupId>org.apache.flink</groupId> >>>>> >>>>> <artifactId>flink-clients</artifactId> >>>>> >>>>> <version>0.10-SNAPSHOT</version> >>>>> <scope>provided</scope> >>>>> </dependency> >>>>> </dependencies> >>>>> </profile> >>>>> </profiles> >>>>> </project> >>>>> >>>>> >>>> >>> >> >