Yes. I added it as Maven project only. I did mvn eclipse:eclipse to create the project and also built the code using mvn clean install -DskipTests.
Regards Ram -----Original Message----- From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf Of Stephan Ewen Sent: Tuesday, January 12, 2016 4:10 PM To: dev@flink.apache.org Subject: Re: Naive question Sorry to hear that it did not work out with Eclipse at all in the end, even with all adjustments. Just making sure: You imported Flink as a Maven project, not manually adding the big Flink dependency JAR? On Tue, Jan 12, 2016 at 5:15 AM, Vasudevan, Ramkrishna S < ramkrishna.s.vasude...@intel.com> wrote: > Thanks to all. I tried with Scala Eclipse IDE with all these > 'change-scala-version.sh'. But in vain. > > So I switched over to Intellij and thing work fine over there. I am > new to Intellij so will try using it. > > Once again thanks for helping me out. > > Regards > Ram > > -----Original Message----- > From: Chiwan Park [mailto:chiwanp...@apache.org] > Sent: Monday, January 11, 2016 4:37 PM > To: dev@flink.apache.org > Subject: Re: Naive question > > Hi Ram, > > If you want to build Flink with Scala 2.10, just checkout Flink > repository from github or download source code from homepage, run `mvn > clean install -DskipTests` and import projects to your IDE. If you > want to build Flink with Scala 2.11, you have to run > `tools/change-scala-version.sh 2.11` before build the project. You can > revert Scala version change by running `tools/change-scala-version.sh 2.10`. > > About IDE, Flink community recommends IntelliJ IDEA because Scala IDE > have some problems in Java/Scala mixed project like Flink. But I > tested importing Flink project with Scala IDE 4.3.0, Scala 2.11.7 and > Flink 0.10.0 source code. Note that you should import the project as maven > project. > > By the way, the community welcomes any questions. Please feel free to > post questions. :) > > > On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S < > ramkrishna.s.vasude...@intel.com> wrote: > > > > Thank you very much for the reply. > > I tried different ways and when I tried setting up the root pom.xml > > to > > 2.11 > > > > <scala.version>2.11.6</scala.version> > > <scala.binary.version>2.11</scala.binary.version> > > > > I got the following error > > [INFO] > > -------------------------------------------------------------------- > > -- > > -- [ERROR] Failed to execute goal on project flink-scala: Could not > > resolve depende ncies for project > > org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find > > artifact > > org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central > > (http://repo.mave > > n.apache.org/maven2) -> [Help 1] > > > > If I leave the scala.binary.verson to be at 2.10 and the scala > > version to be at 2.11.6 then I get the following problem [INFO] > > C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling > > [INFO] Compiling 366 source files to > > C:\flink\flink\flink-runtime\target\test-cl > > asses at 1452508064750 > > [ERROR] > > C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime > > \j > > ob > > manager\JobManagerITCase.scala:700: error: can't expand macros > > compiled by previ ous versions of Scala > > [ERROR] assert(cachedGraph2.isArchived) > > [ERROR] ^ > > > > So am not pretty sure how to proceed with this. If I try to change > > the > version of scala to 2.10 in the IDE then I get lot of compilation issues. > IS there any way to over come this? > > > > Once again thanks a lot and apologies for the naïve question. > > > > Regards > > Ram > > -----Original Message----- > > From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf > > Of Stephan Ewen > > Sent: Friday, January 8, 2016 5:01 PM > > To: dev@flink.apache.org > > Subject: Re: Naive question > > > > Hi! > > > > This looks like a mismatch between the Scala dependency in Flink and > > Scala in your Eclipse. Make sure you use the same for both. By > > default, Flink reference Scala 2.10 > > > > If your IDE is set up for Scala 2.11, set the Scala version variable > > in the Flink root pom.xml also to 2.11 > > > > Greetings, > > Stephan > > > > > > > > > > On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S < > ramkrishna.s.vasude...@intel.com> wrote: > > > >> I have been trying to install, learn and understand Flink. I am > >> using > >> Scala- EclipseIDE as my IDE. > >> > >> I have downloaded the flink source coded, compiled and created the > project. > >> > >> My work laptop is Windows based and I don't have eclipse based > >> workstation but I do have linux boxes for running and testing things. > >> > >> Some of the examples given in Flink source code do run directly > >> from Eclipse but when I try to run the Wordcount example from > >> Eclipse I get this error > >> > >> Exception in thread "main" java.lang.NoSuchMethodError: > >> > scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable > /HashSet; > >> at akka.actor.ActorCell$.<init>(ActorCell.scala:336) > >> at akka.actor.ActorCell$.<clinit>(ActorCell.scala) > >> at akka.actor.RootActorPath.$div(ActorPath.scala:159) > >> at > akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464) > >> at > akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452) > >> at > >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > >> Method) > >> at > >> sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown > >> Source) > >> at > >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown > >> Source) > >> at java.lang.reflect.Constructor.newInstance(Unknown Source) > >> at > >> > akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply( > DynamicAccess.scala:78) > >> at scala.util.Try$.apply(Try.scala:191) > >> at > >> > akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca > la:73) > >> at > >> > akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply( > DynamicAccess.scala:84) > >> at > >> > akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply( > DynamicAccess.scala:84) > >> at scala.util.Success.flatMap(Try.scala:230) > >> at > >> > akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca > la:84) > >> at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585) > >> at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578) > >> at akka.actor.ActorSystem$.apply(ActorSystem.scala:142) > >> at akka.actor.ActorSystem$.apply(ActorSystem.scala:119) > >> at akka.actor.ActorSystem$.create(ActorSystem.scala:67) > >> at > >> > org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.s > cala:84) > >> at > >> > org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerA > ctorSystem(FlinkMiniCluster.scala:196) > >> at > >> > org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSyste > m$lzycompute$1(FlinkMiniCluster.scala:225) > >> at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org > >> > $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$ > 1(FlinkMiniCluster.scala:225) > >> at > >> > org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply > (FlinkMiniCluster.scala:230) > >> at > >> > org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply > (FlinkMiniCluster.scala:228) > >> at > >> > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike. > scala:245) > >> at > >> > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike. > scala:245) > >> at scala.collection.immutable.Range.foreach(Range.scala:166) > >> at > >> scala.collection.TraversableLike$class.map(TraversableLike.scala:245) > >> at scala.collection.AbstractTraversable.map(Traversable.scala:104) > >> at > >> > org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC > luster.scala:228) > >> at > >> > org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC > luster.scala:219) > >> at > >> > org.apache.flink.streaming.api.environment.LocalStreamEnvironment.exec > ute(LocalStreamEnvironment.java:104) > >> at > >> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCo > >> un > >> t > >> .java:80) > >> > >> I know this is a naïve question but I would like to get some help > >> in order to over come this issue. I tried various options like > >> setting > >> scala-2.10 as the compiler for the project (then it shows > >> completely different error) and many of the projects don't even > >> compile. But with > >> 2.11 version I get the above stack trace. Any help here is welcome. > >> > >> Regards > >> Ram > >> > > Regards, > Chiwan Park > > >