Hi Ram, Because there are some Scala IDE (Eclipse) plugins needed, I recommend to avoid `mvn eclipse:eclipse` command. Could you try just run `mvn clean install -DskipTests` and import the project to Scala IDE directly? In middle of importing process, Scala IDE suggests some plugins needed.
And which version of Scala IDE you are using? > On Jan 12, 2016, at 7:58 PM, Vasudevan, Ramkrishna S > <ramkrishna.s.vasude...@intel.com> wrote: > > Yes. I added it as Maven project only. I did mvn eclipse:eclipse to create > the project and also built the code using mvn clean install -DskipTests. > > Regards > Ram > > -----Original Message----- > From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf Of > Stephan Ewen > Sent: Tuesday, January 12, 2016 4:10 PM > To: dev@flink.apache.org > Subject: Re: Naive question > > Sorry to hear that it did not work out with Eclipse at all in the end, even > with all adjustments. > > Just making sure: You imported Flink as a Maven project, not manually adding > the big Flink dependency JAR? > > On Tue, Jan 12, 2016 at 5:15 AM, Vasudevan, Ramkrishna S < > ramkrishna.s.vasude...@intel.com> wrote: > >> Thanks to all. I tried with Scala Eclipse IDE with all these >> 'change-scala-version.sh'. But in vain. >> >> So I switched over to Intellij and thing work fine over there. I am >> new to Intellij so will try using it. >> >> Once again thanks for helping me out. >> >> Regards >> Ram >> >> -----Original Message----- >> From: Chiwan Park [mailto:chiwanp...@apache.org] >> Sent: Monday, January 11, 2016 4:37 PM >> To: dev@flink.apache.org >> Subject: Re: Naive question >> >> Hi Ram, >> >> If you want to build Flink with Scala 2.10, just checkout Flink >> repository from github or download source code from homepage, run `mvn >> clean install -DskipTests` and import projects to your IDE. If you >> want to build Flink with Scala 2.11, you have to run >> `tools/change-scala-version.sh 2.11` before build the project. You can >> revert Scala version change by running `tools/change-scala-version.sh 2.10`. >> >> About IDE, Flink community recommends IntelliJ IDEA because Scala IDE >> have some problems in Java/Scala mixed project like Flink. But I >> tested importing Flink project with Scala IDE 4.3.0, Scala 2.11.7 and >> Flink 0.10.0 source code. Note that you should import the project as maven >> project. >> >> By the way, the community welcomes any questions. Please feel free to >> post questions. :) >> >>> On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S < >> ramkrishna.s.vasude...@intel.com> wrote: >>> >>> Thank you very much for the reply. >>> I tried different ways and when I tried setting up the root pom.xml >>> to >>> 2.11 >>> >>> <scala.version>2.11.6</scala.version> >>> <scala.binary.version>2.11</scala.binary.version> >>> >>> I got the following error >>> [INFO] >>> -------------------------------------------------------------------- >>> -- >>> -- [ERROR] Failed to execute goal on project flink-scala: Could not >>> resolve depende ncies for project >>> org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find >>> artifact >>> org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central >>> (http://repo.mave >>> n.apache.org/maven2) -> [Help 1] >>> >>> If I leave the scala.binary.verson to be at 2.10 and the scala >>> version to be at 2.11.6 then I get the following problem [INFO] >>> C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling >>> [INFO] Compiling 366 source files to >>> C:\flink\flink\flink-runtime\target\test-cl >>> asses at 1452508064750 >>> [ERROR] >>> C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime >>> \j >>> ob >>> manager\JobManagerITCase.scala:700: error: can't expand macros >>> compiled by previ ous versions of Scala >>> [ERROR] assert(cachedGraph2.isArchived) >>> [ERROR] ^ >>> >>> So am not pretty sure how to proceed with this. If I try to change >>> the >> version of scala to 2.10 in the IDE then I get lot of compilation issues. >> IS there any way to over come this? >>> >>> Once again thanks a lot and apologies for the naïve question. >>> >>> Regards >>> Ram >>> -----Original Message----- >>> From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf >>> Of Stephan Ewen >>> Sent: Friday, January 8, 2016 5:01 PM >>> To: dev@flink.apache.org >>> Subject: Re: Naive question >>> >>> Hi! >>> >>> This looks like a mismatch between the Scala dependency in Flink and >>> Scala in your Eclipse. Make sure you use the same for both. By >>> default, Flink reference Scala 2.10 >>> >>> If your IDE is set up for Scala 2.11, set the Scala version variable >>> in the Flink root pom.xml also to 2.11 >>> >>> Greetings, >>> Stephan >>> >>> >>> >>> >>> On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S < >> ramkrishna.s.vasude...@intel.com> wrote: >>> >>>> I have been trying to install, learn and understand Flink. I am >>>> using >>>> Scala- EclipseIDE as my IDE. >>>> >>>> I have downloaded the flink source coded, compiled and created the >> project. >>>> >>>> My work laptop is Windows based and I don't have eclipse based >>>> workstation but I do have linux boxes for running and testing things. >>>> >>>> Some of the examples given in Flink source code do run directly >>>> from Eclipse but when I try to run the Wordcount example from >>>> Eclipse I get this error >>>> >>>> Exception in thread "main" java.lang.NoSuchMethodError: >>>> >> scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable >> /HashSet; >>>> at akka.actor.ActorCell$.<init>(ActorCell.scala:336) >>>> at akka.actor.ActorCell$.<clinit>(ActorCell.scala) >>>> at akka.actor.RootActorPath.$div(ActorPath.scala:159) >>>> at >> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464) >>>> at >> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452) >>>> at >>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>> Method) >>>> at >>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown >>>> Source) >>>> at >>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown >>>> Source) >>>> at java.lang.reflect.Constructor.newInstance(Unknown Source) >>>> at >>>> >> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply( >> DynamicAccess.scala:78) >>>> at scala.util.Try$.apply(Try.scala:191) >>>> at >>>> >> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca >> la:73) >>>> at >>>> >> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply( >> DynamicAccess.scala:84) >>>> at >>>> >> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply( >> DynamicAccess.scala:84) >>>> at scala.util.Success.flatMap(Try.scala:230) >>>> at >>>> >> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca >> la:84) >>>> at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585) >>>> at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578) >>>> at akka.actor.ActorSystem$.apply(ActorSystem.scala:142) >>>> at akka.actor.ActorSystem$.apply(ActorSystem.scala:119) >>>> at akka.actor.ActorSystem$.create(ActorSystem.scala:67) >>>> at >>>> >> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.s >> cala:84) >>>> at >>>> >> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerA >> ctorSystem(FlinkMiniCluster.scala:196) >>>> at >>>> >> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSyste >> m$lzycompute$1(FlinkMiniCluster.scala:225) >>>> at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org >>>> >> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$ >> 1(FlinkMiniCluster.scala:225) >>>> at >>>> >> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply >> (FlinkMiniCluster.scala:230) >>>> at >>>> >> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply >> (FlinkMiniCluster.scala:228) >>>> at >>>> >> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike. >> scala:245) >>>> at >>>> >> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike. >> scala:245) >>>> at scala.collection.immutable.Range.foreach(Range.scala:166) >>>> at >>>> scala.collection.TraversableLike$class.map(TraversableLike.scala:245) >>>> at scala.collection.AbstractTraversable.map(Traversable.scala:104) >>>> at >>>> >> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC >> luster.scala:228) >>>> at >>>> >> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC >> luster.scala:219) >>>> at >>>> >> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.exec >> ute(LocalStreamEnvironment.java:104) >>>> at >>>> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCo >>>> un >>>> t >>>> .java:80) >>>> >>>> I know this is a naïve question but I would like to get some help >>>> in order to over come this issue. I tried various options like >>>> setting >>>> scala-2.10 as the compiler for the project (then it shows >>>> completely different error) and many of the projects don't even >>>> compile. But with >>>> 2.11 version I get the above stack trace. Any help here is welcome. >>>> >>>> Regards >>>> Ram >>>> >> >> Regards, >> Chiwan Park >> Regards, Chiwan Park