This is scala IDE Release 4.4.0. So without doing mvn eclipse:eclipse - how to you import the project directly?
Regards Ram -----Original Message----- From: Chiwan Park [mailto:chiwanp...@apache.org] Sent: Tuesday, January 12, 2016 4:54 PM To: dev@flink.apache.org Subject: Re: Naive question Because I tested with Scala IDE 4.3.0 only, the process in the documentation is slightly different with my experience. > On Jan 12, 2016, at 8:21 PM, Stephan Ewen <se...@apache.org> wrote: > > @Chiwan: Is this still up to date from your experience? > > https://ci.apache.org/projects/flink/flink-docs-release-0.10/internals > /ide_setup.html > > On Tue, Jan 12, 2016 at 12:04 PM, Chiwan Park <chiwanp...@apache.org> wrote: > >> Hi Ram, >> >> Because there are some Scala IDE (Eclipse) plugins needed, I >> recommend to avoid `mvn eclipse:eclipse` command. Could you try just >> run `mvn clean install -DskipTests` and import the project to Scala >> IDE directly? In middle of importing process, Scala IDE suggests some >> plugins needed. >> >> And which version of Scala IDE you are using? >> >>> On Jan 12, 2016, at 7:58 PM, Vasudevan, Ramkrishna S < >> ramkrishna.s.vasude...@intel.com> wrote: >>> >>> Yes. I added it as Maven project only. I did mvn eclipse:eclipse to >> create the project and also built the code using mvn clean install >> -DskipTests. >>> >>> Regards >>> Ram >>> >>> -----Original Message----- >>> From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf >>> Of >> Stephan Ewen >>> Sent: Tuesday, January 12, 2016 4:10 PM >>> To: dev@flink.apache.org >>> Subject: Re: Naive question >>> >>> Sorry to hear that it did not work out with Eclipse at all in the >>> end, >> even with all adjustments. >>> >>> Just making sure: You imported Flink as a Maven project, not >>> manually >> adding the big Flink dependency JAR? >>> >>> On Tue, Jan 12, 2016 at 5:15 AM, Vasudevan, Ramkrishna S < >> ramkrishna.s.vasude...@intel.com> wrote: >>> >>>> Thanks to all. I tried with Scala Eclipse IDE with all these >>>> 'change-scala-version.sh'. But in vain. >>>> >>>> So I switched over to Intellij and thing work fine over there. I am >>>> new to Intellij so will try using it. >>>> >>>> Once again thanks for helping me out. >>>> >>>> Regards >>>> Ram >>>> >>>> -----Original Message----- >>>> From: Chiwan Park [mailto:chiwanp...@apache.org] >>>> Sent: Monday, January 11, 2016 4:37 PM >>>> To: dev@flink.apache.org >>>> Subject: Re: Naive question >>>> >>>> Hi Ram, >>>> >>>> If you want to build Flink with Scala 2.10, just checkout Flink >>>> repository from github or download source code from homepage, run >>>> `mvn clean install -DskipTests` and import projects to your IDE. If >>>> you want to build Flink with Scala 2.11, you have to run >>>> `tools/change-scala-version.sh 2.11` before build the project. You >>>> can revert Scala version change by running >>>> `tools/change-scala-version.sh >> 2.10`. >>>> >>>> About IDE, Flink community recommends IntelliJ IDEA because Scala >>>> IDE have some problems in Java/Scala mixed project like Flink. But >>>> I tested importing Flink project with Scala IDE 4.3.0, Scala 2.11.7 >>>> and Flink 0.10.0 source code. Note that you should import the >>>> project as >> maven project. >>>> >>>> By the way, the community welcomes any questions. Please feel free >>>> to post questions. :) >>>> >>>>> On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S < >>>> ramkrishna.s.vasude...@intel.com> wrote: >>>>> >>>>> Thank you very much for the reply. >>>>> I tried different ways and when I tried setting up the root >>>>> pom.xml to >>>>> 2.11 >>>>> >>>>> <scala.version>2.11.6</scala.version> >>>>> <scala.binary.version>2.11</scala.binary.version> >>>>> >>>>> I got the following error >>>>> [INFO] >>>>> ------------------------------------------------------------------ >>>>> -- >>>>> -- >>>>> -- [ERROR] Failed to execute goal on project flink-scala: Could >>>>> not resolve depende ncies for project >>>>> org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find >>>>> artifact >>>>> org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central >>>>> (http://repo.mave >>>>> n.apache.org/maven2) -> [Help 1] >>>>> >>>>> If I leave the scala.binary.verson to be at 2.10 and the scala >>>>> version to be at 2.11.6 then I get the following problem [INFO] >>>>> C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling >>>>> [INFO] Compiling 366 source files to >>>>> C:\flink\flink\flink-runtime\target\test-cl >>>>> asses at 1452508064750 >>>>> [ERROR] >>>>> C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runti >>>>> me >>>>> \j >>>>> ob >>>>> manager\JobManagerITCase.scala:700: error: can't expand macros >>>>> compiled by previ ous versions of Scala >>>>> [ERROR] assert(cachedGraph2.isArchived) >>>>> [ERROR] ^ >>>>> >>>>> So am not pretty sure how to proceed with this. If I try to change >>>>> the >>>> version of scala to 2.10 in the IDE then I get lot of compilation >> issues. >>>> IS there any way to over come this? >>>>> >>>>> Once again thanks a lot and apologies for the naïve question. >>>>> >>>>> Regards >>>>> Ram >>>>> -----Original Message----- >>>>> From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On >>>>> Behalf Of Stephan Ewen >>>>> Sent: Friday, January 8, 2016 5:01 PM >>>>> To: dev@flink.apache.org >>>>> Subject: Re: Naive question >>>>> >>>>> Hi! >>>>> >>>>> This looks like a mismatch between the Scala dependency in Flink >>>>> and Scala in your Eclipse. Make sure you use the same for both. By >>>>> default, Flink reference Scala 2.10 >>>>> >>>>> If your IDE is set up for Scala 2.11, set the Scala version >>>>> variable in the Flink root pom.xml also to 2.11 >>>>> >>>>> Greetings, >>>>> Stephan >>>>> >>>>> >>>>> >>>>> >>>>> On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S < >>>> ramkrishna.s.vasude...@intel.com> wrote: >>>>> >>>>>> I have been trying to install, learn and understand Flink. I am >>>>>> using >>>>>> Scala- EclipseIDE as my IDE. >>>>>> >>>>>> I have downloaded the flink source coded, compiled and created >>>>>> the >>>> project. >>>>>> >>>>>> My work laptop is Windows based and I don't have eclipse based >>>>>> workstation but I do have linux boxes for running and testing things. >>>>>> >>>>>> Some of the examples given in Flink source code do run directly >>>>>> from Eclipse but when I try to run the Wordcount example from >>>>>> Eclipse I get this error >>>>>> >>>>>> Exception in thread "main" java.lang.NoSuchMethodError: >>>>>> >>>> scala.collection.immutable.HashSet$.empty()Lscala/collection/immuta >>>> ble >>>> /HashSet; >>>>>> at akka.actor.ActorCell$.<init>(ActorCell.scala:336) >>>>>> at akka.actor.ActorCell$.<clinit>(ActorCell.scala) >>>>>> at akka.actor.RootActorPath.$div(ActorPath.scala:159) >>>>>> at >>>> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464) >>>>>> at >>>> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452) >>>>>> at >>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>>>> Method) >>>>>> at >>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown >>>>>> Source) >>>>>> at >>>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown >>>>>> Source) >>>>>> at java.lang.reflect.Constructor.newInstance(Unknown Source) >>>>>> at >>>>>> >>>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.app >>>> ly( >>>> DynamicAccess.scala:78) >>>>>> at scala.util.Try$.apply(Try.scala:191) >>>>>> at >>>>>> >>>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess. >>>> sca >>>> la:73) >>>>>> at >>>>>> >>>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.app >>>> ly( >>>> DynamicAccess.scala:84) >>>>>> at >>>>>> >>>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.app >>>> ly( >>>> DynamicAccess.scala:84) >>>>>> at scala.util.Success.flatMap(Try.scala:230) >>>>>> at >>>>>> >>>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess. >>>> sca >>>> la:84) >>>>>> at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585) >>>>>> at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578) >>>>>> at akka.actor.ActorSystem$.apply(ActorSystem.scala:142) >>>>>> at akka.actor.ActorSystem$.apply(ActorSystem.scala:119) >>>>>> at akka.actor.ActorSystem$.create(ActorSystem.scala:67) >>>>>> at >>>>>> >>>> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtil >>>> s.s >>>> cala:84) >>>>>> at >>>>>> >>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManag >>>> erA >>>> ctorSystem(FlinkMiniCluster.scala:196) >>>>>> at >>>>>> >>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSy >>>> ste >>>> m$lzycompute$1(FlinkMiniCluster.scala:225) >>>>>> at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org >>>>>> >>>> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSyst >>>> em$ >>>> 1(FlinkMiniCluster.scala:225) >>>>>> at >>>>>> >>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.ap >>>> ply >>>> (FlinkMiniCluster.scala:230) >>>>>> at >>>>>> >>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.ap >>>> ply >>>> (FlinkMiniCluster.scala:228) >>>>>> at >>>>>> >>>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike. >>>> scala:245) >>>>>> at >>>>>> >>>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike. >>>> scala:245) >>>>>> at scala.collection.immutable.Range.foreach(Range.scala:166) >>>>>> at >>>>>> scala.collection.TraversableLike$class.map(TraversableLike.scala:245) >>>>>> at scala.collection.AbstractTraversable.map(Traversable.scala:104) >>>>>> at >>>>>> >>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMi >>>> niC >>>> luster.scala:228) >>>>>> at >>>>>> >>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMi >>>> niC >>>> luster.scala:219) >>>>>> at >>>>>> >>>> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.e >>>> xec >>>> ute(LocalStreamEnvironment.java:104) >>>>>> at >>>>>> org.apache.flink.streaming.examples.wordcount.WordCount.main(Word >>>>>> Co >>>>>> un >>>>>> t >>>>>> .java:80) >>>>>> >>>>>> I know this is a naïve question but I would like to get some help >>>>>> in order to over come this issue. I tried various options like >>>>>> setting >>>>>> scala-2.10 as the compiler for the project (then it shows >>>>>> completely different error) and many of the projects don't even >>>>>> compile. But with >>>>>> 2.11 version I get the above stack trace. Any help here is welcome. >>>>>> >>>>>> Regards >>>>>> Ram >>>>>> >>>> >>>> Regards, >>>> Chiwan Park >>>> >> >> Regards, >> Chiwan Park >> Regards, Chiwan Park