Hi Ram,

Because there are some Scala IDE (Eclipse) plugins needed, I recommend to avoid 
`mvn eclipse:eclipse` command. Could you try just run `mvn clean install 
-DskipTests` and import the project to Scala IDE directly? In middle of 
importing process, Scala IDE suggests some plugins needed.

And which version of Scala IDE you are using?

> On Jan 12, 2016, at 7:58 PM, Vasudevan, Ramkrishna S 
> <ramkrishna.s.vasude...@intel.com> wrote:
> 
> Yes. I added it as Maven project only. I did mvn eclipse:eclipse to create 
> the project and also built the code using mvn clean install -DskipTests.
> 
> Regards
> Ram
> 
> -----Original Message-----
> From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf Of 
> Stephan Ewen
> Sent: Tuesday, January 12, 2016 4:10 PM
> To: dev@flink.apache.org
> Subject: Re: Naive question
> 
> Sorry to hear that it did not work out with Eclipse at all in the end, even 
> with all adjustments.
> 
> Just making sure: You imported Flink as a Maven project, not manually adding 
> the big Flink dependency JAR?
> 
> On Tue, Jan 12, 2016 at 5:15 AM, Vasudevan, Ramkrishna S < 
> ramkrishna.s.vasude...@intel.com> wrote:
> 
>> Thanks to all. I tried with Scala Eclipse IDE with all these 
>> 'change-scala-version.sh'. But in vain.
>> 
>> So I switched over to Intellij and thing work fine over there. I am 
>> new to Intellij so will try using it.
>> 
>> Once again thanks for helping me out.
>> 
>> Regards
>> Ram
>> 
>> -----Original Message-----
>> From: Chiwan Park [mailto:chiwanp...@apache.org]
>> Sent: Monday, January 11, 2016 4:37 PM
>> To: dev@flink.apache.org
>> Subject: Re: Naive question
>> 
>> Hi Ram,
>> 
>> If you want to build Flink with Scala 2.10, just checkout Flink 
>> repository from github or download source code from homepage, run `mvn 
>> clean install -DskipTests` and import projects to your IDE. If you 
>> want to build Flink with Scala 2.11, you have to run 
>> `tools/change-scala-version.sh 2.11` before build the project. You can 
>> revert Scala version change by running `tools/change-scala-version.sh 2.10`.
>> 
>> About IDE, Flink community recommends IntelliJ IDEA because Scala IDE 
>> have some problems in Java/Scala mixed project like Flink. But I 
>> tested importing Flink project with Scala IDE 4.3.0, Scala 2.11.7 and 
>> Flink 0.10.0 source code. Note that you should import the project as maven 
>> project.
>> 
>> By the way, the community welcomes any questions. Please feel free to 
>> post questions. :)
>> 
>>> On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S <
>> ramkrishna.s.vasude...@intel.com> wrote:
>>> 
>>> Thank you very much for the reply.
>>> I tried different ways and when I tried setting up the root pom.xml 
>>> to
>>> 2.11
>>> 
>>>              <scala.version>2.11.6</scala.version>
>>>              <scala.binary.version>2.11</scala.binary.version>
>>> 
>>> I got the following error
>>> [INFO]
>>> --------------------------------------------------------------------
>>> --
>>> -- [ERROR] Failed to execute goal on project flink-scala: Could not 
>>> resolve depende ncies for project
>>> org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find 
>>> artifact
>>> org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central 
>>> (http://repo.mave
>>> n.apache.org/maven2) -> [Help 1]
>>> 
>>> If I leave the scala.binary.verson to be at 2.10 and the scala 
>>> version to be at 2.11.6 then I get the following problem [INFO]
>>> C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling 
>>> [INFO] Compiling 366 source files to 
>>> C:\flink\flink\flink-runtime\target\test-cl
>>> asses at 1452508064750
>>> [ERROR]
>>> C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime
>>> \j
>>> ob
>>> manager\JobManagerITCase.scala:700: error: can't expand macros 
>>> compiled by previ ous versions of Scala
>>> [ERROR]               assert(cachedGraph2.isArchived)
>>> [ERROR]                                   ^
>>> 
>>> So am not pretty sure how to proceed with this. If I try to change 
>>> the
>> version of scala to 2.10 in the IDE then I get lot of compilation issues.
>> IS there any way to over come this?
>>> 
>>> Once again thanks a lot and apologies for the naïve question.
>>> 
>>> Regards
>>> Ram
>>> -----Original Message-----
>>> From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf 
>>> Of Stephan Ewen
>>> Sent: Friday, January 8, 2016 5:01 PM
>>> To: dev@flink.apache.org
>>> Subject: Re: Naive question
>>> 
>>> Hi!
>>> 
>>> This looks like a mismatch between the Scala dependency in Flink and 
>>> Scala in your Eclipse. Make sure you use the same for both. By 
>>> default, Flink reference Scala 2.10
>>> 
>>> If your IDE is set up for Scala 2.11, set the Scala version variable 
>>> in the Flink root pom.xml also to 2.11
>>> 
>>> Greetings,
>>> Stephan
>>> 
>>> 
>>> 
>>> 
>>> On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S <
>> ramkrishna.s.vasude...@intel.com> wrote:
>>> 
>>>> I have been trying to install, learn and understand Flink. I am 
>>>> using
>>>> Scala- EclipseIDE as my IDE.
>>>> 
>>>> I have downloaded the flink source coded, compiled and created the
>> project.
>>>> 
>>>> My work laptop is Windows based and I don't have eclipse based 
>>>> workstation but I do have linux boxes for running and testing things.
>>>> 
>>>> Some of the examples given in Flink source code do run directly 
>>>> from Eclipse but when I try to run the Wordcount example from 
>>>> Eclipse I get this error
>>>> 
>>>> Exception in thread "main" java.lang.NoSuchMethodError:
>>>> 
>> scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable
>> /HashSet;
>>>>    at akka.actor.ActorCell$.<init>(ActorCell.scala:336)
>>>>    at akka.actor.ActorCell$.<clinit>(ActorCell.scala)
>>>>    at akka.actor.RootActorPath.$div(ActorPath.scala:159)
>>>>    at
>> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464)
>>>>    at
>> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452)
>>>>    at 
>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>> Method)
>>>>    at 
>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
>>>> Source)
>>>>    at
>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
>>>> Source)
>>>>    at java.lang.reflect.Constructor.newInstance(Unknown Source)
>>>>    at
>>>> 
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply(
>> DynamicAccess.scala:78)
>>>>    at scala.util.Try$.apply(Try.scala:191)
>>>>    at
>>>> 
>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca
>> la:73)
>>>>    at
>>>> 
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(
>> DynamicAccess.scala:84)
>>>>    at
>>>> 
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(
>> DynamicAccess.scala:84)
>>>>    at scala.util.Success.flatMap(Try.scala:230)
>>>>    at
>>>> 
>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca
>> la:84)
>>>>    at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585)
>>>>    at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578)
>>>>    at akka.actor.ActorSystem$.apply(ActorSystem.scala:142)
>>>>    at akka.actor.ActorSystem$.apply(ActorSystem.scala:119)
>>>>    at akka.actor.ActorSystem$.create(ActorSystem.scala:67)
>>>>    at
>>>> 
>> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.s
>> cala:84)
>>>>    at
>>>> 
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerA
>> ctorSystem(FlinkMiniCluster.scala:196)
>>>>    at
>>>> 
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSyste
>> m$lzycompute$1(FlinkMiniCluster.scala:225)
>>>>    at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org
>>>> 
>> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$
>> 1(FlinkMiniCluster.scala:225)
>>>>    at
>>>> 
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply
>> (FlinkMiniCluster.scala:230)
>>>>    at
>>>> 
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply
>> (FlinkMiniCluster.scala:228)
>>>>    at
>>>> 
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.
>> scala:245)
>>>>    at
>>>> 
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.
>> scala:245)
>>>>    at scala.collection.immutable.Range.foreach(Range.scala:166)
>>>>    at
>>>> scala.collection.TraversableLike$class.map(TraversableLike.scala:245)
>>>>    at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>>>>    at
>>>> 
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC
>> luster.scala:228)
>>>>    at
>>>> 
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC
>> luster.scala:219)
>>>>    at
>>>> 
>> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.exec
>> ute(LocalStreamEnvironment.java:104)
>>>>    at
>>>> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCo
>>>> un
>>>> t
>>>> .java:80)
>>>> 
>>>> I know this is a naïve question but I would like to get some help 
>>>> in order to over come this issue. I tried various options like 
>>>> setting
>>>> scala-2.10 as the compiler for the project (then it shows 
>>>> completely different error) and many of the projects don't even 
>>>> compile. But with
>>>> 2.11 version I get the above stack trace. Any help here is welcome.
>>>> 
>>>> Regards
>>>> Ram
>>>> 
>> 
>> Regards,
>> Chiwan Park
>> 

Regards,
Chiwan Park


Reply via email to