Yes. I added it as Maven project only. I did mvn eclipse:eclipse to create the 
project and also built the code using mvn clean install -DskipTests.

Regards
Ram

-----Original Message-----
From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf Of Stephan 
Ewen
Sent: Tuesday, January 12, 2016 4:10 PM
To: dev@flink.apache.org
Subject: Re: Naive question

Sorry to hear that it did not work out with Eclipse at all in the end, even 
with all adjustments.

Just making sure: You imported Flink as a Maven project, not manually adding 
the big Flink dependency JAR?

On Tue, Jan 12, 2016 at 5:15 AM, Vasudevan, Ramkrishna S < 
ramkrishna.s.vasude...@intel.com> wrote:

> Thanks to all. I tried with Scala Eclipse IDE with all these 
> 'change-scala-version.sh'. But in vain.
>
> So I switched over to Intellij and thing work fine over there. I am 
> new to Intellij so will try using it.
>
> Once again thanks for helping me out.
>
> Regards
> Ram
>
> -----Original Message-----
> From: Chiwan Park [mailto:chiwanp...@apache.org]
> Sent: Monday, January 11, 2016 4:37 PM
> To: dev@flink.apache.org
> Subject: Re: Naive question
>
> Hi Ram,
>
> If you want to build Flink with Scala 2.10, just checkout Flink 
> repository from github or download source code from homepage, run `mvn 
> clean install -DskipTests` and import projects to your IDE. If you 
> want to build Flink with Scala 2.11, you have to run 
> `tools/change-scala-version.sh 2.11` before build the project. You can 
> revert Scala version change by running `tools/change-scala-version.sh 2.10`.
>
> About IDE, Flink community recommends IntelliJ IDEA because Scala IDE 
> have some problems in Java/Scala mixed project like Flink. But I 
> tested importing Flink project with Scala IDE 4.3.0, Scala 2.11.7 and 
> Flink 0.10.0 source code. Note that you should import the project as maven 
> project.
>
> By the way, the community welcomes any questions. Please feel free to 
> post questions. :)
>
> > On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S <
> ramkrishna.s.vasude...@intel.com> wrote:
> >
> > Thank you very much for the reply.
> > I tried different ways and when I tried setting up the root pom.xml 
> > to
> > 2.11
> >
> >               <scala.version>2.11.6</scala.version>
> >               <scala.binary.version>2.11</scala.binary.version>
> >
> > I got the following error
> > [INFO]
> > --------------------------------------------------------------------
> > --
> > -- [ERROR] Failed to execute goal on project flink-scala: Could not 
> > resolve depende ncies for project
> > org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find 
> > artifact
> > org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central 
> > (http://repo.mave
> > n.apache.org/maven2) -> [Help 1]
> >
> > If I leave the scala.binary.verson to be at 2.10 and the scala 
> > version to be at 2.11.6 then I get the following problem [INFO]
> > C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling 
> > [INFO] Compiling 366 source files to 
> > C:\flink\flink\flink-runtime\target\test-cl
> > asses at 1452508064750
> > [ERROR]
> > C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime
> > \j
> > ob
> > manager\JobManagerITCase.scala:700: error: can't expand macros 
> > compiled by previ ous versions of Scala
> > [ERROR]               assert(cachedGraph2.isArchived)
> > [ERROR]                                   ^
> >
> > So am not pretty sure how to proceed with this. If I try to change 
> > the
> version of scala to 2.10 in the IDE then I get lot of compilation issues.
> IS there any way to over come this?
> >
> > Once again thanks a lot and apologies for the naïve question.
> >
> > Regards
> > Ram
> > -----Original Message-----
> > From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf 
> > Of Stephan Ewen
> > Sent: Friday, January 8, 2016 5:01 PM
> > To: dev@flink.apache.org
> > Subject: Re: Naive question
> >
> > Hi!
> >
> > This looks like a mismatch between the Scala dependency in Flink and 
> > Scala in your Eclipse. Make sure you use the same for both. By 
> > default, Flink reference Scala 2.10
> >
> > If your IDE is set up for Scala 2.11, set the Scala version variable 
> > in the Flink root pom.xml also to 2.11
> >
> > Greetings,
> > Stephan
> >
> >
> >
> >
> > On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S <
> ramkrishna.s.vasude...@intel.com> wrote:
> >
> >> I have been trying to install, learn and understand Flink. I am 
> >> using
> >> Scala- EclipseIDE as my IDE.
> >>
> >> I have downloaded the flink source coded, compiled and created the
> project.
> >>
> >> My work laptop is Windows based and I don't have eclipse based 
> >> workstation but I do have linux boxes for running and testing things.
> >>
> >> Some of the examples given in Flink source code do run directly 
> >> from Eclipse but when I try to run the Wordcount example from 
> >> Eclipse I get this error
> >>
> >> Exception in thread "main" java.lang.NoSuchMethodError:
> >>
> scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable
> /HashSet;
> >>     at akka.actor.ActorCell$.<init>(ActorCell.scala:336)
> >>     at akka.actor.ActorCell$.<clinit>(ActorCell.scala)
> >>     at akka.actor.RootActorPath.$div(ActorPath.scala:159)
> >>     at
> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464)
> >>     at
> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452)
> >>     at 
> >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> >> Method)
> >>     at 
> >> sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
> >> Source)
> >>     at
> >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
> >> Source)
> >>     at java.lang.reflect.Constructor.newInstance(Unknown Source)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply(
> DynamicAccess.scala:78)
> >>     at scala.util.Try$.apply(Try.scala:191)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca
> la:73)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(
> DynamicAccess.scala:84)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(
> DynamicAccess.scala:84)
> >>     at scala.util.Success.flatMap(Try.scala:230)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.sca
> la:84)
> >>     at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585)
> >>     at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578)
> >>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:142)
> >>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:119)
> >>     at akka.actor.ActorSystem$.create(ActorSystem.scala:67)
> >>     at
> >>
> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.s
> cala:84)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerA
> ctorSystem(FlinkMiniCluster.scala:196)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSyste
> m$lzycompute$1(FlinkMiniCluster.scala:225)
> >>     at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org
> >>
> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$
> 1(FlinkMiniCluster.scala:225)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply
> (FlinkMiniCluster.scala:230)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply
> (FlinkMiniCluster.scala:228)
> >>     at
> >>
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.
> scala:245)
> >>     at
> >>
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.
> scala:245)
> >>     at scala.collection.immutable.Range.foreach(Range.scala:166)
> >>     at
> >> scala.collection.TraversableLike$class.map(TraversableLike.scala:245)
> >>     at scala.collection.AbstractTraversable.map(Traversable.scala:104)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC
> luster.scala:228)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniC
> luster.scala:219)
> >>     at
> >>
> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.exec
> ute(LocalStreamEnvironment.java:104)
> >>     at
> >> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCo
> >> un
> >> t
> >> .java:80)
> >>
> >> I know this is a naïve question but I would like to get some help 
> >> in order to over come this issue. I tried various options like 
> >> setting
> >> scala-2.10 as the compiler for the project (then it shows 
> >> completely different error) and many of the projects don't even 
> >> compile. But with
> >> 2.11 version I get the above stack trace. Any help here is welcome.
> >>
> >> Regards
> >> Ram
> >>
>
> Regards,
> Chiwan Park
>
>
>

Reply via email to