Could you, pls share your code, so that I could try it. -- Be well! Jean Morozov
On Sun, Mar 27, 2016 at 5:20 PM, 吴文超 <wuwenc...@miaozhen.com> wrote: > I am a newbie to spark, when I use IntelliJ idea to write some scala code, > i found it reports error when using spark's implicit conversion.e.g. whe > use the RDD as Pair RDD to get reduceByKey function. However, the project > can run normally in the cluster. > As somebody says it needs import org.apache.spark.SparkContext._ , > http://stackoverflow.com/questions/24084335/reducebykey-method-not-being-found-in-intellij > I did it ,but it still gets error.. > Has anybody encountered the problem and how do you solve it ? > BTY, I have tried both sbt and maven , and the idea version 14.0.3 and > spark version is 1.6.0 >