Whats in your build.sbt? You could be messing with the scala version it
seems.

Thanks
Best Regards

On Fri, Jul 24, 2015 at 2:15 AM, Dan Dong <dongda...@gmail.com> wrote:

> Hi,
>   When I ran with spark-submit the following simple Spark program of:
> import org.apache.spark.SparkContext._
> import org.apache.spark.SparkConf
> import org.apache.spark.rdd.RDD
> import org.apache.spark.SparkContext
> import org.apache.spark._
> import SparkContext._
>
> object TEST2{
> def main(args:Array[String])
> {
>      val conf = new SparkConf().setAppName("TEST")
>      val sc=new SparkContext(conf)
>
>      val list=List(("aa",1),("bb",2),("cc",3))
>      val maps=list.toMap
>   }
>
> }
>
> I got java.lang.NoSuchMethodError for the line of "val maps=list.toMap".
> But in a spark-shell or simply scala, it has no problem:
>
> scala> val list=List(("aa",1),("bb",2),("cc",3))
> list: List[(String, Int)] = List((aa,1), (bb,2), (cc,3))
>
> scala> val maps=list.toMap
> maps: scala.collection.immutable.Map[String,Int] = Map(aa -> 1, bb -> 2,
> cc -> 3)
>
> So to use "toMap" method, what am I missing in spark-submit? I use "sbt
> package" to compile the program and without problem. Thanks!
>
> Cheers,
> Dan
>
>

Reply via email to