Re: distributing Scala Map datatypes to RDD

2014-10-16 Thread Jon Massey
Wow, it really was that easy! The implicit joining works a treat. Many thanks, Jon On 13 October 2014 22:58, Stephen Boesch wrote: > is the following what you are looking for? > > > scala > sc.parallelize(myMap.map{ case (k,v) => (k,v) }.toSeq) > res2: org.apache.spark.rdd.RDD[(String, Int)] =

Re: distributing Scala Map datatypes to RDD

2014-10-13 Thread Sean Owen
Map.toSeq already does that even. You can skip the map. You can put together Maps with ++ too. You should have an RDD of pairs then, but to get the special RDD functions you're looking for remember to import SparkContext._ On Mon, Oct 13, 2014 at 10:58 PM, Stephen Boesch wrote: > is the following

Re: distributing Scala Map datatypes to RDD

2014-10-13 Thread Stephen Boesch
is the following what you are looking for? scala > sc.parallelize(myMap.map{ case (k,v) => (k,v) }.toSeq) res2: org.apache.spark.rdd.RDD[(String, Int)] = ParallelCollectionRDD[0] at parallelize at :21 2014-10-13 14:02 GMT-07:00 jon.g.massey : > Hi guys, > Just starting out with Spark and fol