Wow, it really was that easy! The implicit joining works a treat.
Many thanks,
Jon
On 13 October 2014 22:58, Stephen Boesch wrote:
> is the following what you are looking for?
>
>
> scala > sc.parallelize(myMap.map{ case (k,v) => (k,v) }.toSeq)
> res2: org.apache.spark.rdd.RDD[(String, Int)] =
Map.toSeq already does that even. You can skip the map. You can put
together Maps with ++ too. You should have an RDD of pairs then, but
to get the special RDD functions you're looking for remember to import
SparkContext._
On Mon, Oct 13, 2014 at 10:58 PM, Stephen Boesch wrote:
> is the following
is the following what you are looking for?
scala > sc.parallelize(myMap.map{ case (k,v) => (k,v) }.toSeq)
res2: org.apache.spark.rdd.RDD[(String, Int)] = ParallelCollectionRDD[0] at
parallelize at :21
2014-10-13 14:02 GMT-07:00 jon.g.massey :
> Hi guys,
> Just starting out with Spark and fol