Try this: Import org.apache.spark.SparkContext._ Thanks.
Zhan Zhang On Sep 4, 2014, at 4:36 PM, Veeranagouda Mukkanagoudar <veera...@gmail.com> wrote: > I am planning to use RDD join operation, to test out i was trying to compile > some test code, but am getting following compilation error > > value join is not a member of org.apache.spark.rdd.RDD[(String, Int)] > [error] rddA.join(rddB).map { case (k, (a, b)) => (k, a+b) } > > Code: > > import org.apache.spark.{SparkConf, SparkContext} > import org.apache.spark.rdd.RDD > > def joinTest(rddA: RDD[(String, Int)], rddB: RDD[(String, Int)]) : > RDD[(String, Int)] = { > rddA.join(rddB).map { case (k, (a, b)) => (k, a+b) } > } > > Any help would be great . > > Veera > -- CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You.