specifically, you're picking up the following implicit: import org.apache.spark.SparkContext.rddToPairRDDFunctions
(in case you're a wildcard-phobe like me) On Thu, Sep 4, 2014 at 5:15 PM, Veeranagouda Mukkanagoudar < veera...@gmail.com> wrote: > Thanks a lot, that fixed the issue :) > > > On Thu, Sep 4, 2014 at 4:51 PM, Zhan Zhang <zzh...@hortonworks.com> wrote: > >> Try this: >> Import org.apache.spark.SparkContext._ >> >> Thanks. >> >> Zhan Zhang >> >> >> On Sep 4, 2014, at 4:36 PM, Veeranagouda Mukkanagoudar < >> veera...@gmail.com> wrote: >> >> I am planning to use RDD join operation, to test out i was trying to >> compile some test code, but am getting following compilation error >> >> *value join is not a member of org.apache.spark.rdd.RDD[(String, Int)]* >> *[error] rddA.join(rddB).map { case (k, (a, b)) => (k, a+b) }* >> >> Code: >> >> import org.apache.spark.{SparkConf, SparkContext} >> import org.apache.spark.rdd.RDD >> >> def joinTest(rddA: RDD[(String, Int)], rddB: RDD[(String, Int)]) : >> RDD[(String, Int)] = { >> rddA.join(rddB).map { case (k, (a, b)) => (k, a+b) } >> } >> >> Any help would be great . >> >> Veera >> >> >> >> CONFIDENTIALITY NOTICE >> NOTICE: This message is intended for the use of the individual or entity >> to which it is addressed and may contain information that is confidential, >> privileged and exempt from disclosure under applicable law. If the reader >> of this message is not the intended recipient, you are hereby notified that >> any printing, copying, dissemination, distribution, disclosure or >> forwarding of this communication is strictly prohibited. If you have >> received this communication in error, please contact the sender immediately >> and delete it from your system. Thank You. > > >