I think you mean union(). Yes, you could also simply make an RDD for each
file, and use SparkContext.union() to put them together.

On Wed, Jan 7, 2015 at 9:51 AM, Raghavendra Pandey <
raghavendra.pan...@gmail.com> wrote:

> You can also use join function of rdd. This is actually kind of append
> funtion that add up all the rdds and create one uber rdd.
>
> On Wed, Jan 7, 2015, 14:30 rkgurram <rkgur...@gmail.com> wrote:
>
>> Thank you for the response, sure will try that out.
>>
>> Currently I changed my code such that the first map "files.map" to
>> "files.flatMap", which I guess will do similar what you are saying, it
>> gives
>> me a List[] of elements (in this case LabeledPoints, I could also do RDDs)
>> which I then turned into a mega RDD. The current problem seems to be
>> gone, I
>> no longer get the NPE but further down I am getting a indexOutOfBounds, so
>> trying to figure out if the original problem is manifesting itself as a
>> new
>> one.
>>
>>
>> Regards
>> -Ravi
>>
>>
>>
>>
>> --
>> View this message in context: http://apache-spark-user-list.
>> 1001560.n3.nabble.com/How-to-merge-a-RDD-of-RDDs-into-one-
>> uber-RDD-tp20986p21012.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>

Reply via email to