Sorry for not being clear earlier
how do you want to pass the operations to the spark context?
this is partly what i am looking for . How to access the active spark
context and possible ways to pass operations

Thanks



On Tue, Feb 25, 2014 at 10:02 AM, Mayur Rustagi <mayur.rust...@gmail.com>wrote:

> how do you want to pass the operations to the spark context?
>
>
> Mayur Rustagi
> Ph: +919632149971
> h <https://twitter.com/mayur_rustagi>ttp://www.sigmoidanalytics.com
> https://twitter.com/mayur_rustagi
>
>
>
> On Tue, Feb 25, 2014 at 9:59 AM, abhinav chowdary <
> abhinav.chowd...@gmail.com> wrote:
>
>> Hi,
>>        I am looking for ways to share the sparkContext, meaning i need to
>> be able to perform multiple operations on the same spark context.
>>
>> Below is code of a simple app i am testing
>>
>>  def main(args: Array[String]) {
>>     println("Welcome to example application!")
>>
>>     val sc = new SparkContext("spark://10.128.228.142:7077", "Simple
>> App")
>>
>>     println("Spark context created!")
>>
>>     println("Creating RDD!")
>>
>> Now once this context is created i want to access  this to submit
>> multiple jobs/operations
>>
>> Any help is much appreciated
>>
>> Thanks
>>
>>
>>
>>
>


-- 
Warm Regards
Abhinav Chowdary

Reply via email to