Doesn't the fair scheduler solve this?
Ognen
On 2/25/14, 12:08 PM, abhinav chowdary wrote:
Sorry for not being clear earlier
how do you want to pass the operations to the spark context?
this is partly what i am looking for . How to access the active spark
context and possible ways to pass operations
Thanks
On Tue, Feb 25, 2014 at 10:02 AM, Mayur Rustagi
<mayur.rust...@gmail.com <mailto:mayur.rust...@gmail.com>> wrote:
how do you want to pass the operations to the spark context?
Mayur Rustagi
Ph: +919632149971 <tel:%2B919632149971>
h
<https://twitter.com/mayur_rustagi>ttp://www.sigmoidanalytics.com
<http://www.sigmoidanalytics.com>
https://twitter.com/mayur_rustagi
On Tue, Feb 25, 2014 at 9:59 AM, abhinav chowdary
<abhinav.chowd...@gmail.com <mailto:abhinav.chowd...@gmail.com>>
wrote:
Hi,
I am looking for ways to share the sparkContext,
meaning i need to be able to perform multiple operations on
the same spark context.
Below is code of a simple app i am testing
def main(args: Array[String]) {
println("Welcome to example application!")
val sc = new SparkContext("spark://10.128.228.142:7077
<http://10.128.228.142:7077>", "Simple App")
println("Spark context created!")
println("Creating RDD!")
Now once this context is created i want to access this to
submit multiple jobs/operations
Any help is much appreciated
Thanks
--
Warm Regards
Abhinav Chowdary