Andrew, Thanks, yes, this is what I wanted: basically just to start multiple jobs concurrently in threads.
Alex On Mon, Dec 22, 2014 at 4:04 PM, Andrew Ash <and...@andrewash.com> wrote: > > Hi Alex, > > SparkContext.submitJob() is marked as experimental -- most client programs > shouldn't be using it. What are you looking to do? > > For multiplexing jobs, one thing you can do is have multiple threads in > your client JVM each submit jobs on your SparkContext job. This is > described here in the docs: > http://spark.apache.org/docs/latest/job-scheduling.html#scheduling-within-an-application > > Andrew > > On Mon, Dec 22, 2014 at 1:32 PM, Alessandro Baretta <alexbare...@gmail.com > > wrote: > >> Fellow Sparkers, >> >> I'm rather puzzled at the submitJob API. I can't quite figure out how it >> is >> supposed to be used. Is there any more documentation about it? >> >> Also, is there any simpler way to multiplex jobs on the cluster, such as >> starting multiple computations in as many threads in the driver and >> reaping >> all the results when they are available? >> >> Thanks, >> >> Alex >> > >