You can set the max cores for the first submitted job such that it does not
take all the resources from the master. See
http://spark.apache.org/docs/latest/submitting-applications.html

# Run on a Spark standalone cluster in client deploy mode
./bin/spark-submit \
  --class org.apache.spark.examples.SparkPi \
  --master spark://207.184.161.138:7077 \
  --executor-memory 20G \
  *--total-executor-cores 100 \*
  /path/to/examples.jar \
  1000


On Mon, Oct 19, 2015 at 4:26 PM, Augustus Hong <augus...@branchmetrics.io>
wrote:

> Hi All,
>
> Would it be possible to run multiple spark streaming jobs on a single
> master at the same time?
>
> I currently have one master node and several worker nodes in the
> standalone mode, and I used spark-submit to submit multiple spark streaming
> jobs.
>
> From what I observed, it seems like only the first submitted job would get
> resources and run.  Jobs submitted afterwards will have the status
> "Waiting", and will only run after the first one is finished or killed.
>
> I tried limiting each executor to only 1 core(each worker machine has 8
> cores), but the same things happens that only one job will be run, even
> though there are a lot of idle cores.
>
> Best,
> Augustus
>
>
>
> --
> [image: Branch Metrics mobile deep linking] <http://branch.io/>* Augustus
> Hong*
>  Data Analytics | Branch Metrics
>  m 650-391-3369 | e augus...@branch.io
>

Reply via email to