Hi Oleg,

I set the size of my executors on a standalone cluster when using the shell
like this:

./bin/spark-shell --master $MASTER --total-executor-cores
$CORES_ACROSS_CLUSTER --driver-java-options
"-Dspark.executor.memory=$MEMORY_PER_EXECUTOR"

It doesn't seem particularly clean, but it works.

Andrew


On Thu, Jun 5, 2014 at 2:15 PM, Oleg Proudnikov <oleg.proudni...@gmail.com>
wrote:

> Hi All,
>
> Please help me set Executor JVM memory size. I am using Spark shell and it
> appears that the executors are started with a predefined JVM heap of 512m
> as soon as Spark shell starts. How can I change this setting? I tried
> setting SPARK_EXECUTOR_MEMORY before launching Spark shell:
>
> export SPARK_EXECUTOR_MEMORY=1g
>
> I also tried several other approaches:
>
> 1) setting SPARK_WORKER_MEMORY in conf/spark-env.sh on the worker
> 2)  passing it as -m argument and running bin/start-slave.sh 1 -m 1g on
> the worker
>
> Thank you,
> Oleg
>
>

Reply via email to