Hi Oleg, I set the size of my executors on a standalone cluster when using the shell like this:
./bin/spark-shell --master $MASTER --total-executor-cores $CORES_ACROSS_CLUSTER --driver-java-options "-Dspark.executor.memory=$MEMORY_PER_EXECUTOR" It doesn't seem particularly clean, but it works. Andrew On Thu, Jun 5, 2014 at 2:15 PM, Oleg Proudnikov <oleg.proudni...@gmail.com> wrote: > Hi All, > > Please help me set Executor JVM memory size. I am using Spark shell and it > appears that the executors are started with a predefined JVM heap of 512m > as soon as Spark shell starts. How can I change this setting? I tried > setting SPARK_EXECUTOR_MEMORY before launching Spark shell: > > export SPARK_EXECUTOR_MEMORY=1g > > I also tried several other approaches: > > 1) setting SPARK_WORKER_MEMORY in conf/spark-env.sh on the worker > 2) passing it as -m argument and running bin/start-slave.sh 1 -m 1g on > the worker > > Thank you, > Oleg > >