Thanks. Got it working. On 28 Mar, 2014, at 2:02 pm, Aaron Davidson <ilike...@gmail.com> wrote:
> Assuming you're using a new enough version of Spark, you should use > spark.executor.memory to set the memory for your executors, without changing > the driver memory. See the docs for your version of Spark. > > > On Thu, Mar 27, 2014 at 10:48 PM, Tsai Li Ming <mailingl...@ltsai.com> wrote: > Hi, > > My worker nodes have more memory than the host that I’m submitting my driver > program, but it seems that SPARK_MEM is also setting the Xmx of the spark > shell? > > $ SPARK_MEM=100g MASTER=spark://XXX:7077 bin/spark-shell > > Java HotSpot(TM) 64-Bit Server VM warning: INFO: > os::commit_memory(0x00007f736e130000, 205634994176, 0) failed; error='Cannot > allocate memory' (errno=12) > # > # There is insufficient memory for the Java Runtime Environment to continue. > # Native memory allocation (malloc) failed to allocate 205634994176 bytes for > committing reserved memory. > > I want to allocate at least 100GB of memory per executor. The allocated > memory on the executor seems to depend on the -Xmx heap size of the driver? > > Thanks! > > > >