Hi Egor,

we recommand you to set SPARK_HOME because many spark configurations can be
handled better in this way.
SPARK_SUBMIT_OPTIONS in conf/zeppelin-env.sh is valid only when you set
SPARK_HOME.
Since you get JVM error, if you have any JVM related configuration in your
conf/zeppelin-env.sh can you share it?

If you want to change driver memory without setting SPARK_HOME, you can try
ZEPPELIN_INTP_MEM in conf/zeppelin-env.sh file as below:
export ZEPPELIN_INTP_MEM="-Xmx10g" in conf/zeppelin-env.sh file.
FYI, Zeppelin's default ZEPPELIN_INTP_MEM is "-Xmx1024m
-XX:MaxPermSize=512m".

Though Spark's UI won't display "Storage Memory" as 10g, because it will
shows fraction of java heap to use for memory cache.

I tested both with/without SPARK_HOME setting in yarn-client mode, and
Spark UI shows different value.

   - With SPARK_HOME setting if I export SPARK_SUBMIT_OPTIONS="--driver-momory
   10g", driver storage memory is 5.2G.
   - Without SPARK_HOME setting if I export ZEPPELIN_INT_MEM="-Xmx10g",
   driver storage memory is 4.8G.

The difference here is that first one starts spark interpreter process by
running spark_submit and second one runs with java.

Hope this helps!




On Thu, Nov 12, 2015 at 8:41 AM, Egor Pahomov <pahomov.e...@gmail.com>
wrote:

> I'm trying to configure driver memory size. So far I tied next parameters:
>
>    -
>
>    export JAVA_INTP_OPTS="-Xmx10g"
>
>    -
>
>    export SPARK_SUBMIT_OPTIONS="--driver-memory 10g --executor-memory 10g"
>
>    -
>
>    " -Xmx30000m \
>
>    -
>
>    Change SparkInterpreter.java: conf.set("spark.executor.memory", "10g");
>
>    conf.set("spark.executor.cores", "2");
>    conf.set("spark.driver.memory", "10g");
>    conf.set("spark.shuffle.io.numConnectionsPerPeer", "5");
>    conf.set("spark.sql.autoBroadcastJoinThreshold", "200483647");
>    conf.set("spark.network.timeout", "400s");
>    conf.set("spark.driver.maxResultSize", "3g");
>    conf.set("spark.sql.hive.convertMetastoreParquet", "false");
>    conf.set("spark.kryoserializer.buffer.max", "200m");
>    conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
>    conf.set("spark.dynamicAllocation.enabled", "true");
>    conf.set("spark.shuffle.service.enabled", "true");
>    conf.set("spark.dynamicAllocation.minExecutors", "1");
>    conf.set("spark.dynamicAllocation.maxExecutors", "30");
>    conf.set("spark.dynamicAllocation.executorIdleTimeout", "60s");
>            //.set("spark.sql.hive.metastore.version", "1.1.0")
>    conf.set("spark.dynamicAllocation.cachedExecutorIdleTimeout", "100s");
>
>    -
>
>    I tried setting SPARK_HOME, but it didn't even started, failed with 
> "Incompatible minimum and maximum heap sizes specified"
>
>
> No matter, what I do I get in logs: "INFO [2015-11-11 14:55:24,453]
> ({sparkDriver-akka.actor.default-dispatcher-14} Logging.scala[logInfo]:59)
> - Registering block manager 192.168.12.121:45057 with 530.0 MB RAM,
> BlockManagerId(driver, 192.168.12.121, 45057)" and on my spark UI:
>
> ​Has anyone faced this problem or knows what to do?
>
> --
>
>
> *Sincerely yoursEgor Pakhomov*
>

Reply via email to