Hi,

According to the configuration guide
(http://spark.apache.org/docs/latest/configuration.html#environment-variables),

"Certain Spark settings can be configured through environment variables,
which are read from the conf/spark-env.sh script in the directory where
Spark is installed (or conf/spark-env.cmd on Windows). In Standalone and
Mesos modes, this file can give machine specific information such as
hostnames.* It is also sourced when running local Spark applications or
submission scripts.*"

I am not sure about the last sentence, because when I reset the
SPARK_LOCAL_DIR to some specific paths, I have to restart the cluster, like:
./sbin/stop-all.sh
./sbin/start-all.sh
, otherwise, newly modified env vars(like SPARK_LOCAL_DIR) are not used. The
same for submission scripts and setting SparkConf on Driver program.

It seems that the spark-env.sh is sourced only when the cluster is started,
that means env vars are saved in spark daemon process (which will be used
for future spark App/prog), not in a single spark application/program. We
can not use different SPARK_LOCAL_DIR for different spark applications based
on the same daemon process, since SPARK_LOCAL_DIR takes the one loaded on
booting time of the spark daemon.

Could someone confirm this ? Willing to give some more details, if
necessary.

Thank you.

Hao.





--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Environment-Variables-question-tp11149.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to