Actually I faced it yesterday...

I had to put it in spark-env.sh and take it out from spark-defaults.conf on
1.0.1...Note that this settings should be visible on all workers..

After that I validated that SPARK_LOCAL_DIRS was indeed getting used for
shuffling...


On Thu, Aug 14, 2014 at 10:27 AM, Brad Miller <bmill...@eecs.berkeley.edu>
wrote:

> Hi All,
>
> I'm having some trouble setting the disk spill directory for spark.  The
> following approaches set "spark.local.dir" (according to the "Environment"
> tab of the web UI) but produce the indicated warnings:
>
> *In spark-env.sh:*
> export SPARK_JAVA_OPTS=-Dspark.local.dir=/spark/spill
> *Associated warning:*
> 14/08/14 10:10:39 WARN SparkConf: In Spark 1.0 and later spark.local.dir
> will be overridden by the value set by the cluster manager (via
> SPARK_LOCAL_DIRS in mesos/standalone and LOCAL_DIRS in YARN).
> 14/08/14 10:10:39 WARN SparkConf:
> SPARK_JAVA_OPTS was detected (set to '-Dspark.local.dir=/spark/spill').
> This is deprecated in Spark 1.0+.
> Please instead use...
>
> *In spark-defaults.conf:*
> spark.local.dir                  /spark/spill
> *Associated warning:*
> 14/08/14 10:09:04 WARN SparkConf: In Spark 1.0 and later spark.local.dir
> will be overridden by the value set by the cluster manager (via
> SPARK_LOCAL_DIRS in mesos/standalone and LOCAL_DIRS in YARN).
>
> The following does not produce any warnings, but also produces no sign of
> actually setting "spark.local.dir":
>
> *In spark-env.sh:*
> export SPARK_LOCAL_DIRS=/spark/spill
>
> Does anybody know whether SPARK_LOCAL_DIRS actually works as advertised,
> or if I am perhaps using it incorrectly?
>
> best,
> -Brad
>

Reply via email to