I am using the script in sbin to set it up (spark/sbin/start-all.sh). It works 
fine. The problem is how to configure more than one worker per node (the 
default is one worker only). The documentation for 1.6.1 suggested 
SPARK_WORKER_INSTANCES as the way to do it but the latest documentation has no 
option for it.
The warning does not appear when I start the cluster but rather when I submit 
an application.
I see no way to define  SPARK_WORKER_INSTANCES when starting the cluster but 
not define it when submitting an application.
Assaf.

From: Ofer Eliassaf [mailto:ofer.elias...@gmail.com]
Sent: Thursday, October 06, 2016 9:07 PM
To: Mendelson, Assaf
Cc: user@spark.apache.org
Subject: Re: spark standalone with multiple workers gives a warning

The slaves should connect to the master using the scripts in sbin...
You can read about it here:
http://spark.apache.org/docs/latest/spark-standalone.html

On Thu, Oct 6, 2016 at 6:46 PM, Mendelson, Assaf 
<assaf.mendel...@rsa.com<mailto:assaf.mendel...@rsa.com>> wrote:
Hi,
I have a spark standalone cluster. On it, I am using 3 workers per node.
So I added SPARK_WORKER_INSTANCES set to 3 in spark-env.sh
The problem is, that when I run spark-shell I get the following warning:
WARN SparkConf:
SPARK_WORKER_INSTANCES was detected (set to '3').
This is deprecated in Spark 1.0+.

Please instead use:
- ./spark-submit with --num-executors to specify the number of executors
- Or set SPARK_EXECUTOR_INSTANCES
- spark.executor.instances to configure the number of instances in the spark 
config.

So how would I start a cluster of 3? SPARK_WORKER_INSTANCES is the only way I 
see to start the standalone cluster and the only way I see to define it is in 
spark-env.sh. The spark submit option, SPARK_EXECUTOR_INSTANCES and 
spark.executor.instances are all related to submitting the job.

Any ideas?
Thanks
                Assaf



--
Regards,
Ofer Eliassaf

Reply via email to