Hi! I've been using Spark compiled from 1.0 branch at some point (~2 month
ago). The setup is a standalone cluster with 4 worker machines and 1 master
machine. I used to run spark shell like this:

  ./bin/spark-shell -c 30 -em 20g -dm 10g

Today I've finally updated to Spark 1.0 release. Now I can only run spark
shell like this:

  ./bin/spark-shell --master spark://10.2.1.5:7077 --total-executor-cores 30
--executor-memory 20g --driver-memory 10g

The documentation at
http://spark.apache.org/docs/latest/spark-standalone.html says:

"You can also pass an option --cores <numCores> to control the number of
cores that spark-shell uses on the cluster."
This doesn't work, you need to pass "--total-executor-cores <numCores>"
instead.

"Note that if you are running spark-shell from one of the spark cluster
machines, the bin/spark-shell script will automatically set MASTER from the
SPARK_MASTER_IP and SPARK_MASTER_PORT variables in conf/spark-env.sh."
This is not working for me too. I run the shell from the master machine, and
I do have SPARK_MASTER_IP set up in conf/spark-env.sh like this:
export SPARK_MASTER_IP='10.2.1.5'
But if I omit "--master spark://10.2.1.5:7077" then the console starts but I
can't see it in the UI at http://10.2.1.5:8080. But when I go to
http://10.2.1.5:4040 (the application UI) I see that the app is using only
master as a worker.

My question is: are those just the bugs in the documentation? That there is
no --cores option and that SPARK_MASTER_IP is not used anymore when I run
the Spark shell from the master?



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/issues-with-bin-spark-shell-for-standalone-mode-tp9107.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to