Try -c <numCores> instead, works for me, e.g. bin/spark-shell -c 88
On Tue, Jun 3, 2014 at 8:15 AM, Marek Wiewiorka <marek.wiewio...@gmail.com> wrote: > Hi All, > there is information in 1.0.0 Spark's documentation that > there is an option "--cores" that one can use to set the number of cores > that spark-shell uses on the cluster: > > You can also pass an option --cores <numCores> to control the number of > cores that spark-shell uses on the cluster. > > This option does not seem to work for me. > If run the following command: > ./spark-shell --cores 12 > I'm keep on getting an error: > bad option: '--cores' > > Is there any other way of controlling the total number of cores used by > sparkshell? > > Thanks, > Marek > >