Cool. Thanks! Srikanth
On Wed, Jul 22, 2015 at 3:12 PM, Andrew Or <and...@databricks.com> wrote: > Hi Srikanth, > > I was able to reproduce the issue by setting `spark.cores.max` to a number > greater than the number of cores on a worker. I've filed SPARK-9260 which I > believe is already being fixed in > https://github.com/apache/spark/pull/7274. > > Thanks for reporting the issue! > -Andrew > > 2015-07-22 11:49 GMT-07:00 Andrew Or <and...@databricks.com>: > >> Hi Srikanth, >> >> It does look like a bug. Did you set `spark.executor.cores` in your >> application by any chance? >> >> -Andrew >> >> 2015-07-22 8:05 GMT-07:00 Srikanth <srikanth...@gmail.com>: >> >>> Hello, >>> >>> I've set spark.deploy.spreadOut=false in spark-env.sh. >>> >>>> export SPARK_MASTER_OPTS="-Dspark.deploy.defaultCores=4 >>>> -Dspark.deploy.spreadOut=false" >>> >>> >>> There are 3 workers each with 4 cores. Spark-shell was started with noof >>> cores = 6. >>> Spark UI show that one executor was used with 6 cores. >>> >>> Is this a bug? This is with Spark 1.4. >>> >>> [image: Inline image 1] >>> >>> Srikanth >>> >> >> >