Dan, Even though you may be adding more nodes to the cluster, the Spark application has to be requesting additional executors in order to thus use the added resources. Or the Spark application can be using Dynamic Resource Allocation (http://spark.apache.org/docs/latest/job-scheduling.html#dynamic-resource-allocation) [which may use the resources based on application need and availability]. For example, in EMR release 4.x (http://docs.aws.amazon.com/ElasticMapReduce/latest/ReleaseGuide/emr-spark-configure.html#spark-dynamic-allocation) you can request Spark Dynamic Resource Allocation as the default configuration at cluster creation.
Best regards, Christopher From: Dinesh Ranganathan [mailto:dineshranganat...@gmail.com] Sent: Monday, November 16, 2015 4:57 AM To: Sabarish Sasidharan Cc: user Subject: Re: Spark Expand Cluster Hi Sab, I did not specify number of executors when I submitted the spark application. I was in the impression spark looks at the cluster and figures out the number of executors it can use based on the cluster size automatically, is this what you call dynamic allocation?. I am spark newbie, so apologies if I am missing the obvious. While the application was running I added more core nodes by resizing my EMR instance and I can see the new nodes on the resource manager but my running application did not pick up those machines I've just added. Let me know If i am missing a step here. Thanks, Dan On 16 November 2015 at 12:38, Sabarish Sasidharan <sabarish.sasidha...@manthan.com<mailto:sabarish.sasidha...@manthan.com>> wrote: Spark will use the number of executors you specify in spark-submit. Are you saying that Spark is not able to use more executors after you modify it in spark-submit? Are you using dynamic allocation? Regards Sab On Mon, Nov 16, 2015 at 5:54 PM, dineshranganathan <dineshranganat...@gmail.com<mailto:dineshranganat...@gmail.com>> wrote: I have my Spark application deployed on AWS EMR on yarn cluster mode. When I increase the capacity of my cluster by adding more Core instances on AWS, I don't see Spark picking up the new instances dynamically. Is there anything I can do to tell Spark to pick up the newly added boxes?? Dan -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Expand-Cluster-tp25393.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org<mailto:user-unsubscr...@spark.apache.org> For additional commands, e-mail: user-h...@spark.apache.org<mailto:user-h...@spark.apache.org> -- Architect - Big Data Ph: +91 99805 99458<tel:%2B91%2099805%2099458> Manthan Systems | Company of the year - Analytics (2014 Frost and Sullivan India ICT) +++ -- Dinesh Ranganathan