I'm having 30G per machine This is the first (and only) job I'm trying to submit. So it's weird that for --total-executor-cores=20 it works, and for --total-executor-cores=4 it doesn't
On Tue, Jun 23, 2015 at 10:46 PM, Igor Berman <igor.ber...@gmail.com> wrote: > probably there are already running jobs there > in addition, memory is also a resource, so if you are running 1 > application that took all your memory and then you are trying to run > another application that asks > for the memory the cluster doesn't have then the second app wont be running > > so why are u specifying 22g as executor memory? how much memory you have > for each machine? > > On 23 June 2015 at 09:33, nizang <ni...@windward.eu> wrote: > >> to give a bit more data on what I'm trying to get - >> >> I have many tasks I want to run in parallel, so I want each task to catch >> small part of the cluster (-> only limited part of my 20 cores in the >> cluster) >> >> I have important tasks that I want them to get 10 cores, and I have small >> tasks that I want to run with only 1 or 2 cores) >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/Spark-standalone-cluster-resource-management-tp23444p23445.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >