probably there are already running jobs there
in addition, memory is also a resource, so if you are running 1 application
that took all your memory and then you are trying to run another
application that asks
for the memory the cluster doesn't have then the second app wont be running

so why are u specifying 22g as executor memory? how much memory you have
for each machine?

On 23 June 2015 at 09:33, nizang <ni...@windward.eu> wrote:

> to give a bit more data on what I'm trying to get -
>
> I have many tasks I want to run in parallel, so I want each task to catch
> small part of the cluster (-> only limited part of my 20 cores in the
> cluster)
>
> I have important tasks that I want them to get 10 cores, and I have small
> tasks that I want to run with only 1 or 2 cores)
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Spark-standalone-cluster-resource-management-tp23444p23445.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to