I get an error with a message that state what is max number of cores allowed.
> On Apr 20, 2016, at 11:21 AM, Shushant Arora <shushantaror...@gmail.com> > wrote: > > I am running a spark application on yarn cluster. > > say I have available vcors in cluster as 100.And I start spark application > with --num-executors 200 --num-cores 2 (so I need total 200*2=400 vcores) but > in my cluster only 100 are available. > > What will happen ? Will the job abort or it will be submitted successfully > and 100 vcores will be aallocated to 50 executors and rest executors will be > started as soon as vcores are available ? > > Please note dynamic allocation is not enabled in cluster. I have old version > 1.2. > > Thanks > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org