SIGTERM on YARN generally means the NM is killing your executor because it's running over its requested memory limits. Check your NM logs to make sure. And then take a look at the "memoryOverhead" setting for driver and executors (http://spark.apache.org/docs/latest/running-on-yarn.html).
On Tue, Jul 7, 2015 at 7:43 AM, Kostas Kougios < [email protected]> wrote: > I've recompiled spark deleting the -XX:OnOutOfMemoryError=kill declaration, > but still I am getting a SIGTERM! > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/is-it-possible-to-disable-XX-OnOutOfMemoryError-kill-p-for-the-executors-tp23680p23687.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: [email protected] > For additional commands, e-mail: [email protected] > > -- Marcelo
