Hi Serega, If I understand your problem correctly you would like to kill one executor only and the rest of the app has to be untouched. If that's true yarn -kill is not what you want because it stops the whole application.
I've done similar thing when tested/testing Spark's HA features. - jps -vlm | grep "org.apache.spark.executor.CoarseGrainedExecutorBackend.*applicationid" - kill -9 pidofoneexecutor Be aware if it's a multi-node cluster check whether at least one process runs on a specific node(it's not required). Happy killing... BR, G On Sun, Feb 10, 2019 at 4:19 PM Jörn Franke <jornfra...@gmail.com> wrote: > yarn application -kill applicationid ? > > > Am 10.02.2019 um 13:30 schrieb Serega Sheypak <serega.shey...@gmail.com > >: > > > > Hi there! > > I have weird issue that appears only when tasks fail at specific stage. > I would like to imitate failure on my own. > > The plan is to run problematic app and then kill entire executor or some > tasks when execution reaches certain stage. > > > > Is it do-able? > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >