This is meant to kill the whole driver hosted inside the Master (new feature as of 0.9.0). I assume you are trying to kill a job/task/stage inside the Spark rather than the whole application. Regards Mayur
Mayur Rustagi Ph: +1 (760) 203 3257 http://www.sigmoidanalytics.com @mayur_rustagi <https://twitter.com/mayur_rustagi> On Sun, Mar 16, 2014 at 4:36 PM, Debasish Das <debasish.da...@gmail.com>wrote: > From > http://spark.incubator.apache.org/docs/latest/spark-standalone.html#launching-applications-inside-the-cluster > > ./bin/spark-class org.apache.spark.deploy.Client kill <driverId> > > > does not work / has bugs ? > > > On Sun, Mar 16, 2014 at 1:17 PM, Mayur Rustagi <mayur.rust...@gmail.com>wrote: > >> Thr is a no good way to kill jobs in Spark yet. The closest is >> cancelAllJobs & cancelJobGroup in spark context. I have had bugs using >> both. I am trying to test them out, typically you would start a different >> thread & call these functions on it when you wish to cancel a job. >> Regards >> Mayur >> >> Mayur Rustagi >> Ph: +1 (760) 203 3257 >> http://www.sigmoidanalytics.com >> @mayur_rustagi <https://twitter.com/mayur_rustagi> >> >> >> >> On Sun, Mar 16, 2014 at 2:59 PM, Debasish Das >> <debasish.da...@gmail.com>wrote: >> >>> Are these the right options: >>> >>> 1. If there is a spark script, just do a ctrl-c from spark-shell and the >>> job will be killed property. >>> >>> 2. For spark application also ctrl c will kill the job property on the >>> cluster: >>> >>> Somehow the ctrl-c option did not work for us... >>> >>> Similar option works fine for scalding for example but we see lot of >>> dead nodes if too many jobs are killed abruptly. >>> >>> 3. Use the Client script... >>> >>> /bin/spark-class org.apache.spark.deploy.Client kill spark:// >>> myspark.com:7077 app-20140316142129-0000 >>> Runner java >>> Classpath >>> :/home/debasish/sag_spark/conf:/home/debasish/sag_spark/assembly/target/scala-2.10/spark-assembly-1.0.0-incubating-SNAPSHOT-hadoop2.0.0-mr1-cdh4.5.0.jar >>> Java opts -Djava.library.path= -Xms512m -Xmx512m >>> Options -Dspark.cores.max=16 >>> Sending kill command to spark://myspark.com:7077 >>> Driver app-20140316142129-0000 has already finished or does not exist >>> >>> This option also did not kill the job. I can still see the job running >>> on spark webui... >>> >>> Thanks. >>> Deb >>> >> >> >