I have a spark job running on Mesos in multi-master and supervise mode. If I kill it, it is resilient as expected and respawns on another node. However, I cannot kill it when I need to. I have tried 2 methods:
1) ./bin/spark-class org.apache.spark.deploy.Client kill <masterIp:port???> <driverID> 2) ./bin/spark-submit --master mesos://<masterIp:7077> --kill <driverID> Method 2, accepts the kill request but is respawned on another node. Ultimately, I can't get either method to kill the job. I suspect I have the wrong port for the master URL during the kill request for method 1? I've tried every combination of IP and port I can think of, is there one I am missing? Ports I've tried: 5050 = mesos UI 8080 = marathon 7077 = spark dispatcher 8081 = spark drivers UI 4040 = spark job UI thanks, Duc