Could you post the log of Master? On Mon, Mar 21, 2016 at 9:25 AM, Hao Ren <inv...@gmail.com> wrote:
> Update: > > I am using --supervise flag for fault tolerance. > > > > On Mon, Mar 21, 2016 at 4:16 PM, Hao Ren <inv...@gmail.com> wrote: > >> Using spark 1.6.1 >> Spark Streaming Jobs are submitted via spark-submit (cluster mode) >> >> I tried to kill drivers via webUI, it does not work. These drivers are >> still running. >> I also tried: >> 1. spark-submit --master <master-url> --kill <driver-id> >> 2. ./bin/spark-class org.apache.spark.deploy.Client kill <master url> >> <driver ID> >> >> Neither works. The workaround is to ssh to the driver node, then kill -9 >> ... >> jsp shows the same classname DriverWrapper, so need to pick carefully... >> >> Any idea why this happens ? >> BTW, my streaming job's batch duration is one hour. So do we need to wait >> for job processing to kill kill driver ? >> >> -- >> Hao Ren >> >> Data Engineer @ leboncoin >> >> Paris, France >> > > > > -- > Hao Ren > > Data Engineer @ leboncoin > > Paris, France >