Hello, my Spark streaming v1.3.0 code uses sys.ShutdownHookThread { ssc.stop(stopSparkContext = true, stopGracefully = true) }
to use Ctrl+C in command line to stop it. It returned back to command line after it finished batch but it doesn't with v1.4.0-v.1.5.0. Was the behaviour or required code changed? The last messages are: [2015-09-08 13:02:43,300] INFO Waited for jobs to be processed and checkpoints to be written (org.apache.spark.streaming.scheduler.JobGenerator) [2015-09-08 13:02:43,300] INFO CheckpointWriter executor terminated ? true, waited for 0 ms. (org.apache.spark.streaming.CheckpointWriter) [2015-09-08 13:02:43,301] INFO Stopped JobGenerator (org.apache.spark.streaming.scheduler.JobGenerator) [2015-09-08 13:02:43,302] INFO Stopped JobScheduler (org.apache.spark.streaming.scheduler.JobScheduler) [2015-09-08 13:02:43,303] INFO stopped o.s.j.s.ServletContextHandler{/streaming,null} (org.spark-project.jetty.server.handler.ContextHandler) [2015-09-08 13:02:43,305] INFO stopped o.s.j.s.ServletContextHandler{/streaming/batch,null} (org.spark-project.jetty.server.handler.ContextHandler) [2015-09-08 13:02:43,307] INFO stopped o.s.j.s.ServletContextHandler{/static/streaming,null} (org.spark-project.jetty.server.handler.ContextHandler) Thank you for any explanation, Petr