Hi Natalie, in case we are using Python then listener is not available, we can use the SPARK UI REST API's .
For that to happen we have to start the SPARK session with following settings: spark.sql("SET spark.sql.streaming.metricsEnabled=true") For the UI from which we can use the JSON results from the below: https://spark.apache.org/docs/latest/monitoring.html#rest-api In case you are working on AWS EMR, the UI URL is a bit different, let me know in that case and I will email to you. Hi TD, What will be the best way to stop a long running SPARK streaming job gracefully? I will be really grateful if you could kindly help us out. Regards, Gourav Sengupta On Wed, Sep 4, 2019 at 2:28 AM Tathagata Das <t...@databricks.com> wrote: > > https://spark.apache.org/docs/latest/structured-streaming-programming-guide.html#reporting-metrics-programmatically-using-asynchronous-apis > > On Tue, Sep 3, 2019, 3:26 PM Natalie Ruiz > <natalie.r...@microsoft.com.invalid> wrote: > >> Hello all, >> >> >> >> I’m a beginner, new to Spark and wanted to know if there was an >> equivalent to Spark Streaming’s StreamingListenerBatchCompleted in >> Structured Streaming? I want to add a listener for when a batch is complete >> but the documentation and examples I find are for Spark Streaming and not >> Structured Streaming and registering it to a StreamingContext, which from >> my understanding is just for Spark Streaming >> >> >> >> Thanks! >> >