Re: Fault tolerant broadcast in updateStateByKey

2017-02-07 Thread Amit Sela
I'm updating the Broadcast between batches, but I've ended up doing it in a listener, thanks! On Wed, Feb 8, 2017 at 12:31 AM Tathagata Das wrote: > broadcasts are not saved in checkpoints. so you have to save it externally > yourself, and recover it before restarting the stream from checkpoints

Re: Fault tolerant broadcast in updateStateByKey

2017-02-07 Thread Tathagata Das
broadcasts are not saved in checkpoints. so you have to save it externally yourself, and recover it before restarting the stream from checkpoints. On Tue, Feb 7, 2017 at 3:55 PM, Amit Sela wrote: > I know this approach, only thing is, it relies on the transformation being > an RDD transfomration

Re: Fault tolerant broadcast in updateStateByKey

2017-02-07 Thread Amit Sela
I know this approach, only thing is, it relies on the transformation being an RDD transfomration as well and so could be applied via foreachRDD and using the rdd context to avoid a stale context after recovery/resume. My question is how to void stale context in a DStream-only transformation such as

Re: Fault tolerant broadcast in updateStateByKey

2017-02-07 Thread Shixiong(Ryan) Zhu
It's documented here: http://spark.apache.org/docs/latest/streaming-programming-guide.html#accumulators-broadcast-variables-and-checkpoints On Tue, Feb 7, 2017 at 8:12 AM, Amit Sela wrote: > Hi all, > > I was wondering if anyone ever used a broadcast variable within > an updateStateByKey op. ? U