You also need to ensure you're using checkpointing and support recreating the 
context on driver failure as described in the docs here: 
http://spark.apache.org/docs/latest/streaming-programming-guide.html#failure-of-the-driver-node

From: Matt Narrell <matt.narr...@gmail.com<mailto:matt.narr...@gmail.com>>
Date: Thursday, August 14, 2014 at 10:34 AM
To: Tobias Pfeiffer <t...@preferred.jp<mailto:t...@preferred.jp>>
Cc: salemi <alireza.sal...@udo.edu<mailto:alireza.sal...@udo.edu>>, 
"user@spark.apache.org<mailto:user@spark.apache.org>" 
<user@spark.apache.org<mailto:user@spark.apache.org>>
Subject: Re: spark streaming : what is the best way to make a driver highly 
available

I'd suggest something like Apache YARN, or Apache Mesos with Marathon or 
something similar to allow for management, in particular restart on failure.

mn

On Aug 13, 2014, at 7:15 PM, Tobias Pfeiffer 
<t...@preferred.jp<mailto:t...@preferred.jp>> wrote:

Hi,

On Thu, Aug 14, 2014 at 5:49 AM, salemi 
<alireza.sal...@udo.edu<mailto:alireza.sal...@udo.edu>> wrote:
what is the best way to make a spark streaming driver highly available.

I would also be interested in that. In particular for Streaming applications 
where the Spark driver is running for a long time, this might be important, I 
think.

Thanks
Tobias


Reply via email to