AFAIK spark doesn't restart worker nodes itself. You can have multiple worker nodes and in that case if one worker node goes down, then spark will try to recompute those lost RDDs again with those workers who are alive.
Thanks Best Regards On Sun, Oct 5, 2014 at 5:19 AM, Jahagirdar, Madhu < madhu.jahagir...@philips.com> wrote: > In my spark streaming program I have created kafka utils to receive data > and store data in elastic search and in flume. Storing function is applied > on same dstream. My question what is the behavior of spark if after storing > data in elastic search the worker node dies before storing in flume? Does > it restart worker and then again store the data in elastic search and then > flume or does it only run functions to store in flume. > > Regards > Madhu Jahagirdar > > ________________________________ > The information contained in this message may be confidential and legally > protected under applicable law. The message is intended solely for the > addressee(s). If you are not the intended recipient, you are hereby > notified that any use, forwarding, dissemination, or reproduction of this > message is strictly prohibited and may be unlawful. If you are not the > intended recipient, please contact the sender by return e-mail and destroy > all copies of the original message. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >