Each handle to the RDD holds its lineage information, which means it knows
how it was computed starting from data in a reliable storage or from other
RDDs. RDDs hence can be reconstructed when the node fails.

Best Regards,
Sonal
Nube Technologies <http://www.nubetech.co>

<http://in.linkedin.com/in/sonalgoyal>




On Fri, Mar 28, 2014 at 3:55 PM, colt_colt <weihong9...@hotmail.com> wrote:

> I am curious about Spark fail over scenario, if some executor down,  that
> means the JVM crashed. AM will restart the executor, but how about the RDD
> data in JVM?  if I didn't persist RDD, does Spark will recompute lost RDD
> or
> just let it lose?  there is some description in Spark site: "Each RDD
> remembers the lineage of deterministic operations that were used on a
> fault-tolerant input dataset to create it."
>
> thanks in advance
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/How-does-Spark-handle-executor-down-RDD-in-this-executor-will-be-recomputed-automatically-tp3422.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>

Reply via email to