I'm trying out 2.0, and ran a long job with 10 stages, in spark-shell it seems that after all 10 finished successfully, if I run the last, or the 9th again, spark reruns all the previous stages from scratch, instead of utilizing the partial results.
this is quite serious since I can't experiment while making small changes to the code. any idea what part of the spark framework might have caused this ? thanks Yang