Is this because of versioning issue? can't wait for JDK 9 modular system. I
am not sure if spark plans to leverage it?

On Fri, Jan 20, 2017 at 1:30 PM, kant kodali <kanth...@gmail.com> wrote:

> I get the following exception. I am using Spark 2.0.1 and Scala 2.11.8.
>
> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0
> in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage
> 0.0 (TID 13, 172.31.20.212): java.io.InvalidClassException:
> org.apache.spark.executor.TaskMetrics; local class incompatible: stream
> classdesc serialVersionUID = -2231953621568687904, local class
> serialVersionUID = -6966587383730940799
>

Reply via email to