You seem to have the binary for Hadoop 2, since it was compiled
expecting that TaskAttemptContext is an interface. So the error
indicates that Spark is also seeing Hadoop 1 classes somewhere.

On Wed, Jun 25, 2014 at 4:41 PM, Robert James <srobertja...@gmail.com> wrote:
> After upgrading to Spark 1.0.0, I get this error:
>
>  ERROR org.apache.spark.executor.ExecutorUncaughtExceptionHandler -
> Uncaught exception in thread Thread[Executor task launch
> worker-2,5,main]
> java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected
>
> I thought this was caused by a dependency on Hadoop 1.0.4 (even though
> I downloaded the Spark 1.0.0 for Hadoop 2), but I can't seem to fix
> it.  Any advice?

Reply via email to