You seem to have the binary for Hadoop 2, since it was compiled expecting that TaskAttemptContext is an interface. So the error indicates that Spark is also seeing Hadoop 1 classes somewhere.
On Wed, Jun 25, 2014 at 4:41 PM, Robert James <srobertja...@gmail.com> wrote: > After upgrading to Spark 1.0.0, I get this error: > > ERROR org.apache.spark.executor.ExecutorUncaughtExceptionHandler - > Uncaught exception in thread Thread[Executor task launch > worker-2,5,main] > java.lang.IncompatibleClassChangeError: Found interface > org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected > > I thought this was caused by a dependency on Hadoop 1.0.4 (even though > I downloaded the Spark 1.0.0 for Hadoop 2), but I can't seem to fix > it. Any advice?