Nevermind, it turns out that this is a problem for the Pivotal Hadoop that
we are trying to compile against.


On Wed, Jun 4, 2014 at 4:16 PM, Sung Hwan Chung <coded...@cs.stanford.edu>
wrote:

> When I run sbt/sbt assembly, I get the following exception. Is anyone else
> experiencing a similar problem?
>
>
> ..........
>
> [info] Resolving org.eclipse.jetty.orbit#javax.servlet;3.0.0.v201112011016
> ...
>
> [info] Updating {file:/Users/Sung/Projects/spark_06_04_14/}assembly...
>
> [info] Resolving org.fusesource.jansi#jansi;1.4 ...
>
> [info] Done updating.
>
> [info] Resolving org.eclipse.jetty#jetty-server;8.1.14.v20131031 ...
>
> [info] Updating {file:/Users/Sung/Projects/spark_06_04_14/}examples...
>
> [info] Resolving com.typesafe.genjavadoc#genjavadoc-plugin_2.10.4;0.5 ...
>
> *[error] impossible to get artifacts when data has not been loaded.
> IvyNode = org.slf4j#slf4j-api;1.6.1*
>
> [info] Resolving org.fusesource.jansi#jansi;1.4 ...
>
> [info] Done updating.
>
> [warn]
> /Users/Sung/Projects/spark_06_04_14/core/src/main/scala/org/apache/hadoop/mapred/SparkHadoopMapRedUtil.scala:43:
> constructor TaskAttemptID in class TaskAttemptID is deprecated: see
> corresponding Javadoc for more information.
>
> [warn]     new TaskAttemptID(jtIdentifier, jobId, isMap, taskId,
> attemptId)
>
> [warn]     ^
>
> [warn]
> /Users/Sung/Projects/spark_06_04_14/core/src/main/scala/org/apache/spark/SparkContext.scala:490:
> constructor Job in class Job is deprecated: see corresponding Javadoc for
> more information.
>
> [warn]     val job = new NewHadoopJob(hadoopConfiguration)
>
> [warn]               ^
>
> [warn]
> /Users/Sung/Projects/spark_06_04_14/core/src/main/scala/org/apache/spark/SparkContext.scala:623:
> constructor Job in class Job is deprecated: see corresponding Javadoc for
> more information.
>
> [warn]     val job = new NewHadoopJob(conf)
>

Reply via email to