See this thread: http://search-hadoop.com/m/q3RTtY7aX22B44dB

On Tue, Oct 13, 2015 at 5:53 PM, Jakob Odersky <joder...@gmail.com> wrote:

> I'm having trouble compiling Spark with SBT for Scala 2.11. The command I
> use is:
>
>     dev/change-version-to-2.11.sh
>     build/sbt -Pyarn -Phadoop-2.11 -Dscala-2.11
>
> followed by
>
>     compile
>
> in the sbt shell.
>
> The error I get specifically is:
>
> spark/core/src/main/scala/org/apache/spark/rpc/netty/NettyRpcEnv.scala:308:
> no valid targets for annotation on value conf - it is discarded unused. You
> may specify targets with meta-annotations, e.g. @(transient @param)
> [error] private[netty] class NettyRpcEndpointRef(@transient conf:
> SparkConf)
> [error]
>
> However I am also getting a large amount of deprecation warnings, making
> me wonder if I am supplying some incompatible/unsupported options to sbt. I
> am using Java 1.8 and the latest Spark master sources.
> Does someone know if I am doing anything wrong or is the sbt build broken?
>
> thanks for you help,
> --Jakob
>
>

Reply via email to