See this thread: http://search-hadoop.com/m/q3RTtY7aX22B44dB
On Tue, Oct 13, 2015 at 5:53 PM, Jakob Odersky <joder...@gmail.com> wrote: > I'm having trouble compiling Spark with SBT for Scala 2.11. The command I > use is: > > dev/change-version-to-2.11.sh > build/sbt -Pyarn -Phadoop-2.11 -Dscala-2.11 > > followed by > > compile > > in the sbt shell. > > The error I get specifically is: > > spark/core/src/main/scala/org/apache/spark/rpc/netty/NettyRpcEnv.scala:308: > no valid targets for annotation on value conf - it is discarded unused. You > may specify targets with meta-annotations, e.g. @(transient @param) > [error] private[netty] class NettyRpcEndpointRef(@transient conf: > SparkConf) > [error] > > However I am also getting a large amount of deprecation warnings, making > me wonder if I am supplying some incompatible/unsupported options to sbt. I > am using Java 1.8 and the latest Spark master sources. > Does someone know if I am doing anything wrong or is the sbt build broken? > > thanks for you help, > --Jakob > >