Hi Wang Hao,

This is not removed. We moved it here:
http://spark.apache.org/docs/latest/hadoop-third-party-distributions.html
If you're building with SBT, and you don't specify the
SPARK_HADOOP_VERSION, then it defaults to 1.0.4.

Andrew


2014-06-12 6:24 GMT-07:00 Hao Wang <wh.s...@gmail.com>:

> Hi, all
>
> Why does the Spark 1.0.0 official doc remove how to build Spark with
> corresponding Hadoop version?
>
> It means that if I don't need to specify the Hadoop version with I build
> my Spark 1.0.0 with `sbt/sbt assembly`?
>
>
> Regards,
> Wang Hao(王灏)
>
> CloudTeam | School of Software Engineering
> Shanghai Jiao Tong University
> Address:800 Dongchuan Road, Minhang District, Shanghai, 200240
> Email:wh.s...@gmail.com
>

Reply via email to