I issued the same basic command and it worked fine.

RADTech-MBP:spark $ ./make-distribution.sh --name hadoop-2.6 --tgz -Pyarn
-Phadoop-2.6 -Dhadoop.version=2.6.0 -Phive -Phive-thriftserver -DskipTests

Which created: spark-1.6.0-SNAPSHOT-bin-hadoop-2.6.tgz in the root
directory of the project.

FWIW, the environment was an MBP with OS X 10.10.5 and Java:

java version "1.8.0_51"
Java(TM) SE Runtime Environment (build 1.8.0_51-b16)
Java HotSpot(TM) 64-Bit Server VM (build 25.51-b03, mixed mode)

-Todd

On Tue, Oct 27, 2015 at 12:17 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> I used the following command:
> make-distribution.sh --name custom-spark --tgz -Phadoop-2.4 -Phive
> -Phive-thriftserver -Pyarn
>
> spark-1.6.0-SNAPSHOT-bin-custom-spark.tgz was generated (with patch from
> SPARK-11348)
>
> Can you try above command ?
>
> Thanks
>
> On Tue, Oct 27, 2015 at 7:03 AM, Kayode Odeyemi <drey...@gmail.com> wrote:
>
>> Ted, I switched to this:
>>
>> ./make-distribution.sh --name spark-latest --tgz -Dhadoop.version=2.6.0
>> -Phadoop-2.6 -Phive -Phive-thriftserver -Pyarn -DskipTests clean package -U
>>
>> Same error. No .gz file. Here's the bottom output log:
>>
>> + rm -rf /home/emperor/javaprojects/spark/dist
>> + mkdir -p /home/emperor/javaprojects/spark/dist/lib
>> + echo 'Spark [WARNING] See
>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin (git revision
>> 3689beb) built for Hadoop [WARNING] See
>> http://docs.codehaus.org/display/MAVENUSER/Shade+Pl
>> + echo 'Build flags: -Dhadoop.version=2.6.0' -Phadoop-2.6 -Phive
>> -Phive-thriftserver -Pyarn -DskipTests clean package -U
>> + cp
>> /home/emperor/javaprojects/spark/assembly/target/scala-2.10/spark-assembly-1.6.0-SNAPSHOT-hadoop2.6.0.jar
>> /home/emperor/javaprojects/spark/dist/lib/
>> + cp
>> /home/emperor/javaprojects/spark/examples/target/scala-2.10/spark-examples-1.6.0-SNAPSHOT-hadoop2.6.0.jar
>> /home/emperor/javaprojects/spark/dist/lib/
>> + cp
>> /home/emperor/javaprojects/spark/network/yarn/target/scala-2.10/spark-1.6.0-SNAPSHOT-yarn-shuffle.jar
>> /home/emperor/javaprojects/spark/dist/lib/
>> + mkdir -p /home/emperor/javaprojects/spark/dist/examples/src/main
>> + cp -r /home/emperor/javaprojects/spark/examples/src/main
>> /home/emperor/javaprojects/spark/dist/examples/src/
>> + '[' 1 == 1 ']'
>> + cp
>> /home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-api-jdo-3.2.6.jar
>> /home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-core-3.2.10.jar
>> /home/emperor/javaprojects
>> ed/jars/datanucleus-rdbms-3.2.9.jar
>> /home/emperor/javaprojects/spark/dist/lib/
>> + cp /home/emperor/javaprojects/spark/LICENSE
>> /home/emperor/javaprojects/spark/dist
>> + cp -r /home/emperor/javaprojects/spark/licenses
>> /home/emperor/javaprojects/spark/dist
>> + cp /home/emperor/javaprojects/spark/NOTICE
>> /home/emperor/javaprojects/spark/dist
>> + '[' -e /home/emperor/javaprojects/spark/CHANGES.txt ']'
>> + cp -r /home/emperor/javaprojects/spark/data
>> /home/emperor/javaprojects/spark/dist
>> + mkdir /home/emperor/javaprojects/spark/dist/conf
>> + cp /home/emperor/javaprojects/spark/conf/docker.properties.template
>> /home/emperor/javaprojects/spark/conf/fairscheduler.xml.template
>> /home/emperor/javaprojects/spark/conf/log4j.properties
>> emperor/javaprojects/spark/conf/metrics.properties.template
>> /home/emperor/javaprojects/spark/conf/slaves.template
>> /home/emperor/javaprojects/spark/conf/spark-defaults.conf.template /home/em
>> ts/spark/conf/spark-env.sh.template
>> /home/emperor/javaprojects/spark/dist/conf
>> + cp /home/emperor/javaprojects/spark/README.md
>> /home/emperor/javaprojects/spark/dist
>> + cp -r /home/emperor/javaprojects/spark/bin
>> /home/emperor/javaprojects/spark/dist
>> + cp -r /home/emperor/javaprojects/spark/python
>> /home/emperor/javaprojects/spark/dist
>> + cp -r /home/emperor/javaprojects/spark/sbin
>> /home/emperor/javaprojects/spark/dist
>> + cp -r /home/emperor/javaprojects/spark/ec2
>> /home/emperor/javaprojects/spark/dist
>> + '[' -d /home/emperor/javaprojects/spark/R/lib/SparkR ']'
>> + '[' false == true ']'
>> + '[' true == true ']'
>> + TARDIR_NAME='spark-[WARNING] See
>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
>> + TARDIR='/home/emperor/javaprojects/spark/spark-[WARNING] See
>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
>> + rm -rf '/home/emperor/javaprojects/spark/spark-[WARNING] See
>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
>> + cp -r /home/emperor/javaprojects/spark/dist
>> '/home/emperor/javaprojects/spark/spark-[WARNING] See
>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
>> cp: cannot create directory
>> `/home/emperor/javaprojects/spark/spark-[WARNING] See
>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest':
>> No such file or directory
>>
>>
>> On Tue, Oct 27, 2015 at 2:14 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>
>>> Can you try the same command shown in the pull request ?
>>>
>>> Thanks
>>>
>>> On Oct 27, 2015, at 12:40 AM, Kayode Odeyemi <drey...@gmail.com> wrote:
>>>
>>> Thank you.
>>>
>>> But I'm getting same warnings and it's still preventing the archive from
>>> being generated.
>>>
>>> I've ran this on both OSX Lion and Ubuntu 12. Same error. No .gz file
>>>
>>> On Mon, Oct 26, 2015 at 9:10 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>
>>>> Looks like '-Pyarn' was missing in your command.
>>>>
>>>> On Mon, Oct 26, 2015 at 12:06 PM, Kayode Odeyemi <drey...@gmail.com>
>>>> wrote:
>>>>
>>>>> I used this command which is synonymous to what you have:
>>>>>
>>>>> ./make-distribution.sh --name spark-latest --tgz --mvn mvn
>>>>> -Dhadoop.version=2.6.0 -Phadoop-2.6 -Phive -Phive-thriftserver -DskipTests
>>>>> clean package -U
>>>>>
>>>>> But I still see WARNINGS like this in the output and no .gz file
>>>>> created:
>>>>>
>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/.part-r-00005.gz.parquet.crc:
>>>>> No such file or directory
>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet:
>>>>> No such file or directory
>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>>> No such file or directory
>>>>> cp:
>>>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>>> unable to copy extended attributes to
>>>>> /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>>> No such file or directory
>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>>> No such file or directory
>>>>> cp:
>>>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>>> unable to copy extended attributes to
>>>>> /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>>> No such file or directory
>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/.part-r-00007.gz.parquet.crc:
>>>>> No such file or directory
>>>>>
>>>>> On Mon, Oct 26, 2015 at 8:58 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>>>
>>>>>> If you use the command shown in:
>>>>>> https://github.com/apache/spark/pull/9281
>>>>>>
>>>>>> You should have got the following:
>>>>>>
>>>>>>
>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/part-r-00008.gz.parquet
>>>>>>
>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/part-r-00007.gz.parquet
>>>>>>
>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00004.gz.parquet
>>>>>>
>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00002.gz.parquet
>>>>>>
>>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet
>>>>>>
>>>>>> On Mon, Oct 26, 2015 at 11:47 AM, Kayode Odeyemi <drey...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> I see a lot of stuffs like this after the a successful maven build:
>>>>>>>
>>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/
>>>>>>> part-r-00008.gz.parquet: No such file or directory
>>>>>>>
>>>>>>> Seems it fails when it tries to package the build as an archive.
>>>>>>>
>>>>>>> I'm using the latest code on github master.
>>>>>>>
>>>>>>> Any ideas please?
>>>>>>>
>>>>>>> On Mon, Oct 26, 2015 at 6:20 PM, Yana Kadiyska <
>>>>>>> yana.kadiy...@gmail.com> wrote:
>>>>>>>
>>>>>>>> In 1.4 ./make_distribution produces a .tgz file in the root
>>>>>>>> directory (same directory that make_distribution is in)
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Oct 26, 2015 at 8:46 AM, Kayode Odeyemi <drey...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> The ./make_distribution task completed. However, I can't seem to
>>>>>>>>> locate the
>>>>>>>>> .tar.gz file.
>>>>>>>>>
>>>>>>>>> Where does Spark save this? or should I just work with the dist
>>>>>>>>> directory?
>>>>>>>>>
>>>>>>>>> On Fri, Oct 23, 2015 at 4:23 PM, Kayode Odeyemi <drey...@gmail.com
>>>>>>>>> > wrote:
>>>>>>>>>
>>>>>>>>>> I saw this when I tested manually (without ./make-distribution)
>>>>>>>>>>
>>>>>>>>>> Detected Maven Version: 3.2.2 is not in the allowed range 3.3.3.
>>>>>>>>>>
>>>>>>>>>> So I simply upgraded maven to 3.3.3.
>>>>>>>>>>
>>>>>>>>>> Resolved. Thanks
>>>>>>>>>>
>>>>>>>>>> On Fri, Oct 23, 2015 at 3:17 PM, Sean Owen <so...@cloudera.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> This doesn't show the actual error output from Maven. I have a
>>>>>>>>>>> strong
>>>>>>>>>>> guess that you haven't set MAVEN_OPTS to increase the memory
>>>>>>>>>>> Maven can
>>>>>>>>>>> use.
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Oct 23, 2015 at 6:14 AM, Kayode Odeyemi <
>>>>>>>>>>> drey...@gmail.com> wrote:
>>>>>>>>>>> > Hi,
>>>>>>>>>>> >
>>>>>>>>>>> > I can't seem to get a successful maven build. Please see
>>>>>>>>>>> command output
>>>>>>>>>>> > below:
>>>>>>>>>>> >
>>>>>>>>>>> > bash-3.2$ ./make-distribution.sh --name spark-latest --tgz
>>>>>>>>>>> --mvn mvn
>>>>>>>>>>> > -Dhadoop.version=2.7.0 -Phadoop-2.7 -Phive -Phive-thriftserver
>>>>>>>>>>> -DskipTests
>>>>>>>>>>> > clean package
>>>>>>>>>>> > +++ dirname ./make-distribution.sh
>>>>>>>>>>> > ++ cd .
>>>>>>>>>>> > ++ pwd
>>>>>>>>>>> > + SPARK_HOME=/usr/local/spark-latest
>>>>>>>>>>> > + DISTDIR=/usr/local/spark-latest/dist
>>>>>>>>>>> > + SPARK_TACHYON=false
>>>>>>>>>>> > + TACHYON_VERSION=0.7.1
>>>>>>>>>>> > + TACHYON_TGZ=tachyon-0.7.1-bin.tar.gz
>>>>>>>>>>> > +
>>>>>>>>>>> > TACHYON_URL=
>>>>>>>>>>> https://github.com/amplab/tachyon/releases/download/v0.7.1/tachyon-0.7.1-bin.tar.gz
>>>>>>>>>>> > + MAKE_TGZ=false
>>>>>>>>>>> > + NAME=none
>>>>>>>>>>> > + MVN=/usr/local/spark-latest/build/mvn
>>>>>>>>>>> > + ((  12  ))
>>>>>>>>>>> > + case $1 in
>>>>>>>>>>> > + NAME=spark-latest
>>>>>>>>>>> > + shift
>>>>>>>>>>> > + shift
>>>>>>>>>>> > + ((  10  ))
>>>>>>>>>>> > + case $1 in
>>>>>>>>>>> > + MAKE_TGZ=true
>>>>>>>>>>> > + shift
>>>>>>>>>>> > + ((  9  ))
>>>>>>>>>>> > + case $1 in
>>>>>>>>>>> > + MVN=mvn
>>>>>>>>>>> > + shift
>>>>>>>>>>> > + shift
>>>>>>>>>>> > + ((  7  ))
>>>>>>>>>>> > + case $1 in
>>>>>>>>>>> > + break
>>>>>>>>>>> > + '[' -z
>>>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>>>> > + '[' -z
>>>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>>>> > ++ command -v git
>>>>>>>>>>> > + '[' /usr/bin/git ']'
>>>>>>>>>>> > ++ git rev-parse --short HEAD
>>>>>>>>>>> > + GITREV=487d409
>>>>>>>>>>> > + '[' '!' -z 487d409 ']'
>>>>>>>>>>> > + GITREVSTRING=' (git revision 487d409)'
>>>>>>>>>>> > + unset GITREV
>>>>>>>>>>> > ++ command -v mvn
>>>>>>>>>>> > + '[' '!' /usr/bin/mvn ']'
>>>>>>>>>>> > ++ mvn help:evaluate -Dexpression=project.version
>>>>>>>>>>> -Dhadoop.version=2.7.0
>>>>>>>>>>> > -Phadoop-2.7 -Phive -Phive-thriftserver -DskipTests clean
>>>>>>>>>>> package
>>>>>>>>>>> > ++ grep -v INFO
>>>>>>>>>>> > ++ tail -n 1
>>>>>>>>>>> > + VERSION='[ERROR] [Help 1]
>>>>>>>>>>> >
>>>>>>>>>>> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
>>>>>>>>>>> '
>>>>>>>>>>> >
>>>>>>>>>>> > Same output error with JDK 7
>>>>>>>>>>> >
>>>>>>>>>>> > Appreciate your help.
>>>>>>>>>>> >
>>>>>>>>>>> >
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to