I incremented the version of spark from 1.4.0 to 1.4.0.1 and ran ./make-distribution.sh --tgz -Phadoop-2.4 -Pyarn -Phive -Phive-thriftserver
Build was successful but the script faild. Is there a way to pass the incremented version ? [INFO] BUILD SUCCESS [INFO] ------------------------------------------------------------------------ [INFO] Total time: 09:56 min [INFO] Finished at: 2015-06-28T13:45:29-07:00 [INFO] Final Memory: 84M/902M [INFO] ------------------------------------------------------------------------ + rm -rf /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist + mkdir -p /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib + echo 'Spark 1.4.0.1 built for Hadoop 2.4.0' + echo 'Build flags: -Phadoop-2.4' -Pyarn -Phive -Phive-thriftserver + cp /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/assembly/target/scala-2.10/spark-assembly-1.4.0.1-hadoop2.4.0.jar /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/ + cp /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/examples/target/scala-2.10/spark-examples-1.4.0.1-hadoop2.4.0.jar /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/ + cp /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/network/yarn/target/scala-2.10/spark-1.4.0.1-yarn-shuffle.jar /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/ + mkdir -p /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/examples/src/main + cp -r /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/examples/src/main /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/examples/src/ + '[' 1 == 1 ']' + cp '/Users/dvasthimal/ebay/projects/ep/spark-1.4.0/lib_managed/jars/datanucleus*.jar' /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/dist/lib/ cp: /Users/dvasthimal/ebay/projects/ep/spark-1.4.0/lib_managed/jars/datanucleus*.jar: No such file or directory LM-SJL-00877532:spark-1.4.0 dvasthimal$ ./make-distribution.sh --tgz -Phadoop-2.4 -Pyarn -Phive -Phive-thriftserver On Sun, Jun 28, 2015 at 1:41 PM, Koert Kuipers <ko...@tresata.com> wrote: > you need 1) to publish to inhouse maven, so your application can depend on > your version, and 2) use the spark distribution you compiled to launch your > job (assuming you run with yarn so you can launch multiple versions of > spark on same cluster) > > On Sun, Jun 28, 2015 at 4:33 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> > wrote: > >> How can i import this pre-built spark into my application via maven as i >> want to use the block join API. >> >> On Sun, Jun 28, 2015 at 1:31 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >> wrote: >> >>> I ran this w/o maven options >>> >>> ./make-distribution.sh --tgz -Phadoop-2.4 -Pyarn -Phive >>> -Phive-thriftserver >>> >>> I got this spark-1.4.0-bin-2.4.0.tgz in the same working directory. >>> >>> I hope this is built with 2.4.x hadoop as i did specify -P >>> >>> On Sun, Jun 28, 2015 at 1:10 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> >>> wrote: >>> >>>> ./make-distribution.sh --tgz --*mvn* "-Phadoop-2.4 -Pyarn >>>> -Dhadoop.version=2.4.0 -Phive -Phive-thriftserver -DskipTests clean >>>> package" >>>> >>>> >>>> or >>>> >>>> >>>> ./make-distribution.sh --tgz --*mvn* -Phadoop-2.4 -Pyarn >>>> -Dhadoop.version=2.4.0 -Phive -Phive-thriftserver -DskipTests clean >>>> package" >>>> Both fail with >>>> >>>> + echo -e 'Specify the Maven command with the --mvn flag' >>>> >>>> Specify the Maven command with the --mvn flag >>>> >>>> + exit -1 >>>> >>> >>> >>> >>> -- >>> Deepak >>> >>> >> >> >> -- >> Deepak >> >> > -- Deepak