Can you provide links to the sections that are confusing? My understanding, the HDP1 binaries do not need YARN, while the HDP2 binaries do.
Now, you can also install Hortonworks Spark RPM... For production, in my opinion, RPMs are better for manageability. > On Jul 6, 2014, at 5:39 PM, Konstantin Kudryavtsev > <kudryavtsev.konstan...@gmail.com> wrote: > > Hello, thanks for your message... I'm confused, Hortonworhs suggest install > spark rpm on each node, but on Spark main page said that yarn enough and I > don't need to install it... What the difference? > > sent from my HTC > >> On Jul 6, 2014 8:34 PM, "vs" <vinayshu...@gmail.com> wrote: >> Konstantin, >> >> HWRK provides a Tech Preview of Spark 0.9.1 with HDP 2.1 that you can try >> from >> http://hortonworks.com/wp-content/uploads/2014/05/SparkTechnicalPreview.pdf >> >> Let me know if you see issues with the tech preview. >> >> "spark PI example on HDP 2.0 >> >> I downloaded spark 1.0 pre-build from http://spark.apache.org/downloads.html >> (for HDP2) >> The run example from spark web-site: >> ./bin/spark-submit --class org.apache.spark.examples.SparkPi --master >> yarn-cluster --num-executors 3 --driver-memory 2g --executor-memory 2g >> --executor-cores 1 ./lib/spark-examples-1.0.0-hadoop2.2.0.jar 2 >> >> I got error: >> Application application_1404470405736_0044 failed 3 times due to AM >> Container for appattempt_1404470405736_0044_000003 exited with exitCode: 1 >> due to: Exception from container-launch: >> org.apache.hadoop.util.Shell$ExitCodeException: >> at org.apache.hadoop.util.Shell.runCommand(Shell.java:464) >> at org.apache.hadoop.util.Shell.run(Shell.java:379) >> at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589) >> at >> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195) >> at >> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:283) >> at >> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:79) >> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> at java.lang.Thread.run(Thread.java:744) >> .Failing this attempt.. Failing the application. >> >> Unknown/unsupported param List(--executor-memory, 2048, --executor-cores, 1, >> --num-executors, 3) >> Usage: org.apache.spark.deploy.yarn.ApplicationMaster [options] >> Options: >> --jar JAR_PATH Path to your application's JAR file (required) >> --class CLASS_NAME Name of your application's main class (required) >> ...bla-bla-bla >> " >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/Unable-to-run-Spark-1-0-SparkPi-on-HDP-2-0-tp8802p8873.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com.