>From looking at the exception message that was returned, I would try the following command for running the application:
./bin/spark-submit --class test.etl.RunETL --master yarn-cluster --num-workers 14 --driver-memory 3200m --worker-memory 3g --worker-cores 2 --jar my-etl-1.0-SNAPSHOT-hadoop2.2.0.jar I didn't try this, so it may not work. Best, -Cesar On Sat, Jul 5, 2014 at 2:48 AM, Konstantin Kudryavtsev < kudryavtsev.konstan...@gmail.com> wrote: > Hi all, > > I have cluster with HDP 2.0. I built Spark 1.0 on edge node and trying to > run with a command > ./bin/spark-submit --class test.etl.RunETL --master yarn-cluster > --num-executors 14 --driver-memory 3200m --executor-memory 3g > --executor-cores 2 my-etl-1.0-SNAPSHOT-hadoop2.2.0.jar > > in result I got failed YARN application with following stack trace > > Application application_1404481778533_0068 failed 3 times due to AM > Container for appattempt_1404481778533_0068_000003 exited with exitCode: 1 > due to: Exception from container-launch: > org.apache.hadoop.util.Shell$ExitCodeException: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:464) > at org.apache.hadoop.util.Shell.run(Shell.java:379) > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589) > at > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:283) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:79) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:744) > .Failing this attempt.. Failing the application > > Log Type: stderr > > Log Length: 686 > > Unknown/unsupported param List(--executor-memory, 3072, --executor-cores, 2, > --num-executors, 14) > Usage: org.apache.spark.deploy.yarn.ApplicationMaster [options] > Options: > --jar JAR_PATH Path to your application's JAR file (required) > --class CLASS_NAME Name of your application's main class (required) > --args ARGS Arguments to be passed to your application's main > class. > Mutliple invocations are possible, each will be passed > in order. > --num-workers NUM Number of workers to start (Default: 2) > --worker-cores NUM Number of cores for the workers (Default: 1) > --worker-memory MEM Memory per Worker (e.g. 1000M, 2G) (Default: 1G) > > > Seems like the old spark notation.... any ideas? > > Thank you, > Konstantin Kudryavtsev > -- Cesar Arevalo Software Engineer ❘ Zephyr Health 450 Mission Street, Suite #201 ❘ San Francisco, CA 94105 m: +1 415-571-7687 ❘ s: arevalocesar | t: @zephyrhealth <https://twitter.com/zephyrhealth> o: +1 415-529-7649 ❘ f: +1 415-520-9288 http://www.zephyrhealth.com