@Andrew
Thanks for the reply, did you run this in a Hortonworks or Cloudera cluster?
I suspect the issue is coming from the ​extraJavaOptions as these are
necessary in HDP, the strange thing is that with exactly the same settings
1.5 works.

# jar -tf spark-assembly-1.6.0-SNAPSHOT-hadoop2.7.1.jar | grep
ApplicationMaster.class
org/apache/spark/deploy/yarn/ApplicationMaster.class

----
Exit code: 1
Exception message:
/hadoop/hdfs/disk02/hadoop/yarn/local/usercache/syepes/appcache/application_1445706872927_1593/container_e44_1445706872927_1593_02_000001/launch_container.sh:
line 24:
/usr/hdp/current/hadoop-client/lib/hadoop-lzo-0.6.0.2.3.2.0-2950.jar:$PWD:$PWD/__spark_conf__:$PWD/__spark__.jar:$HADOOP_CONF_DIR:/usr/hdp/current/hadoop-client/*:/usr/hdp/current/hadoop-client/lib/*:/usr/hdp/current/hadoop-hdfs-client/*:/usr/hdp/current/hadoop-hdfs-client/lib/*:/usr/hdp/current/hadoop-yarn-client/*:/usr/hdp/current/hadoop-yarn-client/lib/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:$PWD/mr-framework/hadoop/share/hadoop/tools/lib/*:/usr/hdp/${hdp.version}/hadoop/lib/hadoop-lzo-0.6.0.${hdp.version}.jar:/etc/hadoop/conf/secure:
bad substitution
-----

​Regards,
 Sebastian

On Thu, Dec 17, 2015 at 9:14 PM, Andrew Or <and...@databricks.com> wrote:

> @syepes
>
> I just run Spark 1.6 (881f254) on YARN with Hadoop 2.4.0. I was able to
> run a simple application in cluster mode successfully.
>
> Can you verify whether the org.apache.spark.yarn.ApplicationMaster class
> exists in your assembly jar?
>
> jar -tf assembly.jar | grep ApplicationMaster
>
> -Andrew
>
>
> 2015-12-17 7:44 GMT-08:00 syepes <sye...@gmail.com>:
>
>> -1 (YARN Cluster deployment mode not working)
>>
>> I have just tested 1.6 (d509194b) on our HDP 2.3 platform and the cluster
>> mode does not seem work. It looks like some parameter are not being passed
>> correctly.
>> This example works correctly with 1.5.
>>
>> # spark-submit --master yarn --deploy-mode cluster --num-executors 1
>> --properties-file $PWD/spark-props.conf --class
>> org.apache.spark.examples.SparkPi
>> /opt/spark/lib/spark-examples-1.6.0-SNAPSHOT-hadoop2.7.1.jar
>>
>> Error: Could not find or load main class
>> org.apache.spark.deploy.yarn.ApplicationMaster
>>
>> spark-props.conf
>> -----------------------------
>> spark.driver.
>> ​​
>> extraJavaOptions                -Dhdp.version=2.3.2.0-2950
>> spark.driver.extraLibraryPath
>>
>> /usr/hdp/current/hadoop-client/lib/native:/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64
>> spark.executor.extraJavaOptions              -Dhdp.version=2.3.2.0-2950
>> spark.executor.extraLibraryPath
>>
>> /usr/hdp/current/hadoop-client/lib/native:/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64
>> -----------------------------
>>
>> I will try to do some more debugging on this issue.
>>
>>
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-developers-list.1001551.n3.nabble.com/VOTE-Release-Apache-Spark-1-6-0-RC3-tp15660p15692.html
>> Sent from the Apache Spark Developers List mailing list archive at
>> Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>> For additional commands, e-mail: dev-h...@spark.apache.org
>>
>>
>

Reply via email to