Hi Todd I just made a simple execution with 'yarn-client’ and it seems to work with my hive and spark. I have been using the pre-built version of Spark 1.4.1, so in my case I did not have to build it myself.
<http://www.taiger.com/> SOFIA PANAGIOTIDI www.taiger.com <http://www.taiger.com/> - +34 914497284 sofia.panagiot...@taiger.com - Taiger's everywhere. I'm in Madrid > On 12 Jan 2016, at 03:54, Todd <bit1...@163.com> wrote: > > Hi Sofia, > Thank you for the detailed guide. > I confirmed that there are compatibility issues between Hive 1.2.1 and Spark > 1.5.X. Hive 1.2.1 is by default built against spark 1.3.1 and compatible with > spark 1.4.1,but doesn't work with 1.5.X. > > Basically,I am using somewhat the same configuration as yours,but I am using > yarn-client as the spark.master. > Did you try to use yarn-client as the spark.master? > > I am building spark1.3.1 without hive and hive-thriftserver. I will get back > to you when I do the test, > > Thank you, Sofia. > > > > > At 2016-01-11 23:56:32, "Sofia" <sofia.panagiot...@taiger.com> wrote: > From what I have seen you need to use a Spark version that does not contain > the Hive jars. (maybe build one as described at the page). > Maybe someone can respond to the compatibility issue better than me. > > However, this is how I managed to make hive work with Spark > > I had similar issues with the downloaded Spark 1.5.2 and in the end I used > the prebuilt Spark 1.4.1 with my Hive 1.2.1 (and Hadoop 2.7.1). > When doing the setup I used the configurations below: > > hive-env.sh > > export SPARK_HOME=/usr/local/spark > export HADOOP_HOME=/usr/local/hadoop > > hive-site.xml > > <property> > <name>hive.execution.engine</name> > <value>spark</value> > </property> > <property> > <name>spark.master</name> > <value>spark://spark-master:7077</value> > <spark://spark-master:7077</value>> > </property> > > > spark-env.sh > > export HADOOP_HOME=/usr/local/hadoop > export PATH=$PATH:$HADOOP_HOME/bin > export SPARK_HOME=/usr/local/spark > > export SPARK_DIST_CLASSPATH=`hadoop classpath` > export SPARK_CLASSPATH=$CLASSPATH > > > I don’t know the answer to your issue exactly as at some point I faced the > same, but I hope this might help. > Cheers > > >> On 11 Jan 2016, at 11:39, Todd <bit1...@163.com <mailto:bit1...@163.com>> >> wrote: >> >> >> Thank you, Sofia. >> >> From the log, looks it is java.lang.AbstractMethodError that leads to the >> job failure. >> I am using Hive 1.2.1+Spark 1.5.2, is this a compatibility issue? >> >> >> >> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkBuildPlan >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Map 1 >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO Configuration.deprecation: mapred.task.is.map is deprecated. >> Instead, use mapreduce.task.ismap >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO exec.Utilities: Processing alias t1 >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO exec.Utilities: Adding input file >> hdfs://hadoop.bit.com:9000/user/hive/warehouse/db_1.db/t1 >> <hdfs://hadoop.bit.com:9000/user/hive/warehouse/db_1.db/t1> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: <PERFLOG method=serializePlan >> from=org.apache.hadoop.hive.ql.exec.Utilities> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO exec.Utilities: Serializing MapWork via kryo >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: </PERFLOG method=serializePlan >> start=1452505326307 end=1452505326332 duration=25 >> from=org.apache.hadoop.hive.ql.exec.Utilities> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO storage.MemoryStore: ensureFreeSpace(505800) called with >> curMem=0, maxMem=278302556 >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO storage.MemoryStore: Block broadcast_0 stored as values in >> memory (estimated size 493.9 KB, free 264.9 MB) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO storage.MemoryStore: ensureFreeSpace(42143) called with >> curMem=505800, maxMem=278302556 >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes >> in memory (estimated size 41.2 KB, free 264.9 MB) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory >> on 10.12.167.42:38883 (size: 41.2 KB, free: 265.4 MB) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 ERROR util.Utils: uncaught error in thread SparkListenerBus, >> stopping SparkContext >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: >> java.lang.AbstractMethodError >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.scheduler.SparkListenerBus$class.onPostEvent(SparkListenerBus.scala:62) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:56) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.util.AsynchronousListenerBus.postToAll(AsynchronousListenerBus.scala:37) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(AsynchronousListenerBus.scala:79) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1136) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.util.AsynchronousListenerBus$$anon$1.run(AsynchronousListenerBus.scala:63) >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO spark.SparkContext: Created broadcast 0 from hadoopRDD at >> SparkPlanGenerator.java:188 >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/metrics/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/stages/stage/kill,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/api,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/static,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/executors/threadDump/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/executors/threadDump,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/executors/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/executors,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/environment/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/environment,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/storage/rdd/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/storage/rdd,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/storage/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/storage,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/stages/pool/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/stages/pool,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/stages/stage/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/stages/stage,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/stages/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/stages,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/jobs/job/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/jobs/job,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/jobs/json,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO handler.ContextHandler: stopped >> o.s.j.s.ServletContextHandler{/jobs,null} >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Map 1 >> start=1452505326274 end=1452505326824 duration=550 >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Reducer 2 >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: <PERFLOG method=serializePlan >> from=org.apache.hadoop.hive.ql.exec.Utilities> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO exec.Utilities: Serializing ReduceWork via kryo >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO ui.SparkUI: Stopped Spark web UI at http://10.12.167.42:4040 >> <http://10.12.167.42:4040/> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO scheduler.DAGScheduler: Stopping DAGScheduler >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO cluster.YarnClientSchedulerBackend: Shutting down all executors >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO cluster.YarnClientSchedulerBackend: Interrupting monitor thread >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO cluster.YarnClientSchedulerBackend: Asking each executor to >> shut down >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO cluster.YarnClientSchedulerBackend: Stopped >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: </PERFLOG method=serializePlan >> start=1452505326849 end=1452505326874 duration=25 >> from=org.apache.hadoop.hive.ql.exec.Utilities> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Reducer 2 >> start=1452505326824 end=1452505326936 duration=112 >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkBuildPlan >> start=1452505326273 end=1452505326936 duration=663 >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator> >> 16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkBuildRDDGraph >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan> >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:07 INFO log.PerfLogger: </PERFLOG method=SparkBuildRDDGraph >> start=1452505326936 end=1452505327000 duration=64 >> from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan> >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 >> 17:42:07 INFO client.RemoteDriver: Failed to run job >> b5a8886e-8229-4773-b943-b3c6f3c710bf >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: >> java.lang.IllegalStateException: Cannot call methods on a stopped >> SparkContext >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.SparkContext.org >> <http://org.apache.spark.sparkcontext.org/>$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:104) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.SparkContext.submitJob(SparkContext.scala:1981) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:118) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:116) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.rdd.RDD.withScope(RDD.scala:310) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.rdd.AsyncRDDActions.foreachAsync(AsyncRDDActions.scala:116) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.api.java.JavaRDDLike$class.foreachAsync(JavaRDDLike.scala:690) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.spark.api.java.AbstractJavaRDDLike.foreachAsync(JavaRDDLike.scala:47) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient$JobStatusJob.call(RemoteHiveSparkClient.java:257) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:366) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:335) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> java.util.concurrent.FutureTask.run(FutureTask.java:262) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> 16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at >> java.lang.Thread.run(Thread.java:745) >> 16/01/11 17:42:07 [RPC-Handler-3]: INFO client.SparkClientImpl: Received >> result for b5a8886e-8229-4773-b943-b3c6f3c710bf >> Status: Failed >> 16/01/11 17:42:07 [main]: ERROR status.SparkJobMonitor: Status: Failed >> >> >> >> >> At 2016-01-11 17:17:51, "Sofia" <sofia.panagiot...@taiger.com >> <mailto:sofia.panagiot...@taiger.com>> wrote: >> Check log files to see the exact error. Alternatively run hive like this, to >> get the exact error >> >> hive --hiveconf hive.root.logger=INFO,console >> >> There is a good chance that you are encountering problems between the Hive >> and Spark versions and installation. >> See >> https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started >> >> <https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started> >> >> >>> On 11 Jan 2016, at 08:47, Todd <bit1...@163.com <mailto:bit1...@163.com>> >>> wrote: >>> >>> Hive, >>> >>> I am trying out the Hive on Spark with hive 1.2.1 and spark 1.5.2. Could >>> someone help me on this? Thanks! >>> >>> Following are my steps: >>> 1. build spark 1.5.2 without Hive and Hive Thrift Server. At this point, I >>> can use it to submit application using spark-submit --master yarn-client >>> 2. And the built spark assembly jar into $HIVE_HOME/lib >>> 3. start hive and add the following parameters >>> hive > set spark.master=yarn-client >>> hive > set spark.executor.memory=512M >>> hive > set spark.driver.memory=512M >>> hive > set spark.executor.instances=1 >>> >>> 4. Then I run a simple query : select count(1) from t1; >>> >>> The job fails will following error: >>> >>> =============================================================================== >>> YARN executor launch context: >>> env: >>> CLASSPATH -> >>> {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/* >>> SPARK_LOG_URL_STDERR -> >>> http://hadoop-Aspire-TC-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stderr?start=-4096 >>> >>> <http://hadoop-aspire-tc-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stderr?start=-4096> >>> SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1452320323183_0007 >>> SPARK_YARN_CACHE_FILES_FILE_SIZES -> 142746538 >>> SPARK_USER -> hadoop >>> SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE >>> SPARK_YARN_MODE -> true >>> SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1452496343550 >>> SPARK_LOG_URL_STDOUT -> >>> http://hadoop-Aspire-TC-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stdout?start=-4096 >>> >>> <http://hadoop-aspire-tc-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stdout?start=-4096> >>> SPARK_YARN_CACHE_FILES -> >>> hdfs://hadoop.bit.com:9000/user/hadoop/.sparkStaging/application_1452320323183_0007/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar >>> >>> <hdfs://hadoop.bit.com:9000/user/hadoop/.sparkStaging/application_1452320323183_0007/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar> >>> >>> command: >>> {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' >>> -Xms512m -Xmx512m >>> '-Dhive.spark.log.dir=/home/hadoop/software/bigdata/spark-1.5.2-bin-hadoop2.6.0-withouthive/logs/' >>> -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.driver.port=43675' >>> -Dspark.yarn.app.container.log.dir=<LOG_DIR> >>> org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url >>> akka.tcp://sparkDriver@10.12.167.42:43675/user/CoarseGrainedScheduler >>> <akka.tcp://sparkDriver@10.12.167.42:43675/user/CoarseGrainedScheduler> >>> --executor-id 2 --hostname hadoop-Aspire-TC-606 --cores 1 --app-id >>> application_1452320323183_0007 --user-class-path file:$PWD/__app__.jar 1> >>> <LOG_DIR>/stdout 2> <LOG_DIR>/stderr >>> =============================================================================== >>> >>> 16/01/11 15:12:37 INFO impl.ContainerManagementProtocolProxy: Opening proxy >>> : hadoop-Aspire-TC-606:50804 >>> 16/01/11 15:12:40 INFO yarn.YarnAllocator: Completed container >>> container_1452320323183_0007_01_000003 (state: COMPLETE, exit status: 1) >>> 16/01/11 15:12:40 INFO yarn.YarnAllocator: Container marked as failed: >>> container_1452320323183_0007_01_000003. Exit status: 1. Diagnostics: >>> Exception from container-launch. >>> Container id: container_1452320323183_0007_01_000003 >>> Exit code: 1 >>> Stack trace: ExitCodeException exitCode=1: >>> at org.apache.hadoop.util.Shell.runCommand(Shell.java:538) >>> at org.apache.hadoop.util.Shell.run(Shell.java:455) >>> at >>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715) >>> at >>> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:211) >>> at >>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302) >>> at >>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> >>> >>> Container exited with a non-zero exit code 1 >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >> >