[ https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14225417#comment-14225417 ]
Brock Noland commented on HIVE-8836: ------------------------------------ One of the stuck processes: {noformat} 2014-11-25 14:30:44,874 INFO ql.Driver (SessionState.java:printInfo(828)) - Query ID = hiveptest_20141125143030_c56b5f72-8552-4122-930d-7df9dea96638 2014-11-25 14:30:44,874 INFO ql.Driver (SessionState.java:printInfo(828)) - Total jobs = 1 2014-11-25 14:30:44,874 INFO ql.Driver (SessionState.java:printInfo(828)) - Launching Job 1 out of 1 2014-11-25 14:30:44,875 INFO ql.Driver (Driver.java:launchTask(1643)) - Starting task [Stage-1:MAPRED] in serial mode 2014-11-25 14:30:44,875 INFO exec.Task (SessionState.java:printInfo(828)) - In order to change the average load for a reducer (in bytes): 2014-11-25 14:30:44,876 INFO exec.Task (SessionState.java:printInfo(828)) - set hive.exec.reducers.bytes.per.reducer=<number> 2014-11-25 14:30:44,876 INFO exec.Task (SessionState.java:printInfo(828)) - In order to limit the maximum number of reducers: 2014-11-25 14:30:44,876 INFO exec.Task (SessionState.java:printInfo(828)) - set hive.exec.reducers.max=<number> 2014-11-25 14:30:44,876 INFO exec.Task (SessionState.java:printInfo(828)) - In order to set a constant number of reducers: 2014-11-25 14:30:44,876 INFO exec.Task (SessionState.java:printInfo(828)) - set mapreduce.job.reduces=<number> 2014-11-25 14:30:44,876 INFO spark.HiveSparkClientFactory (HiveSparkClientFactory.java:initiateSparkConf(105)) - load spark configuration from hive configuration (spark.master -> local-cluster[2,1,2048]). 2014-11-25 14:30:44,894 INFO slf4j.Slf4jLogger (Slf4jLogger.scala:applyOrElse(80)) - Slf4jLogger started 2014-11-25 14:30:44,899 INFO Remoting (Slf4jLogger.scala:apply$mcV$sp(74)) - Starting remoting 2014-11-25 14:30:44,907 INFO Remoting (Slf4jLogger.scala:apply$mcV$sp(74)) - Remoting started; listening on addresses :[akka.tcp://d84d934e-e10e-4744-b42c-ed86a49ebbd3@10.227.4.181:56697] 2014-11-25 14:30:44,909 DEBUG client.SparkClientImpl (SparkClientImpl.java:startDriver(252)) - Running client driver with argv: /home/hiveptest/50.18.64.184-hiveptest-1/apache-svn-spark-source/itests/qtest-spark/../../itests/qtest-spark/target/spark/bin/spark-submit --properties-file /home/hiveptest/50.18.64.184-hiveptest-1/apache-svn-spark-source/itests/qtest-spark/target/tmp/spark-submit.8721943354350626566.properties --class org.apache.hive.spark.client.RemoteDriver /home/hiveptest/50.18.64.184-hiveptest-1/maven/org/apache/hive/hive-exec/0.15.0-SNAPSHOT/hive-exec-0.15.0-SNAPSHOT.jar --remote akka.tcp://d84d934e-e10e-4744-b42c-ed86a49ebbd3@10.227.4.181:56697/user/SparkClient-aa5b3525-a031-41d7-ab04-8a18b0aa3fcf 2014-11-25 14:30:48,058 INFO client.SparkClientImpl (SparkClientImpl.java:onReceive(312)) - Received hello from akka.tcp://92e75da1-125e-4576-b63e-dc3166653dbe@10.227.4.181:35948/user/RemoteDriver 2014-11-25 14:30:48,059 DEBUG session.SparkSessionManagerImpl (SparkSessionManagerImpl.java:getSession(126)) - New session (1743e7e0-1fb3-4766-9249-cc138f88a2a7) is created. 2014-11-25 14:30:48,085 INFO ql.Context (Context.java:getMRScratchDir(266)) - New scratch dir is file:/home/hiveptest/50.18.64.184-hiveptest-1/apache-svn-spark-source/itests/qtest-spark/target/tmp/scratchdir/hiveptest/cf1f8ff8-a03d-413c-9922-d7c5c3c25e18/hive_2014-11-25_14-30-44_827_2636220880717371319-1 2014-11-25 14:30:48,283 INFO client.SparkClientImpl (SparkClientImpl.java:onReceive(329)) - Received result for fa96fc57-3999-4103-9bbd-a4bab346a324 2014-11-25 14:30:48,535 INFO client.SparkClientImpl (SparkClientImpl.java:onReceive(329)) - Received result for 00716575-c666-4b2a-bffe-682780684df8 2014-11-25 14:47:28,067 INFO transport.ProtocolStateActor (Slf4jLogger.scala:apply$mcV$sp(74)) - No response from remote. Handshake timed out or transport failure detector triggered. 2014-11-25 14:47:28,071 WARN remote.ReliableDeliverySupervisor (Slf4jLogger.scala:apply$mcV$sp(71)) - Association with remote system [akka.tcp://92e75da1-125e-4576-b63e-dc3166653dbe@10.227.4.181:35948] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. {noformat} > Enable automatic tests with remote spark client.[Spark Branch] > -------------------------------------------------------------- > > Key: HIVE-8836 > URL: https://issues.apache.org/jira/browse/HIVE-8836 > Project: Hive > Issue Type: Sub-task > Components: Spark > Affects Versions: spark-branch > Reporter: Chengxiang Li > Assignee: Rui Li > Labels: Spark-M3 > Fix For: spark-branch > > Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch, > HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch > > > In real production environment, remote spark client should be used to submit > spark job for Hive mostly, we should enable automatic test with remote spark > client to make sure the Hive feature workable with it. -- This message was sent by Atlassian JIRA (v6.3.4#6332)