JoneZhang created HIVE-12649: -------------------------------- Summary: Hive on Spark will resubmitted application when not enough resouces to launch yarn application master Key: HIVE-12649 URL: https://issues.apache.org/jira/browse/HIVE-12649 Project: Hive Issue Type: Bug Affects Versions: 1.2.1, 1.1.1 Reporter: JoneZhang Assignee: Xuefu Zhang
Hive on spark will estimate reducer number when the query is not set reduce number,which cause a application submit.The application will pending if the yarn queue's resources is insufficient. So there are more than one pending applications probably because there are more than one estimate call.The failure is soft, so it doesn't prevent subsequent processings. We can make that a hard failure That code is found in 728237 at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:112) 728238 at org.apache.hadoop.hive.ql.optimizer.spark.SetSparkReducerParallelism.process(SetSparkReducerParallelism.java:115) -- This message was sent by Atlassian JIRA (v6.3.4#6332)