[ https://issues.apache.org/jira/browse/HIVE-16799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16036167#comment-16036167 ]
Lefty Leverenz commented on HIVE-16799: --------------------------------------- Doc note: This adds *hive.spark.stage.max.tasks* to HiveConf.java, so it needs to be documented in the Spark section of Configuration Properties. * [ConfigurationProperties -- Spark | https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-Spark] * [hive.spark.stage.max.tasks | https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties#ConfigurationProperties-hive.spark.stage.max.tasks] (This link won't work until the documentation is done.) Thanks for the TODOC3.0 label, Xuefu. > Control the max number of task for a stage in a spark job > --------------------------------------------------------- > > Key: HIVE-16799 > URL: https://issues.apache.org/jira/browse/HIVE-16799 > Project: Hive > Issue Type: Improvement > Reporter: Xuefu Zhang > Assignee: Xuefu Zhang > Labels: TODOC3.0 > Fix For: 3.0.0 > > Attachments: HIVE-16799.1.patch, HIVE-16799.patch > > > HIVE-16552 gives admin an option to control the maximum number of tasks a > Spark job may have. However, this may not be sufficient as this tends to > penalize jobs that have many stages while favoring jobs that has fewer > stages. Ideally, we should also limit the number of tasks in a stage, which > is closer to the maximum number of mappers or reducers in a MR job. -- This message was sent by Atlassian JIRA (v6.3.15#6346)