[ 
https://issues.apache.org/jira/browse/HIVE-7567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chengxiang Li updated HIVE-7567:
--------------------------------

    Attachment: HIVE-7567.1-spark.patch

support automatic adjusting reducer number same as MR, configure through 3 
following parameters:
# In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
# In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
# In order to set a constant number of reducers:
  set mapreduce.job.reduces=<number>

Currently Tez specified optimization(committed with 
[HIVE-7158|https://issues.apache.org/jira/browse/HIVE-7158]) is not supported 
as it's based on Tez feature, refer to [dynamic graph 
reconfiguration|http://hortonworks.com/blog/apache-tez-dynamic-graph-reconfiguration/]
 for details.

> support automatic calculating reduce task number
> ------------------------------------------------
>
>                 Key: HIVE-7567
>                 URL: https://issues.apache.org/jira/browse/HIVE-7567
>             Project: Hive
>          Issue Type: Task
>          Components: Spark
>            Reporter: Chengxiang Li
>            Assignee: Chengxiang Li
>         Attachments: HIVE-7567.1-spark.patch
>
>
> Hive have its own machenism to calculate reduce task number, we need to 
> implement it on spark job.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to