[ https://issues.apache.org/jira/browse/HIVE-10458?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14539299#comment-14539299 ]
Rui Li commented on HIVE-10458: ------------------------------- Hi [~xuefuz], I've looked at some of the failures. Most of them is due to how we handle {{order by + limit}} queries. When order by is followed by a limit, hive assumes that there's only 1 reducer. So we won't create an extra shuffle to do the limit (otherwise there'll be an extra shuffle with only 1 reducer for the limit clause). To solve this, we can either disable parallel order by when there's a limit, or we can force an extra shuffle for the limit when running on spark. What's your opinion? > Enable parallel order by for spark [Spark Branch] > ------------------------------------------------- > > Key: HIVE-10458 > URL: https://issues.apache.org/jira/browse/HIVE-10458 > Project: Hive > Issue Type: Sub-task > Components: Spark > Reporter: Rui Li > Assignee: Rui Li > Attachments: HIVE-10458.1-spark.patch > > > We don't have to force reducer# to 1 as spark supports parallel sorting. -- This message was sent by Atlassian JIRA (v6.3.4#6332)