You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Jimmy Xiang (JIRA)" <ji...@apache.org> on 2014/11/06 03:27:34 UTC

[jira] [Commented] (HIVE-8649) Increase level of parallelism in reduce phase [Spark Branch]

    [ https://issues.apache.org/jira/browse/HIVE-8649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14199651#comment-14199651 ] 

Jimmy Xiang commented on HIVE-8649:
-----------------------------------

This patch assumes information doesn't change. If we do support changes like increasing # of executors, we need a listener for such events.

> Increase level of parallelism in reduce phase [Spark Branch]
> ------------------------------------------------------------
>
>                 Key: HIVE-8649
>                 URL: https://issues.apache.org/jira/browse/HIVE-8649
>             Project: Hive
>          Issue Type: Sub-task
>          Components: Spark
>            Reporter: Brock Noland
>            Assignee: Jimmy Xiang
>             Fix For: spark-branch
>
>         Attachments: HIVE-8649.1-spark.patch
>
>
> We calculate the number of reducers based on the same code for MapReduce. However, reducers are vastly cheaper in Spark and it's generally recommended we have many more reducers than in MR.
> Sandy Ryza who works on Spark has some ideas about a heuristic.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)