You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-issues@hadoop.apache.org by "zhihai xu (JIRA)" <ji...@apache.org> on 2016/05/16 08:05:12 UTC

[jira] [Commented] (MAPREDUCE-6696) Add a configuration to limit the number of map tasks allowed per job.

    [ https://issues.apache.org/jira/browse/MAPREDUCE-6696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15284250#comment-15284250 ] 

zhihai xu commented on MAPREDUCE-6696:
--------------------------------------

I attached a patch MAPREDUCE-6696.000.patch, which added a configuration "mapreduce.job.max.map" to limit the number of map tasks. The default value is -1, which means no limit.

> Add a configuration to limit the number of map tasks allowed per job.
> ---------------------------------------------------------------------
>
>                 Key: MAPREDUCE-6696
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6696
>             Project: Hadoop Map/Reduce
>          Issue Type: Improvement
>          Components: job submission
>    Affects Versions: 2.8.0
>            Reporter: zhihai xu
>         Attachments: MAPREDUCE-6696.000.patch
>
>
> Add a configuration "mapreduce.job.max.map" to limit the number of map tasks allowed per job. It will be useful for Hadoop admin to save Hadoop cluster resource by preventing users from submitting big mapreduce jobs. A mapredeuce job with too many mappers may fail with OOM after running for long time. It will be a big waste.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: mapreduce-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: mapreduce-issues-help@hadoop.apache.org