You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-issues@hadoop.apache.org by "Maysam Yabandeh (JIRA)" <ji...@apache.org> on 2015/09/23 17:46:04 UTC

[jira] [Created] (MAPREDUCE-6489) Fail fast rogue tasks that write too much to local disk

Maysam Yabandeh created MAPREDUCE-6489:
------------------------------------------

             Summary: Fail fast rogue tasks that write too much to local disk
                 Key: MAPREDUCE-6489
                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6489
             Project: Hadoop Map/Reduce
          Issue Type: Improvement
          Components: task
    Affects Versions: 2.7.1
            Reporter: Maysam Yabandeh


Tasks of the rogue jobs can write too much to local disk, negatively affecting the jobs running in collocated containers. Ideally YARN will be able to limit amount of local disk used by each task: YARN-4011. Until then, the mapreduce task can fail fast if the task is writing too much (above a configured threshold) to local disk.

As we discussed [here|https://issues.apache.org/jira/browse/YARN-4011?focusedCommentId=14902750&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14902750] the suggested approach is that the MapReduce task checks for BYTES_WRITTEN counter for the local disk and throws an exception when it goes beyond a configured value.  It is true that written bytes is larger than the actual used disk space, but to detect a rogue task the exact value is not required and a very large value for written bytes to local disk is a good indicative that the task is misbehaving.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)