You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Chris Douglas (JIRA)" <ji...@apache.org> on 2009/04/10 05:53:12 UTC

[jira] Updated: (HADOOP-5652) Reduce does not respect in-memory segment memory limit when number of on disk segments == io.sort.factor

     [ https://issues.apache.org/jira/browse/HADOOP-5652?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Chris Douglas updated HADOOP-5652:
----------------------------------

    Attachment: 5652-0.patch

Easy fix

> Reduce does not respect in-memory segment memory limit when number of on disk segments == io.sort.factor
> --------------------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-5652
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5652
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: mapred
>    Affects Versions: 0.20.0
>            Reporter: Chris Douglas
>            Priority: Minor
>         Attachments: 5652-0.patch
>
>
> If the number of on-disk segments is exactly {{io.sort.factor}}, then map output segments may be left in memory for the reduce contrary to the specification in {{mapred.job.reduce.input.buffer.percent}}.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.