You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Ruyue Ma (JIRA)" <ji...@apache.org> on 2009/04/11 09:43:15 UTC

[jira] Commented: (HADOOP-5600) mapred.jobtracker.retirejob.interval killing long running reduce task

    [ https://issues.apache.org/jira/browse/HADOOP-5600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12698061#action_12698061 ] 

Ruyue Ma commented on HADOOP-5600:
----------------------------------

This is related to mapred.userlog.retain.hours. 

Current, every task jvm tries to clean up user logs in hadoop/logs/userlogs dir. The standard is 
return file.lastModified() < purgeTimeStamp. This 'file' is the attempt dir. But the dir lastModified time doesn't change. so the change is 
+      File indexFile = new File(file, "log.index");
+      if (indexFile.exists()){
+    	  return indexFile.lastModified() < purgeTimeStamp;
+      }else{
+    	  return file.lastModified() < purgeTimeStamp;
+      }  

> mapred.jobtracker.retirejob.interval killing long running reduce task
> ---------------------------------------------------------------------
>
>                 Key: HADOOP-5600
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5600
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: mapred
>    Affects Versions: 0.19.2
>         Environment: 0.19.2-dev, r753365 
>            Reporter: Billy Pearson
>         Attachments: hadoop-5600.patch
>
>
> Can verify by changing the mapred.jobtracker.retirejob.interval to < then your normal map time and watch the reduce task fail
> more info on closed ticket HADOOP-5591

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.