You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Ruyue Ma (JIRA)" <ji...@apache.org> on 2009/04/11 09:43:15 UTC
[jira] Commented: (HADOOP-5600)
mapred.jobtracker.retirejob.interval killing long running reduce task
[ https://issues.apache.org/jira/browse/HADOOP-5600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12698061#action_12698061 ]
Ruyue Ma commented on HADOOP-5600:
----------------------------------
This is related to mapred.userlog.retain.hours.
Current, every task jvm tries to clean up user logs in hadoop/logs/userlogs dir. The standard is
return file.lastModified() < purgeTimeStamp. This 'file' is the attempt dir. But the dir lastModified time doesn't change. so the change is
+ File indexFile = new File(file, "log.index");
+ if (indexFile.exists()){
+ return indexFile.lastModified() < purgeTimeStamp;
+ }else{
+ return file.lastModified() < purgeTimeStamp;
+ }
> mapred.jobtracker.retirejob.interval killing long running reduce task
> ---------------------------------------------------------------------
>
> Key: HADOOP-5600
> URL: https://issues.apache.org/jira/browse/HADOOP-5600
> Project: Hadoop Core
> Issue Type: Bug
> Components: mapred
> Affects Versions: 0.19.2
> Environment: 0.19.2-dev, r753365
> Reporter: Billy Pearson
> Attachments: hadoop-5600.patch
>
>
> Can verify by changing the mapred.jobtracker.retirejob.interval to < then your normal map time and watch the reduce task fail
> more info on closed ticket HADOOP-5591
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.