You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-issues@hadoop.apache.org by "Akira AJISAKA (JIRA)" <ji...@apache.org> on 2013/07/15 20:45:19 UTC

[jira] [Commented] (MAPREDUCE-1729) Distributed cache should provide an option to fail the job or not, if cache file gets modified on the fly.

    [ https://issues.apache.org/jira/browse/MAPREDUCE-1729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13708801#comment-13708801 ] 

Akira AJISAKA commented on MAPREDUCE-1729:
------------------------------------------

Hi, I need this option.

A long-term job failed in our production environment because the file which was used as distributed cache was modified at fixed intervals. As the output of the job would be better if distributed cache is newer, we don't want to fail the job if cache file gets modified on the fly. Our workaround is, copy original file to tmpfile and use tmpfile as distributed cache. If the option exists, we don't need to copy original file before the job begin.
                
> Distributed cache should provide an option to fail the job or not, if cache file gets modified on the fly.
> ----------------------------------------------------------------------------------------------------------
>
>                 Key: MAPREDUCE-1729
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-1729
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>          Components: distributed-cache
>            Reporter: Amareshwari Sriramadasu
>
> Currently, distributed cache fails the job if the cache file gets modified on the fly. But there should be an option to fail a job or not.
> See discussions in MAPREDUCE-1288.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira