You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Arun C Murthy (JIRA)" <ji...@apache.org> on 2007/07/13 11:33:05 UTC

[jira] Updated: (HADOOP-1084) updating a hdfs file, doesn't cause the distributed file cache to update itself

     [ https://issues.apache.org/jira/browse/HADOOP-1084?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Arun C Murthy updated HADOOP-1084:
----------------------------------

    Fix Version/s: 0.14.0
         Priority: Blocker  (was: Major)

With HADOOP-1134 slated for 0.14, we need to fix this since crcs are going away... hence I'm marking this as a BLOCKER.

> updating a hdfs file, doesn't cause the distributed file cache to update itself
> -------------------------------------------------------------------------------
>
>                 Key: HADOOP-1084
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1084
>             Project: Hadoop
>          Issue Type: Bug
>          Components: mapred
>    Affects Versions: 0.12.0
>            Reporter: Owen O'Malley
>            Assignee: Arun C Murthy
>            Priority: Blocker
>             Fix For: 0.14.0
>
>
> If I delete and upload a new version of a file /user/owen/foo to HDFS and start my job with hdfs://user/owen/foo as a cached file, it will use the previous contents.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.