You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Sameer Paranjpye (JIRA)" <ji...@apache.org> on 2006/03/24 22:33:28 UTC

[jira] Updated: (HADOOP-35) Files missing chunks can cause mapred runs to get stuck

     [ http://issues.apache.org/jira/browse/HADOOP-35?page=all ]

Sameer Paranjpye updated HADOOP-35:
-----------------------------------

    Fix Version: 0.1
        Version: 0.1

> Files missing chunks can cause mapred runs to get stuck
> -------------------------------------------------------
>
>          Key: HADOOP-35
>          URL: http://issues.apache.org/jira/browse/HADOOP-35
>      Project: Hadoop
>         Type: Bug
>   Components: dfs
>     Versions: 0.1
>  Environment: ~20 datanode DFS cluster
>     Reporter: Bryan Pendleton
>      Fix For: 0.1
>  Attachments: dfsshell.health.patch.txt
>
> I've now several times run into a problem where a large run gets stalled as a result of a missing data block. The latest was a stall in the Summer - ie, the data might've all been there, but it was impossible to proceed because the CRC file was missing a block. It would be nice to:
> 1) Have a "health check" running on a map reduce. If any data isn't available, emmit periodic warnings, and maybe have a timeout for if the data never comes back. Such warnings *should* specify which file(s) are affected by the missing blocks.
> 2) Have a utility, possible part of the existing dfs utility, which can check for dfs files with unlocatable blocks. Possibly, even show a 'health' of a file - ie, what percentage of its blocks are currently at the desired replication level. Currently, there's no way that I know of to find out if a file in DFS is going to be unreadable.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
   http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see:
   http://www.atlassian.com/software/jira