You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Andrew Ryan (JIRA)" <ji...@apache.org> on 2010/07/07 00:13:49 UTC

[jira] Created: (HDFS-1282) namenode should reject datanodes which send impossible block reports

namenode should reject datanodes which send impossible block reports
--------------------------------------------------------------------

                 Key: HDFS-1282
                 URL: https://issues.apache.org/jira/browse/HDFS-1282
             Project: Hadoop HDFS
          Issue Type: Bug
          Components: data-node, name-node
    Affects Versions: 0.20.1
            Reporter: Andrew Ryan


Over the past few weeks we've had several datanodes with bad disks that suffer ext3 corruption, and consequently start reporting impossible values for how full they are. This particular node, for example, has a configured capacity of 10.86TB but reports 1733.95TB used, for a total of 15973.57% utilization.

Node 	 Last Contact 	Admin State 	Configured Capacity (TB) 	Used (TB) 	Non DFS  Used (TB) 	Remaining  (TB) 	Used  (%) 	Used  (%) 	Remaining (%) 	Blocks 
hadoop2254	 44	In Service 	10.86	1733.95	0	5.24	15973.57   48.25	65602 

If we can avoid generating such bogus data on the datanode that would be great.  But if the namenode receives such an impossible block report, it should definitely consider that datanode to be not trustworthy, and in my opinion, make it dead.

The "fix" in our case was either to fsck or replace the bad disk.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.