You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Hairong Kuang (JIRA)" <ji...@apache.org> on 2009/10/12 19:14:31 UTC

[jira] Created: (HDFS-694) Add a test to make sure that node decomission does not get blocked by underreplicated blocks in an unclosed file

Add a test to make sure that node decomission does not get blocked by underreplicated blocks in an unclosed file
----------------------------------------------------------------------------------------------------------------

                 Key: HDFS-694
                 URL: https://issues.apache.org/jira/browse/HDFS-694
             Project: Hadoop HDFS
          Issue Type: Test
    Affects Versions: 0.21.0
            Reporter: Hairong Kuang
             Fix For: 0.21.0


We have a cluster that took  much longer time to decommission datanodes than usual. It turned out to be caused by an open file. In HDFS 0.20 when a file is open, NN does not schedule any block belonged to this file to replicate. So if an open file has an under-replicated block with a replica on a decommission datanode, the decommission won't complete until the file is closed.

The new append implementation changed the replication strategy that any complete block in an unclosed file will be scheduled to be replicated if it becomes under-replicated. This jira aims to add a test to make sure that node decommission does not get blocked by under-replicated blocks in an open file.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.