You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Todd Lipcon (JIRA)" <ji...@apache.org> on 2010/03/22 02:45:27 UTC
[jira] Created: (HDFS-1059) completeFile loops forever if the
block's only replica has become corrupt
completeFile loops forever if the block's only replica has become corrupt
-------------------------------------------------------------------------
Key: HDFS-1059
URL: https://issues.apache.org/jira/browse/HDFS-1059
Project: Hadoop HDFS
Issue Type: Bug
Affects Versions: 0.21.0, 0.22.0
Reporter: Todd Lipcon
If a writer is appending to a block with replication factor 1, and that block has become corrupt, a reader will report the corruption to the NN. Then when the writer tries to complete the file, it will loop forever with an error like:
[junit] 2010-03-21 17:40:08,093 INFO namenode.FSNamesystem (FSNamesystem.java:checkFileProgress(1613)) - BLOCK* NameSystem.checkFileProgress: block blk_-4256412191814117589_1001{blockUCState=COMMITTED, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[127.0.0.1:56782|RBW]]} has not reached minimal replication 1
[junit] 2010-03-21 17:40:08,495 INFO hdfs.DFSClient (DFSOutputStream.java:completeFile(1435)) - Could not complete file /TestReadWhileWriting/file1 retrying...
Should add tests that cover the case of a writer appending to a block that is corrupt while a reader accesses it.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.