You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Kihwal Lee (JIRA)" <ji...@apache.org> on 2016/03/17 22:48:33 UTC

[jira] [Created] (HDFS-10178) Permanent write failures can happen if pipeline recoveries occur for the first packet

Kihwal Lee created HDFS-10178:
---------------------------------

             Summary: Permanent write failures can happen if pipeline recoveries occur for the first packet
                 Key: HDFS-10178
                 URL: https://issues.apache.org/jira/browse/HDFS-10178
             Project: Hadoop HDFS
          Issue Type: Bug
            Reporter: Kihwal Lee
            Priority: Critical


We have observed that write fails permanently if the first packet doesn't go through properly and pipeline recovery happens. If the packet header is sent out, but the data portion of the packet does not reach one or more datanodes in time, the pipeline recovery will be done against the 0-byte partial block.  

If additional datanodes are added, the block is transferred to the new nodes.  After the transfer, each node will have a meta file containing the header and 0-length data block file. The pipeline recovery seems to work correctly up to this point, but write fails when actual data packet is resent. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)