You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Todd Lipcon (JIRA)" <ji...@apache.org> on 2009/12/18 03:00:18 UTC
[jira] Resolved: (HDFS-795) DFS Write pipeline does not detect
defective datanode correctly in some cases (HADOOP-3339)
[ https://issues.apache.org/jira/browse/HDFS-795?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Todd Lipcon resolved HDFS-795.
------------------------------
Resolution: Duplicate
HDFS-101 duplicates this, and fix is under way there.
> DFS Write pipeline does not detect defective datanode correctly in some cases (HADOOP-3339)
> -------------------------------------------------------------------------------------------
>
> Key: HDFS-795
> URL: https://issues.apache.org/jira/browse/HDFS-795
> Project: Hadoop HDFS
> Issue Type: Bug
> Components: hdfs client
> Affects Versions: 0.20.1
> Reporter: Raghu Angadi
> Priority: Critical
> Fix For: 0.20.2
>
> Attachments: toreproduce-5796.patch
>
>
> HDFS write pipeline does not select the correct datanode in some error cases. One example : say DN2 is the second datanode and write to it times out since it is in a bad state.. pipeline actually removes the first datanode. If such a datanode happens to be the last one in the pipeline, write is aborted completely with a hard error.
> Essentially the error occurs when writing to a downstream datanode fails rather than reading. This bug was actually fixed in 0.18 (HADOOP-3339). But HADOOP-1700 essentially reverted it. I am not sure why.
> It is absolutely essential for HDFS to handle failures on subset of datanodes in a pipeline. We should not have at least known bugs that lead to hard failures.
> I will attach patch for a hack that illustrates this problem. Still thinking of how an automated test would look like for this one.
> My preferred target for this fix is 0.20.1.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.