You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Xiao Chen (JIRA)" <ji...@apache.org> on 2018/08/02 16:46:00 UTC
[jira] [Created] (HDFS-13788) Update EC documentation about rack
fault tolerance
Xiao Chen created HDFS-13788:
--------------------------------
Summary: Update EC documentation about rack fault tolerance
Key: HDFS-13788
URL: https://issues.apache.org/jira/browse/HDFS-13788
Project: Hadoop HDFS
Issue Type: Task
Components: documentation, erasure-coding
Affects Versions: 3.0.0
Reporter: Xiao Chen
Assignee: Kitti Nanasi
From http://hadoop.apache.org/docs/r3.0.0/hadoop-project-dist/hadoop-hdfs/HDFSErasureCoding.html:
{quote}
For rack fault-tolerance, it is also important to have at least as many racks as the configured EC stripe width. For EC policy RS (6,3), this means minimally 9 racks, and ideally 10 or 11 to handle planned and unplanned outages. For clusters with fewer racks than the stripe width, HDFS cannot maintain rack fault-tolerance, but will still attempt to spread a striped file across multiple nodes to preserve node-level fault-tolerance.
{quote}
Theoretical minimum is 3 racks, and ideally 9 or more, so the document should be updated.
(I didn't check timestamps, but this is probably due to {{BlockPlacementPolicyRackFaultTolerant}} isn't completely done when HDFS-9088 introduced this doc. Later there's also examples in {{TestErasureCodingMultipleRacks}} to test this explicitly.)
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-dev-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-help@hadoop.apache.org