You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Virajith Jalaparti (JIRA)" <ji...@apache.org> on 2018/01/30 19:52:00 UTC
[jira] [Created] (HDFS-13088) Allow HDFS files/blocks to be
over-replicated.
Virajith Jalaparti created HDFS-13088:
-----------------------------------------
Summary: Allow HDFS files/blocks to be over-replicated.
Key: HDFS-13088
URL: https://issues.apache.org/jira/browse/HDFS-13088
Project: Hadoop HDFS
Issue Type: Sub-task
Reporter: Virajith Jalaparti
This JIRA is to add a per-file "over-replication" factor to HDFS. As mentioned in HDFS-13069, the over-replication factor will be the excess replicas that will be allowed to exist for a file or block. This is beneficial if the application deems additional replicas for a file are needed. In the case of HDFS-13069, it would allow copies of data in PROVIDED storage to be cached locally in HDFS in a read-through manner.
The Namenode will not proactively meet the over-replication i.e., it does not schedule replications if the number of replicas for a block is less than (replication factor + over-replication factor) as long as they are more than the replication factor of the file.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-dev-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-help@hadoop.apache.org