You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Virajith Jalaparti (JIRA)" <ji...@apache.org> on 2018/01/30 19:52:00 UTC

[jira] [Created] (HDFS-13088) Allow HDFS files/blocks to be over-replicated.

Virajith Jalaparti created HDFS-13088:
-----------------------------------------

             Summary: Allow HDFS files/blocks to be over-replicated.
                 Key: HDFS-13088
                 URL: https://issues.apache.org/jira/browse/HDFS-13088
             Project: Hadoop HDFS
          Issue Type: Sub-task
            Reporter: Virajith Jalaparti


This JIRA is to add a per-file "over-replication" factor to HDFS. As mentioned in HDFS-13069, the over-replication factor will be the excess replicas that will be allowed to exist for a file or block. This is beneficial if the application deems additional replicas for a file are needed. In the case of  HDFS-13069, it would allow copies of data in PROVIDED storage to be cached locally in HDFS in a read-through manner.

The Namenode will not proactively meet the over-replication i.e., it does not schedule replications if the number of replicas for a block is less than (replication factor + over-replication factor) as long as they are more than the replication factor of the file.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-dev-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-help@hadoop.apache.org