You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-commits@hadoop.apache.org by sz...@apache.org on 2009/09/17 03:24:36 UTC
svn commit: r816022 [2/2] - in /hadoop/hdfs/branches/HDFS-265: ./
src/contrib/hdfsproxy/ src/docs/src/documentation/content/xdocs/
src/docs/src/documentation/resources/images/ src/java/
src/java/org/apache/hadoop/hdfs/ src/java/org/apache/hadoop/hdfs/p...
Modified: hadoop/hdfs/branches/HDFS-265/src/test/hdfs/org/apache/hadoop/hdfs/server/namenode/TestReplicationPolicy.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/branches/HDFS-265/src/test/hdfs/org/apache/hadoop/hdfs/server/namenode/TestReplicationPolicy.java?rev=816022&r1=816021&r2=816022&view=diff
==============================================================================
--- hadoop/hdfs/branches/HDFS-265/src/test/hdfs/org/apache/hadoop/hdfs/server/namenode/TestReplicationPolicy.java (original)
+++ hadoop/hdfs/branches/HDFS-265/src/test/hdfs/org/apache/hadoop/hdfs/server/namenode/TestReplicationPolicy.java Thu Sep 17 01:24:35 2009
@@ -38,7 +38,8 @@
private static final Configuration CONF = new Configuration();
private static final NetworkTopology cluster;
private static final NameNode namenode;
- private static final ReplicationTargetChooser replicator;
+ private static final BlockPlacementPolicy replicator;
+ private static final String filename = "/dummyfile.txt";
private static final DatanodeDescriptor dataNodes[] =
new DatanodeDescriptor[] {
new DatanodeDescriptor(new DatanodeID("h1:5020"), "/d1/r1"),
@@ -91,30 +92,30 @@
FSConstants.MIN_BLOCKS_FOR_WRITE*BLOCK_SIZE, 4); // overloaded
DatanodeDescriptor[] targets;
- targets = replicator.chooseTarget(
- 0, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 0, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 0);
- targets = replicator.chooseTarget(
- 1, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 1);
assertEquals(targets[0], dataNodes[0]);
- targets = replicator.chooseTarget(
- 2, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 2);
assertEquals(targets[0], dataNodes[0]);
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
- targets = replicator.chooseTarget(
- 3, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 3, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 3);
assertEquals(targets[0], dataNodes[0]);
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
assertTrue(cluster.isOnSameRack(targets[1], targets[2]));
- targets = replicator.chooseTarget(
- 4, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 4, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 4);
assertEquals(targets[0], dataNodes[0]);
assertTrue(cluster.isOnSameRack(targets[1], targets[2]) ||
@@ -137,41 +138,47 @@
public void testChooseTarget2() throws Exception {
HashMap<Node, Node> excludedNodes;
DatanodeDescriptor[] targets;
+ BlockPlacementPolicyDefault repl = (BlockPlacementPolicyDefault)replicator;
+ List<DatanodeDescriptor> chosenNodes = new ArrayList<DatanodeDescriptor>();
excludedNodes = new HashMap<Node, Node>();
excludedNodes.put(dataNodes[1], dataNodes[1]);
- targets = replicator.chooseTarget(
- 0, dataNodes[0], excludedNodes, BLOCK_SIZE);
+ targets = repl.chooseTarget(
+ 0, dataNodes[0], chosenNodes, excludedNodes, BLOCK_SIZE);
assertEquals(targets.length, 0);
excludedNodes.clear();
+ chosenNodes.clear();
excludedNodes.put(dataNodes[1], dataNodes[1]);
- targets = replicator.chooseTarget(
- 1, dataNodes[0], excludedNodes, BLOCK_SIZE);
+ targets = repl.chooseTarget(
+ 1, dataNodes[0], chosenNodes, excludedNodes, BLOCK_SIZE);
assertEquals(targets.length, 1);
assertEquals(targets[0], dataNodes[0]);
excludedNodes.clear();
+ chosenNodes.clear();
excludedNodes.put(dataNodes[1], dataNodes[1]);
- targets = replicator.chooseTarget(
- 2, dataNodes[0], excludedNodes, BLOCK_SIZE);
+ targets = repl.chooseTarget(
+ 2, dataNodes[0], chosenNodes, excludedNodes, BLOCK_SIZE);
assertEquals(targets.length, 2);
assertEquals(targets[0], dataNodes[0]);
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
excludedNodes.clear();
+ chosenNodes.clear();
excludedNodes.put(dataNodes[1], dataNodes[1]);
- targets = replicator.chooseTarget(
- 3, dataNodes[0], excludedNodes, BLOCK_SIZE);
+ targets = repl.chooseTarget(
+ 3, dataNodes[0], chosenNodes, excludedNodes, BLOCK_SIZE);
assertEquals(targets.length, 3);
assertEquals(targets[0], dataNodes[0]);
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
assertTrue(cluster.isOnSameRack(targets[1], targets[2]));
excludedNodes.clear();
+ chosenNodes.clear();
excludedNodes.put(dataNodes[1], dataNodes[1]);
- targets = replicator.chooseTarget(
- 4, dataNodes[0], excludedNodes, BLOCK_SIZE);
+ targets = repl.chooseTarget(
+ 4, dataNodes[0], chosenNodes, excludedNodes, BLOCK_SIZE);
assertEquals(targets.length, 4);
assertEquals(targets[0], dataNodes[0]);
for(int i=1; i<4; i++) {
@@ -197,30 +204,30 @@
(FSConstants.MIN_BLOCKS_FOR_WRITE-1)*BLOCK_SIZE, 0); // no space
DatanodeDescriptor[] targets;
- targets = replicator.chooseTarget(
- 0, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 0, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 0);
- targets = replicator.chooseTarget(
- 1, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 1);
assertEquals(targets[0], dataNodes[1]);
- targets = replicator.chooseTarget(
- 2, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 2);
assertEquals(targets[0], dataNodes[1]);
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
- targets = replicator.chooseTarget(
- 3, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 3, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 3);
assertEquals(targets[0], dataNodes[1]);
assertTrue(cluster.isOnSameRack(targets[1], targets[2]));
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
- targets = replicator.chooseTarget(
- 4, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 4, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 4);
assertEquals(targets[0], dataNodes[1]);
for(int i=1; i<4; i++) {
@@ -252,23 +259,23 @@
}
DatanodeDescriptor[] targets;
- targets = replicator.chooseTarget(
- 0, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 0, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 0);
- targets = replicator.chooseTarget(
- 1, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 1);
assertFalse(cluster.isOnSameRack(targets[0], dataNodes[0]));
- targets = replicator.chooseTarget(
- 2, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 2);
assertFalse(cluster.isOnSameRack(targets[0], dataNodes[0]));
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
- targets = replicator.chooseTarget(
- 3, dataNodes[0], null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 3, dataNodes[0], BLOCK_SIZE);
assertEquals(targets.length, 3);
for(int i=0; i<3; i++) {
assertFalse(cluster.isOnSameRack(targets[i], dataNodes[0]));
@@ -292,21 +299,21 @@
*/
public void testChooseTarget5() throws Exception {
DatanodeDescriptor[] targets;
- targets = replicator.chooseTarget(
- 0, NODE, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 0, NODE, BLOCK_SIZE);
assertEquals(targets.length, 0);
- targets = replicator.chooseTarget(
- 1, NODE, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, NODE, BLOCK_SIZE);
assertEquals(targets.length, 1);
- targets = replicator.chooseTarget(
- 2, NODE, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, NODE, BLOCK_SIZE);
assertEquals(targets.length, 2);
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
- targets = replicator.chooseTarget(
- 3, NODE, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 3, NODE, BLOCK_SIZE);
assertEquals(targets.length, 3);
assertTrue(cluster.isOnSameRack(targets[1], targets[2]));
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
@@ -324,23 +331,23 @@
chosenNodes.add(dataNodes[0]);
DatanodeDescriptor[] targets;
- targets = replicator.chooseTarget(
- 0, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 0, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 0);
- targets = replicator.chooseTarget(
- 1, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 1);
assertFalse(cluster.isOnSameRack(dataNodes[0], targets[0]));
- targets = replicator.chooseTarget(
- 2, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 2);
assertTrue(cluster.isOnSameRack(dataNodes[0], targets[0]));
assertFalse(cluster.isOnSameRack(targets[0], targets[1]));
- targets = replicator.chooseTarget(
- 3, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 3, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 3);
assertTrue(cluster.isOnSameRack(dataNodes[0], targets[0]));
assertFalse(cluster.isOnSameRack(targets[0], targets[2]));
@@ -359,17 +366,17 @@
chosenNodes.add(dataNodes[1]);
DatanodeDescriptor[] targets;
- targets = replicator.chooseTarget(
- 0, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 0, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 0);
- targets = replicator.chooseTarget(
- 1, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 1);
assertFalse(cluster.isOnSameRack(dataNodes[0], targets[0]));
- targets = replicator.chooseTarget(
- 2, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 2);
assertFalse(cluster.isOnSameRack(dataNodes[0], targets[0]));
assertFalse(cluster.isOnSameRack(dataNodes[0], targets[1]));
@@ -388,29 +395,29 @@
chosenNodes.add(dataNodes[2]);
DatanodeDescriptor[] targets;
- targets = replicator.chooseTarget(
- 0, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 0, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 0);
- targets = replicator.chooseTarget(
- 1, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 1);
assertTrue(cluster.isOnSameRack(dataNodes[0], targets[0]));
assertFalse(cluster.isOnSameRack(dataNodes[2], targets[0]));
- targets = replicator.chooseTarget(
- 1, dataNodes[2], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 1, dataNodes[2], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 1);
assertTrue(cluster.isOnSameRack(dataNodes[2], targets[0]));
assertFalse(cluster.isOnSameRack(dataNodes[0], targets[0]));
- targets = replicator.chooseTarget(
- 2, dataNodes[0], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, dataNodes[0], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 2);
assertTrue(cluster.isOnSameRack(dataNodes[0], targets[0]));
- targets = replicator.chooseTarget(
- 2, dataNodes[2], chosenNodes, null, BLOCK_SIZE);
+ targets = replicator.chooseTarget(filename,
+ 2, dataNodes[2], chosenNodes, BLOCK_SIZE);
assertEquals(targets.length, 2);
assertTrue(cluster.isOnSameRack(dataNodes[2], targets[0]));
}
Propchange: hadoop/hdfs/branches/HDFS-265/src/webapps/datanode/
------------------------------------------------------------------------------
--- svn:mergeinfo (original)
+++ svn:mergeinfo Thu Sep 17 01:24:35 2009
@@ -1,3 +1,3 @@
/hadoop/core/branches/branch-0.19/hdfs/src/webapps/datanode:713112
/hadoop/core/trunk/src/webapps/datanode:776175-784663
-/hadoop/hdfs/trunk/src/webapps/datanode:796829-800617,800619-803337,804756-805652,808672-809439,811495-813103,813105-813630
+/hadoop/hdfs/trunk/src/webapps/datanode:796829-800617,800619-803337,804756-805652,808672-809439,811495-813103,813105-813630,814223-815964
Propchange: hadoop/hdfs/branches/HDFS-265/src/webapps/hdfs/
------------------------------------------------------------------------------
--- svn:mergeinfo (original)
+++ svn:mergeinfo Thu Sep 17 01:24:35 2009
@@ -1,3 +1,3 @@
/hadoop/core/branches/branch-0.19/hdfs/src/webapps/hdfs:713112
/hadoop/core/trunk/src/webapps/hdfs:776175-784663
-/hadoop/hdfs/trunk/src/webapps/hdfs:796829-800617,800619-803337,804756-805652,808672-809439,811495-813103,813105-813630
+/hadoop/hdfs/trunk/src/webapps/hdfs:796829-800617,800619-803337,804756-805652,808672-809439,811495-813103,813105-813630,814223-815964
Propchange: hadoop/hdfs/branches/HDFS-265/src/webapps/secondary/
------------------------------------------------------------------------------
--- svn:mergeinfo (original)
+++ svn:mergeinfo Thu Sep 17 01:24:35 2009
@@ -1,3 +1,3 @@
/hadoop/core/branches/branch-0.19/hdfs/src/webapps/secondary:713112
/hadoop/core/trunk/src/webapps/secondary:776175-784663
-/hadoop/hdfs/trunk/src/webapps/secondary:796829-800617,800619-803337,804756-805652,808672-809439,811495-813103,813105-813630
+/hadoop/hdfs/trunk/src/webapps/secondary:796829-800617,800619-803337,804756-805652,808672-809439,811495-813103,813105-813630,814223-815964