You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by um...@apache.org on 2016/10/17 23:14:36 UTC
[17/50] hadoop git commit: HDFS-10949. DiskBalancer: deprecate
TestDiskBalancer#setVolumeCapacity. Contributed by Xiaobing Zhou.
HDFS-10949. DiskBalancer: deprecate TestDiskBalancer#setVolumeCapacity. Contributed by Xiaobing Zhou.
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/b371c563
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/b371c563
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/b371c563
Branch: refs/heads/HDFS-10285
Commit: b371c56365c14bbab0f5cdfffc0becaabfde8145
Parents: 1291254
Author: Anu Engineer <ae...@apache.org>
Authored: Thu Oct 13 10:26:07 2016 -0700
Committer: Anu Engineer <ae...@apache.org>
Committed: Thu Oct 13 10:26:07 2016 -0700
----------------------------------------------------------------------
.../server/diskbalancer/TestDiskBalancer.java | 44 +++++---------------
1 file changed, 11 insertions(+), 33 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/b371c563/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
index d911e74..9985210 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/TestDiskBalancer.java
@@ -44,7 +44,6 @@ import org.apache.hadoop.hdfs.server.diskbalancer.connectors.ClusterConnector;
import org.apache.hadoop.hdfs.server.diskbalancer.connectors.ConnectorFactory;
import org.apache.hadoop.hdfs.server.diskbalancer.datamodel.DiskBalancerCluster;
import org.apache.hadoop.hdfs.server.diskbalancer.datamodel.DiskBalancerDataNode;
-import org.apache.hadoop.hdfs.server.diskbalancer.datamodel.DiskBalancerVolume;
import org.apache.hadoop.hdfs.server.diskbalancer.planner.NodePlan;
import org.apache.hadoop.test.GenericTestUtils;
import org.apache.hadoop.util.Time;
@@ -137,6 +136,7 @@ public class TestDiskBalancer {
final int dataNodeCount = 1;
final int dataNodeIndex = 0;
final int sourceDiskIndex = 0;
+ final long cap = blockSize * 2L * blockCount;
MiniDFSCluster cluster = new ClusterBuilder()
.setBlockCount(blockCount)
@@ -144,6 +144,7 @@ public class TestDiskBalancer {
.setDiskCount(diskCount)
.setNumDatanodes(dataNodeCount)
.setConf(conf)
+ .setCapacities(new long[] {cap, cap})
.build();
try {
DataMover dataMover = new DataMover(cluster, dataNodeIndex,
@@ -174,7 +175,7 @@ public class TestDiskBalancer {
final int dataNodeCount = 1;
final int dataNodeIndex = 0;
final int sourceDiskIndex = 0;
-
+ final long cap = blockSize * 2L * blockCount;
MiniDFSCluster cluster = new ClusterBuilder()
.setBlockCount(blockCount)
@@ -182,9 +183,9 @@ public class TestDiskBalancer {
.setDiskCount(diskCount)
.setNumDatanodes(dataNodeCount)
.setConf(conf)
+ .setCapacities(new long[] {cap, cap, cap})
.build();
-
try {
DataMover dataMover = new DataMover(cluster, dataNodeIndex,
sourceDiskIndex, conf, blockSize, blockCount);
@@ -221,6 +222,7 @@ public class TestDiskBalancer {
final int dataNodeCount = 1;
final int dataNodeIndex = 0;
final int sourceDiskIndex = 0;
+ final long cap = blockSize * 2L * blockCount;
MiniDFSCluster cluster = new ClusterBuilder()
.setBlockCount(blockCount)
@@ -228,6 +230,7 @@ public class TestDiskBalancer {
.setDiskCount(diskCount)
.setNumDatanodes(dataNodeCount)
.setConf(conf)
+ .setCapacities(new long[] {cap, cap})
.build();
try {
@@ -246,24 +249,6 @@ public class TestDiskBalancer {
}
/**
- * Sets alll Disks capacity to size specified.
- *
- * @param cluster - DiskBalancerCluster
- * @param size - new size of the disk
- */
- private void setVolumeCapacity(DiskBalancerCluster cluster, long size,
- String diskType) {
- Preconditions.checkNotNull(cluster);
- for (DiskBalancerDataNode node : cluster.getNodes()) {
- for (DiskBalancerVolume vol :
- node.getVolumeSets().get(diskType).getVolumes()) {
- vol.setCapacity(size);
- }
- node.getVolumeSets().get(diskType).computeVolumeDataDensity();
- }
- }
-
- /**
* Helper class that allows us to create different kinds of MiniDFSClusters
* and populate data.
*/
@@ -274,6 +259,7 @@ public class TestDiskBalancer {
private int fileLen;
private int blockCount;
private int diskCount;
+ private long[] capacities;
public ClusterBuilder setConf(Configuration conf) {
this.conf = conf;
@@ -300,13 +286,9 @@ public class TestDiskBalancer {
return this;
}
- private long[] getCapacities(int diskCount, int bSize, int fSize) {
- Preconditions.checkState(diskCount > 0);
- long[] capacities = new long[diskCount];
- for (int x = 0; x < diskCount; x++) {
- capacities[x] = diskCount * bSize * fSize * 2L;
- }
- return capacities;
+ private ClusterBuilder setCapacities(final long[] caps) {
+ this.capacities = caps;
+ return this;
}
private StorageType[] getStorageTypes(int diskCount) {
@@ -338,7 +320,7 @@ public class TestDiskBalancer {
// Write a file and restart the cluster
MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
.numDataNodes(numDatanodes)
- .storageCapacities(getCapacities(diskCount, blockSize, fileLen))
+ .storageCapacities(capacities)
.storageTypes(getStorageTypes(diskCount))
.storagesPerDatanode(diskCount)
.build();
@@ -447,10 +429,6 @@ public class TestDiskBalancer {
diskBalancerCluster.readClusterInfo();
List<DiskBalancerDataNode> nodesToProcess = new LinkedList<>();
- // Rewrite the capacity in the model to show that disks need
- // re-balancing.
- setVolumeCapacity(diskBalancerCluster, blockSize * 2L * blockCount,
- "DISK");
// Pick a node to process.
nodesToProcess.add(diskBalancerCluster.getNodeByUUID(
node.getDatanodeUuid()));
---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-commits-help@hadoop.apache.org