You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by ra...@apache.org on 2008/10/01 23:08:06 UTC
svn commit: r700923 - in /hadoop/core/trunk: CHANGES.txt
src/test/org/apache/hadoop/hdfs/TestReplication.java
src/test/org/apache/hadoop/hdfs/server/datanode/SimulatedFSDataset.java
Author: rangadi
Date: Wed Oct 1 14:08:05 2008
New Revision: 700923
URL: http://svn.apache.org/viewvc?rev=700923&view=rev
Log:
HADOOP-4314. Simulated datanodes should not include blocks that are still
being written in their block report. (Raghu Angadi)
Modified:
hadoop/core/trunk/CHANGES.txt
hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestReplication.java
hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/server/datanode/SimulatedFSDataset.java
Modified: hadoop/core/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/CHANGES.txt?rev=700923&r1=700922&r2=700923&view=diff
==============================================================================
--- hadoop/core/trunk/CHANGES.txt (original)
+++ hadoop/core/trunk/CHANGES.txt Wed Oct 1 14:08:05 2008
@@ -831,6 +831,9 @@
HADOOP-3614. Fix a bug that Datanode may use an old GenerationStamp to get
meta file. (szetszwo)
+ HADOOP-4314. Simulated datanodes should not include blocks that are still
+ being written in their block report. (Raghu Angadi)
+
Release 0.18.1 - 2008-09-17
IMPROVEMENTS
Modified: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestReplication.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestReplication.java?rev=700923&r1=700922&r2=700923&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestReplication.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestReplication.java Wed Oct 1 14:08:05 2008
@@ -35,8 +35,6 @@
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.ipc.RPC;
-import org.apache.hadoop.net.NetUtils;
/**
* This class tests the replication of a DFS file.
@@ -71,27 +69,11 @@
Configuration conf = fileSys.getConf();
ClientProtocol namenode = DFSClient.createNamenode(conf);
- LocatedBlocks locations;
- boolean isReplicationDone;
- do {
- locations = namenode.getBlockLocations(name.toString(),0,Long.MAX_VALUE);
- isReplicationDone = true;
- for (LocatedBlock blk : locations.getLocatedBlocks()) {
- DatanodeInfo[] datanodes = blk.getLocations();
- if (Math.min(numDatanodes, repl) != datanodes.length) {
- isReplicationDone=false;
- LOG.warn("File has "+datanodes.length+" replicas, expecting "
- +Math.min(numDatanodes, repl));
- try {
- Thread.sleep(15000L);
- } catch (InterruptedException e) {
- // nothing
- }
- break;
- }
- }
- } while(!isReplicationDone);
-
+ waitForBlockReplication(name.toString(), namenode,
+ Math.min(numDatanodes, repl), -1);
+
+ LocatedBlocks locations = namenode.getBlockLocations(name.toString(),0,
+ Long.MAX_VALUE);
boolean isOnSameRack = true, isNotOnSameRack = true;
for (LocatedBlock blk : locations.getLocatedBlocks()) {
DatanodeInfo[] datanodes = blk.getLocations();
Modified: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/server/datanode/SimulatedFSDataset.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/server/datanode/SimulatedFSDataset.java?rev=700923&r1=700922&r2=700923&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/server/datanode/SimulatedFSDataset.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/server/datanode/SimulatedFSDataset.java Wed Oct 1 14:08:05 2008
@@ -20,6 +20,7 @@
import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStream;
+import java.util.Arrays;
import java.util.HashMap;
import java.util.Random;
@@ -279,9 +280,14 @@
public synchronized Block[] getBlockReport() {
Block[] blockTable = new Block[blockMap.size()];
- int i = 0;
- for (Block b: blockMap.keySet()) {
- blockTable[i++] = blockMap.get(b).theBlock;
+ int count = 0;
+ for (BInfo b : blockMap.values()) {
+ if (b.isFinalized()) {
+ blockTable[count++] = b.theBlock;
+ }
+ }
+ if (count != blockTable.length) {
+ blockTable = Arrays.copyOf(blockTable, count);
}
return blockTable;
}