You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by ar...@apache.org on 2015/03/18 20:35:58 UTC

[3/3] hadoop git commit: HDFS-7950. Fix TestFsDatasetImpl#testAddVolumes failure on Windows. (Contributed by Xiaoyu Yao)

HDFS-7950. Fix TestFsDatasetImpl#testAddVolumes failure on Windows. (Contributed by Xiaoyu Yao)


Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/ae6d273d
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/ae6d273d
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/ae6d273d

Branch: refs/heads/branch-2.7
Commit: ae6d273d63fdd8234bd420f63009496b57fd9dc5
Parents: 502c040
Author: Arpit Agarwal <ar...@apache.org>
Authored: Wed Mar 18 12:33:59 2015 -0700
Committer: Arpit Agarwal <ar...@apache.org>
Committed: Wed Mar 18 12:35:41 2015 -0700

----------------------------------------------------------------------
 hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt              |  3 +++
 .../datanode/fsdataset/impl/TestFsDatasetImpl.java       | 11 +++++++----
 2 files changed, 10 insertions(+), 4 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hadoop/blob/ae6d273d/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
index 0ee0ab8..0234241 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
+++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
@@ -858,6 +858,9 @@ Release 2.7.0 - UNRELEASED
     HDFS-7948. TestDataNodeHotSwapVolumes#testAddVolumeFailures failed on
     Windows. (Xiaoyu Yao via Arpit Agarwal)
 
+    HDFS-7950. Fix TestFsDatasetImpl#testAddVolumes failure on Windows.
+    (Xiaoyu Yao via Arpit Agarwal)
+
     BREAKDOWN OF HDFS-7584 SUBTASKS AND RELATED JIRAS
 
       HDFS-7720. Quota by Storage Type API, tools and ClientNameNode

http://git-wip-us.apache.org/repos/asf/hadoop/blob/ae6d273d/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestFsDatasetImpl.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestFsDatasetImpl.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestFsDatasetImpl.java
index 8654773..73fcd19 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestFsDatasetImpl.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestFsDatasetImpl.java
@@ -21,6 +21,7 @@ import com.google.common.collect.Lists;
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystemTestHelper;
+import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.fs.StorageType;
 import org.apache.hadoop.hdfs.DFSConfigKeys;
 import org.apache.hadoop.hdfs.HdfsConfiguration;
@@ -111,7 +112,7 @@ public class TestFsDatasetImpl {
     List<String> dirStrings = new ArrayList<String>();
     for (int i = 0; i < numDirs; i++) {
       File loc = new File(BASE_DIR + "/data" + i);
-      dirStrings.add(loc.toString());
+      dirStrings.add(new Path(loc.toString()).toUri().toString());
       loc.mkdirs();
       dirs.add(createStorageDirectory(loc));
       when(storage.getStorageDir(i)).thenReturn(dirs.get(i));
@@ -158,8 +159,9 @@ public class TestFsDatasetImpl {
     }
     for (int i = 0; i < numNewVolumes; i++) {
       String path = BASE_DIR + "/newData" + i;
-      expectedVolumes.add(path);
-      StorageLocation loc = StorageLocation.parse(path);
+      String pathUri = new Path(path).toUri().toString();
+      expectedVolumes.add(new File(pathUri).toString());
+      StorageLocation loc = StorageLocation.parse(pathUri);
       Storage.StorageDirectory sd = createStorageDirectory(new File(path));
       DataStorage.VolumeBuilder builder =
           new DataStorage.VolumeBuilder(storage, sd);
@@ -178,7 +180,8 @@ public class TestFsDatasetImpl {
       actualVolumes.add(
           dataset.getVolumes().get(numExistingVolumes + i).getBasePath());
     }
-    assertEquals(actualVolumes, expectedVolumes);
+    assertEquals(actualVolumes.size(), expectedVolumes.size());
+    assertTrue(actualVolumes.containsAll(expectedVolumes));
   }
 
   @Test(timeout = 30000)