You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by cu...@apache.org on 2007/08/29 19:43:36 UTC

svn commit: r570875 - in /lucene/hadoop/trunk: CHANGES.txt src/test/org/apache/hadoop/dfs/UpgradeUtilities.java

Author: cutting
Date: Wed Aug 29 10:43:34 2007
New Revision: 570875

URL: http://svn.apache.org/viewvc?rev=570875&view=rev
Log:
HADOOP-1749.  Change TestDFSUpgrade to sort files, fixing sporadic test failures.  Contributed by Enis.

Modified:
    lucene/hadoop/trunk/CHANGES.txt
    lucene/hadoop/trunk/src/test/org/apache/hadoop/dfs/UpgradeUtilities.java

Modified: lucene/hadoop/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/CHANGES.txt?rev=570875&r1=570874&r2=570875&view=diff
==============================================================================
--- lucene/hadoop/trunk/CHANGES.txt (original)
+++ lucene/hadoop/trunk/CHANGES.txt Wed Aug 29 10:43:34 2007
@@ -61,6 +61,10 @@
     HADOOP-1772.  Fix hadoop-daemon.sh script to get correct hostname
     under Cygwin.  (Tsz Wo (Nicholas), SZE via cutting)
 
+    HADOOP-1749.  Change TestDFSUpgrade to sort files, fixing sporadic
+    test failures.  (Enis Soztutar via cutting)
+
+
   IMPROVEMENTS
 
     HADOOP-1756. Add toString() to some Writable-s. (ab)

Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/dfs/UpgradeUtilities.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/dfs/UpgradeUtilities.java?rev=570875&r1=570874&r2=570875&view=diff
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/dfs/UpgradeUtilities.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/dfs/UpgradeUtilities.java Wed Aug 29 10:43:34 2007
@@ -25,6 +25,7 @@
 import java.io.IOException;
 import java.io.OutputStream;
 import java.io.RandomAccessFile;
+import java.util.Arrays;
 import java.util.Random;
 import java.util.zip.CRC32;
 import org.apache.hadoop.conf.Configuration;
@@ -200,6 +201,7 @@
                                          "Given argument is not a directory:" + dir);
     }
     File[] list = dir.listFiles();
+    Arrays.sort(list);
     CRC32 checksum = new CRC32();
     for (int i = 0; i < list.length; i++) {
       if (list[i].isFile()) {