You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by cu...@apache.org on 2007/04/19 23:34:53 UTC

svn commit: r530556 [12/12] - in /lucene/hadoop/trunk: ./ src/contrib/abacus/src/java/org/apache/hadoop/abacus/ src/contrib/hbase/src/java/org/apache/hadoop/hbase/ src/contrib/hbase/src/test/org/apache/hadoop/hbase/ src/contrib/streaming/src/java/org/a...

Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestMiniMRLocalFS.java Thu Apr 19 14:34:41 2007
@@ -196,8 +196,8 @@
     
     public InputSplit[] getSplits(JobConf job, 
                                   int numSplits) throws IOException {
-      return new MySplit[]{new MySplit(0,1), new MySplit(1,3),
-                           new MySplit(4,2)};
+      return new MySplit[]{new MySplit(0, 1), new MySplit(1, 3),
+                           new MySplit(4, 2)};
     }
 
     public RecordReader getRecordReader(InputSplit split,

Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/mapred/TestSequenceFileInputFilter.java Thu Apr 19 14:34:41 2007
@@ -73,7 +73,7 @@
     InputFormat format = new SequenceFileInputFilter();
     Text key = new Text();
     BytesWritable value = new BytesWritable();
-    if(numSplits==0) {
+    if (numSplits==0) {
       numSplits =
         random.nextInt(MAX_LENGTH/(SequenceFile.SYNC_INTERVAL/20))+1;
     }
@@ -138,7 +138,7 @@
       int count = countRecords(1);
       LOG.info("Accepted "+count+" records");
       int expectedCount = length/1000;
-      if(expectedCount*1000!=length)
+      if (expectedCount*1000!=length)
         expectedCount++;
       assertEquals(count, expectedCount);
     }

Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/net/TestNetworkTopology.java Thu Apr 19 14:34:41 2007
@@ -21,15 +21,15 @@
   
   static {
     for(int i=0; i<dataNodes.length; i++) {
-      cluster.add( dataNodes[i] );
+      cluster.add(dataNodes[i]);
     }
   }
   
   public void testContains() {
     for(int i=0; i<dataNodes.length; i++) {
-      assertTrue( cluster.contains(dataNodes[i]));
+      assertTrue(cluster.contains(dataNodes[i]));
     }
-    assertFalse( cluster.contains( NODE ));
+    assertFalse(cluster.contains(NODE));
   }
   
   public void testNumOfChildren() throws Exception {
@@ -55,14 +55,14 @@
 
   public void testRemove() throws Exception {
     for(int i=0; i<dataNodes.length; i++) {
-      cluster.remove( dataNodes[i] );
+      cluster.remove(dataNodes[i]);
     }
     for(int i=0; i<dataNodes.length; i++) {
-      assertFalse( cluster.contains( dataNodes[i] ) );
+      assertFalse(cluster.contains(dataNodes[i]));
     }
     assertEquals(0, cluster.getNumOfLeaves());
     for(int i=0; i<dataNodes.length; i++) {
-      cluster.add( dataNodes[i] );
+      cluster.add(dataNodes[i]);
     }
   }
 }

Modified: lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java?view=diff&rev=530556&r1=530555&r2=530556
==============================================================================
--- lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java (original)
+++ lucene/hadoop/trunk/src/test/org/apache/hadoop/record/TestRecordMR.java Thu Apr 19 14:34:41 2007
@@ -287,7 +287,7 @@
     fs.delete(randomOuts);
 
 
-    JobConf genJob = new JobConf(conf,TestRecordMR.class);
+    JobConf genJob = new JobConf(conf, TestRecordMR.class);
     genJob.setInputPath(randomIns);
     genJob.setInputKeyClass(RecInt.class);
     genJob.setInputValueClass(RecInt.class);
@@ -334,7 +334,7 @@
     int intermediateReduces = 10;
     Path intermediateOuts = new Path(testdir, "intermediateouts");
     fs.delete(intermediateOuts);
-    JobConf checkJob = new JobConf(conf,TestRecordMR.class);
+    JobConf checkJob = new JobConf(conf, TestRecordMR.class);
     checkJob.setInputPath(randomOuts);
     checkJob.setInputKeyClass(RecInt.class);
     checkJob.setInputValueClass(RecString.class);
@@ -359,7 +359,7 @@
     //
     Path finalOuts = new Path(testdir, "finalouts");        
     fs.delete(finalOuts);
-    JobConf mergeJob = new JobConf(conf,TestRecordMR.class);
+    JobConf mergeJob = new JobConf(conf, TestRecordMR.class);
     mergeJob.setInputPath(intermediateOuts);
     mergeJob.setInputKeyClass(RecInt.class);
     mergeJob.setInputValueClass(RecString.class);
@@ -392,12 +392,12 @@
         if (dist[i] == 0) {
           continue;
         }
-        if (! in.next(key, val)) {
+        if (!in.next(key, val)) {
           System.err.println("Cannot read entry " + i);
           success = false;
           break;
         } else {
-          if ( !((key.getData() == i ) && (val.getData() == dist[i]))) {
+          if (!((key.getData() == i) && (val.getData() == dist[i]))) {
             System.err.println("Mismatch!  Pos=" + key.getData() + ", i=" + i + ", val=" + val.getData() + ", dist[i]=" + dist[i]);
             success = false;
           }