You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by ae...@apache.org on 2016/03/28 19:36:11 UTC

[43/48] hadoop git commit: Revert "MAPREDUCE-6543. Migrate MR client test cases part 2. Contributed by Dustin Cote."

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSequenceFileInputFormat.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSequenceFileInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSequenceFileInputFormat.java
index 338e91d..575ed53 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSequenceFileInputFormat.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSequenceFileInputFormat.java
@@ -18,28 +18,22 @@
 
 package org.apache.hadoop.mapred;
 
-import org.apache.commons.logging.Log;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.BytesWritable;
-import org.apache.hadoop.io.IntWritable;
-import org.apache.hadoop.io.SequenceFile;
-import org.junit.Test;
-
-import java.util.BitSet;
-import java.util.Random;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-
-public class TestSequenceFileInputFormat {
+import java.io.*;
+import java.util.*;
+import junit.framework.TestCase;
+
+import org.apache.commons.logging.*;
+
+import org.apache.hadoop.fs.*;
+import org.apache.hadoop.io.*;
+import org.apache.hadoop.conf.*;
+
+public class TestSequenceFileInputFormat extends TestCase {
   private static final Log LOG = FileInputFormat.LOG;
 
   private static int MAX_LENGTH = 10000;
   private static Configuration conf = new Configuration();
 
-  @Test
   public void testFormat() throws Exception {
     JobConf job = new JobConf(conf);
     FileSystem fs = FileSystem.getLocal(conf);
@@ -116,4 +110,8 @@ public class TestSequenceFileInputFormat {
 
     }
   }
+
+  public static void main(String[] args) throws Exception {
+    new TestSequenceFileInputFormat().testFormat();
+  }
 }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSortedRanges.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSortedRanges.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSortedRanges.java
index 82d1d2d..ad4d4ce 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSortedRanges.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSortedRanges.java
@@ -17,20 +17,18 @@
  */
 package org.apache.hadoop.mapred;
 
+import java.util.Iterator;
+
+import junit.framework.TestCase;
+
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.mapred.SortedRanges.Range;
-import org.junit.Test;
-
-import java.util.Iterator;
-
-import static org.junit.Assert.assertEquals;
 
-public class TestSortedRanges {
-  private static final Log LOG =
+public class TestSortedRanges extends TestCase {
+  private static final Log LOG = 
     LogFactory.getLog(TestSortedRanges.class);
-
-  @Test
+  
   public void testAdd() {
     SortedRanges sr = new SortedRanges();
     sr.add(new Range(2,9));
@@ -68,8 +66,7 @@ public class TestSortedRanges {
     assertEquals(77, it.next().longValue());
     
   }
-
-  @Test
+  
   public void testRemove() {
     SortedRanges sr = new SortedRanges();
     sr.add(new Range(2,19));

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSpecialCharactersInOutputPath.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSpecialCharactersInOutputPath.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSpecialCharactersInOutputPath.java
index b9e3275..426686f 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSpecialCharactersInOutputPath.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestSpecialCharactersInOutputPath.java
@@ -18,6 +18,12 @@
 
 package org.apache.hadoop.mapred;
 
+import java.io.DataOutputStream;
+import java.io.IOException;
+import java.net.URI;
+
+import junit.framework.TestCase;
+
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
@@ -28,20 +34,14 @@ import org.apache.hadoop.io.LongWritable;
 import org.apache.hadoop.io.Text;
 import org.apache.hadoop.mapred.lib.IdentityMapper;
 import org.apache.hadoop.mapred.lib.IdentityReducer;
+import org.apache.hadoop.mapreduce.MRConfig;
+import org.apache.hadoop.mapreduce.server.jobtracker.JTConfig;
 import org.apache.hadoop.util.Progressable;
-import org.junit.Test;
-
-import java.io.DataOutputStream;
-import java.io.IOException;
-import java.net.URI;
-
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
 
 /**
  * A JUnit test to test that jobs' output filenames are not HTML-encoded (cf HADOOP-1795).
  */
-public class TestSpecialCharactersInOutputPath {
+public class TestSpecialCharactersInOutputPath extends TestCase {
   private static final Log LOG =
     LogFactory.getLog(TestSpecialCharactersInOutputPath.class.getName());
   
@@ -96,8 +96,7 @@ public class TestSpecialCharactersInOutputPath {
     LOG.info("job is complete: " + runningJob.isSuccessful());
     return (runningJob.isSuccessful());
   }
-
-  @Test
+  
   public void testJobWithDFS() throws IOException {
     String namenode = null;
     MiniDFSCluster dfs = null;

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestStatisticsCollector.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestStatisticsCollector.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestStatisticsCollector.java
index 8a83e81..12568d0 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestStatisticsCollector.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestStatisticsCollector.java
@@ -19,18 +19,14 @@ package org.apache.hadoop.mapred;
 
 import java.util.Map;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.mapred.StatisticsCollector.TimeWindow;
 import org.apache.hadoop.mapred.StatisticsCollector.Stat;
-import org.junit.Test;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertNull;
 
-public class TestStatisticsCollector {
+public class TestStatisticsCollector extends TestCase{
 
   @SuppressWarnings("rawtypes")
-  @Test
   public void testMovingWindow() throws Exception {
     StatisticsCollector collector = new StatisticsCollector(1);
     TimeWindow window = new TimeWindow("test", 6, 2);

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestUserDefinedCounters.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestUserDefinedCounters.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestUserDefinedCounters.java
index 2d67edc..3c2cf21 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestUserDefinedCounters.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestUserDefinedCounters.java
@@ -17,15 +17,6 @@
  */
 package org.apache.hadoop.mapred;
 
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.FileUtil;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.LongWritable;
-import org.apache.hadoop.io.Text;
-import org.apache.hadoop.mapred.lib.IdentityMapper;
-import org.apache.hadoop.mapred.lib.IdentityReducer;
-import org.junit.Test;
-
 import java.io.BufferedReader;
 import java.io.File;
 import java.io.IOException;
@@ -35,10 +26,18 @@ import java.io.OutputStream;
 import java.io.OutputStreamWriter;
 import java.io.Writer;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
+import junit.framework.TestCase;
 
-public class TestUserDefinedCounters {
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.FileUtil;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.io.LongWritable;
+import org.apache.hadoop.io.Text;
+import org.apache.hadoop.mapred.lib.IdentityMapper;
+import org.apache.hadoop.mapred.lib.IdentityReducer;
+
+public class TestUserDefinedCounters extends TestCase {
+  
   private static String TEST_ROOT_DIR =
     new File(System.getProperty("test.build.data", "/tmp")).toURI()
     .toString().replace(' ', '+')
@@ -76,7 +75,6 @@ public class TestUserDefinedCounters {
     wr.close();
   }
 
-  @Test
   public void testMapReduceJob() throws Exception {
 
     JobConf conf = new JobConf(TestUserDefinedCounters.class);

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestWritableJobConf.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestWritableJobConf.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestWritableJobConf.java
index 82c68db..2c0cedc 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestWritableJobConf.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestWritableJobConf.java
@@ -18,6 +18,12 @@
 
 package org.apache.hadoop.mapred;
 
+import java.util.HashMap;
+import java.util.Iterator;
+import java.util.Map;
+
+import junit.framework.TestCase;
+
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.io.DataInputBuffer;
 import org.apache.hadoop.io.DataOutputBuffer;
@@ -25,15 +31,8 @@ import org.apache.hadoop.io.serializer.Deserializer;
 import org.apache.hadoop.io.serializer.SerializationFactory;
 import org.apache.hadoop.io.serializer.Serializer;
 import org.apache.hadoop.util.GenericsUtil;
-import org.junit.Test;
-
-import java.util.HashMap;
-import java.util.Iterator;
-import java.util.Map;
-
-import static org.junit.Assert.assertTrue;
 
-public class TestWritableJobConf {
+public class TestWritableJobConf extends TestCase {
 
   private static final Configuration CONF = new Configuration();
 
@@ -79,17 +78,15 @@ public class TestWritableJobConf {
       }
     }
 
-    assertTrue(map1.equals(map2));
+    assertEquals(map1, map2);
   }
 
-  @Test
   public void testEmptyConfiguration() throws Exception {
     JobConf conf = new JobConf();
     Configuration deser = serDeser(conf);
     assertEquals(conf, deser);
   }
 
-  @Test
   public void testNonEmptyConfiguration() throws Exception {
     JobConf conf = new JobConf();
     conf.set("a", "A");
@@ -98,7 +95,6 @@ public class TestWritableJobConf {
     assertEquals(conf, deser);
   }
 
-  @Test
   public void testConfigurationWithDefaults() throws Exception {
     JobConf conf = new JobConf(false);
     conf.set("a", "A");

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestYARNRunner.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestYARNRunner.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestYARNRunner.java
index abf2e72..0e34042 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestYARNRunner.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestYARNRunner.java
@@ -18,10 +18,6 @@
 
 package org.apache.hadoop.mapred;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertTrue;
 import static org.mockito.Matchers.any;
 import static org.mockito.Mockito.doAnswer;
 import static org.mockito.Mockito.doReturn;
@@ -42,6 +38,8 @@ import java.security.PrivilegedExceptionAction;
 import java.util.List;
 import java.util.Map;
 
+import junit.framework.TestCase;
+
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
@@ -115,7 +113,7 @@ import org.mockito.stubbing.Answer;
  * Test YarnRunner and make sure the client side plugin works
  * fine
  */
-public class TestYARNRunner {
+public class TestYARNRunner extends TestCase {
   private static final Log LOG = LogFactory.getLog(TestYARNRunner.class);
   private static final RecordFactory recordFactory = RecordFactoryProvider.getRecordFactory(null);
 

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestDatamerge.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestDatamerge.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestDatamerge.java
index a306676..15cea69 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestDatamerge.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestDatamerge.java
@@ -22,6 +22,11 @@ import java.io.DataOutput;
 import java.io.IOException;
 import java.util.Iterator;
 
+import junit.framework.Test;
+import junit.framework.TestCase;
+import junit.framework.TestSuite;
+import junit.extensions.TestSetup;
+
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileStatus;
 import org.apache.hadoop.fs.Path;
@@ -49,27 +54,23 @@ import org.apache.hadoop.mapred.Utils;
 import org.apache.hadoop.mapred.lib.IdentityMapper;
 import org.apache.hadoop.mapred.lib.IdentityReducer;
 import org.apache.hadoop.util.ReflectionUtils;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.assertFalse;
 
-public class TestDatamerge {
+public class TestDatamerge extends TestCase {
 
   private static MiniDFSCluster cluster = null;
-
-  @Before
-  public void setUp() throws Exception {
-    Configuration conf = new Configuration();
-    cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
-  }
-  @After
-  public void tearDown() throws Exception {
-    if (cluster != null) {
-      cluster.shutdown();
-    }
+  public static Test suite() {
+    TestSetup setup = new TestSetup(new TestSuite(TestDatamerge.class)) {
+      protected void setUp() throws Exception {
+        Configuration conf = new Configuration();
+        cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
+      }
+      protected void tearDown() throws Exception {
+        if (cluster != null) {
+          cluster.shutdown();
+        }
+      }
+    };
+    return setup;
   }
 
   private static SequenceFile.Writer[] createWriters(Path testdir,
@@ -245,22 +246,18 @@ public class TestDatamerge {
     base.getFileSystem(job).delete(base, true);
   }
 
-  @Test
   public void testSimpleInnerJoin() throws Exception {
     joinAs("inner", InnerJoinChecker.class);
   }
 
-  @Test
   public void testSimpleOuterJoin() throws Exception {
     joinAs("outer", OuterJoinChecker.class);
   }
 
-  @Test
   public void testSimpleOverride() throws Exception {
     joinAs("override", OverrideChecker.class);
   }
 
-  @Test
   public void testNestedJoin() throws Exception {
     // outer(inner(S1,...,Sn),outer(S1,...Sn))
     final int SOURCES = 3;
@@ -353,7 +350,6 @@ public class TestDatamerge {
 
   }
 
-  @Test
   public void testEmptyJoin() throws Exception {
     JobConf job = new JobConf();
     Path base = cluster.getFileSystem().makeQualified(new Path("/empty"));

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestTupleWritable.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestTupleWritable.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestTupleWritable.java
index 5687155..e421ede 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestTupleWritable.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestTupleWritable.java
@@ -26,6 +26,8 @@ import java.io.IOException;
 import java.util.Arrays;
 import java.util.Random;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.io.BooleanWritable;
 import org.apache.hadoop.io.BytesWritable;
 import org.apache.hadoop.io.FloatWritable;
@@ -34,12 +36,8 @@ import org.apache.hadoop.io.LongWritable;
 import org.apache.hadoop.io.Text;
 import org.apache.hadoop.io.Writable;
 import org.apache.hadoop.io.WritableUtils;
-import org.junit.Test;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
 
-public class TestTupleWritable {
+public class TestTupleWritable extends TestCase {
 
   private TupleWritable makeTuple(Writable[] writs) {
     Writable[] sub1 = { writs[1], writs[2] };
@@ -102,7 +100,6 @@ public class TestTupleWritable {
     return i;
   }
 
-  @Test
   public void testIterable() throws Exception {
     Random r = new Random();
     Writable[] writs = {
@@ -124,7 +121,6 @@ public class TestTupleWritable {
     verifIter(writs, t, 0);
   }
 
-  @Test
   public void testNestedIterable() throws Exception {
     Random r = new Random();
     Writable[] writs = {
@@ -143,7 +139,6 @@ public class TestTupleWritable {
     assertTrue("Bad count", writs.length == verifIter(writs, sTuple, 0));
   }
 
-  @Test
   public void testWritable() throws Exception {
     Random r = new Random();
     Writable[] writs = {
@@ -167,7 +162,6 @@ public class TestTupleWritable {
     assertTrue("Failed to write/read tuple", sTuple.equals(dTuple));
   }
 
-  @Test
   public void testWideWritable() throws Exception {
     Writable[] manyWrits = makeRandomWritables(131);
     
@@ -186,8 +180,7 @@ public class TestTupleWritable {
     assertTrue("Failed to write/read tuple", sTuple.equals(dTuple));
     assertEquals("All tuple data has not been read from the stream",-1,in.read());
   }
-
-  @Test
+  
   public void testWideWritable2() throws Exception {
     Writable[] manyWrits = makeRandomWritables(71);
     
@@ -209,7 +202,6 @@ public class TestTupleWritable {
    * Tests a tuple writable with more than 64 values and the values set written
    * spread far apart.
    */
-  @Test
   public void testSparseWideWritable() throws Exception {
     Writable[] manyWrits = makeRandomWritables(131);
     
@@ -228,7 +220,7 @@ public class TestTupleWritable {
     assertTrue("Failed to write/read tuple", sTuple.equals(dTuple));
     assertEquals("All tuple data has not been read from the stream",-1,in.read());
   }
-  @Test
+  
   public void testWideTuple() throws Exception {
     Text emptyText = new Text("Should be empty");
     Writable[] values = new Writable[64];
@@ -248,7 +240,7 @@ public class TestTupleWritable {
       }
     }
   }
-  @Test
+  
   public void testWideTuple2() throws Exception {
     Text emptyText = new Text("Should be empty");
     Writable[] values = new Writable[64];
@@ -272,7 +264,6 @@ public class TestTupleWritable {
   /**
    * Tests that we can write more than 64 values.
    */
-  @Test
   public void testWideTupleBoundary() throws Exception {
     Text emptyText = new Text("Should not be set written");
     Writable[] values = new Writable[65];
@@ -296,7 +287,6 @@ public class TestTupleWritable {
   /**
    * Tests compatibility with pre-0.21 versions of TupleWritable
    */
-  @Test
   public void testPreVersion21Compatibility() throws Exception {
     Writable[] manyWrits = makeRandomWritables(64);
     PreVersion21TupleWritable oldTuple = new PreVersion21TupleWritable(manyWrits);
@@ -314,7 +304,7 @@ public class TestTupleWritable {
     assertTrue("Tuple writable is unable to read pre-0.21 versions of TupleWritable", oldTuple.isCompatible(dTuple));
     assertEquals("All tuple data has not been read from the stream",-1,in.read());
   }
-  @Test
+  
   public void testPreVersion21CompatibilityEmptyTuple() throws Exception {
     Writable[] manyWrits = new Writable[0];
     PreVersion21TupleWritable oldTuple = new PreVersion21TupleWritable(manyWrits);

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestWrappedRecordReaderClassloader.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestWrappedRecordReaderClassloader.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestWrappedRecordReaderClassloader.java
index ae5572f..3ca175a 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestWrappedRecordReaderClassloader.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/join/TestWrappedRecordReaderClassloader.java
@@ -21,6 +21,8 @@ import java.io.DataInput;
 import java.io.DataOutput;
 import java.io.IOException;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.NullWritable;
@@ -33,16 +35,13 @@ import org.apache.hadoop.mapred.JobConfigurable;
 import org.apache.hadoop.mapred.RecordReader;
 import org.apache.hadoop.mapred.Reporter;
 import org.apache.hadoop.util.ReflectionUtils;
-import org.junit.Test;
-import static org.junit.Assert.assertTrue;
 
-public class TestWrappedRecordReaderClassloader {
+public class TestWrappedRecordReaderClassloader extends TestCase {
   /**
    * Tests the class loader set by {@link JobConf#setClassLoader(ClassLoader)}
    * is inherited by any {@link WrappedRecordReader}s created by
    * {@link CompositeRecordReader}
    */
-  @Test
   public void testClassLoader() throws Exception {
     JobConf job = new JobConf();
     Fake_ClassLoader classLoader = new Fake_ClassLoader();

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestDelegatingInputFormat.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestDelegatingInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestDelegatingInputFormat.java
index b916026..8bd8554 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestDelegatingInputFormat.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestDelegatingInputFormat.java
@@ -20,6 +20,8 @@ package org.apache.hadoop.mapred.lib;
 import java.io.DataOutputStream;
 import java.io.IOException;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hdfs.MiniDFSCluster;
@@ -30,12 +32,9 @@ import org.apache.hadoop.mapred.Mapper;
 import org.apache.hadoop.mapred.OutputCollector;
 import org.apache.hadoop.mapred.Reporter;
 import org.apache.hadoop.mapred.TextInputFormat;
-import org.junit.Test;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.assertEquals;
 
-public class TestDelegatingInputFormat {
-  @Test
+public class TestDelegatingInputFormat extends TestCase {
+
   public void testSplitting() throws Exception {
     JobConf conf = new JobConf();
     MiniDFSCluster dfs = null;

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestLineInputFormat.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestLineInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestLineInputFormat.java
index 388de0f..db9c219 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestLineInputFormat.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestLineInputFormat.java
@@ -20,14 +20,13 @@ package org.apache.hadoop.mapred.lib;
 
 import java.io.*;
 import java.util.*;
+import junit.framework.TestCase;
 
 import org.apache.hadoop.fs.*;
 import org.apache.hadoop.io.*;
 import org.apache.hadoop.mapred.*;
-import org.junit.Test;
-import static org.junit.Assert.assertEquals;
 
-public class TestLineInputFormat {
+public class TestLineInputFormat extends TestCase {
   private static int MAX_LENGTH = 200;
   
   private static JobConf defaultConf = new JobConf();
@@ -44,7 +43,7 @@ public class TestLineInputFormat {
   private static Path workDir = 
     new Path(new Path(System.getProperty("test.build.data", "."), "data"),
              "TestLineInputFormat");
-  @Test
+  
   public void testFormat() throws Exception {
     JobConf job = new JobConf();
     Path file = new Path(workDir, "test.txt");

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestMultipleInputs.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestMultipleInputs.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestMultipleInputs.java
index 115a6f7..3a9cb9e 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestMultipleInputs.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/TestMultipleInputs.java
@@ -36,6 +36,7 @@ import static org.junit.Assert.assertEquals;
  * @see TestDelegatingInputFormat
  */
 public class TestMultipleInputs {
+
   @Test
   public void testAddInputPathWithFormat() {
     final JobConf conf = new JobConf();
@@ -48,6 +49,7 @@ public class TestMultipleInputs {
     assertEquals(KeyValueTextInputFormat.class, inputs.get(new Path("/bar"))
        .getClass());
   }
+
   @Test
   public void testAddInputPathWithMapper() {
     final JobConf conf = new JobConf();

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/aggregate/TestAggregates.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/aggregate/TestAggregates.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/aggregate/TestAggregates.java
index f33f83c..6da96ce 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/aggregate/TestAggregates.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/aggregate/TestAggregates.java
@@ -22,14 +22,13 @@ import org.apache.hadoop.io.*;
 import org.apache.hadoop.mapred.*;
 import org.apache.hadoop.mapred.lib.*;
 import org.apache.hadoop.mapreduce.MapReduceTestUtil;
-import org.junit.Test;
-import static org.junit.Assert.assertEquals;
 
+import junit.framework.TestCase;
 import java.io.*;
 import java.util.*;
 import java.text.NumberFormat;
 
-public class TestAggregates {
+public class TestAggregates extends TestCase {
 
   private static NumberFormat idFormat = NumberFormat.getInstance();
     static {
@@ -37,7 +36,7 @@ public class TestAggregates {
       idFormat.setGroupingUsed(false);
   }
 
-  @Test
+
   public void testAggregates() throws Exception {
     launch();
   }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/db/TestConstructQuery.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/db/TestConstructQuery.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/db/TestConstructQuery.java
index 203da4e..968bb06 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/db/TestConstructQuery.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/lib/db/TestConstructQuery.java
@@ -19,13 +19,13 @@ package org.apache.hadoop.mapred.lib.db;
 
 import java.io.IOException;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.io.NullWritable;
 import org.apache.hadoop.mapred.JobConf;
-import org.junit.Test;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertNull;
 
-public class TestConstructQuery {
+public class TestConstructQuery extends TestCase {
+  
   private String[] fieldNames = new String[] { "id", "name", "value" };
   private String[] nullFieldNames = new String[] { null, null, null };
   private String expected = "INSERT INTO hadoop_output (id,name,value) VALUES (?,?,?);";
@@ -33,15 +33,15 @@ public class TestConstructQuery {
   
   private DBOutputFormat<DBWritable, NullWritable> format 
     = new DBOutputFormat<DBWritable, NullWritable>();
-  @Test
-  public void testConstructQuery() {
+  
+  public void testConstructQuery() {  
     String actual = format.constructQuery("hadoop_output", fieldNames);
     assertEquals(expected, actual);
-
+    
     actual = format.constructQuery("hadoop_output", nullFieldNames);
     assertEquals(nullExpected, actual);
   }
-  @Test
+  
   public void testSetOutput() throws IOException {
     JobConf job = new JobConf();
     DBOutputFormat.setOutput(job, "hadoop_output", fieldNames);

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/pipes/TestPipes.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/pipes/TestPipes.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/pipes/TestPipes.java
index 34b1d75..dd7817d 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/pipes/TestPipes.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/pipes/TestPipes.java
@@ -44,13 +44,10 @@ import org.apache.hadoop.mapreduce.MapReduceTestUtil;
 import org.apache.hadoop.util.StringUtils;
 import org.apache.hadoop.util.ToolRunner;
 import org.junit.Ignore;
-import org.junit.Test;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.assertFalse;
 
+import junit.framework.TestCase;
 @Ignore
-public class TestPipes {
+public class TestPipes extends TestCase {
   private static final Log LOG =
     LogFactory.getLog(TestPipes.class.getName());
   
@@ -69,7 +66,7 @@ public class TestPipes {
     fs.delete(p, true);
     assertFalse("output not cleaned up", fs.exists(p));
   }
-  @Test
+
   public void testPipes() throws IOException {
     if (System.getProperty("compile.c++") == null) {
       LOG.info("compile.c++ is not defined, so skipping TestPipes");

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestLocalRunner.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestLocalRunner.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestLocalRunner.java
index 8177ecd..29640c8 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestLocalRunner.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestLocalRunner.java
@@ -17,42 +17,36 @@
  */
 package org.apache.hadoop.mapreduce;
 
+import java.io.BufferedReader;
+import java.io.BufferedWriter;
+import java.io.InputStream;
+import java.io.InputStreamReader;
+import java.io.OutputStream;
+import java.io.OutputStreamWriter;
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FSDataInputStream;
-import org.apache.hadoop.fs.FileStatus;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.LongWritable;
 import org.apache.hadoop.io.NullWritable;
 import org.apache.hadoop.io.Text;
+import org.apache.hadoop.fs.*;
 import org.apache.hadoop.mapred.LocalJobRunner;
 import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
 import org.apache.hadoop.mapreduce.lib.input.FileSplit;
 import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
 import org.apache.hadoop.util.ReflectionUtils;
-import org.junit.Test;
 
-import java.io.BufferedReader;
-import java.io.BufferedWriter;
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.InputStreamReader;
-import java.io.OutputStream;
-import java.io.OutputStreamWriter;
-import java.util.ArrayList;
-import java.util.List;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertTrue;
+import org.junit.Test;
+import junit.framework.TestCase;
 
 /**
  * Stress tests for the LocalJobRunner
  */
-public class TestLocalRunner {
+public class TestLocalRunner extends TestCase {
 
   private static final Log LOG = LogFactory.getLog(TestLocalRunner.class);
 

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMRJobClient.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMRJobClient.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMRJobClient.java
index 8fe9078..6f45b5f 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMRJobClient.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMRJobClient.java
@@ -17,23 +17,6 @@
  */
 package org.apache.hadoop.mapreduce;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.LocatedFileStatus;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.fs.RemoteIterator;
-import org.apache.hadoop.mapred.ClusterMapReduceTestCase;
-import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
-import org.apache.hadoop.mapreduce.tools.CLI;
-import org.apache.hadoop.util.ExitUtil;
-import org.apache.hadoop.util.Tool;
-import org.apache.hadoop.util.ToolRunner;
-import org.codehaus.jettison.json.JSONException;
-import org.codehaus.jettison.json.JSONObject;
-import org.junit.Test;
-
 import java.io.BufferedReader;
 import java.io.ByteArrayInputStream;
 import java.io.ByteArrayOutputStream;
@@ -48,11 +31,23 @@ import java.io.PipedOutputStream;
 import java.io.PrintStream;
 import java.util.Arrays;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
+import org.apache.hadoop.fs.LocatedFileStatus;
+import org.apache.hadoop.fs.RemoteIterator;
+import org.codehaus.jettison.json.JSONException;
+import org.codehaus.jettison.json.JSONObject;
+import org.junit.Assert;
+
+import org.apache.commons.logging.Log;
+import org.apache.commons.logging.LogFactory;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.mapred.ClusterMapReduceTestCase;
+import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
+import org.apache.hadoop.mapreduce.tools.CLI;
+import org.apache.hadoop.util.ExitUtil;
+import org.apache.hadoop.util.Tool;
+import org.apache.hadoop.util.ToolRunner;
 
 /**
  test CLI class. CLI class implemented  the Tool interface. 
@@ -108,7 +103,7 @@ public class TestMRJobClient extends ClusterMapReduceTestCase {
       throw new IOException();
     }
   }
-  @Test
+  
   public void testJobSubmissionSpecsAndFiles() throws Exception {
     Configuration conf = createJobConf();
     Job job = MapReduceTestUtil.createJob(conf, getInputDir(), getOutputDir(),
@@ -132,7 +127,7 @@ public class TestMRJobClient extends ClusterMapReduceTestCase {
   /**
    * main test method
    */
-  @Test
+
   public void testJobClient() throws Exception {
     Configuration conf = createJobConf();
     Job job = runJob(conf);
@@ -185,7 +180,8 @@ public class TestMRJobClient extends ClusterMapReduceTestCase {
 
     runTool(conf, jc, new String[] { "-fail-task", taid.toString() }, out);
     String answer = new String(out.toByteArray(), "UTF-8");
-    assertTrue(answer.contains("Killed task " + taid + " by failing it"));
+    Assert
+      .assertTrue(answer.contains("Killed task " + taid + " by failing it"));
   }
 
   /**
@@ -203,7 +199,7 @@ public class TestMRJobClient extends ClusterMapReduceTestCase {
 
     runTool(conf, jc, new String[] { "-kill-task", taid.toString() }, out);
     String answer = new String(out.toByteArray(), "UTF-8");
-    assertTrue(answer.contains("Killed task " + taid));
+    Assert.assertTrue(answer.contains("Killed task " + taid));
   }
   
   /**
@@ -690,7 +686,6 @@ public class TestMRJobClient extends ClusterMapReduceTestCase {
    * Test -list option displays job name.
    * The name is capped to 20 characters for display.
    */
-  @Test
   public void testJobName() throws Exception {
     Configuration conf = createJobConf();
     CLI jc = createJobClient();

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMapReduceLazyOutput.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMapReduceLazyOutput.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMapReduceLazyOutput.java
index a69e06e..1e4f4de 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMapReduceLazyOutput.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestMapReduceLazyOutput.java
@@ -25,6 +25,8 @@ import java.io.Writer;
 import java.util.Arrays;
 import java.util.List;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FileUtil;
@@ -40,16 +42,13 @@ import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
 import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
 import org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat;
 import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
-import org.junit.Test;
-
-import static org.junit.Assert.assertTrue;
 
 /**
  * A JUnit test to test the Map-Reduce framework's feature to create part
  * files only if there is an explicit output.collect. This helps in preventing
  * 0 byte files
  */
-public class TestMapReduceLazyOutput {
+public class TestMapReduceLazyOutput extends TestCase {
   private static final int NUM_HADOOP_SLAVES = 3;
   private static final int NUM_MAPS_PER_NODE = 2;
   private static final Path INPUT = new Path("/testlazy/input");
@@ -123,7 +122,7 @@ public class TestMapReduceLazyOutput {
     }
   }
 
-  @Test
+
   public void testLazyOutput() throws Exception {
     MiniDFSCluster dfs = null;
     MiniMRCluster mr = null;

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestValueIterReset.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestValueIterReset.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestValueIterReset.java
index b757fb2..5cf0899 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestValueIterReset.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestValueIterReset.java
@@ -27,6 +27,8 @@ import java.io.Writer;
 import java.util.ArrayList;
 import java.util.StringTokenizer;
 
+import junit.framework.TestCase;
+
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
@@ -41,15 +43,12 @@ import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
 import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
 import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
 import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
-import org.junit.Test;
-
-import static org.junit.Assert.assertTrue;
 
 /**
  * A JUnit test to test the Map-Reduce framework's support for the
  * "mark-reset" functionality in Reduce Values Iterator
  */
-public class TestValueIterReset {
+public class TestValueIterReset extends TestCase {
   private static final int NUM_MAPS = 1;
   private static final int NUM_TESTS = 4;
   private static final int NUM_VALUES = 40;
@@ -519,7 +518,6 @@ public class TestValueIterReset {
     }
   }
 
-  @Test
   public void testValueIterReset() {
     try {
       Configuration conf = new Configuration();

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestYarnClientProtocolProvider.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestYarnClientProtocolProvider.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestYarnClientProtocolProvider.java
index 308b777..4d84fa9 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestYarnClientProtocolProvider.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/TestYarnClientProtocolProvider.java
@@ -18,7 +18,6 @@
 
 package org.apache.hadoop.mapreduce;
 
-import static org.junit.Assert.assertTrue;
 import static org.mockito.Matchers.any;
 import static org.mockito.Mockito.mock;
 import static org.mockito.Mockito.when;
@@ -27,6 +26,7 @@ import static org.mockito.Mockito.doNothing;
 
 import java.io.IOException;
 import java.nio.ByteBuffer;
+import junit.framework.TestCase;
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.io.Text;
@@ -44,7 +44,8 @@ import org.apache.hadoop.yarn.factories.RecordFactory;
 import org.apache.hadoop.yarn.factory.providers.RecordFactoryProvider;
 import org.junit.Test;
 
-public class TestYarnClientProtocolProvider {
+public class TestYarnClientProtocolProvider extends TestCase {
+  
   private static final RecordFactory recordFactory = RecordFactoryProvider.
       getRecordFactory(null);
   

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/aggregate/TestMapReduceAggregates.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/aggregate/TestMapReduceAggregates.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/aggregate/TestMapReduceAggregates.java
index 789ed98..f24dffe 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/aggregate/TestMapReduceAggregates.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/aggregate/TestMapReduceAggregates.java
@@ -18,24 +18,22 @@
 package org.apache.hadoop.mapreduce.lib.aggregate;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FSDataOutputStream;
+import org.apache.hadoop.fs.*;
 import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.Text;
+import org.apache.hadoop.io.*;
+import org.apache.hadoop.mapred.Utils;
 import org.apache.hadoop.mapreduce.Job;
 import org.apache.hadoop.mapreduce.MapReduceTestUtil;
 import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
 import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
 import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
 import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
-import org.junit.Test;
 
+import junit.framework.TestCase;
+import java.io.*;
 import java.text.NumberFormat;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-
-public class TestMapReduceAggregates {
+public class TestMapReduceAggregates extends TestCase {
 
   private static NumberFormat idFormat = NumberFormat.getInstance();
     static {
@@ -43,7 +41,7 @@ public class TestMapReduceAggregates {
       idFormat.setGroupingUsed(false);
   }
 
-  @Test
+
   public void testAggregates() throws Exception {
     launch();
   }
@@ -124,4 +122,11 @@ public class TestMapReduceAggregates {
     fs.delete(OUTPUT_DIR, true);
     fs.delete(INPUT_DIR, true);
   }
+  
+  /**
+   * Launches all the tasks in order.
+   */
+  public static void main(String[] argv) throws Exception {
+    launch();
+  }
 }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestDBOutputFormat.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestDBOutputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestDBOutputFormat.java
index 014855f..bff25d2 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestDBOutputFormat.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestDBOutputFormat.java
@@ -19,15 +19,14 @@ package org.apache.hadoop.mapreduce.lib.db;
 
 import java.io.IOException;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.io.NullWritable;
 import org.apache.hadoop.mapreduce.Job;
-import org.junit.Test;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertNull;
 
-public class TestDBOutputFormat {
+public class TestDBOutputFormat extends TestCase {
+  
   private String[] fieldNames = new String[] { "id", "name", "value" };
   private String[] nullFieldNames = new String[] { null, null, null };
   private String expected = "INSERT INTO hadoop_output " +
@@ -36,17 +35,15 @@ public class TestDBOutputFormat {
   
   private DBOutputFormat<DBWritable, NullWritable> format 
     = new DBOutputFormat<DBWritable, NullWritable>();
-
-  @Test
-  public void testConstructQuery() {
+  
+  public void testConstructQuery() {  
     String actual = format.constructQuery("hadoop_output", fieldNames);
     assertEquals(expected, actual);
     
     actual = format.constructQuery("hadoop_output", nullFieldNames);
     assertEquals(nullExpected, actual);
   }
-
-  @Test
+  
   public void testSetOutput() throws IOException {
     Job job = Job.getInstance(new Configuration());
     DBOutputFormat.setOutput(job, "hadoop_output", fieldNames);

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestIntegerSplitter.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestIntegerSplitter.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestIntegerSplitter.java
index 8b5d907..e50aba4 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestIntegerSplitter.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestIntegerSplitter.java
@@ -17,15 +17,15 @@
  */
 package org.apache.hadoop.mapreduce.lib.db;
 
-import org.junit.Test;
-
+import java.io.IOException;
+import java.math.BigDecimal;
 import java.sql.SQLException;
+import java.util.ArrayList;
 import java.util.List;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.fail;
+import junit.framework.TestCase;
 
-public class TestIntegerSplitter {
+public class TestIntegerSplitter extends TestCase {
   private long [] toLongArray(List<Long> in) {
     long [] out = new long[in.size()];
     for (int i = 0; i < in.size(); i++) {
@@ -70,14 +70,12 @@ public class TestIntegerSplitter {
     }
   }
 
-  @Test
   public void testEvenSplits() throws SQLException {
     List<Long> splits = new IntegerSplitter().split(10, 0, 100);
     long [] expected = { 0, 10, 20, 30, 40, 50, 60, 70, 80, 90, 100 };
     assertLongArrayEquals(expected, toLongArray(splits));
   }
 
-  @Test
   public void testOddSplits() throws SQLException {
     List<Long> splits = new IntegerSplitter().split(10, 0, 95);
     long [] expected = { 0, 9, 18, 27, 36, 45, 54, 63, 72, 81, 90, 95 };
@@ -85,14 +83,12 @@ public class TestIntegerSplitter {
 
   }
 
-  @Test
   public void testSingletonSplit() throws SQLException {
     List<Long> splits = new IntegerSplitter().split(1, 5, 5);
     long [] expected = { 5, 5 };
     assertLongArrayEquals(expected, toLongArray(splits));
   }
 
-  @Test
   public void testSingletonSplit2() throws SQLException {
     // Same test, but overly-high numSplits
     List<Long> splits = new IntegerSplitter().split(5, 5, 5);
@@ -100,7 +96,6 @@ public class TestIntegerSplitter {
     assertLongArrayEquals(expected, toLongArray(splits));
   }
 
-  @Test
   public void testTooManySplits() throws SQLException {
     List<Long> splits = new IntegerSplitter().split(5, 3, 5);
     long [] expected = { 3, 4, 5 };

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestTextSplitter.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestTextSplitter.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestTextSplitter.java
index e16f423..045e3a1 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestTextSplitter.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/db/TestTextSplitter.java
@@ -17,16 +17,15 @@
  */
 package org.apache.hadoop.mapreduce.lib.db;
 
-import org.junit.Test;
-
+import java.io.IOException;
 import java.math.BigDecimal;
 import java.sql.SQLException;
+import java.util.ArrayList;
 import java.util.List;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.fail;
+import junit.framework.TestCase;
 
-public class TestTextSplitter {
+public class TestTextSplitter extends TestCase {
 
   public String formatArray(Object [] ar) {
     StringBuilder sb = new StringBuilder();
@@ -63,56 +62,48 @@ public class TestTextSplitter {
     }
   }
 
-  @Test
   public void testStringConvertEmpty() {
     TextSplitter splitter = new TextSplitter();
     BigDecimal emptyBigDec = splitter.stringToBigDecimal("");
     assertEquals(BigDecimal.ZERO, emptyBigDec);
   }
 
-  @Test
   public void testBigDecConvertEmpty() {
     TextSplitter splitter = new TextSplitter();
     String emptyStr = splitter.bigDecimalToString(BigDecimal.ZERO);
     assertEquals("", emptyStr);
   }
 
-  @Test
   public void testConvertA() {
     TextSplitter splitter = new TextSplitter();
     String out = splitter.bigDecimalToString(splitter.stringToBigDecimal("A"));
     assertEquals("A", out);
   }
 
-  @Test
   public void testConvertZ() {
     TextSplitter splitter = new TextSplitter();
     String out = splitter.bigDecimalToString(splitter.stringToBigDecimal("Z"));
     assertEquals("Z", out);
   }
 
-  @Test
   public void testConvertThreeChars() {
     TextSplitter splitter = new TextSplitter();
     String out = splitter.bigDecimalToString(splitter.stringToBigDecimal("abc"));
     assertEquals("abc", out);
   }
 
-  @Test
   public void testConvertStr() {
     TextSplitter splitter = new TextSplitter();
     String out = splitter.bigDecimalToString(splitter.stringToBigDecimal("big str"));
     assertEquals("big str", out);
   }
 
-  @Test
   public void testConvertChomped() {
     TextSplitter splitter = new TextSplitter();
     String out = splitter.bigDecimalToString(splitter.stringToBigDecimal("AVeryLongStringIndeed"));
     assertEquals("AVeryLon", out);
   }
 
-  @Test
   public void testAlphabetSplit() throws SQLException {
     // This should give us 25 splits, one per letter.
     TextSplitter splitter = new TextSplitter();
@@ -122,7 +113,6 @@ public class TestTextSplitter {
     assertArrayEquals(expected, splits.toArray(new String [0]));
   }
 
-  @Test
   public void testCommonPrefix() throws SQLException {
     // Splits between 'Hand' and 'Hardy'
     TextSplitter splitter = new TextSplitter();

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/fieldsel/TestMRFieldSelection.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/fieldsel/TestMRFieldSelection.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/fieldsel/TestMRFieldSelection.java
index 6f9183a..91070f8 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/fieldsel/TestMRFieldSelection.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/fieldsel/TestMRFieldSelection.java
@@ -18,19 +18,15 @@
 package org.apache.hadoop.mapreduce.lib.fieldsel;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.Text;
+import org.apache.hadoop.fs.*;
+import org.apache.hadoop.io.*;
 import org.apache.hadoop.mapreduce.Job;
 import org.apache.hadoop.mapreduce.MapReduceTestUtil;
-import org.junit.Test;
 
+import junit.framework.TestCase;
 import java.text.NumberFormat;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-
-public class TestMRFieldSelection {
+public class TestMRFieldSelection extends TestCase {
 
 private static NumberFormat idFormat = NumberFormat.getInstance();
   static {
@@ -38,7 +34,6 @@ private static NumberFormat idFormat = NumberFormat.getInstance();
     idFormat.setGroupingUsed(false);
   }
 
-  @Test
   public void testFieldSelection() throws Exception {
     launch();
   }
@@ -119,4 +114,11 @@ private static NumberFormat idFormat = NumberFormat.getInstance();
     System.out.println("ExpectedData:");
     System.out.println(expectedOutput.toString());
   }
+  
+  /**
+   * Launches all the tasks in order.
+   */
+  public static void main(String[] argv) throws Exception {
+    launch();
+  }
 }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsBinaryInputFormat.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsBinaryInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsBinaryInputFormat.java
index cbf9d18..f0b3d57 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsBinaryInputFormat.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsBinaryInputFormat.java
@@ -18,12 +18,11 @@
 
 package org.apache.hadoop.mapreduce.lib.input;
 
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.BytesWritable;
-import org.apache.hadoop.io.DataInputBuffer;
-import org.apache.hadoop.io.SequenceFile;
-import org.apache.hadoop.io.Text;
+import java.io.IOException;
+import java.util.Random;
+
+import org.apache.hadoop.fs.*;
+import org.apache.hadoop.io.*;
 import org.apache.hadoop.mapreduce.InputFormat;
 import org.apache.hadoop.mapreduce.InputSplit;
 import org.apache.hadoop.mapreduce.Job;
@@ -32,18 +31,12 @@ import org.apache.hadoop.mapreduce.MapReduceTestUtil;
 import org.apache.hadoop.mapreduce.RecordReader;
 import org.apache.hadoop.mapreduce.TaskAttemptContext;
 import org.apache.hadoop.mapreduce.task.MapContextImpl;
-import org.junit.Test;
-
-import java.io.IOException;
-import java.util.Random;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
+import junit.framework.TestCase;
 
-public class TestMRSequenceFileAsBinaryInputFormat {
+public class TestMRSequenceFileAsBinaryInputFormat extends TestCase {
   private static final int RECORDS = 10000;
 
-  @Test
   public void testBinary() throws IOException, InterruptedException {
     Job job = Job.getInstance();
     FileSystem fs = FileSystem.getLocal(job.getConfiguration());

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsTextInputFormat.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsTextInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsTextInputFormat.java
index 335ce05..2d03c2d 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsTextInputFormat.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileAsTextInputFormat.java
@@ -18,13 +18,11 @@
 
 package org.apache.hadoop.mapreduce.lib.input;
 
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.IntWritable;
-import org.apache.hadoop.io.LongWritable;
-import org.apache.hadoop.io.SequenceFile;
-import org.apache.hadoop.io.Text;
+import java.util.*;
+import junit.framework.TestCase;
+
+import org.apache.hadoop.fs.*;
+import org.apache.hadoop.io.*;
 import org.apache.hadoop.mapreduce.InputFormat;
 import org.apache.hadoop.mapreduce.InputSplit;
 import org.apache.hadoop.mapreduce.Job;
@@ -33,19 +31,12 @@ import org.apache.hadoop.mapreduce.MapReduceTestUtil;
 import org.apache.hadoop.mapreduce.RecordReader;
 import org.apache.hadoop.mapreduce.TaskAttemptContext;
 import org.apache.hadoop.mapreduce.task.MapContextImpl;
-import org.junit.Test;
-
-import java.util.BitSet;
-import java.util.Random;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
+import org.apache.hadoop.conf.*;
 
-public class TestMRSequenceFileAsTextInputFormat {
+public class TestMRSequenceFileAsTextInputFormat extends TestCase {
   private static int MAX_LENGTH = 10000;
   private static Configuration conf = new Configuration();
 
-  @Test
   public void testFormat() throws Exception {
     Job job = Job.getInstance(conf);
     FileSystem fs = FileSystem.getLocal(conf);
@@ -121,4 +112,8 @@ public class TestMRSequenceFileAsTextInputFormat {
 
     }
   }
+
+  public static void main(String[] args) throws Exception {
+    new TestMRSequenceFileAsTextInputFormat().testFormat();
+  }
 }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileInputFilter.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileInputFilter.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileInputFilter.java
index 89aa7b2..edf7e1a 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileInputFilter.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestMRSequenceFileInputFilter.java
@@ -18,14 +18,14 @@
 
 package org.apache.hadoop.mapreduce.lib.input;
 
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.BytesWritable;
-import org.apache.hadoop.io.SequenceFile;
-import org.apache.hadoop.io.Text;
+import java.io.*;
+import java.util.*;
+import junit.framework.TestCase;
+
+import org.apache.commons.logging.*;
+
+import org.apache.hadoop.fs.*;
+import org.apache.hadoop.io.*;
 import org.apache.hadoop.mapreduce.InputFormat;
 import org.apache.hadoop.mapreduce.InputSplit;
 import org.apache.hadoop.mapreduce.Job;
@@ -34,15 +34,10 @@ import org.apache.hadoop.mapreduce.MapReduceTestUtil;
 import org.apache.hadoop.mapreduce.RecordReader;
 import org.apache.hadoop.mapreduce.TaskAttemptContext;
 import org.apache.hadoop.mapreduce.task.MapContextImpl;
-import org.junit.Test;
-
-import java.io.IOException;
-import java.util.Random;
+import org.apache.hadoop.conf.*;
 
-import static org.junit.Assert.assertEquals;
-
-public class TestMRSequenceFileInputFilter {
-  private static final Log LOG =
+public class TestMRSequenceFileInputFilter extends TestCase {
+  private static final Log LOG = 
     LogFactory.getLog(TestMRSequenceFileInputFilter.class.getName());
 
   private static final int MAX_LENGTH = 15000;
@@ -118,8 +113,7 @@ public class TestMRSequenceFileInputFilter {
     }
     return count;
   }
-
-  @Test
+  
   public void testRegexFilter() throws Exception {
     // set the filter class
     LOG.info("Testing Regex Filter with patter: \\A10*");
@@ -144,7 +138,6 @@ public class TestMRSequenceFileInputFilter {
     fs.delete(inDir, true);
   }
 
-  @Test
   public void testPercentFilter() throws Exception {
     LOG.info("Testing Percent Filter with frequency: 1000");
     // set the filter class
@@ -172,8 +165,7 @@ public class TestMRSequenceFileInputFilter {
     // clean up
     fs.delete(inDir, true);
   }
-
-  @Test
+  
   public void testMD5Filter() throws Exception {
     // set the filter class
     LOG.info("Testing MD5 Filter with frequency: 1000");
@@ -195,4 +187,9 @@ public class TestMRSequenceFileInputFilter {
     // clean up
     fs.delete(inDir, true);
   }
+
+  public static void main(String[] args) throws Exception {
+    TestMRSequenceFileInputFilter filter = new TestMRSequenceFileInputFilter();
+    filter.testRegexFilter();
+  }
 }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestNLineInputFormat.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestNLineInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestNLineInputFormat.java
index 477866f..7b3878d 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestNLineInputFormat.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestNLineInputFormat.java
@@ -18,28 +18,17 @@
 
 package org.apache.hadoop.mapreduce.lib.input;
 
+import java.io.*;
+import java.util.*;
+import junit.framework.TestCase;
+
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.io.LongWritable;
-import org.apache.hadoop.io.Text;
-import org.apache.hadoop.mapreduce.InputSplit;
-import org.apache.hadoop.mapreduce.Job;
-import org.apache.hadoop.mapreduce.MapContext;
-import org.apache.hadoop.mapreduce.MapReduceTestUtil;
-import org.apache.hadoop.mapreduce.RecordReader;
-import org.apache.hadoop.mapreduce.TaskAttemptContext;
+import org.apache.hadoop.fs.*;
+import org.apache.hadoop.io.*;
+import org.apache.hadoop.mapreduce.*;
 import org.apache.hadoop.mapreduce.task.MapContextImpl;
-import org.junit.Test;
-
-import java.io.IOException;
-import java.io.OutputStreamWriter;
-import java.io.Writer;
-import java.util.List;
 
-import static org.junit.Assert.assertEquals;
-
-public class TestNLineInputFormat {
+public class TestNLineInputFormat extends TestCase {
   private static int MAX_LENGTH = 200;
   
   private static Configuration conf = new Configuration();
@@ -56,8 +45,7 @@ public class TestNLineInputFormat {
   private static Path workDir = 
     new Path(new Path(System.getProperty("test.build.data", "."), "data"),
              "TestNLineInputFormat");
-
-  @Test
+  
   public void testFormat() throws Exception {
     Job job = Job.getInstance(conf);
     Path file = new Path(workDir, "test.txt");
@@ -128,4 +116,8 @@ public class TestNLineInputFormat {
       }
     }
   }
+  
+  public static void main(String[] args) throws Exception {
+    new TestNLineInputFormat().testFormat();
+  }
 }

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinDatamerge.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinDatamerge.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinDatamerge.java
index 1173ea4..d245bfd 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinDatamerge.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinDatamerge.java
@@ -19,6 +19,11 @@ package org.apache.hadoop.mapreduce.lib.join;
 
 import java.io.IOException;
 
+import junit.framework.Test;
+import junit.framework.TestCase;
+import junit.framework.TestSuite;
+import junit.extensions.TestSetup;
+
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileStatus;
 import org.apache.hadoop.fs.Path;
@@ -32,31 +37,23 @@ import org.apache.hadoop.mapreduce.*;
 import org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat;
 import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
 import org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat;
-import org.junit.After;
-import org.junit.AfterClass;
-import org.junit.Before;
-import org.junit.BeforeClass;
-import org.junit.Test;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
 
-public class TestJoinDatamerge {
+public class TestJoinDatamerge extends TestCase {
 
   private static MiniDFSCluster cluster = null;
-
-  @BeforeClass
-  public static void setUp() throws Exception {
-    Configuration conf = new Configuration();
-    cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
-  }
-
-  @AfterClass
-  public static void tearDown() throws Exception {
-    if (cluster != null) {
-      cluster.shutdown();
-    }
+  public static Test suite() {
+    TestSetup setup = new TestSetup(new TestSuite(TestJoinDatamerge.class)) {
+      protected void setUp() throws Exception {
+        Configuration conf = new Configuration();
+        cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
+      }
+      protected void tearDown() throws Exception {
+        if (cluster != null) {
+          cluster.shutdown();
+        }
+      }
+    };
+    return setup;
   }
 
   private static SequenceFile.Writer[] createWriters(Path testdir,
@@ -114,7 +111,7 @@ public class TestJoinDatamerge {
       extends Mapper<IntWritable, V, IntWritable, IntWritable>{
     protected final static IntWritable one = new IntWritable(1);
     int srcs;
-
+    
     public void setup(Context context) {
       srcs = context.getConfiguration().getInt("testdatamerge.sources", 0);
       assertTrue("Invalid src count: " + srcs, srcs > 0);
@@ -126,7 +123,7 @@ public class TestJoinDatamerge {
     protected final static IntWritable one = new IntWritable(1);
 
     int srcs;
-
+    
     public void setup(Context context) {
       srcs = context.getConfiguration().getInt("testdatamerge.sources", 0);
       assertTrue("Invalid src count: " + srcs, srcs > 0);
@@ -275,12 +272,10 @@ public class TestJoinDatamerge {
     base.getFileSystem(conf).delete(base, true);
   }
 
-  @Test
   public void testSimpleInnerJoin() throws Exception {
     joinAs("inner", InnerJoinMapChecker.class, InnerJoinReduceChecker.class);
   }
 
-  @Test
   public void testSimpleOuterJoin() throws Exception {
     joinAs("outer", OuterJoinMapChecker.class, OuterJoinReduceChecker.class);
   }
@@ -327,13 +322,11 @@ public class TestJoinDatamerge {
     }
     return product;
   }
-
-  @Test
+  
   public void testSimpleOverride() throws Exception {
     joinAs("override", OverrideMapChecker.class, OverrideReduceChecker.class);
   }
 
-  @Test
   public void testNestedJoin() throws Exception {
     // outer(inner(S1,...,Sn),outer(S1,...Sn))
     final int SOURCES = 3;
@@ -429,7 +422,6 @@ public class TestJoinDatamerge {
 
   }
 
-  @Test
   public void testEmptyJoin() throws Exception {
     Configuration conf = new Configuration();
     Path base = cluster.getFileSystem().makeQualified(new Path("/empty"));

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinProperties.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinProperties.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinProperties.java
index b6e7606..151bc87 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinProperties.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinProperties.java
@@ -20,6 +20,11 @@ package org.apache.hadoop.mapreduce.lib.join;
 import java.io.IOException;
 import java.util.List;
 
+import junit.framework.Test;
+import junit.framework.TestCase;
+import junit.framework.TestSuite;
+import junit.extensions.TestSetup;
+
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hdfs.MiniDFSCluster;
@@ -31,14 +36,8 @@ import org.apache.hadoop.io.WritableComparable;
 import org.apache.hadoop.mapreduce.*;
 import org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat;
 import org.apache.hadoop.mapreduce.task.MapContextImpl;
-import org.junit.AfterClass;
-import org.junit.BeforeClass;
-import org.junit.Test;
-
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
 
-public class TestJoinProperties {
+public class TestJoinProperties extends TestCase {
 
   private static MiniDFSCluster cluster = null;
   final static int SOURCES = 3;
@@ -47,19 +46,21 @@ public class TestJoinProperties {
   static Path[] src;
   static Path base;
 
-  @BeforeClass
-  public static void setUp() throws Exception {
-    Configuration conf = new Configuration();
-    cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
-    base = cluster.getFileSystem().makeQualified(new Path("/nested"));
-    src = generateSources(conf);
-  }
-
-  @AfterClass
-  public static void tearDown() throws Exception {
-    if (cluster != null) {
-      cluster.shutdown();
-    }
+  public static Test suite() {
+    TestSetup setup = new TestSetup(new TestSuite(TestJoinProperties.class)) {
+      protected void setUp() throws Exception {
+        Configuration conf = new Configuration();
+        cluster = new MiniDFSCluster.Builder(conf).numDataNodes(2).build();
+        base = cluster.getFileSystem().makeQualified(new Path("/nested"));
+        src = generateSources(conf);
+      }
+      protected void tearDown() throws Exception {
+        if (cluster != null) {
+          cluster.shutdown();
+        }
+      }
+    };
+    return setup;
   }
 
   // Sources from 0 to srcs-2 have IntWritable key and IntWritable value
@@ -232,7 +233,6 @@ public class TestJoinProperties {
   }
 
   // outer(outer(A, B), C) == outer(A,outer(B, C)) == outer(A, B, C)
-  @Test
   public void testOuterAssociativity() throws Exception {
     Configuration conf = new Configuration();
     testExpr1(conf, "outer", TestType.OUTER_ASSOCIATIVITY, 33);
@@ -241,7 +241,6 @@ public class TestJoinProperties {
   }
  
   // inner(inner(A, B), C) == inner(A,inner(B, C)) == inner(A, B, C)
-  @Test
   public void testInnerAssociativity() throws Exception {
     Configuration conf = new Configuration();
     testExpr1(conf, "inner", TestType.INNER_ASSOCIATIVITY, 2);
@@ -250,7 +249,6 @@ public class TestJoinProperties {
   }
 
   // override(inner(A, B), A) == A
-  @Test
   public void testIdentity() throws Exception {
     Configuration conf = new Configuration();
     testExpr4(conf);

http://git-wip-us.apache.org/repos/asf/hadoop/blob/115be193/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinTupleWritable.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinTupleWritable.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinTupleWritable.java
index 093da26..d35941f 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinTupleWritable.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/join/TestJoinTupleWritable.java
@@ -24,6 +24,8 @@ import java.io.DataOutputStream;
 import java.util.Arrays;
 import java.util.Random;
 
+import junit.framework.TestCase;
+
 import org.apache.hadoop.io.BooleanWritable;
 import org.apache.hadoop.io.BytesWritable;
 import org.apache.hadoop.io.FloatWritable;
@@ -31,13 +33,8 @@ import org.apache.hadoop.io.IntWritable;
 import org.apache.hadoop.io.LongWritable;
 import org.apache.hadoop.io.Text;
 import org.apache.hadoop.io.Writable;
-import org.junit.Test;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
 
-public class TestJoinTupleWritable {
+public class TestJoinTupleWritable extends TestCase {
 
   private TupleWritable makeTuple(Writable[] writs) {
     Writable[] sub1 = { writs[1], writs[2] };
@@ -100,7 +97,6 @@ public class TestJoinTupleWritable {
     return i;
   }
 
-  @Test
   public void testIterable() throws Exception {
     Random r = new Random();
     Writable[] writs = {
@@ -122,7 +118,6 @@ public class TestJoinTupleWritable {
     verifIter(writs, t, 0);
   }
 
-  @Test
   public void testNestedIterable() throws Exception {
     Random r = new Random();
     Writable[] writs = {
@@ -141,7 +136,6 @@ public class TestJoinTupleWritable {
     assertTrue("Bad count", writs.length == verifIter(writs, sTuple, 0));
   }
 
-  @Test
   public void testWritable() throws Exception {
     Random r = new Random();
     Writable[] writs = {
@@ -165,7 +159,6 @@ public class TestJoinTupleWritable {
     assertTrue("Failed to write/read tuple", sTuple.equals(dTuple));
   }
 
-  @Test
   public void testWideWritable() throws Exception {
     Writable[] manyWrits = makeRandomWritables(131);
     
@@ -185,8 +178,7 @@ public class TestJoinTupleWritable {
     assertEquals("All tuple data has not been read from the stream", 
       -1, in.read());
   }
-
-  @Test
+  
   public void testWideWritable2() throws Exception {
     Writable[] manyWrits = makeRandomWritables(71);
     
@@ -209,7 +201,6 @@ public class TestJoinTupleWritable {
    * Tests a tuple writable with more than 64 values and the values set written
    * spread far apart.
    */
-  @Test
   public void testSparseWideWritable() throws Exception {
     Writable[] manyWrits = makeRandomWritables(131);
     
@@ -229,8 +220,7 @@ public class TestJoinTupleWritable {
     assertEquals("All tuple data has not been read from the stream", 
       -1, in.read());
   }
-
-  @Test
+  
   public void testWideTuple() throws Exception {
     Text emptyText = new Text("Should be empty");
     Writable[] values = new Writable[64];
@@ -251,8 +241,7 @@ public class TestJoinTupleWritable {
       }
     }
   }
-
-  @Test
+  
   public void testWideTuple2() throws Exception {
     Text emptyText = new Text("Should be empty");
     Writable[] values = new Writable[64];
@@ -277,7 +266,6 @@ public class TestJoinTupleWritable {
   /**
    * Tests that we can write more than 64 values.
    */
-  @Test
   public void testWideTupleBoundary() throws Exception {
     Text emptyText = new Text("Should not be set written");
     Writable[] values = new Writable[65];