You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by jm...@apache.org on 2012/04/25 02:07:41 UTC

svn commit: r1330072 - /hbase/trunk/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHLogRecordReader.java

Author: jmhsieh
Date: Wed Apr 25 00:07:40 2012
New Revision: 1330072

URL: http://svn.apache.org/viewvc?rev=1330072&view=rev
Log:
HBASE-5861 Hadoop 23 compile broken due to tests introduced in HBASE-5064

Modified:
    hbase/trunk/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHLogRecordReader.java

Modified: hbase/trunk/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHLogRecordReader.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHLogRecordReader.java?rev=1330072&r1=1330071&r2=1330072&view=diff
==============================================================================
--- hbase/trunk/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHLogRecordReader.java (original)
+++ hbase/trunk/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHLogRecordReader.java Wed Apr 25 00:07:40 2012
@@ -18,8 +18,8 @@
 package org.apache.hadoop.hbase.mapreduce;
 
 import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
 import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
 
 import java.util.List;
 
@@ -40,9 +40,7 @@ import org.apache.hadoop.hbase.regionser
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.mapreduce.InputSplit;
 import org.apache.hadoop.mapreduce.JobContext;
-import org.apache.hadoop.mapreduce.JobID;
-import org.apache.hadoop.mapreduce.TaskAttemptContext;
-import org.apache.hadoop.mapreduce.TaskAttemptID;
+import org.apache.hadoop.mapreduce.MapReduceTestUtil;
 import org.junit.AfterClass;
 import org.junit.Before;
 import org.junit.BeforeClass;
@@ -144,13 +142,13 @@ public class TestHLogRecordReader {
     jobConf.setLong(HLogInputFormat.END_TIME_KEY, ts);
 
     // only 1st file is considered, and only its 1st entry is used
-    List<InputSplit> splits = input.getSplits(new JobContext(jobConf, new JobID()));
+    List<InputSplit> splits = input.getSplits(MapreduceTestingShim.createJobContext(jobConf));
     assertEquals(1, splits.size());
     testSplit(splits.get(0), Bytes.toBytes("1"));
 
     jobConf.setLong(HLogInputFormat.START_TIME_KEY, ts+1);
     jobConf.setLong(HLogInputFormat.END_TIME_KEY, ts1+1);
-    splits = input.getSplits(new JobContext(jobConf, new JobID()));
+    splits = input.getSplits(MapreduceTestingShim.createJobContext(jobConf));
     // both files need to be considered
     assertEquals(2, splits.size());
     // only the 2nd entry from the 1st file is used
@@ -191,7 +189,7 @@ public class TestHLogRecordReader {
     jobConf.set("mapred.input.dir", logDir.toString());
 
     // make sure both logs are found
-    List<InputSplit> splits = input.getSplits(new JobContext(jobConf, new JobID()));
+    List<InputSplit> splits = input.getSplits(MapreduceTestingShim.createJobContext(jobConf));
     assertEquals(2, splits.size());
 
     // should return exactly one KV
@@ -203,14 +201,14 @@ public class TestHLogRecordReader {
 
     // set an endtime, the 2nd log file can be ignored completely.
     jobConf.setLong(HLogInputFormat.END_TIME_KEY, secondTs-1);
-    splits = input.getSplits(new JobContext(jobConf, new JobID()));
+    splits = input.getSplits(MapreduceTestingShim.createJobContext(jobConf));
     assertEquals(1, splits.size());
     testSplit(splits.get(0), Bytes.toBytes("1"));
 
     // now set a start time
     jobConf.setLong(HLogInputFormat.END_TIME_KEY, Long.MAX_VALUE);
     jobConf.setLong(HLogInputFormat.START_TIME_KEY, thirdTs);
-    splits = input.getSplits(new JobContext(jobConf, new JobID()));
+    splits = input.getSplits(MapreduceTestingShim.createJobContext(jobConf));
     // both logs need to be considered
     assertEquals(2, splits.size());
     // but both readers skip all edits
@@ -223,7 +221,7 @@ public class TestHLogRecordReader {
    */
   private void testSplit(InputSplit split, byte[]... columns) throws Exception {
     HLogRecordReader reader = new HLogRecordReader();
-    reader.initialize(split, new TaskAttemptContext(conf, new TaskAttemptID()));
+    reader.initialize(split, MapReduceTestUtil.createDummyMapTaskAttemptContext(conf));
 
     for (byte[] column : columns) {
       assertTrue(reader.nextKeyValue());