You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by ap...@apache.org on 2013/04/30 00:11:46 UTC

svn commit: r1477375 - in /hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase: HBaseTestingUtility.java mapreduce/TestImportExport.java

Author: apurtell
Date: Mon Apr 29 22:11:46 2013
New Revision: 1477375

URL: http://svn.apache.org/r1477375
Log:
HBASE-8453. TestImportExport failing again due to configuration issues

Modified:
    hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
    hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java

Modified: hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java?rev=1477375&r1=1477374&r2=1477375&view=diff
==============================================================================
--- hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java (original)
+++ hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java Mon Apr 29 22:11:46 2013
@@ -1663,6 +1663,10 @@ public class HBaseTestingUtility {
     this.dfsCluster = cluster;
   }
 
+  public MiniMRCluster getMRCluster() {
+    return mrCluster;
+  }
+
   public FileSystem getTestFileSystem() throws IOException {
     return HFileSystem.get(conf);
   }

Modified: hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java?rev=1477375&r1=1477374&r2=1477375&view=diff
==============================================================================
--- hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java (original)
+++ hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java Mon Apr 29 22:11:46 2013
@@ -26,6 +26,7 @@ import java.io.IOException;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hbase.HBaseConfiguration;
 import org.apache.hadoop.hbase.HBaseTestingUtility;
 import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HTableDescriptor;
@@ -42,6 +43,7 @@ import org.apache.hadoop.hbase.client.Sc
 import org.apache.hadoop.hbase.filter.Filter;
 import org.apache.hadoop.hbase.filter.PrefixFilter;
 import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.hadoop.mapred.JobConf;
 import org.apache.hadoop.mapreduce.Job;
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.junit.After;
@@ -64,12 +66,11 @@ public class TestImportExport {
   private static final byte[] QUAL = Bytes.toBytes("q");
   private static final String OUTPUT_DIR = "outputdir";
 
-  private static MiniHBaseCluster cluster;
   private static long now = System.currentTimeMillis();
 
   @BeforeClass
   public static void beforeClass() throws Exception {
-    cluster = UTIL.startMiniCluster();
+    UTIL.startMiniCluster();
     UTIL.startMiniMapReduceCluster();
   }
 
@@ -111,16 +112,16 @@ public class TestImportExport {
         "1000"
     };
 
-    GenericOptionsParser opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+    GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
     Configuration conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    Job job = Export.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    JobConf jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    Job job = Export.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
     assertTrue(job.isSuccessful());
 
-
     String IMPORT_TABLE = "importTableSimpleCase";
     t = UTIL.createTable(Bytes.toBytes(IMPORT_TABLE), FAMILYB);
     args = new String[] {
@@ -129,12 +130,13 @@ public class TestImportExport {
         OUTPUT_DIR
     };
 
-    opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+    opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
     conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    job = Import.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    job = Import.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
     assertTrue(job.isSuccessful());
 
@@ -158,12 +160,13 @@ public class TestImportExport {
     String EXPORT_TABLE = ".META.";
     String[] args = new String[] { EXPORT_TABLE, OUTPUT_DIR, "1", "0", "0" };
     GenericOptionsParser opts = new GenericOptionsParser(new Configuration(
-        cluster.getConfiguration()), args);
+        UTIL.getConfiguration()), args);
     Configuration conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    Job job = Export.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    JobConf jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    Job job = Export.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
     assertTrue(job.isSuccessful());
   }
@@ -200,12 +203,13 @@ public class TestImportExport {
         "1000"
     };
 
-    GenericOptionsParser opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+    GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
     Configuration conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    Job job = Export.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    JobConf jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    Job job = Export.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
     assertTrue(job.isSuccessful());
 
@@ -224,12 +228,13 @@ public class TestImportExport {
         OUTPUT_DIR
     };
 
-    opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+    opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
     conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    job = Import.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    job = Import.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
     assertTrue(job.isSuccessful());
 
@@ -268,12 +273,13 @@ public class TestImportExport {
     String[] args = new String[] { EXPORT_TABLE, OUTPUT_DIR, "1000" };
 
     GenericOptionsParser opts = new GenericOptionsParser(new Configuration(
-        cluster.getConfiguration()), args);
+        UTIL.getConfiguration()), args);
     Configuration conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    Job job = Export.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    JobConf jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    Job job = Export.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
     assertTrue(job.isSuccessful());
 
@@ -287,12 +293,13 @@ public class TestImportExport {
         "-D" + Import.FILTER_ARGS_CONF_KEY + "=" + Bytes.toString(ROW1), IMPORT_TABLE, OUTPUT_DIR,
         "1000" };
 
-    opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+    opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
     conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    job = Import.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    job = Import.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
     assertTrue(job.isSuccessful());
 
@@ -310,14 +317,15 @@ public class TestImportExport {
         "-D" + Import.FILTER_ARGS_CONF_KEY + "=" + Bytes.toString(ROW1) + "", EXPORT_TABLE,
         OUTPUT_DIR, "1000" };
 
-    opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+    opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
     conf = opts.getConfiguration();
     args = opts.getRemainingArgs();
 
-    job = Import.createSubmittableJob(conf, args);
-    job.getConfiguration().set("mapreduce.framework.name", "yarn");
+    jobConf = UTIL.getMRCluster().createJobConf();
+    HBaseConfiguration.merge(jobConf, conf);
+    job = Import.createSubmittableJob(jobConf, args);
     job.waitForCompletion(false);
-    assertFalse("Job succeeedd, but it had a non-instantiable filter!", job.isSuccessful());
+    assertFalse("Job succeeded, but it had a non-instantiable filter!", job.isSuccessful());
 
     // cleanup
     exportTable.close();