You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by ap...@apache.org on 2013/04/30 00:11:46 UTC
svn commit: r1477375 - in
/hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase:
HBaseTestingUtility.java mapreduce/TestImportExport.java
Author: apurtell
Date: Mon Apr 29 22:11:46 2013
New Revision: 1477375
URL: http://svn.apache.org/r1477375
Log:
HBASE-8453. TestImportExport failing again due to configuration issues
Modified:
hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java
Modified: hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java?rev=1477375&r1=1477374&r2=1477375&view=diff
==============================================================================
--- hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java (original)
+++ hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java Mon Apr 29 22:11:46 2013
@@ -1663,6 +1663,10 @@ public class HBaseTestingUtility {
this.dfsCluster = cluster;
}
+ public MiniMRCluster getMRCluster() {
+ return mrCluster;
+ }
+
public FileSystem getTestFileSystem() throws IOException {
return HFileSystem.get(conf);
}
Modified: hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java?rev=1477375&r1=1477374&r2=1477375&view=diff
==============================================================================
--- hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java (original)
+++ hbase/branches/0.94/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java Mon Apr 29 22:11:46 2013
@@ -26,6 +26,7 @@ import java.io.IOException;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HBaseTestingUtility;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HTableDescriptor;
@@ -42,6 +43,7 @@ import org.apache.hadoop.hbase.client.Sc
import org.apache.hadoop.hbase.filter.Filter;
import org.apache.hadoop.hbase.filter.PrefixFilter;
import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.util.GenericOptionsParser;
import org.junit.After;
@@ -64,12 +66,11 @@ public class TestImportExport {
private static final byte[] QUAL = Bytes.toBytes("q");
private static final String OUTPUT_DIR = "outputdir";
- private static MiniHBaseCluster cluster;
private static long now = System.currentTimeMillis();
@BeforeClass
public static void beforeClass() throws Exception {
- cluster = UTIL.startMiniCluster();
+ UTIL.startMiniCluster();
UTIL.startMiniMapReduceCluster();
}
@@ -111,16 +112,16 @@ public class TestImportExport {
"1000"
};
- GenericOptionsParser opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+ GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
Configuration conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- Job job = Export.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ JobConf jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ Job job = Export.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
assertTrue(job.isSuccessful());
-
String IMPORT_TABLE = "importTableSimpleCase";
t = UTIL.createTable(Bytes.toBytes(IMPORT_TABLE), FAMILYB);
args = new String[] {
@@ -129,12 +130,13 @@ public class TestImportExport {
OUTPUT_DIR
};
- opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+ opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- job = Import.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ job = Import.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
assertTrue(job.isSuccessful());
@@ -158,12 +160,13 @@ public class TestImportExport {
String EXPORT_TABLE = ".META.";
String[] args = new String[] { EXPORT_TABLE, OUTPUT_DIR, "1", "0", "0" };
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(
- cluster.getConfiguration()), args);
+ UTIL.getConfiguration()), args);
Configuration conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- Job job = Export.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ JobConf jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ Job job = Export.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
assertTrue(job.isSuccessful());
}
@@ -200,12 +203,13 @@ public class TestImportExport {
"1000"
};
- GenericOptionsParser opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+ GenericOptionsParser opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
Configuration conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- Job job = Export.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ JobConf jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ Job job = Export.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
assertTrue(job.isSuccessful());
@@ -224,12 +228,13 @@ public class TestImportExport {
OUTPUT_DIR
};
- opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+ opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- job = Import.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ job = Import.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
assertTrue(job.isSuccessful());
@@ -268,12 +273,13 @@ public class TestImportExport {
String[] args = new String[] { EXPORT_TABLE, OUTPUT_DIR, "1000" };
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(
- cluster.getConfiguration()), args);
+ UTIL.getConfiguration()), args);
Configuration conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- Job job = Export.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ JobConf jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ Job job = Export.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
assertTrue(job.isSuccessful());
@@ -287,12 +293,13 @@ public class TestImportExport {
"-D" + Import.FILTER_ARGS_CONF_KEY + "=" + Bytes.toString(ROW1), IMPORT_TABLE, OUTPUT_DIR,
"1000" };
- opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+ opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- job = Import.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ job = Import.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
assertTrue(job.isSuccessful());
@@ -310,14 +317,15 @@ public class TestImportExport {
"-D" + Import.FILTER_ARGS_CONF_KEY + "=" + Bytes.toString(ROW1) + "", EXPORT_TABLE,
OUTPUT_DIR, "1000" };
- opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
+ opts = new GenericOptionsParser(new Configuration(UTIL.getConfiguration()), args);
conf = opts.getConfiguration();
args = opts.getRemainingArgs();
- job = Import.createSubmittableJob(conf, args);
- job.getConfiguration().set("mapreduce.framework.name", "yarn");
+ jobConf = UTIL.getMRCluster().createJobConf();
+ HBaseConfiguration.merge(jobConf, conf);
+ job = Import.createSubmittableJob(jobConf, args);
job.waitForCompletion(false);
- assertFalse("Job succeeedd, but it had a non-instantiable filter!", job.isSuccessful());
+ assertFalse("Job succeeded, but it had a non-instantiable filter!", job.isSuccessful());
// cleanup
exportTable.close();