You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@bigtop.apache.org by rv...@apache.org on 2017/03/23 17:27:45 UTC

[14/50] [abbrv] bigtop git commit: A few other last minute changes to make it sort of work with hard wired values.

A few other last minute changes to make it sort of work with hard wired values.

(cherry picked from commit 9bac863d4181a7bb964fa793770b359ac5cd97b0)


Project: http://git-wip-us.apache.org/repos/asf/bigtop/repo
Commit: http://git-wip-us.apache.org/repos/asf/bigtop/commit/ab4f414a
Tree: http://git-wip-us.apache.org/repos/asf/bigtop/tree/ab4f414a
Diff: http://git-wip-us.apache.org/repos/asf/bigtop/diff/ab4f414a

Branch: refs/heads/master
Commit: ab4f414a945bcbf3924f4c4390b569b3b9175c0d
Parents: f9c6c65
Author: Alan Gates <ga...@hortonworks.com>
Authored: Mon Nov 7 16:39:01 2016 -0800
Committer: Roman Shaposhnik <rv...@apache.org>
Committed: Thu Mar 23 10:27:12 2017 -0700

----------------------------------------------------------------------
 .../org/odpi/specs/runtime/hive/HiveHelper.java | 12 ++-
 .../odpi/specs/runtime/hive/TestHCatalog.java   | 86 ++------------------
 2 files changed, 19 insertions(+), 79 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/bigtop/blob/ab4f414a/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/HiveHelper.java
----------------------------------------------------------------------
diff --git a/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/HiveHelper.java b/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/HiveHelper.java
index a4477ff..3e56224 100644
--- a/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/HiveHelper.java
+++ b/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/HiveHelper.java
@@ -36,8 +36,13 @@ import org.apache.commons.logging.LogFactory;
 public class HiveHelper {
 	
 	private static final Log LOG = LogFactory.getLog(HiveHelper.class.getName());
-		
+
 	public static Map<String, String> execCommand(CommandLine commandline) {
+		return execCommand(commandline, null);
+	}
+
+	public static Map<String, String> execCommand(CommandLine commandline,
+																								Map<String, String> envVars) {
 		
 		System.out.println("Executing command:");
 		System.out.println(commandline.toString());
@@ -50,6 +55,11 @@ public class HiveHelper {
 			LOG.debug("Failed to get process environment: "+ e1.getMessage());
 			e1.printStackTrace();
 		}
+		if (envVars != null) {
+			for (String key : envVars.keySet()) {
+				env.put(key, envVars.get(key));
+			}
+		}
 
 		DefaultExecuteResultHandler resultHandler = new DefaultExecuteResultHandler();
 		ByteArrayOutputStream outputStream = new ByteArrayOutputStream();

http://git-wip-us.apache.org/repos/asf/bigtop/blob/ab4f414a/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/TestHCatalog.java
----------------------------------------------------------------------
diff --git a/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/TestHCatalog.java b/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/TestHCatalog.java
index 4b61131..bb237d8 100644
--- a/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/TestHCatalog.java
+++ b/bigtop-tests/spec-tests/runtime/src/test/java/org/odpi/specs/runtime/hive/TestHCatalog.java
@@ -32,20 +32,8 @@ import org.apache.hadoop.hive.metastore.api.MetaException;
 import org.apache.hadoop.hive.metastore.api.SerDeInfo;
 import org.apache.hadoop.hive.metastore.api.StorageDescriptor;
 import org.apache.hadoop.hive.metastore.api.Table;
-import org.apache.hadoop.io.IntWritable;
-import org.apache.hadoop.io.Text;
-import org.apache.hadoop.io.WritableComparable;
-import org.apache.hadoop.mapreduce.Job;
-import org.apache.hadoop.mapreduce.Mapper;
-import org.apache.hadoop.mapreduce.Reducer;
-import org.apache.hive.hcatalog.data.DefaultHCatRecord;
-import org.apache.hive.hcatalog.data.HCatRecord;
 import org.apache.hive.hcatalog.data.schema.HCatFieldSchema;
 import org.apache.hive.hcatalog.data.schema.HCatSchema;
-import org.apache.hive.hcatalog.data.schema.HCatSchemaUtils;
-import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 import org.apache.thrift.TException;
 import org.junit.Assert;
 import org.junit.Assume;
@@ -54,14 +42,12 @@ import org.junit.BeforeClass;
 import org.junit.Test;
 
 import java.io.IOException;
-import java.net.URI;
 import java.net.URISyntaxException;
 import java.util.Arrays;
 import java.util.Collections;
 import java.util.HashMap;
 import java.util.Map;
 import java.util.Random;
-import java.util.StringTokenizer;
 
 
 public class TestHCatalog {
@@ -136,10 +122,6 @@ public class TestHCatalog {
     // Write some stuff into a file in the location of the table
     table = client.getTable("default", inputTable);
     String inputFile = table.getSd().getLocation() + "/input";
-    /*
-    String inputFile = JdbcConnector.getProperty(JdbcConnector.LOCATION,
-        "Directory to write a file in ") + "/odpi_hcat_input_" + rand.nextInt(Integer.MAX_VALUE);
-        */
     Path inputPath = new Path(inputFile);
     FileSystem fs = FileSystem.get(conf);
     FSDataOutputStream out = fs.create(inputPath);
@@ -149,76 +131,24 @@ public class TestHCatalog {
     out.writeChars("the lamb was sure to go\n");
     out.close();
 
-    Map<String, String> results = HiveHelper.execCommand(new CommandLine("hadoop")
+    Map<String, String> env = new HashMap<>();
+    env.put("HADOOP_HOME","/Users/gates/grid/odpi-testing/hadoop-2.7.3");
+    env.put("HADOOP_CLASSPATH", "/Users/gates/grid/odpi-testing/apache-hive-1.2.1-bin/hcatalog/share/hcatalog/hive-hcatalog-core-1.2.1.jar");
+    env.put("HIVE_HOME", "/Users/gates/grid/odpi-testing/apache-hive-1.2.1-bin");
+    Map<String, String> results = HiveHelper.execCommand(new CommandLine("/Users/gates/grid/odpi-testing/apache-hive-1.2.1-bin/bin/hive")
+        .addArgument("--service")
         .addArgument("jar")
         .addArgument("/Users/gates/git/bigtop/runtime-1.2.0-SNAPSHOT.jar")
         .addArgument(HCatalogMR.class.getName())
         .addArgument(inputTable)
         .addArgument(outputTable)
         .addArgument(inputSchema.getSchemaAsTypeString())
-        .addArgument(outputSchema.getSchemaAsTypeString()));
+        .addArgument(outputSchema.getSchemaAsTypeString()), env);
+    LOG.info(results.toString());
     Assert.assertEquals("HCat job failed", 0, Integer.parseInt(results.get("exitValue")));
 
-
-
-    /*
-    Job job = new Job(conf, "odpi_hcat_test");
-    HCatInputFormat.setInput(job, "default", inputTable);
-
-    job.setInputFormatClass(HCatInputFormat.class);
-    job.setJarByClass(TestHCatalog.class);
-    job.setMapperClass(Map.class);
-    job.setReducerClass(Reduce.class);
-    job.setMapOutputKeyClass(Text.class);
-    job.setMapOutputValueClass(IntWritable.class);
-    job.setOutputKeyClass(WritableComparable.class);
-    job.setOutputValueClass(HCatRecord.class);
-    HCatOutputFormat.setOutput(job, OutputJobInfo.create("default", outputTable, null));
-    HCatOutputFormat.setSchema(job, outputSchema);
-    job.setOutputFormatClass(HCatOutputFormat.class);
-
-    job.addCacheArchive(new URI("hdfs:/user/gates/hive-hcatalog-core-1.2.1.jar"));
-    job.addCacheArchive(new URI("hdfs:/user/gates/hive-metastore-1.2.1.jar"));
-    job.addCacheArchive(new URI("hdfs:/user/gates/hive-exec-1.2.1.jar"));
-
-    Assert.assertTrue(job.waitForCompletion(true));
-    */
-
     client.dropTable("default", inputTable);
     client.dropTable("default", outputTable);
   }
 
-  /*
-  public static class Map extends Mapper<WritableComparable,
-        HCatRecord, Text, IntWritable> {
-    private final static IntWritable one = new IntWritable(1);
-    private Text word = new Text();
-
-    @Override
-    protected void map(WritableComparable key, HCatRecord value, Context context)
-        throws IOException, InterruptedException {
-      String line = value.getString("line", inputSchema);
-      StringTokenizer tokenizer = new StringTokenizer(line);
-      while (tokenizer.hasMoreTokens()) {
-        word.set(tokenizer.nextToken());
-        context.write(word, one);
-      }
-    }
-  }
-
-  public static class Reduce extends Reducer<Text, IntWritable, WritableComparable, HCatRecord> {
-    @Override
-    protected void reduce(Text key, Iterable<IntWritable> values, Context context) throws
-        IOException, InterruptedException {
-      int sum = 0;
-      for (IntWritable i : values) {
-        sum += i.get();
-      }
-      HCatRecord output = new DefaultHCatRecord(2);
-      output.set("word", outputSchema, key);
-      output.set("count", outputSchema, sum);
-      context.write(null, output);
-    }
-  }
-  */
 }