You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by xu...@apache.org on 2015/12/01 19:49:16 UTC
hive git commit: HIVE-12554: Fix Spark branch build after merge
[Spark Branch] (Rui via Xuefu)
Repository: hive
Updated Branches:
refs/heads/spark 79035f1c5 -> 1a87bcc0f
HIVE-12554: Fix Spark branch build after merge [Spark Branch] (Rui via Xuefu)
Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/1a87bcc0
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/1a87bcc0
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/1a87bcc0
Branch: refs/heads/spark
Commit: 1a87bcc0f27e5a819035ac67fd68ace4c41301e9
Parents: 79035f1
Author: Xuefu Zhang <xz...@Cloudera.com>
Authored: Tue Dec 1 10:49:04 2015 -0800
Committer: Xuefu Zhang <xz...@Cloudera.com>
Committed: Tue Dec 1 10:49:04 2015 -0800
----------------------------------------------------------------------
.../apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java | 3 ++-
ql/src/test/results/clientpositive/gen_udf_example_add10.q.out | 1 +
.../test/results/clientpositive/spark/gen_udf_example_add10.q.out | 1 +
3 files changed, 4 insertions(+), 1 deletion(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hive/blob/1a87bcc0/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
index d215873..ec0fdea 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
@@ -27,6 +27,7 @@ import java.util.Properties;
import java.util.Set;
import org.apache.commons.compress.utils.CharsetNames;
+import org.apache.hadoop.fs.CommonConfigurationKeysPublic;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.apache.hadoop.hive.conf.HiveConf;
@@ -133,7 +134,7 @@ public class HiveSparkClientFactory {
LOG.info(String.format(
"load yarn property from hive configuration in %s mode (%s -> %s).",
sparkMaster, propertyName, value));
- } else if (propertyName.equals(HiveConf.ConfVars.HADOOPFS.varname)) {
+ } else if (propertyName.equals(CommonConfigurationKeysPublic.FS_DEFAULT_NAME_KEY)) {
String value = hiveConf.get(propertyName);
if (value != null && !value.isEmpty()) {
sparkConf.put("spark.hadoop." + propertyName, value);
http://git-wip-us.apache.org/repos/asf/hive/blob/1a87bcc0/ql/src/test/results/clientpositive/gen_udf_example_add10.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/gen_udf_example_add10.q.out b/ql/src/test/results/clientpositive/gen_udf_example_add10.q.out
index 984554d..cab2ec8 100644
--- a/ql/src/test/results/clientpositive/gen_udf_example_add10.q.out
+++ b/ql/src/test/results/clientpositive/gen_udf_example_add10.q.out
@@ -43,6 +43,7 @@ STAGE PLANS:
key expressions: _col0 (type: int), _col1 (type: double)
sort order: -+
Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE
+ TopN Hash Memory Usage: 0.1
Reduce Operator Tree:
Select Operator
expressions: KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: double)
http://git-wip-us.apache.org/repos/asf/hive/blob/1a87bcc0/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out b/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out
index 05ec1f5..493d0a4 100644
--- a/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out
+++ b/ql/src/test/results/clientpositive/spark/gen_udf_example_add10.q.out
@@ -48,6 +48,7 @@ STAGE PLANS:
key expressions: _col0 (type: int), _col1 (type: double)
sort order: -+
Statistics: Num rows: 2 Data size: 30 Basic stats: COMPLETE Column stats: NONE
+ TopN Hash Memory Usage: 0.1
Reducer 2
Reduce Operator Tree:
Select Operator