You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by st...@apache.org on 2018/02/05 18:44:00 UTC
hive git commit: HIVE-18596: Synchronize value of
hive.spark.client.connect.timeout across unit tests (Sahil Takiar,
reviewed by Peter Vary)
Repository: hive
Updated Branches:
refs/heads/master 3a677303f -> e0eee6bef
HIVE-18596: Synchronize value of hive.spark.client.connect.timeout across unit tests (Sahil Takiar, reviewed by Peter Vary)
Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/e0eee6be
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/e0eee6be
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/e0eee6be
Branch: refs/heads/master
Commit: e0eee6bef3fa5816b8ca60b7e737a12a99dcecfa
Parents: 3a67730
Author: Sahil Takiar <ta...@gmail.com>
Authored: Mon Feb 5 11:43:26 2018 -0600
Committer: Sahil Takiar <st...@cloudera.com>
Committed: Mon Feb 5 11:44:03 2018 -0600
----------------------------------------------------------------------
data/conf/spark/standalone/hive-site.xml | 5 +++++
.../org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java | 1 +
.../hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java | 1 +
.../test/java/org/apache/hive/spark/client/TestSparkClient.java | 4 ++++
4 files changed, 11 insertions(+)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/data/conf/spark/standalone/hive-site.xml
----------------------------------------------------------------------
diff --git a/data/conf/spark/standalone/hive-site.xml b/data/conf/spark/standalone/hive-site.xml
index 1e5bd65..7095979 100644
--- a/data/conf/spark/standalone/hive-site.xml
+++ b/data/conf/spark/standalone/hive-site.xml
@@ -255,6 +255,11 @@
<description>Internal marker for test. Used for masking env-dependent values</description>
</property>
+<property>
+ <name>hive.spark.client.connect.timeout</name>
+ <value>30000ms</value>
+</property>
+
<!-- remove this after HIVE-18139 -->
<property>
<name>hive.stats.column.autogather</name>
http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
----------------------------------------------------------------------
diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
index 0fee3c6..2a4da20 100644
--- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
+++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
@@ -70,6 +70,7 @@ public class TestJdbcWithLocalClusterSpark {
conf.set("hive.execution.engine", "spark");
conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
conf.set("spark.master", "local-cluster[2,2,1024]");
+ conf.set("hive.spark.client.connect.timeout", "30000ms");
// FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout
// while spark2 is still using Hadoop2.
// Spark requires Hive to support Hadoop3 first then Spark can start
http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
----------------------------------------------------------------------
diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
index 93fe776..9f72e51 100644
--- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
+++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
@@ -79,6 +79,7 @@ public class TestMultiSessionsHS2WithLocalClusterSpark {
conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
conf.set("spark.master", "local-cluster[2,2,1024]");
conf.set("spark.deploy.defaultCores", "2");
+ conf.set("hive.spark.client.connect.timeout", "30000ms");
// FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout
// while spark2 is still using Hadoop2.
// Spark requires Hive to support Hadoop3 first then Spark can start
http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
----------------------------------------------------------------------
diff --git a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
index 23df792..579ca07 100644
--- a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
+++ b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
@@ -68,6 +68,10 @@ public class TestSparkClient {
private static final long TIMEOUT = 20;
private static final HiveConf HIVECONF = new HiveConf();
+ static {
+ HIVECONF.set("hive.spark.client.connect.timeout", "30000ms");
+ }
+
private Map<String, String> createConf() {
Map<String, String> conf = new HashMap<String, String>();