You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by st...@apache.org on 2018/02/05 18:44:00 UTC

hive git commit: HIVE-18596: Synchronize value of hive.spark.client.connect.timeout across unit tests (Sahil Takiar, reviewed by Peter Vary)

Repository: hive
Updated Branches:
  refs/heads/master 3a677303f -> e0eee6bef


HIVE-18596: Synchronize value of hive.spark.client.connect.timeout across unit tests (Sahil Takiar, reviewed by Peter Vary)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/e0eee6be
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/e0eee6be
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/e0eee6be

Branch: refs/heads/master
Commit: e0eee6bef3fa5816b8ca60b7e737a12a99dcecfa
Parents: 3a67730
Author: Sahil Takiar <ta...@gmail.com>
Authored: Mon Feb 5 11:43:26 2018 -0600
Committer: Sahil Takiar <st...@cloudera.com>
Committed: Mon Feb 5 11:44:03 2018 -0600

----------------------------------------------------------------------
 data/conf/spark/standalone/hive-site.xml                        | 5 +++++
 .../org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java     | 1 +
 .../hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java    | 1 +
 .../test/java/org/apache/hive/spark/client/TestSparkClient.java | 4 ++++
 4 files changed, 11 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/data/conf/spark/standalone/hive-site.xml
----------------------------------------------------------------------
diff --git a/data/conf/spark/standalone/hive-site.xml b/data/conf/spark/standalone/hive-site.xml
index 1e5bd65..7095979 100644
--- a/data/conf/spark/standalone/hive-site.xml
+++ b/data/conf/spark/standalone/hive-site.xml
@@ -255,6 +255,11 @@
   <description>Internal marker for test. Used for masking env-dependent values</description>
 </property>
 
+<property>
+  <name>hive.spark.client.connect.timeout</name>
+  <value>30000ms</value>
+</property>
+
 <!-- remove this after HIVE-18139 -->
 <property>
   <name>hive.stats.column.autogather</name>

http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
----------------------------------------------------------------------
diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
index 0fee3c6..2a4da20 100644
--- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
+++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java
@@ -70,6 +70,7 @@ public class TestJdbcWithLocalClusterSpark {
     conf.set("hive.execution.engine", "spark");
     conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
     conf.set("spark.master", "local-cluster[2,2,1024]");
+    conf.set("hive.spark.client.connect.timeout", "30000ms");
     // FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout
     // while spark2 is still using Hadoop2.
     // Spark requires Hive to support Hadoop3 first then Spark can start

http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
----------------------------------------------------------------------
diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
index 93fe776..9f72e51 100644
--- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
+++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java
@@ -79,6 +79,7 @@ public class TestMultiSessionsHS2WithLocalClusterSpark {
     conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer");
     conf.set("spark.master", "local-cluster[2,2,1024]");
     conf.set("spark.deploy.defaultCores", "2");
+    conf.set("hive.spark.client.connect.timeout", "30000ms");
     // FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout
     // while spark2 is still using Hadoop2.
     // Spark requires Hive to support Hadoop3 first then Spark can start

http://git-wip-us.apache.org/repos/asf/hive/blob/e0eee6be/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
----------------------------------------------------------------------
diff --git a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
index 23df792..579ca07 100644
--- a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
+++ b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java
@@ -68,6 +68,10 @@ public class TestSparkClient {
   private static final long TIMEOUT = 20;
   private static final HiveConf HIVECONF = new HiveConf();
 
+  static {
+    HIVECONF.set("hive.spark.client.connect.timeout", "30000ms");
+  }
+
   private Map<String, String> createConf() {
     Map<String, String> conf = new HashMap<String, String>();