You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@impala.apache.org by jo...@apache.org on 2019/08/08 16:05:30 UTC
[impala] 02/02: Bump CDP_BUILD_NUMBER to 1318335
This is an automated email from the ASF dual-hosted git repository.
joemcdonnell pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/impala.git
commit a0c00e508f08542b55f140e4c329bf498141e921
Author: Csaba Ringhofer <cs...@cloudera.com>
AuthorDate: Thu Aug 1 16:27:06 2019 +0200
Bump CDP_BUILD_NUMBER to 1318335
The main reason for bumping is to include HIVE-21838.
Also skips / fixes some tests.
Change-Id: I432e8c02dbd349a3507bfabfef2727914537652c
Reviewed-on: http://gerrit.cloudera.org:8080/14005
Reviewed-by: Impala Public Jenkins <im...@cloudera.com>
Tested-by: Impala Public Jenkins <im...@cloudera.com>
---
bin/impala-config.sh | 10 +++++-----
tests/custom_cluster/test_kudu.py | 2 ++
tests/query_test/test_insert.py | 15 +++++++++------
3 files changed, 16 insertions(+), 11 deletions(-)
diff --git a/bin/impala-config.sh b/bin/impala-config.sh
index e5f52a2..effddc4 100755
--- a/bin/impala-config.sh
+++ b/bin/impala-config.sh
@@ -168,19 +168,19 @@ fi
export IMPALA_TOOLCHAIN_HOST
export CDH_MAJOR_VERSION=6
export CDH_BUILD_NUMBER=1173663
-export CDP_BUILD_NUMBER=1268086
+export CDP_BUILD_NUMBER=1318335
export CDH_HADOOP_VERSION=3.0.0-cdh6.x-SNAPSHOT
-export CDP_HADOOP_VERSION=3.1.1.7.0.0.0-280
+export CDP_HADOOP_VERSION=3.1.1.7.0.0.0-365
export IMPALA_HBASE_VERSION=2.1.0-cdh6.x-SNAPSHOT
export IMPALA_SENTRY_VERSION=2.1.0-cdh6.x-SNAPSHOT
-export IMPALA_RANGER_VERSION=1.2.0.7.0.0.0-280
+export IMPALA_RANGER_VERSION=1.2.0.7.0.0.0-365
export IMPALA_PARQUET_VERSION=1.9.0-cdh6.x-SNAPSHOT
export IMPALA_AVRO_JAVA_VERSION=1.8.2-cdh6.x-SNAPSHOT
export IMPALA_LLAMA_MINIKDC_VERSION=1.0.0
export IMPALA_KITE_VERSION=1.0.0-cdh6.x-SNAPSHOT
export IMPALA_KUDU_JAVA_VERSION=1.10.0-cdh6.x-SNAPSHOT
export CDH_HIVE_VERSION=2.1.1-cdh6.x-SNAPSHOT
-export CDP_HIVE_VERSION=3.1.0.7.0.0.0-280
+export CDP_HIVE_VERSION=3.1.0.7.0.0.0-365
# When IMPALA_(CDH_COMPONENT)_URL are overridden, they may contain '$(platform_label)'
# which will be substituted for the CDH platform label in bootstrap_toolchain.py
@@ -207,7 +207,7 @@ if $USE_CDP_HIVE; then
# When USE_CDP_HIVE is set we use the CDP hive version to build as well as deploy in
# the minicluster
export IMPALA_HIVE_VERSION=${CDP_HIVE_VERSION}
- export IMPALA_TEZ_VERSION=0.9.1.7.0.0.0-280
+ export IMPALA_TEZ_VERSION=0.9.1.7.0.0.0-365
export IMPALA_HADOOP_VERSION=${CDP_HADOOP_VERSION}
export HADOOP_HOME="$CDP_COMPONENTS_HOME/hadoop-${CDP_HADOOP_VERSION}/"
else
diff --git a/tests/custom_cluster/test_kudu.py b/tests/custom_cluster/test_kudu.py
index c4049ca..17ea63a 100644
--- a/tests/custom_cluster/test_kudu.py
+++ b/tests/custom_cluster/test_kudu.py
@@ -147,6 +147,7 @@ class TestKuduHMSIntegration(CustomClusterTestSuite, KuduTestSuite):
self.run_test_case('QueryTest/kudu_create', vector, use_db=unique_database)
@pytest.mark.execute_serially
+ @SkipIfHive3.kudu_hms_notifications_not_supported
def test_implicit_external_table_props(self, cursor, kudu_client):
"""Check that table properties added internally for external table during
table creation are as expected.
@@ -172,6 +173,7 @@ class TestKuduHMSIntegration(CustomClusterTestSuite, KuduTestSuite):
in table_desc
@pytest.mark.execute_serially
+ @SkipIfHive3.kudu_hms_notifications_not_supported
def test_implicit_managed_table_props(self, cursor, kudu_client, unique_database):
"""Check that table properties added internally for managed table during table
creation are as expected.
diff --git a/tests/query_test/test_insert.py b/tests/query_test/test_insert.py
index 41cce30..be298a7 100644
--- a/tests/query_test/test_insert.py
+++ b/tests/query_test/test_insert.py
@@ -147,12 +147,15 @@ class TestInsertQueries(ImpalaTestSuite):
# need to drop and create such tables, and table properties are preserved during
# those operations and this is enough for the tests (A table is ACID if it has the
# relevant table properties).
- capability_check = self.hive_client.getMetaConf("metastore.client.capability.check")
- self.hive_client.setMetaConf("metastore.client.capability.check", "false")
- self.run_test_case('QueryTest/acid-insert', vector,
- multiple_impalad=vector.get_value('exec_option')['sync_ddl'] == 1)
- # Reset original state.
- self.hive_client.setMetaConf("metastore.client.capability.check", capability_check)
+ CAPABILITY_CHECK_CONF = "hive.metastore.client.capability.check"
+ capability_check = self.hive_client.getMetaConf(CAPABILITY_CHECK_CONF)
+ try:
+ self.hive_client.setMetaConf(CAPABILITY_CHECK_CONF, "false")
+ self.run_test_case('QueryTest/acid-insert', vector,
+ multiple_impalad=vector.get_value('exec_option')['sync_ddl'] == 1)
+ finally:
+ # Reset original state.
+ self.hive_client.setMetaConf(CAPABILITY_CHECK_CONF, capability_check)
@SkipIfHive2.acid
@UniqueDatabase.parametrize(sync_ddl=True)