You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@impala.apache.org by jo...@apache.org on 2022/11/04 22:19:38 UTC
[impala] branch master updated: IMPALA-7092: Restore tests after HDFS fixes
This is an automated email from the ASF dual-hosted git repository.
joemcdonnell pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/impala.git
The following commit(s) were added to refs/heads/master by this push:
new eed92b223 IMPALA-7092: Restore tests after HDFS fixes
eed92b223 is described below
commit eed92b223fb7f582ba0f78974fed28a19204ab62
Author: Michael Smith <mi...@cloudera.com>
AuthorDate: Fri Oct 21 13:23:04 2022 -0700
IMPALA-7092: Restore tests after HDFS fixes
Restore EC tests that were disabled until HDFS-13539 and HDFS-13540 were
fixed, as the fixes are available in the current version of Hadoop we
test.
Testing: ran these tests with EC enabled.
Change-Id: I8b0bbc604601e6fab742f145c1adfb3c47b3fb6e
Reviewed-on: http://gerrit.cloudera.org:8080/19159
Reviewed-by: Joe McDonnell <jo...@cloudera.com>
Tested-by: Joe McDonnell <jo...@cloudera.com>
---
tests/common/skip.py | 1 -
tests/query_test/test_insert_parquet.py | 1 -
tests/query_test/test_queries.py | 4 ----
tests/query_test/test_query_mem_limit.py | 1 -
4 files changed, 7 deletions(-)
diff --git a/tests/common/skip.py b/tests/common/skip.py
index 7b8118dca..02c3cc021 100644
--- a/tests/common/skip.py
+++ b/tests/common/skip.py
@@ -141,7 +141,6 @@ class SkipIfBuildType:
reason="Test depends on running against a local Impala cluster")
class SkipIfEC:
- oom = pytest.mark.skipif(IS_EC, reason="Probably broken by HDFS-13540.")
contain_full_explain = pytest.mark.skipif(IS_EC, reason="Contain full explain output "
"for hdfs tables.")
different_scan_split = pytest.mark.skipif(IS_EC, reason="Scan split of row "
diff --git a/tests/query_test/test_insert_parquet.py b/tests/query_test/test_insert_parquet.py
index b9231a43c..6113664f0 100644
--- a/tests/query_test/test_insert_parquet.py
+++ b/tests/query_test/test_insert_parquet.py
@@ -114,7 +114,6 @@ class TestInsertParquetQueries(ImpalaTestSuite):
cls.ImpalaTestMatrix.add_constraint(
lambda v: v.get_value('table_format').compression_codec == 'none')
- @SkipIfEC.oom
@SkipIfLocal.multiple_impalad
@UniqueDatabase.parametrize(sync_ddl=True)
def test_insert_parquet(self, vector, unique_database):
diff --git a/tests/query_test/test_queries.py b/tests/query_test/test_queries.py
index 7b19dd23a..db311e913 100644
--- a/tests/query_test/test_queries.py
+++ b/tests/query_test/test_queries.py
@@ -234,14 +234,12 @@ class TestQueriesTextTables(ImpalaTestSuite):
vector.get_value('exec_option')['num_nodes'] = 1
self.run_test_case('QueryTest/distinct-estimate', vector)
- @SkipIfEC.oom
def test_random(self, vector):
# These results will vary slightly depending on how the values get split up
# so only run with 1 node and on text.
vector.get_value('exec_option')['num_nodes'] = 1
self.run_test_case('QueryTest/random', vector)
- @SkipIfEC.oom
def test_values(self, vector):
self.run_test_case('QueryTest/values', vector)
@@ -258,7 +256,6 @@ class TestQueriesParquetTables(ImpalaTestSuite):
def get_workload(cls):
return 'functional-query'
- @SkipIfEC.oom
@pytest.mark.execute_serially
def test_very_large_strings(self, vector):
"""Regression test for IMPALA-1619. Doesn't need to be run on all file formats.
@@ -294,7 +291,6 @@ class TestHdfsQueries(ImpalaTestSuite):
def get_workload(cls):
return 'functional-query'
- @SkipIfEC.oom
def test_hdfs_scan_node(self, vector):
self.run_test_case('QueryTest/hdfs-scan-node', vector)
diff --git a/tests/query_test/test_query_mem_limit.py b/tests/query_test/test_query_mem_limit.py
index 6de7c1f4f..2586d39bc 100644
--- a/tests/query_test/test_query_mem_limit.py
+++ b/tests/query_test/test_query_mem_limit.py
@@ -88,7 +88,6 @@ class TestQueryMemLimit(ImpalaTestSuite):
cls.ImpalaTestMatrix.add_constraint(
lambda v: v.get_value('exec_option')['batch_size'] == 0)
- @SkipIfEC.oom
@pytest.mark.execute_serially
def test_mem_limit(self, vector):
mem_limit = copy(vector.get_value('mem_limit'))