You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@impala.apache.org by "Michael Brown (JIRA)" <ji...@apache.org> on 2017/10/11 17:05:00 UTC
[jira] [Created] (IMPALA-6040) test_multi_compression_types uses
hive in incompatible environments
Michael Brown created IMPALA-6040:
-------------------------------------
Summary: test_multi_compression_types uses hive in incompatible environments
Key: IMPALA-6040
URL: https://issues.apache.org/jira/browse/IMPALA-6040
Project: IMPALA
Issue Type: Bug
Components: Infrastructure
Affects Versions: Impala 2.11.0
Reporter: Michael Brown
Assignee: Michael Brown
Priority: Blocker
On Isilon and Local filesystem so far:
{noformat}
=================================== FAILURES ===================================
TestParquet.test_multi_compression_types[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: parquet/none]
[gw3] linux2 -- Python 2.6.6 /data/jenkins/workspace/impala-umbrella-build-and-test/repos/Impala/bin/../infra/python/env/bin/python
query_test/test_scanners.py:345: in test_multi_compression_types
check_call(['hive', '-e', hql_format.format(codec="snappy", year=2010, month=1)])
/usr/lib64/python2.6/subprocess.py:505: in check_call
raise CalledProcessError(retcode, cmd)
E CalledProcessError: Command '['hive', '-e', 'set parquet.compression=snappy;insert into table test_multi_compression_types_cc30cc12.alltypes_multi_compression partition (year = 2010, month = 1) select id, bool_col, tinyint_col, smallint_col, int_col, bigint_col, float_col, double_col,date_string_col,string_col,timestamp_col from functional_parquet.alltypes where year = 2010 and month = 1']' returned non-zero exit status 10
---------------------------- Captured stderr setup -----------------------------
SET sync_ddl=False;
-- executing against localhost:21000
DROP DATABASE IF EXISTS `test_multi_compression_types_cc30cc12` CASCADE;
SET sync_ddl=False;
-- executing against localhost:21000
CREATE DATABASE `test_multi_compression_types_cc30cc12`;
MainThread: Created database "test_multi_compression_types_cc30cc12" for test ID "query_test/test_scanners.py::TestParquet::()::test_multi_compression_types[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: parquet/none]"
----------------------------- Captured stderr call -----------------------------
-- executing against localhost:21000
create table test_multi_compression_types_cc30cc12.alltypes_multi_compression like functional_parquet.alltypes;
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/data/jenkins/workspace/impala-umbrella-build-and-test/Impala-Toolchain/cdh_components/hbase-1.2.0-cdh5.14.0-SNAPSHOT/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/data/jenkins/workspace/impala-umbrella-build-and-test/Impala-Toolchain/cdh_components/hadoop-2.6.0-cdh5.14.0-SNAPSHOT/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
17/10/11 03:21:43 WARN conf.HiveConf: HiveConf of name hive.access.conf.url does not exist
Logging initialized using configuration in file:/data/jenkins/workspace/impala-umbrella-build-and-test/repos/Impala/fe/src/test/resources/hive-log4j.properties
Unable to acquire IMPLICIT, SHARED lock functional_parquet after 100 attempts.
Error in acquireLocks...
FAILED: Error in acquiring locks: Locks on the underlying objects cannot be acquired. retry after some time
{noformat}
Fix is to skip when using these and others.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)