You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@impala.apache.org by "Tim Armstrong (JIRA)" <ji...@apache.org> on 2018/01/31 22:12:00 UTC

[jira] [Resolved] (IMPALA-6455) test_partition_metadata_compatibility flaky failures

     [ https://issues.apache.org/jira/browse/IMPALA-6455?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Tim Armstrong resolved IMPALA-6455.
-----------------------------------
       Resolution: Fixed
    Fix Version/s: Impala 2.12.0



IMPALA-6455: unique tmpdirs for test_partition_metadata_compatibility

Concurrent hive statements running in local mode can race to modify
the contents of temporary directories - see IMPALA-6108. This applies
the workaround for IMPALA-6108 to the run_stmt_in_hive() utility
function, which is used by test_partition_metadata_compatibility.

Testing:
I wasn't able to reproduce the race locally, but I ran the test and
confirmed that it still passed. I also confirmed that the temporary
directories /tmp/impala-tests-* were created using "ls" while the
tests were running.

Change-Id: Ibabff859d19ddbb2a3048ecc02897a611d8ddb20
Reviewed-on: http://gerrit.cloudera.org:8080/9165
Reviewed-by: Philip Zeyliger <ph...@cloudera.com>
Tested-by: Impala Public Jenkins


> test_partition_metadata_compatibility flaky failures
> ----------------------------------------------------
>
>                 Key: IMPALA-6455
>                 URL: https://issues.apache.org/jira/browse/IMPALA-6455
>             Project: IMPALA
>          Issue Type: Bug
>          Components: Infrastructure
>    Affects Versions: Impala 2.11.0
>            Reporter: Sailesh Mukil
>            Assignee: Tim Armstrong
>            Priority: Blocker
>              Labels: broken-build
>             Fix For: Impala 2.12.0
>
>
> This test has been observed to fail fairly frequently on GVO jobs.
> https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/1095/consoleFull
> https://jenkins.impala.io/job/ubuntu-16.04-from-scratch/1101/consoleFull
> {code:java}
> 07:54:35 [gw5] FAILED metadata/test_partition_metadata.py::TestPartitionMetadata::test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: seq/snap/block] 07:54:35 [gw2] FAILED metadata/test_partition_metadata.py::TestPartitionMetadata::test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]
> {code}
> {code:java}
> 08:35:07 FAIL metadata/test_partition_metadata.py::TestPartitionMetadata::()::test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: seq/snap/block]
> 08:35:07 FAIL metadata/test_partition_metadata.py::TestPartitionMetadata::()::test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]
> 08:35:07 =================================== FAILURES ===================================
> 08:35:07  TestPartitionMetadata.test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: seq/snap/block] 
> 08:35:07 [gw5] linux2 -- Python 2.7.12 /home/ubuntu/Impala/bin/../infra/python/env/bin/python
> 08:35:07 metadata/test_partition_metadata.py:127: in test_partition_metadata_compatibility
> 08:35:07     "insert into %s partition (x) values(1,1)" % FQ_TBL_HIVE)
> 08:35:07 common/impala_test_suite.py:665: in run_stmt_in_hive
> 08:35:07     raise RuntimeError(stderr)
> 08:35:07 E   RuntimeError: SLF4J: Class path contains multiple SLF4J bindings.
> 08:35:07 E   SLF4J: Found binding in [jar:file:/home/ubuntu/Impala/toolchain/cdh_components/hbase-1.2.0-cdh5.15.0-SNAPSHOT/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> 08:35:07 E   SLF4J: Found binding in [jar:file:/home/ubuntu/Impala/toolchain/cdh_components/hadoop-2.6.0-cdh5.15.0-SNAPSHOT/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> 08:35:07 E   SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
> 08:35:07 E   SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 08:35:07 E   scan complete in 2ms
> 08:35:07 E   Connecting to jdbc:hive2://localhost:11050
> 08:35:07 E   Connected to: Apache Hive (version 1.1.0-cdh5.15.0-SNAPSHOT)
> 08:35:07 E   Driver: Hive JDBC (version 1.1.0-cdh5.15.0-SNAPSHOT)
> 08:35:07 E   Transaction isolation: TRANSACTION_REPEATABLE_READ
> 08:35:07 E   No rows affected (0.042 seconds)
> 08:35:07 E   INFO  : Compiling command(queryId=ubuntu_20180130075454_4bf94127-1f71-4a2e-be7e-9086b1b1f34e): insert into test_partition_metadata_compatibility_cca30e65.part_parquet_tbl_hive partition (x) values(1,1)
> 08:35:07 E   INFO  : Semantic Analysis Completed
> 08:35:07 E   INFO  : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:_col0, type:int, comment:null), FieldSchema(name:_col1, type:string, comment:null)], properties:null)
> 08:35:07 E   INFO  : Completed compiling command(queryId=ubuntu_20180130075454_4bf94127-1f71-4a2e-be7e-9086b1b1f34e); Time taken: 0.105 seconds
> 08:35:07 E   INFO  : Executing command(queryId=ubuntu_20180130075454_4bf94127-1f71-4a2e-be7e-9086b1b1f34e): insert into test_partition_metadata_compatibility_cca30e65.part_parquet_tbl_hive partition (x) values(1,1)
> 08:35:07 E   INFO  : Query ID = ubuntu_20180130075454_4bf94127-1f71-4a2e-be7e-9086b1b1f34e
> 08:35:07 E   INFO  : Total jobs = 3
> 08:35:07 E   INFO  : Launching Job 1 out of 3
> 08:35:07 E   INFO  : Starting task [Stage-1:MAPRED] in serial mode
> 08:35:07 E   INFO  : Number of reduce tasks is set to 0 since there's no reduce operator
> 08:35:07 E   INFO  : number of splits:1
> 08:35:07 E   INFO  : Submitting tokens for job: job_local512703045_0002
> 08:35:07 E   INFO  : Cleaning up the staging area file:/tmp/hadoop-ubuntu/mapred/staging/ubuntu512703045/.staging/job_local512703045_0002
> 08:35:07 E   ERROR : Job Submission failed with exception 'java.io.IOException(java.util.concurrent.ExecutionException: java.io.FileNotFoundException: File /tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp does not exist)'
> 08:35:07 E   java.io.IOException: java.util.concurrent.ExecutionException: java.io.FileNotFoundException: File /tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp does not exist
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:143)
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:171)
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:758)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:244)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1307)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1304)
> 08:35:07 E   	at java.security.AccessController.doPrivileged(Native Method)
> 08:35:07 E   	at javax.security.auth.Subject.doAs(Subject.java:422)
> 08:35:07 E   	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.Job.submit(Job.java:1304)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:578)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:573)
> 08:35:07 E   	at java.security.AccessController.doPrivileged(Native Method)
> 08:35:07 E   	at javax.security.auth.Subject.doAs(Subject.java:422)
> 08:35:07 E   	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:573)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:564)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:436)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:142)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:214)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:99)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2052)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1748)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1501)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1285)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1280)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:236)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation.access$300(SQLOperation.java:89)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation$3$1.run(SQLOperation.java:301)
> 08:35:07 E   	at java.security.AccessController.doPrivileged(Native Method)
> 08:35:07 E   	at javax.security.auth.Subject.doAs(Subject.java:422)
> 08:35:07 E   	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation$3.run(SQLOperation.java:314)
> 08:35:07 E   	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> 08:35:07 E   	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 08:35:07 E   	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> 08:35:07 E   	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> 08:35:07 E   	at java.lang.Thread.run(Thread.java:748)
> 08:35:07 E   Caused by: java.util.concurrent.ExecutionException: java.io.FileNotFoundException: File /tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp does not exist
> 08:35:07 E   	at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> 08:35:07 E   	at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:139)
> 08:35:07 E   	... 37 more
> 08:35:07 E   Caused by: java.io.FileNotFoundException: File /tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp does not exist
> 08:35:07 E   	at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:598)
> 08:35:07 E   	at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:811)
> 08:35:07 E   	at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:588)
> 08:35:07 E   	at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileLinkStatusInternal(RawLocalFileSystem.java:827)
> 08:35:07 E   	at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:813)
> 08:35:07 E   	at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatus(RawLocalFileSystem.java:784)
> 08:35:07 E   	at org.apache.hadoop.fs.DelegateToFileSystem.getFileLinkStatus(DelegateToFileSystem.java:132)
> 08:35:07 E   	at org.apache.hadoop.fs.AbstractFileSystem.renameInternal(AbstractFileSystem.java:701)
> 08:35:07 E   	at org.apache.hadoop.fs.FilterFs.renameInternal(FilterFs.java:236)
> 08:35:07 E   	at org.apache.hadoop.fs.AbstractFileSystem.rename(AbstractFileSystem.java:674)
> 08:35:07 E   	at org.apache.hadoop.fs.FileContext.rename(FileContext.java:932)
> 08:35:07 E   	at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:369)
> 08:35:07 E   	at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60)
> 08:35:07 E   	... 4 more
> 08:35:07 E   
> 08:35:07 E   ERROR : FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> 08:35:07 E   INFO  : Completed executing command(queryId=ubuntu_20180130075454_4bf94127-1f71-4a2e-be7e-9086b1b1f34e); Time taken: 0.479 seconds
> 08:35:07 E   Error: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask (state=08S01,code=1)
> 08:35:07 E   Closing: 0: jdbc:hive2://localhost:11050
> 08:35:07 ---------------------------- Captured stderr setup -----------------------------
> 08:35:07 -- connecting to: localhost:21000
> 08:35:07 SET sync_ddl=False;
> 08:35:07 -- executing against localhost:21000
> 08:35:07 DROP DATABASE IF EXISTS `test_partition_metadata_compatibility_cca30e65` CASCADE;
> 08:35:07 
> 08:35:07 SET sync_ddl=False;
> 08:35:07 -- executing against localhost:21000
> 08:35:07 CREATE DATABASE `test_partition_metadata_compatibility_cca30e65`;
> 08:35:07 
> 08:35:07 MainThread: Created database "test_partition_metadata_compatibility_cca30e65" for test ID "metadata/test_partition_metadata.py::TestPartitionMetadata::()::test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: seq/snap/block]"
> 08:35:07  TestPartitionMetadata.test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block] 
> 08:35:07 [gw2] linux2 -- Python 2.7.12 /home/ubuntu/Impala/bin/../infra/python/env/bin/python
> 08:35:07 metadata/test_partition_metadata.py:127: in test_partition_metadata_compatibility
> 08:35:07     "insert into %s partition (x) values(1,1)" % FQ_TBL_HIVE)
> 08:35:07 common/impala_test_suite.py:665: in run_stmt_in_hive
> 08:35:07     raise RuntimeError(stderr)
> 08:35:07 E   RuntimeError: SLF4J: Class path contains multiple SLF4J bindings.
> 08:35:07 E   SLF4J: Found binding in [jar:file:/home/ubuntu/Impala/toolchain/cdh_components/hbase-1.2.0-cdh5.15.0-SNAPSHOT/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> 08:35:07 E   SLF4J: Found binding in [jar:file:/home/ubuntu/Impala/toolchain/cdh_components/hadoop-2.6.0-cdh5.15.0-SNAPSHOT/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> 08:35:07 E   SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
> 08:35:07 E   SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 08:35:07 E   scan complete in 2ms
> 08:35:07 E   Connecting to jdbc:hive2://localhost:11050
> 08:35:07 E   Connected to: Apache Hive (version 1.1.0-cdh5.15.0-SNAPSHOT)
> 08:35:07 E   Driver: Hive JDBC (version 1.1.0-cdh5.15.0-SNAPSHOT)
> 08:35:07 E   Transaction isolation: TRANSACTION_REPEATABLE_READ
> 08:35:07 E   No rows affected (0.085 seconds)
> 08:35:07 E   INFO  : Compiling command(queryId=ubuntu_20180130075454_9972bada-1343-4927-97f3-d12a2b550629): insert into test_partition_metadata_compatibility_b2ac5e.part_parquet_tbl_hive partition (x) values(1,1)
> 08:35:07 E   INFO  : Semantic Analysis Completed
> 08:35:07 E   INFO  : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:_col0, type:int, comment:null), FieldSchema(name:_col1, type:string, comment:null)], properties:null)
> 08:35:07 E   INFO  : Completed compiling command(queryId=ubuntu_20180130075454_9972bada-1343-4927-97f3-d12a2b550629); Time taken: 0.292 seconds
> 08:35:07 E   INFO  : Executing command(queryId=ubuntu_20180130075454_9972bada-1343-4927-97f3-d12a2b550629): insert into test_partition_metadata_compatibility_b2ac5e.part_parquet_tbl_hive partition (x) values(1,1)
> 08:35:07 E   INFO  : Query ID = ubuntu_20180130075454_9972bada-1343-4927-97f3-d12a2b550629
> 08:35:07 E   INFO  : Total jobs = 3
> 08:35:07 E   INFO  : Launching Job 1 out of 3
> 08:35:07 E   INFO  : Starting task [Stage-1:MAPRED] in serial mode
> 08:35:07 E   INFO  : Number of reduce tasks is set to 0 since there's no reduce operator
> 08:35:07 E   INFO  : number of splits:1
> 08:35:07 E   INFO  : Submitting tokens for job: job_local1043698527_0001
> 08:35:07 E   INFO  : Cleaning up the staging area file:/tmp/hadoop-ubuntu/mapred/staging/ubuntu1043698527/.staging/job_local1043698527_0001
> 08:35:07 E   ERROR : Job Submission failed with exception 'java.io.IOException(java.util.concurrent.ExecutionException: java.io.IOException: Unable to rename file: [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/tmp_postgresql-9.0-801.jdbc4.jar] to [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/postgresql-9.0-801.jdbc4.jar])'
> 08:35:07 E   java.io.IOException: java.util.concurrent.ExecutionException: java.io.IOException: Unable to rename file: [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/tmp_postgresql-9.0-801.jdbc4.jar] to [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/postgresql-9.0-801.jdbc4.jar]
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:143)
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:171)
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:758)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:244)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1307)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1304)
> 08:35:07 E   	at java.security.AccessController.doPrivileged(Native Method)
> 08:35:07 E   	at javax.security.auth.Subject.doAs(Subject.java:422)
> 08:35:07 E   	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
> 08:35:07 E   	at org.apache.hadoop.mapreduce.Job.submit(Job.java:1304)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:578)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:573)
> 08:35:07 E   	at java.security.AccessController.doPrivileged(Native Method)
> 08:35:07 E   	at javax.security.auth.Subject.doAs(Subject.java:422)
> 08:35:07 E   	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:573)
> 08:35:07 E   	at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:564)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:436)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:142)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:214)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:99)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2052)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1748)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1501)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1285)
> 08:35:07 E   	at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1280)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:236)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation.access$300(SQLOperation.java:89)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation$3$1.run(SQLOperation.java:301)
> 08:35:07 E   	at java.security.AccessController.doPrivileged(Native Method)
> 08:35:07 E   	at javax.security.auth.Subject.doAs(Subject.java:422)
> 08:35:07 E   	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
> 08:35:07 E   	at org.apache.hive.service.cli.operation.SQLOperation$3.run(SQLOperation.java:314)
> 08:35:07 E   	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> 08:35:07 E   	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 08:35:07 E   	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> 08:35:07 E   	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> 08:35:07 E   	at java.lang.Thread.run(Thread.java:748)
> 08:35:07 E   Caused by: java.util.concurrent.ExecutionException: java.io.IOException: Unable to rename file: [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/tmp_postgresql-9.0-801.jdbc4.jar] to [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/postgresql-9.0-801.jdbc4.jar]
> 08:35:07 E   	at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> 08:35:07 E   	at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> 08:35:07 E   	at org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:139)
> 08:35:07 E   	... 37 more
> 08:35:07 E   Caused by: java.io.IOException: Unable to rename file: [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/tmp_postgresql-9.0-801.jdbc4.jar] to [/tmp/hadoop-ubuntu/mapred/local/1517298875469_tmp/postgresql-9.0-801.jdbc4.jar]
> 08:35:07 E   	at org.apache.hadoop.yarn.util.FSDownload.unpack(FSDownload.java:327)
> 08:35:07 E   	at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:367)
> 08:35:07 E   	at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60)
> 08:35:07 E   	... 4 more
> 08:35:07 E   
> 08:35:07 E   ERROR : FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> 08:35:07 E   INFO  : Completed executing command(queryId=ubuntu_20180130075454_9972bada-1343-4927-97f3-d12a2b550629); Time taken: 0.587 seconds
> 08:35:07 E   Error: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask (state=08S01,code=1)
> 08:35:07 E   Closing: 0: jdbc:hive2://localhost:11050
> 08:35:07 ---------------------------- Captured stderr setup -----------------------------
> 08:35:07 -- connecting to: localhost:21000
> 08:35:07 SET sync_ddl=False;
> 08:35:07 -- executing against localhost:21000
> 08:35:07 DROP DATABASE IF EXISTS `test_partition_metadata_compatibility_b2ac5e` CASCADE;
> 08:35:07 
> 08:35:07 SET sync_ddl=False;
> 08:35:07 -- executing against localhost:21000
> 08:35:07 CREATE DATABASE `test_partition_metadata_compatibility_b2ac5e`;
> 08:35:07 
> 08:35:07 MainThread: Created database "test_partition_metadata_compatibility_b2ac5e" for test ID "metadata/test_partition_metadata.py::TestPartitionMetadata::()::test_partition_metadata_compatibility[exec_option: {'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 5000, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: avro/snap/block]"
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)