You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@drill.apache.org by "Abhishek Girish (Jira)" <ji...@apache.org> on 2020/08/30 20:55:00 UTC

[jira] [Created] (DRILL-7785) Hive Clustered Bucketed ORC transactional table fails with UndeclaredThrowableException

Abhishek Girish created DRILL-7785:
--------------------------------------

             Summary: Hive Clustered Bucketed ORC transactional table fails with UndeclaredThrowableException
                 Key: DRILL-7785
                 URL: https://issues.apache.org/jira/browse/DRILL-7785
             Project: Apache Drill
          Issue Type: Bug
          Components: Execution - Flow, Storage - Hive
    Affects Versions: 1.18.0
            Reporter: Abhishek Girish
            Assignee: Vova Vysotskyi


Query: 
{code}
Functional/hive/hive_storage/fileformats/orc/transactional/orc_table_clustered_bucketed.sql
select * from hive_orc_transactional.orc_table_clustered_bucketed
{code}
Exception:
{code}
java.sql.SQLException: EXECUTION_ERROR ERROR: java.lang.reflect.UndeclaredThrowableException

Failed to setup reader: HiveDefaultRecordReader
Fragment: 0:0

[Error Id: 323434cc-7bd2-4551-94d4-a5925f6a66af on drill80:31010]

  (org.apache.drill.common.exceptions.ExecutionSetupException) java.lang.reflect.UndeclaredThrowableException
    org.apache.drill.common.exceptions.ExecutionSetupException.fromThrowable():30
    org.apache.drill.exec.store.hive.readers.HiveDefaultRecordReader.setup():257
    org.apache.drill.exec.physical.impl.ScanBatch.getNextReaderIfHas():331
    org.apache.drill.exec.physical.impl.ScanBatch.internalNext():227
    org.apache.drill.exec.physical.impl.ScanBatch.next():298
    org.apache.drill.exec.record.AbstractRecordBatch.next():119
    org.apache.drill.exec.record.AbstractRecordBatch.next():111
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():59
    org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():85
    org.apache.drill.exec.record.AbstractRecordBatch.next():170
    org.apache.drill.exec.physical.impl.BaseRootExec.next():103
    org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():81
    org.apache.drill.exec.physical.impl.BaseRootExec.next():93
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():323
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():310
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1669
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():310
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (java.util.concurrent.ExecutionException) java.lang.reflect.UndeclaredThrowableException
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.AbstractFuture.getDoneValue():553
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.AbstractFuture.get():534
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.FluentFuture$TrustedFuture.get():88
    org.apache.drill.exec.store.hive.readers.HiveDefaultRecordReader.setup():252
    org.apache.drill.exec.physical.impl.ScanBatch.getNextReaderIfHas():331
    org.apache.drill.exec.physical.impl.ScanBatch.internalNext():227
    org.apache.drill.exec.physical.impl.ScanBatch.next():298
    org.apache.drill.exec.record.AbstractRecordBatch.next():119
    org.apache.drill.exec.record.AbstractRecordBatch.next():111
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():59
    org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():85
    org.apache.drill.exec.record.AbstractRecordBatch.next():170
    org.apache.drill.exec.physical.impl.BaseRootExec.next():103
    org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():81
    org.apache.drill.exec.physical.impl.BaseRootExec.next():93
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():323
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():310
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1669
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():310
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (java.lang.reflect.UndeclaredThrowableException) null
    org.apache.hadoop.security.UserGroupInformation.doAs():1687
    org.apache.drill.exec.ops.OperatorContextImpl$1.call():101
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly():125
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.InterruptibleTask.run():69
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.TrustedListenableFutureTask.run():78
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (org.apache.drill.common.exceptions.ExecutionSetupException) Failed to get o.a.hadoop.mapred.RecordReader from Hive InputFormat
    org.apache.drill.exec.store.hive.readers.HiveDefaultRecordReader.initNextReader():509
    org.apache.drill.exec.store.hive.readers.HiveDefaultRecordReader.lambda$getInitTask$1():367
    org.apache.drill.exec.ops.OperatorContextImpl$1$1.run():104
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1669
    org.apache.drill.exec.ops.OperatorContextImpl$1.call():101
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly():125
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.InterruptibleTask.run():69
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.TrustedListenableFutureTask.run():78
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
  Caused By (java.io.IOException) Open failed for file: /user/hive/warehouse/hive_orc_transactional.db/orc_table_clustered_bucketed, error: Invalid argument (22)
    com.mapr.fs.MapRClientImpl.open():338
    com.mapr.fs.MapRFileSystem.open():1018
    org.apache.hadoop.fs.FileSystem.open():807
    org.apache.orc.impl.ReaderImpl.extractFileTail():492
    org.apache.orc.impl.ReaderImpl.<init>():378
    org.apache.hadoop.hive.ql.io.orc.ReaderImpl.<init>():63
    org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader():90
    org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getRecordReader():1848
    org.apache.drill.exec.store.hive.readers.HiveDefaultRecordReader.initNextReader():505
    org.apache.drill.exec.store.hive.readers.HiveDefaultRecordReader.lambda$getInitTask$1():367
    org.apache.drill.exec.ops.OperatorContextImpl$1$1.run():104
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1669
    org.apache.drill.exec.ops.OperatorContextImpl$1.call():101
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly():125
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.InterruptibleTask.run():69
    org.apache.drill.shaded.guava.com.google.common.util.concurrent.TrustedListenableFutureTask.run():78
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)