You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@drill.apache.org by "Abhishek Girish (JIRA)" <ji...@apache.org> on 2018/10/03 18:04:00 UTC
[jira] [Updated] (DRILL-6771) Queries on Hive 2.3.x fails with
SYSTEM ERROR: ArrayIndexOutOfBoundsException
[ https://issues.apache.org/jira/browse/DRILL-6771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Abhishek Girish updated DRILL-6771:
-----------------------------------
Description:
Query: Functional/partition_pruning/hive/general/plan/orc1.q
{code}
select * from hive.orc_create_people_dp where state = 'Ca'
java.sql.SQLException: SYSTEM ERROR: ArrayIndexOutOfBoundsException: 6
{code}
Stack Trace:
{code}
(org.apache.drill.exec.work.foreman.ForemanException) Unexpected exception during fragment initialization: Error while applying rule Prel.ScanPrule, args [rel#2103503:DrillScanRel.LOGICAL.ANY([]).[](table=[hive, orc_create_people_dp],groupscan=HiveScan [table=Table(dbName:default, tableName:orc_create_people_dp), columns=[`id`, `first_name`, `last_name`, `address`, `state`, `**`], numPartitions=1, partitions= [Partition(values:[Ca])], inputDirectories=[maprfs:/drill/testdata/hive_storage/orc_create_people_dp/state=Ca], confProperties={}])]
org.apache.drill.exec.work.foreman.Foreman.run():300
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) Error while applying rule Prel.ScanPrule, args [rel#2103503:DrillScanRel.LOGICAL.ANY([]).[](table=[hive, orc_create_people_dp],groupscan=HiveScan [table=Table(dbName:default, tableName:orc_create_people_dp), columns=[`id`, `first_name`, `last_name`, `address`, `state`, `**`], numPartitions=1, partitions= [Partition(values:[Ca])], inputDirectories=[maprfs:/drill/testdata/hive_storage/orc_create_people_dp/state=Ca], confProperties={}])]
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():236
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (org.apache.drill.common.exceptions.DrillRuntimeException) Failed to get InputSplits
org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():182
org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) ORC split generation failed with exception: java.lang.ArrayIndexOutOfBoundsException: 6
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo():1579
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits():1665
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$splitInputWithUGI$2():258
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1595
org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():247
org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():156
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$getInputSplits$0():177
java.util.stream.ReferencePipeline$7$1.accept():267
java.util.ArrayList$ArrayListSpliterator.forEachRemaining():1382
java.util.stream.AbstractPipeline.copyInto():481
java.util.stream.AbstractPipeline.wrapAndCopyInto():471
java.util.stream.ReduceOps$ReduceOp.evaluateSequential():708
java.util.stream.AbstractPipeline.evaluate():234
java.util.stream.ReferencePipeline.collect():499
org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():178
org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.util.concurrent.ExecutionException) java.lang.ArrayIndexOutOfBoundsException: 6
java.util.concurrent.FutureTask.report():122
java.util.concurrent.FutureTask.get():192
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo():1573
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits():1665
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$splitInputWithUGI$2():258
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1595
org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():247
org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():156
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$getInputSplits$0():177
java.util.stream.ReferencePipeline$7$1.accept():267
java.util.ArrayList$ArrayListSpliterator.forEachRemaining():1382
java.util.stream.AbstractPipeline.copyInto():481
java.util.stream.AbstractPipeline.wrapAndCopyInto():471
java.util.stream.ReduceOps$ReduceOp.evaluateSequential():708
java.util.stream.AbstractPipeline.evaluate():234
java.util.stream.ReferencePipeline.collect():499
org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():178
org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.ArrayIndexOutOfBoundsException) 6
org.apache.orc.OrcFile$WriterVersion.from():145
org.apache.orc.impl.OrcTail.getWriterVersion():74
org.apache.orc.impl.ReaderImpl.<init>():385
org.apache.hadoop.hive.ql.io.orc.ReaderImpl.<init>():62
org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader():89
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails():1380
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal():1266
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2600():1064
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run():1246
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run():1243
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1595
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call():1243
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call():1064
java.util.concurrent.FutureTask.run():266
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
at org.apache.drill.jdbc.impl.DrillCursor.nextRowInternally(DrillCursor.java:528)
at org.apache.drill.jdbc.impl.DrillCursor.loadInitialSchema(DrillCursor.java:600)
at org.apache.drill.jdbc.impl.DrillResultSetImpl.execute(DrillResultSetImpl.java:1288)
at org.apache.drill.jdbc.impl.DrillResultSetImpl.execute(DrillResultSetImpl.java:61)
at oadd.org.apache.calcite.avatica.AvaticaConnection$1.execute(AvaticaConnection.java:667)
at org.apache.drill.jdbc.impl.DrillMetaImpl.prepareAndExecute(DrillMetaImpl.java:1109)
at org.apache.drill.jdbc.impl.DrillMetaImpl.prepareAndExecute(DrillMetaImpl.java:1120)
at oadd.org.apache.calcite.avatica.AvaticaConnection.prepareAndExecuteInternal(AvaticaConnection.java:675)
at org.apache.drill.jdbc.impl.DrillConnectionImpl.prepareAndExecuteInternal(DrillConnectionImpl.java:196)
at oadd.org.apache.calcite.avatica.AvaticaStatement.executeInternal(AvaticaStatement.java:156)
at oadd.org.apache.calcite.avatica.AvaticaStatement.executeQuery(AvaticaStatement.java:227)
at org.apache.drill.test.framework.DrillTestJdbc.executeQuery(DrillTestJdbc.java:210)
at org.apache.drill.test.framework.DrillTestJdbc.run(DrillTestJdbc.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: oadd.org.apache.drill.common.exceptions.UserRemoteException: SYSTEM ERROR: ArrayIndexOutOfBoundsException: 6
{code}
was:
Query: Functional/partition_pruning/hive/general/plan/orc1.q
{code}
select * from hive.orc_create_people_dp where state = 'Ca'
java.sql.SQLException: SYSTEM ERROR: ArrayIndexOutOfBoundsException: 6
{code}
Stack Trace:
{code}
(org.apache.drill.exec.work.foreman.ForemanException) Unexpected exception during fragment initialization: Error while applying rule Prel.ScanPrule, args [rel#2103503:DrillScanRel.LOGICAL.ANY([]).[](table=[hive, orc_create_people_dp],groupscan=HiveScan [table=Table(dbName:default, tableName:orc_create_people_dp), columns=[`id`, `first_name`, `last_name`, `address`, `state`, `**`], numPartitions=1, partitions= [Partition(values:[Ca])], inputDirectories=[maprfs:/drill/testdata/hive_storage/orc_create_people_dp/state=Ca], confProperties={}])]
org.apache.drill.exec.work.foreman.Foreman.run():300
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) Error while applying rule Prel.ScanPrule, args [rel#2103503:DrillScanRel.LOGICAL.ANY([]).[](table=[hive, orc_create_people_dp],groupscan=HiveScan [table=Table(dbName:default, tableName:orc_create_people_dp), columns=[`id`, `first_name`, `last_name`, `address`, `state`, `**`], numPartitions=1, partitions= [Partition(values:[Ca])], inputDirectories=[maprfs:/drill/testdata/hive_storage/orc_create_people_dp/state=Ca], confProperties={}])]
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():236
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (org.apache.drill.common.exceptions.DrillRuntimeException) Failed to get InputSplits
org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():182
org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.RuntimeException) ORC split generation failed with exception: java.lang.ArrayIndexOutOfBoundsException: 6
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo():1579
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits():1665
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$splitInputWithUGI$2():258
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1595
org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():247
org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():156
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$getInputSplits$0():177
java.util.stream.ReferencePipeline$7$1.accept():267
java.util.ArrayList$ArrayListSpliterator.forEachRemaining():1382
java.util.stream.AbstractPipeline.copyInto():481
java.util.stream.AbstractPipeline.wrapAndCopyInto():471
java.util.stream.ReduceOps$ReduceOp.evaluateSequential():708
java.util.stream.AbstractPipeline.evaluate():234
java.util.stream.ReferencePipeline.collect():499
org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():178
org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.util.concurrent.ExecutionException) java.lang.ArrayIndexOutOfBoundsException: 6
java.util.concurrent.FutureTask.report():122
java.util.concurrent.FutureTask.get():192
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo():1573
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits():1665
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$splitInputWithUGI$2():258
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1595
org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():247
org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():156
org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$getInputSplits$0():177
java.util.stream.ReferencePipeline$7$1.accept():267
java.util.ArrayList$ArrayListSpliterator.forEachRemaining():1382
java.util.stream.AbstractPipeline.copyInto():481
java.util.stream.AbstractPipeline.wrapAndCopyInto():471
java.util.stream.ReduceOps$ReduceOp.evaluateSequential():708
java.util.stream.AbstractPipeline.evaluate():234
java.util.stream.ReferencePipeline.collect():499
org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():178
org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
org.apache.calcite.tools.Programs$RuleSetProgram.run():339
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
org.apache.drill.exec.work.foreman.Foreman.runSQL():584
org.apache.drill.exec.work.foreman.Foreman.run():272
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
Caused By (java.lang.ArrayIndexOutOfBoundsException) 6
org.apache.orc.OrcFile$WriterVersion.from():145
org.apache.orc.impl.OrcTail.getWriterVersion():74
org.apache.orc.impl.ReaderImpl.<init>():385
org.apache.hadoop.hive.ql.io.orc.ReaderImpl.<init>():62
org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader():89
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails():1380
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal():1266
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2600():1064
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run():1246
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run():1243
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():422
org.apache.hadoop.security.UserGroupInformation.doAs():1595
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call():1243
org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call():1064
java.util.concurrent.FutureTask.run():266
java.util.concurrent.ThreadPoolExecutor.runWorker():1149
java.util.concurrent.ThreadPoolExecutor$Worker.run():624
java.lang.Thread.run():748
at org.apache.drill.jdbc.impl.DrillCursor.nextRowInternally(DrillCursor.java:528)
at org.apache.drill.jdbc.impl.DrillCursor.loadInitialSchema(DrillCursor.java:600)
at org.apache.drill.jdbc.impl.DrillResultSetImpl.execute(DrillResultSetImpl.java:1288)
at org.apache.drill.jdbc.impl.DrillResultSetImpl.execute(DrillResultSetImpl.java:61)
at oadd.org.apache.calcite.avatica.AvaticaConnection$1.execute(AvaticaConnection.java:667)
at org.apache.drill.jdbc.impl.DrillMetaImpl.prepareAndExecute(DrillMetaImpl.java:1109)
at org.apache.drill.jdbc.impl.DrillMetaImpl.prepareAndExecute(DrillMetaImpl.java:1120)
at oadd.org.apache.calcite.avatica.AvaticaConnection.prepareAndExecuteInternal(AvaticaConnection.java:675)
at org.apache.drill.jdbc.impl.DrillConnectionImpl.prepareAndExecuteInternal(DrillConnectionImpl.java:196)
at oadd.org.apache.calcite.avatica.AvaticaStatement.executeInternal(AvaticaStatement.java:156)
at oadd.org.apache.calcite.avatica.AvaticaStatement.executeQuery(AvaticaStatement.java:227)
at org.apache.drill.test.framework.DrillTestJdbc.executeQuery(DrillTestJdbc.java:210)
at org.apache.drill.test.framework.DrillTestJdbc.run(DrillTestJdbc.java:115)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: oadd.org.apache.drill.common.exceptions.UserRemoteException: SYSTEM ERROR: ArrayIndexOutOfBoundsException: 6
> Queries on Hive 2.3.x fails with SYSTEM ERROR: ArrayIndexOutOfBoundsException
> -----------------------------------------------------------------------------
>
> Key: DRILL-6771
> URL: https://issues.apache.org/jira/browse/DRILL-6771
> Project: Apache Drill
> Issue Type: Bug
> Components: Query Planning & Optimization, Storage - Hive
> Affects Versions: 1.15.0
> Environment: Hive 2.3.3
> MapR 6.1.0
> Reporter: Abhishek Girish
> Assignee: Bohdan Kazydub
> Priority: Critical
> Fix For: 1.15.0
>
>
> Query: Functional/partition_pruning/hive/general/plan/orc1.q
> {code}
> select * from hive.orc_create_people_dp where state = 'Ca'
> java.sql.SQLException: SYSTEM ERROR: ArrayIndexOutOfBoundsException: 6
> {code}
> Stack Trace:
> {code}
> (org.apache.drill.exec.work.foreman.ForemanException) Unexpected exception during fragment initialization: Error while applying rule Prel.ScanPrule, args [rel#2103503:DrillScanRel.LOGICAL.ANY([]).[](table=[hive, orc_create_people_dp],groupscan=HiveScan [table=Table(dbName:default, tableName:orc_create_people_dp), columns=[`id`, `first_name`, `last_name`, `address`, `state`, `**`], numPartitions=1, partitions= [Partition(values:[Ca])], inputDirectories=[maprfs:/drill/testdata/hive_storage/orc_create_people_dp/state=Ca], confProperties={}])]
> org.apache.drill.exec.work.foreman.Foreman.run():300
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (java.lang.RuntimeException) Error while applying rule Prel.ScanPrule, args [rel#2103503:DrillScanRel.LOGICAL.ANY([]).[](table=[hive, orc_create_people_dp],groupscan=HiveScan [table=Table(dbName:default, tableName:orc_create_people_dp), columns=[`id`, `first_name`, `last_name`, `address`, `state`, `**`], numPartitions=1, partitions= [Partition(values:[Ca])], inputDirectories=[maprfs:/drill/testdata/hive_storage/orc_create_people_dp/state=Ca], confProperties={}])]
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():236
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
> org.apache.calcite.tools.Programs$RuleSetProgram.run():339
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
> org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
> org.apache.drill.exec.work.foreman.Foreman.runSQL():584
> org.apache.drill.exec.work.foreman.Foreman.run():272
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (org.apache.drill.common.exceptions.DrillRuntimeException) Failed to get InputSplits
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():182
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
> org.apache.calcite.tools.Programs$RuleSetProgram.run():339
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
> org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
> org.apache.drill.exec.work.foreman.Foreman.runSQL():584
> org.apache.drill.exec.work.foreman.Foreman.run():272
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (java.lang.RuntimeException) ORC split generation failed with exception: java.lang.ArrayIndexOutOfBoundsException: 6
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo():1579
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits():1665
> org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$splitInputWithUGI$2():258
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1595
> org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():247
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():156
> org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$getInputSplits$0():177
> java.util.stream.ReferencePipeline$7$1.accept():267
> java.util.ArrayList$ArrayListSpliterator.forEachRemaining():1382
> java.util.stream.AbstractPipeline.copyInto():481
> java.util.stream.AbstractPipeline.wrapAndCopyInto():471
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential():708
> java.util.stream.AbstractPipeline.evaluate():234
> java.util.stream.ReferencePipeline.collect():499
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():178
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
> org.apache.calcite.tools.Programs$RuleSetProgram.run():339
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
> org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
> org.apache.drill.exec.work.foreman.Foreman.runSQL():584
> org.apache.drill.exec.work.foreman.Foreman.run():272
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (java.util.concurrent.ExecutionException) java.lang.ArrayIndexOutOfBoundsException: 6
> java.util.concurrent.FutureTask.report():122
> java.util.concurrent.FutureTask.get():192
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo():1573
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits():1665
> org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$splitInputWithUGI$2():258
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1595
> org.apache.drill.exec.store.hive.HiveMetadataProvider.splitInputWithUGI():247
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getPartitionInputSplits():156
> org.apache.drill.exec.store.hive.HiveMetadataProvider.lambda$getInputSplits$0():177
> java.util.stream.ReferencePipeline$7$1.accept():267
> java.util.ArrayList$ArrayListSpliterator.forEachRemaining():1382
> java.util.stream.AbstractPipeline.copyInto():481
> java.util.stream.AbstractPipeline.wrapAndCopyInto():471
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential():708
> java.util.stream.AbstractPipeline.evaluate():234
> java.util.stream.ReferencePipeline.collect():499
> org.apache.drill.exec.store.hive.HiveMetadataProvider.getInputSplits():178
> org.apache.drill.exec.store.hive.HiveScan.getInputSplits():288
> org.apache.drill.exec.store.hive.HiveScan.getMaxParallelizationWidth():197
> org.apache.drill.exec.planner.physical.ScanPrule.onMatch():42
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch():212
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp():648
> org.apache.calcite.tools.Programs$RuleSetProgram.run():339
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform():425
> org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToPrel():455
> org.apache.drill.exec.planner.sql.handlers.ExplainHandler.getPlan():68
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan():145
> org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan():83
> org.apache.drill.exec.work.foreman.Foreman.runSQL():584
> org.apache.drill.exec.work.foreman.Foreman.run():272
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> Caused By (java.lang.ArrayIndexOutOfBoundsException) 6
> org.apache.orc.OrcFile$WriterVersion.from():145
> org.apache.orc.impl.OrcTail.getWriterVersion():74
> org.apache.orc.impl.ReaderImpl.<init>():385
> org.apache.hadoop.hive.ql.io.orc.ReaderImpl.<init>():62
> org.apache.hadoop.hive.ql.io.orc.OrcFile.createReader():89
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.populateAndCacheStripeDetails():1380
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.callInternal():1266
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.access$2600():1064
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run():1246
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator$1.run():1243
> java.security.AccessController.doPrivileged():-2
> javax.security.auth.Subject.doAs():422
> org.apache.hadoop.security.UserGroupInformation.doAs():1595
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call():1243
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$SplitGenerator.call():1064
> java.util.concurrent.FutureTask.run():266
> java.util.concurrent.ThreadPoolExecutor.runWorker():1149
> java.util.concurrent.ThreadPoolExecutor$Worker.run():624
> java.lang.Thread.run():748
> at org.apache.drill.jdbc.impl.DrillCursor.nextRowInternally(DrillCursor.java:528)
> at org.apache.drill.jdbc.impl.DrillCursor.loadInitialSchema(DrillCursor.java:600)
> at org.apache.drill.jdbc.impl.DrillResultSetImpl.execute(DrillResultSetImpl.java:1288)
> at org.apache.drill.jdbc.impl.DrillResultSetImpl.execute(DrillResultSetImpl.java:61)
> at oadd.org.apache.calcite.avatica.AvaticaConnection$1.execute(AvaticaConnection.java:667)
> at org.apache.drill.jdbc.impl.DrillMetaImpl.prepareAndExecute(DrillMetaImpl.java:1109)
> at org.apache.drill.jdbc.impl.DrillMetaImpl.prepareAndExecute(DrillMetaImpl.java:1120)
> at oadd.org.apache.calcite.avatica.AvaticaConnection.prepareAndExecuteInternal(AvaticaConnection.java:675)
> at org.apache.drill.jdbc.impl.DrillConnectionImpl.prepareAndExecuteInternal(DrillConnectionImpl.java:196)
> at oadd.org.apache.calcite.avatica.AvaticaStatement.executeInternal(AvaticaStatement.java:156)
> at oadd.org.apache.calcite.avatica.AvaticaStatement.executeQuery(AvaticaStatement.java:227)
> at org.apache.drill.test.framework.DrillTestJdbc.executeQuery(DrillTestJdbc.java:210)
> at org.apache.drill.test.framework.DrillTestJdbc.run(DrillTestJdbc.java:115)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: oadd.org.apache.drill.common.exceptions.UserRemoteException: SYSTEM ERROR: ArrayIndexOutOfBoundsException: 6
> {code}
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)