You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@drill.apache.org by "Anton Gozhiy (JIRA)" <ji...@apache.org> on 2019/04/15 15:42:00 UTC

[jira] [Updated] (DRILL-7181) [Text V3 Reader] Exception with inadequate message is thrown if select columns as array with extractHeader set to true

     [ https://issues.apache.org/jira/browse/DRILL-7181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Anton Gozhiy updated DRILL-7181:
--------------------------------
    Description: 
*Prerequisites:*
 # Create a simple .csv file with header, like this:
{noformat}
col1,col2,col3
1,2,3
4,5,6
7,8,9
{noformat}

 # Set exec.storage.enable_v3_text_reader
 # Set "extractHeader": true for csv format in dfs storage plugin.

*Query:*
{code:sql}
select columns[0] from dfs.tmp.`/test.csv`
{code}
*Expected result:* Exception should happen, here is the message from V2 reader:
{noformat}
UNSUPPORTED_OPERATION ERROR: Drill Remote Exception



  (java.lang.Exception) UNSUPPORTED_OPERATION ERROR: With extractHeader enabled, only header names are supported

column name columns
column index
Fragment 0:0

[Error Id: 5affa696-1dbd-43d7-ac14-72d235c00f43 on userf87d-pc:31010]
    org.apache.drill.common.exceptions.UserException$Builder.build():630
    org.apache.drill.exec.store.easy.text.compliant.FieldVarCharOutput.<init>():106
    org.apache.drill.exec.store.easy.text.compliant.CompliantTextRecordReader.setup():139
    org.apache.drill.exec.physical.impl.ScanBatch.getNextReaderIfHas():321
    org.apache.drill.exec.physical.impl.ScanBatch.internalNext():216
    org.apache.drill.exec.physical.impl.ScanBatch.next():271
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.physical.impl.BaseRootExec.next():104
    org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():83
    org.apache.drill.exec.physical.impl.BaseRootExec.next():94
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():296
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():283
    .......():0
    org.apache.hadoop.security.UserGroupInformation.doAs():1746
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():283
    org.apache.drill.common.SelfCleaningRunnable.run():38
    .......():0
{noformat}
*Actual result:* The exception message is inadequate:
{noformat}
org.apache.drill.common.exceptions.UserRemoteException: EXECUTION_ERROR ERROR: Table schema must have exactly one column.

Exception thrown from org.apache.drill.exec.physical.impl.scan.ScanOperatorExec
Fragment 0:0

[Error Id: a76a1576-419a-413f-840f-088157167a6d on userf87d-pc:31010]

  (java.lang.IllegalStateException) Table schema must have exactly one column.
    org.apache.drill.exec.physical.impl.scan.columns.ColumnsArrayManager.resolveColumn():108
    org.apache.drill.exec.physical.impl.scan.project.ReaderLevelProjection.resolveSpecial():91
    org.apache.drill.exec.physical.impl.scan.project.ExplicitSchemaProjection.resolveRootTuple():62
    org.apache.drill.exec.physical.impl.scan.project.ExplicitSchemaProjection.<init>():52
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.doExplicitProjection():223
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.reviseOutputProjection():155
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.endBatch():117
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.defineSchema():94
    org.apache.drill.exec.physical.impl.scan.framework.ShimBatchReader.defineSchema():105
    org.apache.drill.exec.physical.impl.scan.ReaderState.buildSchema():300
    org.apache.drill.exec.physical.impl.scan.ScanOperatorExec.nextAction():182
    org.apache.drill.exec.physical.impl.scan.ScanOperatorExec.buildSchema():122
    org.apache.drill.exec.physical.impl.protocol.OperatorDriver.start():160
    org.apache.drill.exec.physical.impl.protocol.OperatorDriver.next():112
    org.apache.drill.exec.physical.impl.protocol.OperatorRecordBatch.next():147
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.physical.impl.BaseRootExec.next():104
    org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():83
    org.apache.drill.exec.physical.impl.BaseRootExec.next():94
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():296
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():283
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1746
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():283
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
{noformat}

  was:
*Prerequisites:*
# Create a simple .csv file with header, like this:
{noformat}
col1,col2,col3
1,2,3
4,5,6
7,8,9
{noformat}
# exec.storage.enable_v3_text_reader
# Set "extractHeader": true for csv format in dfs storage plugin.

*Query:*
{code:sql}
select columns[0] from dfs.tmp.`/test.csv`
{code}

*Expected result:* Exception should happen, here is the message from V2 reader:
{noformat}
UNSUPPORTED_OPERATION ERROR: Drill Remote Exception



  (java.lang.Exception) UNSUPPORTED_OPERATION ERROR: With extractHeader enabled, only header names are supported

column name columns
column index
Fragment 0:0

[Error Id: 5affa696-1dbd-43d7-ac14-72d235c00f43 on userf87d-pc:31010]
    org.apache.drill.common.exceptions.UserException$Builder.build():630
    org.apache.drill.exec.store.easy.text.compliant.FieldVarCharOutput.<init>():106
    org.apache.drill.exec.store.easy.text.compliant.CompliantTextRecordReader.setup():139
    org.apache.drill.exec.physical.impl.ScanBatch.getNextReaderIfHas():321
    org.apache.drill.exec.physical.impl.ScanBatch.internalNext():216
    org.apache.drill.exec.physical.impl.ScanBatch.next():271
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.physical.impl.BaseRootExec.next():104
    org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():83
    org.apache.drill.exec.physical.impl.BaseRootExec.next():94
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():296
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():283
    .......():0
    org.apache.hadoop.security.UserGroupInformation.doAs():1746
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():283
    org.apache.drill.common.SelfCleaningRunnable.run():38
    .......():0
{noformat}

*Actual result:* The exception message is inadequate:
{noformat}
org.apache.drill.common.exceptions.UserRemoteException: EXECUTION_ERROR ERROR: Table schema must have exactly one column.

Exception thrown from org.apache.drill.exec.physical.impl.scan.ScanOperatorExec
Fragment 0:0

[Error Id: a76a1576-419a-413f-840f-088157167a6d on userf87d-pc:31010]

  (java.lang.IllegalStateException) Table schema must have exactly one column.
    org.apache.drill.exec.physical.impl.scan.columns.ColumnsArrayManager.resolveColumn():108
    org.apache.drill.exec.physical.impl.scan.project.ReaderLevelProjection.resolveSpecial():91
    org.apache.drill.exec.physical.impl.scan.project.ExplicitSchemaProjection.resolveRootTuple():62
    org.apache.drill.exec.physical.impl.scan.project.ExplicitSchemaProjection.<init>():52
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.doExplicitProjection():223
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.reviseOutputProjection():155
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.endBatch():117
    org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.defineSchema():94
    org.apache.drill.exec.physical.impl.scan.framework.ShimBatchReader.defineSchema():105
    org.apache.drill.exec.physical.impl.scan.ReaderState.buildSchema():300
    org.apache.drill.exec.physical.impl.scan.ScanOperatorExec.nextAction():182
    org.apache.drill.exec.physical.impl.scan.ScanOperatorExec.buildSchema():122
    org.apache.drill.exec.physical.impl.protocol.OperatorDriver.start():160
    org.apache.drill.exec.physical.impl.protocol.OperatorDriver.next():112
    org.apache.drill.exec.physical.impl.protocol.OperatorRecordBatch.next():147
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.record.AbstractRecordBatch.next():126
    org.apache.drill.exec.record.AbstractRecordBatch.next():116
    org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
    org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
    org.apache.drill.exec.record.AbstractRecordBatch.next():186
    org.apache.drill.exec.physical.impl.BaseRootExec.next():104
    org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():83
    org.apache.drill.exec.physical.impl.BaseRootExec.next():94
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():296
    org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():283
    java.security.AccessController.doPrivileged():-2
    javax.security.auth.Subject.doAs():422
    org.apache.hadoop.security.UserGroupInformation.doAs():1746
    org.apache.drill.exec.work.fragment.FragmentExecutor.run():283
    org.apache.drill.common.SelfCleaningRunnable.run():38
    java.util.concurrent.ThreadPoolExecutor.runWorker():1149
    java.util.concurrent.ThreadPoolExecutor$Worker.run():624
    java.lang.Thread.run():748
{noformat}


> [Text V3 Reader] Exception with inadequate message is thrown if select columns as array with extractHeader set to true
> ----------------------------------------------------------------------------------------------------------------------
>
>                 Key: DRILL-7181
>                 URL: https://issues.apache.org/jira/browse/DRILL-7181
>             Project: Apache Drill
>          Issue Type: Bug
>            Reporter: Anton Gozhiy
>            Priority: Major
>
> *Prerequisites:*
>  # Create a simple .csv file with header, like this:
> {noformat}
> col1,col2,col3
> 1,2,3
> 4,5,6
> 7,8,9
> {noformat}
>  # Set exec.storage.enable_v3_text_reader
>  # Set "extractHeader": true for csv format in dfs storage plugin.
> *Query:*
> {code:sql}
> select columns[0] from dfs.tmp.`/test.csv`
> {code}
> *Expected result:* Exception should happen, here is the message from V2 reader:
> {noformat}
> UNSUPPORTED_OPERATION ERROR: Drill Remote Exception
>   (java.lang.Exception) UNSUPPORTED_OPERATION ERROR: With extractHeader enabled, only header names are supported
> column name columns
> column index
> Fragment 0:0
> [Error Id: 5affa696-1dbd-43d7-ac14-72d235c00f43 on userf87d-pc:31010]
>     org.apache.drill.common.exceptions.UserException$Builder.build():630
>     org.apache.drill.exec.store.easy.text.compliant.FieldVarCharOutput.<init>():106
>     org.apache.drill.exec.store.easy.text.compliant.CompliantTextRecordReader.setup():139
>     org.apache.drill.exec.physical.impl.ScanBatch.getNextReaderIfHas():321
>     org.apache.drill.exec.physical.impl.ScanBatch.internalNext():216
>     org.apache.drill.exec.physical.impl.ScanBatch.next():271
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.physical.impl.BaseRootExec.next():104
>     org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():83
>     org.apache.drill.exec.physical.impl.BaseRootExec.next():94
>     org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():296
>     org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():283
>     .......():0
>     org.apache.hadoop.security.UserGroupInformation.doAs():1746
>     org.apache.drill.exec.work.fragment.FragmentExecutor.run():283
>     org.apache.drill.common.SelfCleaningRunnable.run():38
>     .......():0
> {noformat}
> *Actual result:* The exception message is inadequate:
> {noformat}
> org.apache.drill.common.exceptions.UserRemoteException: EXECUTION_ERROR ERROR: Table schema must have exactly one column.
> Exception thrown from org.apache.drill.exec.physical.impl.scan.ScanOperatorExec
> Fragment 0:0
> [Error Id: a76a1576-419a-413f-840f-088157167a6d on userf87d-pc:31010]
>   (java.lang.IllegalStateException) Table schema must have exactly one column.
>     org.apache.drill.exec.physical.impl.scan.columns.ColumnsArrayManager.resolveColumn():108
>     org.apache.drill.exec.physical.impl.scan.project.ReaderLevelProjection.resolveSpecial():91
>     org.apache.drill.exec.physical.impl.scan.project.ExplicitSchemaProjection.resolveRootTuple():62
>     org.apache.drill.exec.physical.impl.scan.project.ExplicitSchemaProjection.<init>():52
>     org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.doExplicitProjection():223
>     org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.reviseOutputProjection():155
>     org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.endBatch():117
>     org.apache.drill.exec.physical.impl.scan.project.ReaderSchemaOrchestrator.defineSchema():94
>     org.apache.drill.exec.physical.impl.scan.framework.ShimBatchReader.defineSchema():105
>     org.apache.drill.exec.physical.impl.scan.ReaderState.buildSchema():300
>     org.apache.drill.exec.physical.impl.scan.ScanOperatorExec.nextAction():182
>     org.apache.drill.exec.physical.impl.scan.ScanOperatorExec.buildSchema():122
>     org.apache.drill.exec.physical.impl.protocol.OperatorDriver.start():160
>     org.apache.drill.exec.physical.impl.protocol.OperatorDriver.next():112
>     org.apache.drill.exec.physical.impl.protocol.OperatorRecordBatch.next():147
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.physical.impl.limit.LimitRecordBatch.innerNext():101
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.record.AbstractRecordBatch.next():126
>     org.apache.drill.exec.record.AbstractRecordBatch.next():116
>     org.apache.drill.exec.record.AbstractUnaryRecordBatch.innerNext():63
>     org.apache.drill.exec.physical.impl.project.ProjectRecordBatch.innerNext():141
>     org.apache.drill.exec.record.AbstractRecordBatch.next():186
>     org.apache.drill.exec.physical.impl.BaseRootExec.next():104
>     org.apache.drill.exec.physical.impl.ScreenCreator$ScreenRoot.innerNext():83
>     org.apache.drill.exec.physical.impl.BaseRootExec.next():94
>     org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():296
>     org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():283
>     java.security.AccessController.doPrivileged():-2
>     javax.security.auth.Subject.doAs():422
>     org.apache.hadoop.security.UserGroupInformation.doAs():1746
>     org.apache.drill.exec.work.fragment.FragmentExecutor.run():283
>     org.apache.drill.common.SelfCleaningRunnable.run():38
>     java.util.concurrent.ThreadPoolExecutor.runWorker():1149
>     java.util.concurrent.ThreadPoolExecutor$Worker.run():624
>     java.lang.Thread.run():748
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)