You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@drill.apache.org by "Stefan Hammer (Jira)" <ji...@apache.org> on 2020/02/12 13:22:00 UTC

[jira] [Comment Edited] (DRILL-5733) Unable to SELECT from parquet file with Hadoop 2.7.4

    [ https://issues.apache.org/jira/browse/DRILL-5733?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17035350#comment-17035350 ] 

Stefan Hammer edited comment on DRILL-5733 at 2/12/20 1:21 PM:
---------------------------------------------------------------

Hello [~ihuzenko] and [~volodymyr] ,

Thanks for fixing that issue. Unfortunately, your changes cause a fatal regression on our systems.

When back-porting your changes ([https://github.com/apache/drill/pull/1969/files)] to the release 1.17.0 the server fails to start with an AuthorizationException while the release 1.17.0 works perfectly (except the issue mentioned here):

 
{code:java}
Exception in thread "main" org.apache.drill.exec.exception.DrillbitStartupException: Failure during initial startup of Drillbit.
     at org.apache.drill.exec.server.Drillbit.start(Drillbit.java:584)
     at org.apache.drill.exec.server.Drillbit.start(Drillbit.java:550)
     at org.apache.drill.exec.server.Drillbit.main(Drillbit.java:546)
Caused by: org.apache.drill.common.exceptions.DrillRuntimeException: Error during udf area creation [/user/user/drll/udf/registry] on file system [hdfs://cluster]
     at org.apache.drill.common.exceptions.DrillRuntimeException.format(DrillRuntimeException.java:49)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.createArea(RemoteFunctionRegistry.java:280)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.prepareAreas(RemoteFunctionRegistry.java:237)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.init(RemoteFunctionRegistry.java:107)
     at org.apache.drill.exec.server.Drillbit.run(Drillbit.java:227)
     at org.apache.drill.exec.server.Drillbit.start(Drillbit.java:580)
     ... 2 more
Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: USER@REALM is not allowed to impersonate user@REALM
     at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1545)
     at org.apache.hadoop.ipc.Client.call(Client.java:1491)
     at org.apache.hadoop.ipc.Client.call(Client.java:1388)
     at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233)
     at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:118)
     at com.sun.proxy.$Proxy48.mkdirs(Unknown Source)
     at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:660)
     at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
     at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     at java.base/java.lang.reflect.Method.invoke(Method.java:566)
     at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
     at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
     at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
     at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
     at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
     at com.sun.proxy.$Proxy49.mkdirs(Unknown Source)
     at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2425)
     at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2401)
     at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1318)
     at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1315)
     at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
     at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1332)
     at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1307)
     at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:2275)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.createArea(RemoteFunctionRegistry.java:257)
     ... 6 more
Apache Drill terminated
{code}
Should I open a new issue for that? If you need more information or tests, I am more than glad to help here!

Thanks for your support!

 


was (Author: s.h.):
Hello [~ihuzenko] ! Thanks for fixing that issue. Unfortunately, your changes cause a fatal regression on our systems.

When back-porting your changes ([https://github.com/apache/drill/pull/1969/files)] to the release 1.17.0 the server fails to start with an AuthorizationException while the release 1.17.0 works perfectly (except the issue mentioned here):

 
{code:java}
Exception in thread "main" org.apache.drill.exec.exception.DrillbitStartupException: Failure during initial startup of Drillbit.
     at org.apache.drill.exec.server.Drillbit.start(Drillbit.java:584)
     at org.apache.drill.exec.server.Drillbit.start(Drillbit.java:550)
     at org.apache.drill.exec.server.Drillbit.main(Drillbit.java:546)
Caused by: org.apache.drill.common.exceptions.DrillRuntimeException: Error during udf area creation [/user/user/drll/udf/registry] on file system [hdfs://cluster]
     at org.apache.drill.common.exceptions.DrillRuntimeException.format(DrillRuntimeException.java:49)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.createArea(RemoteFunctionRegistry.java:280)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.prepareAreas(RemoteFunctionRegistry.java:237)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.init(RemoteFunctionRegistry.java:107)
     at org.apache.drill.exec.server.Drillbit.run(Drillbit.java:227)
     at org.apache.drill.exec.server.Drillbit.start(Drillbit.java:580)
     ... 2 more
Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: USER@REALM is not allowed to impersonate user@REALM
     at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1545)
     at org.apache.hadoop.ipc.Client.call(Client.java:1491)
     at org.apache.hadoop.ipc.Client.call(Client.java:1388)
     at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233)
     at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:118)
     at com.sun.proxy.$Proxy48.mkdirs(Unknown Source)
     at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:660)
     at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
     at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
     at java.base/java.lang.reflect.Method.invoke(Method.java:566)
     at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
     at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
     at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
     at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
     at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
     at com.sun.proxy.$Proxy49.mkdirs(Unknown Source)
     at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2425)
     at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2401)
     at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1318)
     at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1315)
     at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
     at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1332)
     at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1307)
     at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:2275)
     at org.apache.drill.exec.expr.fn.registry.RemoteFunctionRegistry.createArea(RemoteFunctionRegistry.java:257)
     ... 6 more
Apache Drill terminated
{code}
Should I open a new issue for that? If you need more information or tests, I am more than glad to help here!

Thanks for your support!

 

> Unable to SELECT from parquet file with Hadoop 2.7.4
> ----------------------------------------------------
>
>                 Key: DRILL-5733
>                 URL: https://issues.apache.org/jira/browse/DRILL-5733
>             Project: Apache Drill
>          Issue Type: Bug
>    Affects Versions: 1.11.0
>            Reporter: Michele Lamarca
>            Assignee: Vova Vysotskyi
>            Priority: Major
>              Labels: ready-to-commit
>             Fix For: 1.18.0
>
>
> {{SELECT * FROM hdfs.`/user/drill/nation.parquet`;}} fails with Hadoop 2.7.4 with {noformat}
> 1/2          SELECT * FROM hdfs.`/user/drill/nation.parquet`;
> Error: SYSTEM ERROR: RemoteException: /user/drill/nation.parquet (is
> not a directory)
>         at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkTraverse(FSPermissionChecker.java:272)
>         at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:215)
>         at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:199)
>         at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1752)
>         at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:100)
>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:3820)
>         at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1012)
>         at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:855)
>         at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2217)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2213)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:422)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2213)
> {noformat}
> Query correctly executes with Hadoop 2.7.3, while it fails with:
> - Hadoop 2.7.4 with Drill 1.11 (default pom.xml)
> - Hadoop 2.7.4 with Drill 1.11 (with -Dhadoop.version=2.7.4)
> - Hadoop 2.8.0 with Drill 1.11 (default pom.xml)
> - Hadoop 3.0.0-alpha4 with Drill 1.11 (default pom.xml)
> thus looking related to https://issues.apache.org/jira/browse/HDFS-10673
> Temporary workaround consists in querying on an enclosing directory, as suggested by [~kkhatua] on drill-user mailinglist.
> Relevant stacktrace from drillbit log
> {noformat}
> 2017-08-19 09:00:45,570 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.drill.exec.work.foreman.Foreman - Query text for query id 26681de9-2b48-2c3a-cc7c-2c7ceeb1beae: SELECT * FROM hdfs.`/user/drill/nation.parquet`
> 2017-08-19 09:00:45,571 [UserServer-1] WARN  o.a.drill.exec.rpc.user.UserServer - Message of mode REQUEST of rpc type 3 took longer than 500ms.  Actual duration was 7137ms.
> 2017-08-19 09:00:45,617 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 7 classes for org.apache.drill.exec.store.dfs.FormatPlugin took 0ms
> 2017-08-19 09:00:45,618 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 8 classes for org.apache.drill.common.logical.FormatPluginConfig took 0ms
> 2017-08-19 09:00:45,619 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 8 classes for org.apache.drill.common.logical.FormatPluginConfig took 0ms
> 2017-08-19 09:00:45,619 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 8 classes for org.apache.drill.common.logical.FormatPluginConfig took 0ms
> 2017-08-19 09:00:45,648 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 7 classes for org.apache.drill.exec.store.dfs.FormatPlugin took 0ms
> 2017-08-19 09:00:45,649 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 8 classes for org.apache.drill.common.logical.FormatPluginConfig took 0ms
> 2017-08-19 09:00:45,649 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 8 classes for org.apache.drill.common.logical.FormatPluginConfig took 0ms
> 2017-08-19 09:00:45,650 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.c.s.persistence.ScanResult - loading 8 classes for org.apache.drill.common.logical.FormatPluginConfig took 0ms
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,726 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] INFO  o.a.d.exec.store.dfs.FileSelection - FileSelection.getStatuses() took 0 ms, numFiles: 1
> 2017-08-19 09:00:45,775 [26681de9-2b48-2c3a-cc7c-2c7ceeb1beae:foreman] ERROR o.a.drill.exec.work.foreman.Foreman - SYSTEM ERROR: RemoteException: /user/drill/nation.parquet (is not a directory)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkTraverse(FSPermissionChecker.java:272)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:215)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:199)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1752)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:100)
>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:3820)
>     at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1012)
>     at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:855)
>     at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>     at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2217)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2213)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:422)
>     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2213)
> [Error Id: 8f351c63-d3f7-4b61-a5e6-1a09c6c2ba8d on node001.cm.cluster:31010]
> org.apache.drill.common.exceptions.UserException: SYSTEM ERROR: RemoteException: /user/drill/nation.parquet (is not a directory)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkTraverse(FSPermissionChecker.java:272)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:215)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:199)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1752)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:100)
>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:3820)
>     at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1012)
>     at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:855)
>     at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>     at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2217)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2213)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:422)
>     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2213)
> [Error Id: 8f351c63-d3f7-4b61-a5e6-1a09c6c2ba8d on node001.cm.cluster:31010]
>     at org.apache.drill.common.exceptions.UserException$Builder.build(UserException.java:550) ~[drill-common-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.work.foreman.Foreman$ForemanResult.close(Foreman.java:847) [drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.work.foreman.Foreman.moveToState(Foreman.java:977) [drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:297) [drill-java-exec-1.11.0.jar:1.11.0]
>     at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [na:1.8.0_141]
>     at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [na:1.8.0_141]
>     at java.lang.Thread.run(Thread.java:748) [na:1.8.0_141]
> Caused by: org.apache.drill.exec.work.foreman.ForemanException: Unexpected exception during fragment initialization: Internal error: Error while applying rule DrillTableRule, args [rel#152:EnumerableTableScan.ENUMERABLE.ANY([]).[](table=[hdfs, /user/drill/nation.parquet])]
>     ... 4 common frames omitted
> Caused by: java.lang.AssertionError: Internal error: Error while applying rule DrillTableRule, args [rel#152:EnumerableTableScan.ENUMERABLE.ANY([]).[](table=[hdfs, /user/drill/nation.parquet])]
>     at org.apache.calcite.util.Util.newInternal(Util.java:792) ~[calcite-core-1.4.0-drill-r21.jar:1.4.0-drill-r21]
>     at org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:251) ~[calcite-core-1.4.0-drill-r21.jar:1.4.0-drill-r21]
>     at org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:811) ~[calcite-core-1.4.0-drill-r21.jar:1.4.0-drill-r21]
>     at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:310) ~[calcite-core-1.4.0-drill-r21.jar:1.4.0-drill-r21]
>     at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:401) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.transform(DefaultSqlHandler.java:343) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToRawDrel(DefaultSqlHandler.java:242) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.convertToDrel(DefaultSqlHandler.java:292) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.sql.handlers.DefaultSqlHandler.getPlan(DefaultSqlHandler.java:169) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.sql.DrillSqlWorker.getQueryPlan(DrillSqlWorker.java:131) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.sql.DrillSqlWorker.getPlan(DrillSqlWorker.java:79) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.work.foreman.Foreman.runSQL(Foreman.java:1050) [drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.work.foreman.Foreman.run(Foreman.java:280) [drill-java-exec-1.11.0.jar:1.11.0]
>     ... 3 common frames omitted
> Caused by: org.apache.drill.common.exceptions.DrillRuntimeException: Failure creating scan.
>     at org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:92) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:70) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:63) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.logical.DrillScanRule.onMatch(DrillScanRule.java:37) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:228) ~[calcite-core-1.4.0-drill-r21.jar:1.4.0-drill-r21]
>     ... 14 common frames omitted
> Caused by: org.apache.hadoop.security.AccessControlException: /user/drill/nation.parquet (is not a directory)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkTraverse(FSPermissionChecker.java:272)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:215)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:199)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1752)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:100)
>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:3820)
>     at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1012)
>     at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:855)
>     at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>     at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2217)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2213)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:422)
>     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2213)
>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[na:1.8.0_141]
>     at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) ~[na:1.8.0_141]
>     at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[na:1.8.0_141]
>     at java.lang.reflect.Constructor.newInstance(Constructor.java:423) ~[na:1.8.0_141]
>     at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) ~[hadoop-common-2.7.1.jar:na]
>     at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73) ~[hadoop-common-2.7.1.jar:na]
>     at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2118) ~[hadoop-hdfs-2.7.1.jar:na]
>     at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305) ~[hadoop-hdfs-2.7.1.jar:na]
>     at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301) ~[hadoop-hdfs-2.7.1.jar:na]
>     at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-2.7.1.jar:na]
>     at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1317) ~[hadoop-hdfs-2.7.1.jar:na]
>     at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1424) ~[hadoop-common-2.7.1.jar:na]
>     at org.apache.drill.exec.store.dfs.DrillFileSystem.exists(DrillFileSystem.java:603) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.store.parquet.ParquetGroupScan.expandIfNecessary(ParquetGroupScan.java:270) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:207) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.store.parquet.ParquetGroupScan.<init>(ParquetGroupScan.java:186) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:170) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.store.parquet.ParquetFormatPlugin.getGroupScan(ParquetFormatPlugin.java:66) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.store.dfs.FileSystemPlugin.getPhysicalScan(FileSystemPlugin.java:144) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.store.AbstractStoragePlugin.getPhysicalScan(AbstractStoragePlugin.java:100) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.logical.DrillTable.getGroupScan(DrillTable.java:85) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     at org.apache.drill.exec.planner.logical.DrillScanRel.<init>(DrillScanRel.java:90) ~[drill-java-exec-1.11.0.jar:1.11.0]
>     ... 18 common frames omitted
> Caused by: org.apache.hadoop.ipc.RemoteException: /user/drill/nation.parquet (is not a directory)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkTraverse(FSPermissionChecker.java:272)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:215)
>     at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:199)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1752)
>     at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getFileInfo(FSDirStatAndListingOp.java:100)
>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:3820)
>     at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getFileInfo(NameNodeRpcServer.java:1012)
>     at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getFileInfo(ClientNamenodeProtocolServerSideTranslatorPB.java:855)
>     at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>     at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
>     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2217)
>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2213)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:422)
>     at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1746)
>     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2213)
>     at org.apache.hadoop.ipc.Client.call(Client.java:1476) ~[hadoop-common-2.7.1.jar:na]
>     at org.apache.hadoop.ipc.Client.call(Client.java:1407) ~[hadoop-common-2.7.1.jar:na]
>     at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229) ~[hadoop-common-2.7.1.jar:na]
>     at com.sun.proxy.$Proxy65.getFileInfo(Unknown Source) ~[na:na]
>     at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:771) ~[hadoop-hdfs-2.7.1.jar:na]
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_141]
>     at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_141]
>     at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_141]
>     at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_141]
>     at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187) ~[hadoop-common-2.7.1.jar:na]
>     at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) ~[hadoop-common-2.7.1.jar:na]
>     at com.sun.proxy.$Proxy66.getFileInfo(Unknown Source) ~[na:na]
>     at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2116) ~[hadoop-hdfs-2.7.1.jar:na]
>     ... 33 common frames omitted
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)