You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@ambari.apache.org by "Gaurav Nagar (JIRA)" <ji...@apache.org> on 2017/03/01 08:39:45 UTC
[jira] [Updated] (AMBARI-20236) HiveView 2.0: save as HDFS shows
success message on error
[ https://issues.apache.org/jira/browse/AMBARI-20236?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Gaurav Nagar updated AMBARI-20236:
----------------------------------
Status: Patch Available (was: Open)
> HiveView 2.0: save as HDFS shows success message on error
> ---------------------------------------------------------
>
> Key: AMBARI-20236
> URL: https://issues.apache.org/jira/browse/AMBARI-20236
> Project: Ambari
> Issue Type: Bug
> Components: ambari-views
> Affects Versions: 2.5.0
> Reporter: Anusha Bilgi
> Assignee: Gaurav Nagar
> Priority: Critical
> Fix For: 2.5.0
>
> Attachments: AMBARI-20236_branch-2.5.patch
>
>
> {code:java}
> Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=admin, access=WRITE, inode="/d1/d1.csv":hdfs:hdfs:drwxr-xr-x
> at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)
> at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:292)
> at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:213)
> at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)
> at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1955)
> at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1939)
> at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1922)
> at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:2736)
> at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2671)
> at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2555)
> at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:735)
> at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:408)
> at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1833)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2345)
> at org.apache.hadoop.hdfs.web.JsonUtil.toRemoteException(JsonUtil.java:118)
> at org.apache.hadoop.hdfs.web.WebHdfsFileSystem.validateResponse(WebHdfsFileSystem.java:480)
> ... 8 more
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)