You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@falcon.apache.org by "Shwetha G S (JIRA)" <ji...@apache.org> on 2014/08/20 08:54:28 UTC

[jira] [Updated] (FALCON-616) cluster submission should fail when shared libs copy fail

     [ https://issues.apache.org/jira/browse/FALCON-616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Shwetha G S updated FALCON-616:
-------------------------------

    Attachment: FALCON-616.patch

> cluster submission should fail when shared libs copy fail 
> ----------------------------------------------------------
>
>                 Key: FALCON-616
>                 URL: https://issues.apache.org/jira/browse/FALCON-616
>             Project: Falcon
>          Issue Type: Bug
>          Components: general
>    Affects Versions: 0.5
>         Environment: QA
>            Reporter: Samarth Gupta
>            Assignee: Shwetha G S
>         Attachments: FALCON-616.patch
>
>
> during cluster submission falcon copies shared libs to hdfs. If coz of some reason if this copy fails , cluster submission should also if. Currently submission is successful even if shared lib are not copied. 
> following exception can be seen in application logs :
> (for example when user dont have access to write to hdfs) 
> {code}
>         at org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542)
>         at org.mortbay.jetty.HttpConnection$RequestHandler.content(HttpConnection.java:945)
>         at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:756)
>         at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:218)
>         at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404)
>         at org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:228)
>         at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
> Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=dataqa, access=WRITE, inode="/":hdfs:supergroup:drwxr-xr-x
>         at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:224)
>         at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:204)
>         at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:149)
>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:4716)
>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:4698)
>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:4672)
>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInternal(FSNamesystem.java:3035)
>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirsInt(FSNamesystem.java:2999)
>         at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:2980)
>         at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:648)
>         at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:419)
>         at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:44970)
>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1002)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1701)
>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1697)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:416)
>         at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1695)
>         at org.apache.hadoop.ipc.Client.call(Client.java:1225)
>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
>         at com.sun.proxy.$Proxy61.mkdirs(Unknown Source)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:622)
>         at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
>         at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
>         at com.sun.proxy.$Proxy61.mkdirs(Unknown Source)
>         at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:426)
>         at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2151)
>         ... 53 more
> 2014-08-19 07:43:00,819 INFO  - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ Adding lineage for entity: cluster-ua2, type: CLUSTER (MetadataMappingService:189)
> 2014-08-19 07:43:00,819 INFO  - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ Adding cluster entity: cluster-ua2 (EntityRelationshipGraphBuilder:51)
> 2014-08-19 07:43:00,822 DEBUG - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ Finding vertex for: name=cluster-ua2, type=CLUSTER_ENTITY (RelationshipGraphBuilder:87)
> 2014-08-19 07:43:00,830 DEBUG - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ Found an existing vertex for: name=cluster-ua2, type=CLUSTER_ENTITY (RelationshipGraphBuilder:69)
> 2014-08-19 07:43:00,831 DEBUG - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ Finding vertex for: name=ua2, type=COLO (RelationshipGraphBuilder:87)
> 2014-08-19 07:43:00,832 DEBUG - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ Found an existing vertex for: name=ua2, type=COLO (RelationshipGraphBuilder:69)
> 2014-08-19 07:43:00,835 INFO  - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ CLUSTER/cluster-ua2 is published into config store (AUDIT:176)
> 2014-08-19 07:43:00,836 INFO  - [600515448@qtp-1571280869-0:dataqa:POST//sync/submit/cluster 5e6ba9aa-fe06-4952-a30d-2f8ac5d1735e] ~ Submit successful: (cluster): cluster-ua2 (AbstractEntityManager:342)
> 2014-08-19 07:43:17,790 INFO  - [721286742@qtp-1571280869-2:dataqa:POST//sync/submit/feed 7045d7ac-748c-465c-a8e3-5a230fab7468] ~ Logging in dataqa (CurrentUser:53)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)