You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@ambari.apache.org by "Srikanth Janardhan (JIRA)" <ji...@apache.org> on 2018/05/23 04:12:00 UTC
[jira] [Resolved] (AMBARI-23933) YARN service check failure with no
permission to submit jobs for ambari-qa user
[ https://issues.apache.org/jira/browse/AMBARI-23933?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Srikanth Janardhan resolved AMBARI-23933.
-----------------------------------------
Resolution: Fixed
> YARN service check failure with no permission to submit jobs for ambari-qa user
> -------------------------------------------------------------------------------
>
> Key: AMBARI-23933
> URL: https://issues.apache.org/jira/browse/AMBARI-23933
> Project: Ambari
> Issue Type: Bug
> Components: ambari-server
> Affects Versions: 2.7.0
> Reporter: Srikanth Janardhan
> Assignee: Dmitry Lysnichenko
> Priority: Blocker
> Fix For: 2.7.0
>
>
> Service check for YARN failed:
> {code}
> stderr:
> Traceback (most recent call last):
> File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/service_check.py", line 185, in
> ServiceCheck().execute()
> File "/usr/lib/ambari-agent/lib/resource_management/libraries/script/script.py", line 353, in execute
> method(env)
> File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/service_check.py", line 121, in service_check
> user=params.smokeuser,
> File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 72, in inner
> result = function(command, **kwargs)
> File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 102, in checked_call
> tries=tries, try_sleep=try_sleep, timeout_kill_strategy=timeout_kill_strategy, returns=returns)
> File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 150, in _call_wrapper
> result = _call(command, **kwargs_copy)
> File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 308, in _call
> raise ExecutionFailed(err_msg, code, out, err)
> resource_management.core.exceptions.ExecutionFailed: Execution of '/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa@EXAMPLE.COM; yarn org.apache.hadoop.yarn.applications.distributedshell.Client -shell_command ls -num_containers 1 -jar /usr/hdp/current/hadoop-yarn-client/hadoop-yarn-applications-distributedshell.jar -timeout 300000 --queue default' returned 1. ######## Hortonworks #############
> This is MOTD message, added for testing in qe infra
> 18/05/18 11:33:09 INFO distributedshell.Client: Initializing Client
> 18/05/18 11:33:09 INFO distributedshell.Client: Running Client
> 18/05/18 11:33:09 INFO client.AHSProxy: Connecting to Application History server at ctr-e138-1518143905142-317995-01-000004.hwx.site/172.27.56.82:10200
> 18/05/18 11:33:09 INFO client.ConfiguredRMFailoverProxyProvider: Failing over to rm2
> 18/05/18 11:33:10 INFO distributedshell.Client: Got Cluster metric info from ASM, numNodeManagers=3
> 18/05/18 11:33:10 INFO distributedshell.Client: Got Cluster node info from ASM
> 18/05/18 11:33:10 INFO distributedshell.Client: Got node report from ASM for, nodeId=ctr-e138-1518143905142-317995-01-000009.hwx.site:25454, nodeAddress=ctr-e138-1518143905142-317995-01-000009.hwx.site:8042, nodeRackName=/default-rack, nodeNumContainers=1
> 18/05/18 11:33:10 INFO distributedshell.Client: Got node report from ASM for, nodeId=ctr-e138-1518143905142-317995-01-000006.hwx.site:25454, nodeAddress=ctr-e138-1518143905142-317995-01-000006.hwx.site:8042, nodeRackName=/default-rack, nodeNumContainers=0
> 18/05/18 11:33:10 INFO distributedshell.Client: Got node report from ASM for, nodeId=ctr-e138-1518143905142-317995-01-000007.hwx.site:25454, nodeAddress=ctr-e138-1518143905142-317995-01-000007.hwx.site:8042, nodeRackName=/default-rack, nodeNumContainers=0
> 18/05/18 11:33:10 INFO distributedshell.Client: Queue info, queueName=default, queueCurrentCapacity=0.027777778, queueMaxCapacity=1.0, queueApplicationCount=0, queueChildQueueCount=0
> 18/05/18 11:33:10 INFO distributedshell.Client: Max mem capability of resources in this cluster 12288
> 18/05/18 11:33:10 INFO distributedshell.Client: Max virtual cores capability of resources in this cluster 1
> 18/05/18 11:33:10 WARN distributedshell.Client: AM Memory not specified, use 100 mb as AM memory
> 18/05/18 11:33:10 WARN distributedshell.Client: AM vcore not specified, use 1 mb as AM vcores
> 18/05/18 11:33:10 WARN distributedshell.Client: AM Resource capability=
> 18/05/18 11:33:10 INFO distributedshell.Client: Copy App Master jar from local filesystem and add to local environment
> 18/05/18 11:33:11 INFO distributedshell.Client: Set the environment for the application master
> 18/05/18 11:33:11 INFO distributedshell.Client: Setting up app master command
> 18/05/18 11:33:11 INFO distributedshell.Client: Completed setting up app master command {{JAVA_HOME}}/bin/java -Xmx100m org.apache.hadoop.yarn.applications.distributedshell.ApplicationMaster --container_type GUARANTEED --container_memory 10 --container_vcores 1 --num_containers 1 --priority 0 1>/AppMaster.stdout 2>/AppMaster.stderr
> 18/05/18 11:33:11 INFO hdfs.DFSClient: Created token for ambari-qa: HDFS_DELEGATION_TOKEN owner=ambari-qa@EXAMPLE.COM, renewer=cstm-yarn, realUser=, issueDate=1526643191972, maxDate=1527247991972, sequenceNumber=27, masterKeyId=4 on ha-hdfs:mycluster
> 18/05/18 11:33:12 INFO distributedshell.Client: Got dt for hdfs://mycluster; Kind: HDFS_DELEGATION_TOKEN, Service: ha-hdfs:mycluster, Ident: (token for ambari-qa: HDFS_DELEGATION_TOKEN owner=ambari-qa@EXAMPLE.COM, renewer=cstm-yarn, realUser=, issueDate=1526643191972, maxDate=1527247991972, sequenceNumber=27, masterKeyId=4)
> 18/05/18 11:33:12 INFO distributedshell.Client: Submitting application to ASM
> 18/05/18 11:33:12 INFO impl.TimelineClientImpl: Timeline service address: ctr-e138-1518143905142-317995-01-000004.hwx.site:8188
> 18/05/18 11:33:12 ERROR distributedshell.Client: Error running Client
> org.apache.hadoop.yarn.exceptions.YarnException: org.apache.hadoop.security.AccessControlException: User ambari-qa does not have permission to submit application_1526608768177_0003 to queue default
> at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38)
> at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:436)
> at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:321)
> at org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645)
> at org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277)
> at org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563)
> at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678)
> Caused by: org.apache.hadoop.security.AccessControlException: User ambari-qa does not have permission to submit application_1526608768177_0003 to queue default
> at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:438)
> ... 12 more
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
> at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53)
> at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateYarnException(RPCUtil.java:75)
> at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:116)
> at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:292)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
> at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
> at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
> at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
> at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
> at com.sun.proxy.$Proxy9.submitApplication(Unknown Source)
> at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.submitApplication(YarnClientImpl.java:286)
> at org.apache.hadoop.yarn.applications.distributedshell.Client.run(Client.java:981)
> at org.apache.hadoop.yarn.applications.distributedshell.Client.main(Client.java:265)
> Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.YarnException): org.apache.hadoop.security.AccessControlException: User ambari-qa does not have permission to submit application_1526608768177_0003 to queue default
> at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38)
> at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:436)
> at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:321)
> at org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645)
> at org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277)
> at org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563)
> at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678)
> Caused by: org.apache.hadoop.security.AccessControlException: User ambari-qa does not have permission to submit application_1526608768177_0003 to queue default
> at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:438)
> ... 12 more
> at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1491)
> at org.apache.hadoop.ipc.Client.call(Client.java:1437)
> at org.apache.hadoop.ipc.Client.call(Client.java:1347)
> at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228)
> at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
> at com.sun.proxy.$Proxy8.submitApplication(Unknown Source)
> at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:289)
> ... 13 more
> stdout:
> 2018-05-18 11:33:06,916 - Using hadoop conf dir: /usr/hdp/3.0.0.0-1356/hadoop/conf
> 2018-05-18 11:33:06,917 - Stack Feature Version Info: Cluster Stack=3.0, Command Stack=None, Command Version=3.0.0.0-1356 -> 3.0.0.0-1356
> 2018-05-18 11:33:07,046 - Using hadoop conf dir: /usr/hdp/3.0.0.0-1356/hadoop/conf
> 2018-05-18 11:33:07,107 - HdfsResource['/user/ambari-qa'] {'security_enabled': True, 'hadoop_bin_dir': '/usr/hdp/3.0.0.0-1356/hadoop/bin', 'keytab': '/etc/security/keytabs/hdfs.headless.keytab', 'dfs_type': '', 'default_fs': 'hdfs://mycluster', 'hdfs_resource_ignore_file': '/var/lib/ambari-agent/data/.hdfs_resource_ignore', 'hdfs_site': ..., 'kinit_path_local': '/usr/bin/kinit', 'principal_name': 'cstm-hdfs@EXAMPLE.COM', 'user': 'cstm-hdfs', 'owner': 'ambari-qa', 'hadoop_conf_dir': '/usr/hdp/3.0.0.0-1356/hadoop/conf', 'type': 'directory', 'action': ['create_on_execute'], 'immutable_paths': [u'/apps/hive/warehouse', u'/mr-history/done', u'/app-logs', u'/tmp'], 'mode': 0770}
> 2018-05-18 11:33:07,122 - Execute['/usr/bin/kinit -kt /etc/security/keytabs/hdfs.headless.keytab cstm-hdfs@EXAMPLE.COM'] {'user': 'cstm-hdfs'}
> 2018-05-18 11:33:07,233 - call['ambari-sudo.sh su cstm-hdfs -l -s /bin/bash -c 'curl --negotiate -u : -s '"'"'http://ctr-e138-1518143905142-317995-01-000004.hwx.site:20070/jmx?qry=Hadoop:service=NameNode,name=FSNamesystem'"'"' 1>/tmp/tmpH0Q_aW 2>/tmp/tmpmFbDsW''] {'quiet': False}
> 2018-05-18 11:33:07,340 - call returned (0, '######## Hortonworks #############\nThis is MOTD message, added for testing in qe infra')
> 2018-05-18 11:33:07,341 - get_user_call_output returned (0, u'{\n "beans" : [ {\n "name" : "Hadoop:service=NameNode,name=FSNamesystem",\n "modelerType" : "FSNamesystem",\n "tag.Context" : "dfs",\n "tag.HAState" : "standby",\n "tag.TotalSyncTimes" : "",\n "tag.Hostname" : "ctr-e138-1518143905142-317995-01-000004.hwx.site",\n "MissingBlocks" : 0,\n "MissingReplOneBlocks" : 0,\n "ExpiredHeartbeats" : 0,\n "TransactionsSinceLastCheckpoint" : 5054,\n "TransactionsSinceLastLogRoll" : 0,\n "LastWrittenTransactionId" : 6338,\n "LastCheckpointTime" : 1526630224438,\n "CapacityTotal" : 1656383242752,\n "CapacityTotalGB" : 1543.0,\n "CapacityUsed" : 3494182380,\n "CapacityUsedGB" : 3.0,\n "CapacityRemaining" : 1549721537687,\n "ProvidedCapacityTotal" : 0,\n "CapacityRemainingGB" : 1443.0,\n "CapacityUsedNonDFS" : 0,\n "TotalLoad" : 68,\n "SnapshottableDirectories" : 0,\n "Snapshots" : 0,\n "NumEncryptionZones" : 0,\n "LockQueueLength" : 0,\n "BlocksTotal" : 258,\n "NumFilesUnderConstruction" : 11,\n "NumActiveClients" : 9,\n "FilesTotal" : 1233,\n "PendingReplicationBlocks" : 0,\n "PendingReconstructionBlocks" : 0,\n "UnderReplicatedBlocks" : 0,\n "LowRedundancyBlocks" : 0,\n "CorruptBlocks" : 0,\n "ScheduledReplicationBlocks" : 0,\n "PendingDeletionBlocks" : 0,\n "LowRedundancyReplicatedBlocks" : 0,\n "CorruptReplicatedBlocks" : 0,\n "MissingReplicatedBlocks" : 0,\n "MissingReplicationOneBlocks" : 0,\n "BytesInFutureReplicatedBlocks" : 0,\n "PendingDeletionReplicatedBlocks" : 0,\n "TotalReplicatedBlocks" : 258,\n "LowRedundancyECBlockGroups" : 0,\n "CorruptECBlockGroups" : 0,\n "MissingECBlockGroups" : 0,\n "BytesInFutureECBlockGroups" : 0,\n "PendingDeletionECBlocks" : 0,\n "TotalECBlockGroups" : 0,\n "ExcessBlocks" : 0,\n "NumTimedOutPendingReconstructions" : 0,\n "PostponedMisreplicatedBlocks" : 0,\n "PendingDataNodeMessageCount" : 1,\n "MillisSinceLastLoadedEdits" : 71633,\n "BlockCapacity" : 4194304,\n "NumLiveDataNodes" : 7,\n "NumDeadDataNodes" : 0,\n "NumDecomLiveDataNodes" : 0,\n "NumDecomDeadDataNodes" : 0,\n "VolumeFailuresTotal" : 0,\n "EstimatedCapacityLostTotal" : 0,\n "NumDecommissioningDataNodes" : 0,\n "StaleDataNodes" : 0,\n "NumStaleStorages" : 0,\n "TotalSyncCount" : 0,\n "NumInMaintenanceLiveDataNodes" : 0,\n "NumInMaintenanceDeadDataNodes" : 0,\n "NumEnteringMaintenanceDataNodes" : 0\n } ]\n}', u'')
> 2018-05-18 11:33:07,344 - call['ambari-sudo.sh su cstm-hdfs -l -s /bin/bash -c 'curl --negotiate -u : -s '"'"'http://ctr-e138-1518143905142-317995-01-000005.hwx.site:20070/jmx?qry=Hadoop:service=NameNode,name=FSNamesystem'"'"' 1>/tmp/tmpj1d3wc 2>/tmp/tmpQ4AEly''] {'quiet': False}
> 2018-05-18 11:33:07,906 - call returned (0, '######## Hortonworks #############\nThis is MOTD message, added for testing in qe infra')
> 2018-05-18 11:33:07,906 - get_user_call_output returned (0, u'{\n "beans" : [ {\n "name" : "Hadoop:service=NameNode,name=FSNamesystem",\n "modelerType" : "FSNamesystem",\n "tag.Context" : "dfs",\n "tag.HAState" : "active",\n "tag.TotalSyncTimes" : "43 17 ",\n "tag.Hostname" : "ctr-e138-1518143905142-317995-01-000005.hwx.site",\n "MissingBlocks" : 0,\n "MissingReplOneBlocks" : 0,\n "ExpiredHeartbeats" : 0,\n "TransactionsSinceLastCheckpoint" : 5082,\n "TransactionsSinceLastLogRoll" : 28,\n "LastWrittenTransactionId" : 20810,\n "LastCheckpointTime" : 1526630224790,\n "CapacityTotal" : 1656383242752,\n "CapacityTotalGB" : 1543.0,\n "CapacityUsed" : 3494182380,\n "CapacityUsedGB" : 3.0,\n "CapacityRemaining" : 1549721537687,\n "ProvidedCapacityTotal" : 0,\n "CapacityRemainingGB" : 1443.0,\n "CapacityUsedNonDFS" : 0,\n "TotalLoad" : 68,\n "SnapshottableDirectories" : 0,\n "Snapshots" : 0,\n "NumEncryptionZones" : 0,\n "LockQueueLength" : 0,\n "BlocksTotal" : 258,\n "NumFilesUnderConstruction" : 11,\n "NumActiveClients" : 9,\n "FilesTotal" : 1233,\n "PendingReplicationBlocks" : 0,\n "PendingReconstructionBlocks" : 0,\n "UnderReplicatedBlocks" : 0,\n "LowRedundancyBlocks" : 0,\n "CorruptBlocks" : 0,\n "ScheduledReplicationBlocks" : 0,\n "PendingDeletionBlocks" : 0,\n "LowRedundancyReplicatedBlocks" : 0,\n "CorruptReplicatedBlocks" : 0,\n "MissingReplicatedBlocks" : 0,\n "MissingReplicationOneBlocks" : 0,\n "BytesInFutureReplicatedBlocks" : 0,\n "PendingDeletionReplicatedBlocks" : 0,\n "TotalReplicatedBlocks" : 258,\n "LowRedundancyECBlockGroups" : 0,\n "CorruptECBlockGroups" : 0,\n "MissingECBlockGroups" : 0,\n "BytesInFutureECBlockGroups" : 0,\n "PendingDeletionECBlocks" : 0,\n "TotalECBlockGroups" : 0,\n "ExcessBlocks" : 0,\n "NumTimedOutPendingReconstructions" : 31,\n "PostponedMisreplicatedBlocks" : 0,\n "PendingDataNodeMessageCount" : 0,\n "MillisSinceLastLoadedEdits" : 0,\n "BlockCapacity" : 4194304,\n "NumLiveDataNodes" : 7,\n "NumDeadDataNodes" : 0,\n "NumDecomLiveDataNodes" : 0,\n "NumDecomDeadDataNodes" : 0,\n "VolumeFailuresTotal" : 0,\n "EstimatedCapacityLostTotal" : 0,\n "NumDecommissioningDataNodes" : 0,\n "StaleDataNodes" : 0,\n "NumStaleStorages" : 0,\n "TotalSyncCount" : 28,\n "NumInMaintenanceLiveDataNodes" : 0,\n "NumInMaintenanceDeadDataNodes" : 0,\n "NumEnteringMaintenanceDataNodes" : 0\n } ]\n}', u'')
> 2018-05-18 11:33:07,908 - NameNode HA states: active_namenodes = [(u'nn2', 'ctr-e138-1518143905142-317995-01-000005.hwx.site:20070')], standby_namenodes = [(u'nn1', 'ctr-e138-1518143905142-317995-01-000004.hwx.site:20070')], unknown_namenodes = []
> 2018-05-18 11:33:07,909 - call['ambari-sudo.sh su cstm-hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET --negotiate -u : '"'"'http://ctr-e138-1518143905142-317995-01-000005.hwx.site:20070/webhdfs/v1/user/ambari-qa?op=GETFILESTATUS'"'"' 1>/tmp/tmpyVY03k 2>/tmp/tmpL2L6US''] {'logoutput': None, 'quiet': False}
> 2018-05-18 11:33:08,024 - call returned (0, '######## Hortonworks #############\nThis is MOTD message, added for testing in qe infra')
> 2018-05-18 11:33:08,025 - get_user_call_output returned (0, u'{"FileStatus":{"accessTime":0,"blockSize":0,"childrenNum":1,"fileId":16388,"group":"hdfs","length":0,"modificationTime":1526609917254,"owner":"ambari-qa","pathSuffix":"","permission":"770","replication":0,"storagePolicy":0,"type":"DIRECTORY"}}200', u'')
> 2018-05-18 11:33:08,029 - checked_call['/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa@EXAMPLE.COM; yarn org.apache.hadoop.yarn.applications.distributedshell.Client -shell_command ls -num_containers 1 -jar /usr/hdp/current/hadoop-yarn-client/hadoop-yarn-applications-distributedshell.jar -timeout 300000 --queue default'] {'path': '/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin', 'user': 'ambari-qa'}
> Command failed after 1 tries
> {code}
> STR:
> # Add a new Host in Ambari
> # Install and start NodeManager to the new host
> # Run Yarn Service Check
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)