You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Alejandro Fernandez (JIRA)" <ji...@apache.org> on 2015/02/05 21:10:35 UTC

[jira] [Updated] (AMBARI-9504) Configuration parameter 'dfs.journalnode.kerberos.principal' was not found in configurations dictionary. Kerberized cluster

     [ https://issues.apache.org/jira/browse/AMBARI-9504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Alejandro Fernandez updated AMBARI-9504:
----------------------------------------
    Attachment: AMBARI-9504.patch

> Configuration parameter 'dfs.journalnode.kerberos.principal' was not found in configurations dictionary. Kerberized cluster
> ---------------------------------------------------------------------------------------------------------------------------
>
>                 Key: AMBARI-9504
>                 URL: https://issues.apache.org/jira/browse/AMBARI-9504
>             Project: Ambari
>          Issue Type: Bug
>          Components: ambari-server
>    Affects Versions: 2.0.0
>            Reporter: Alejandro Fernandez
>            Assignee: Alejandro Fernandez
>             Fix For: 2.0.0
>
>         Attachments: AMBARI-9504.patch
>
>
> Cluster uses Blueprint
> HDFS_CLIENT install and DATANODE restart failed on secured cluster with error:
> Fail: Configuration parameter 'dfs.journalnode.kerberos.principal' was not found in configurations dictionary!
> {code}
> 2015-02-04 07:16:24,098 - Error while executing command 'restart':
> Traceback (most recent call last):
>   File \"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py\", line 184, in execute
>     method(env)
>   File \"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py\", line 338, in restart
>     self.stop(env)
>   File \"/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/datanode.py\", line 67, in stop
>     import params
>   File \"/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/params.py\", line 245, in <module>
>     _jn_principal_name = _jn_principal_name.replace('_HOST', hostname.lower())
>   File \"/usr/lib/python2.6/site-packages/resource_management/libraries/script/config_dictionary.py\", line 79, in __getattr__
>     raise Fail(\"Configuration parameter '\" + self.name + \"' was not found in configurations dictionary!\")
> Fail: Configuration parameter 'dfs.journalnode.kerberos.principal' was not found in configurations dictionary!",
>     "stdout" : "2015-02-04 07:16:22,978 - u\"Group['hadoop']\" {'ignore_failures': False}
> 2015-02-04 07:16:22,980 - Modifying group hadoop
> 2015-02-04 07:16:23,067 - u\"Group['users']\" {'ignore_failures': False}
> 2015-02-04 07:16:23,067 - Modifying group users
> 2015-02-04 07:16:23,132 - u\"Group['hdfs']\" {'ignore_failures': False}
> 2015-02-04 07:16:23,133 - Modifying group hdfs
> 2015-02-04 07:16:23,221 - u\"Group['knox']\" {'ignore_failures': False}
> 2015-02-04 07:16:23,221 - Modifying group knox
> 2015-02-04 07:16:23,284 - u\"User['hive']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,285 - Modifying user hive
> 2015-02-04 07:16:23,304 - u\"User['oozie']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
> 2015-02-04 07:16:23,305 - Modifying user oozie
> 2015-02-04 07:16:23,321 - u\"User['ambari-qa']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
> 2015-02-04 07:16:23,321 - Modifying user ambari-qa
> 2015-02-04 07:16:23,340 - u\"User['flume']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,341 - Modifying user flume
> 2015-02-04 07:16:23,363 - u\"User['hdfs']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hdfs']}
> 2015-02-04 07:16:23,364 - Modifying user hdfs
> 2015-02-04 07:16:23,383 - u\"User['knox']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,384 - Modifying user knox
> 2015-02-04 07:16:23,396 - u\"User['storm']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,396 - Modifying user storm
> 2015-02-04 07:16:23,413 - u\"User['mapred']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,413 - Modifying user mapred
> 2015-02-04 07:16:23,432 - u\"User['hbase']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,433 - Modifying user hbase
> 2015-02-04 07:16:23,452 - u\"User['tez']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'users']}
> 2015-02-04 07:16:23,454 - Modifying user tez
> 2015-02-04 07:16:23,473 - u\"User['zookeeper']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,473 - Modifying user zookeeper
> 2015-02-04 07:16:23,491 - u\"User['falcon']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,492 - Modifying user falcon
> 2015-02-04 07:16:23,511 - u\"User['sqoop']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,511 - Modifying user sqoop
> 2015-02-04 07:16:23,532 - u\"User['yarn']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,532 - Modifying user yarn
> 2015-02-04 07:16:23,552 - u\"User['hcat']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,553 - Modifying user hcat
> 2015-02-04 07:16:23,572 - u\"User['ams']\" {'gid': 'hadoop', 'ignore_failures': False, 'groups': [u'hadoop']}
> 2015-02-04 07:16:23,573 - Modifying user ams
> 2015-02-04 07:16:23,586 - u\"File['/var/lib/ambari-agent/data/tmp/changeUid.sh']\" {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
> 2015-02-04 07:16:23,603 - u\"Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa']\" {'not_if': 'test $(id -u ambari-qa) -gt 1000'}
> 2015-02-04 07:16:23,609 - Skipping u\"Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa']\" due to not_if
> 2015-02-04 07:16:23,610 - u\"Directory['/hadoop/hbase']\" {'owner': 'hbase', 'recursive': True, 'mode': 0775, 'cd_access': 'a'}
> 2015-02-04 07:16:23,645 - u\"File['/var/lib/ambari-agent/data/tmp/changeUid.sh']\" {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
> 2015-02-04 07:16:23,663 - u\"Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/hadoop/hbase']\" {'not_if': 'test $(id -u hbase) -gt 1000'}
> 2015-02-04 07:16:23,669 - Skipping u\"Execute['/var/lib/ambari-agent/data/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/hadoop/hbase']\" due to not_if
> 2015-02-04 07:16:23,669 - u\"Directory['/etc/hadoop']\" {'mode': 0755}
> 2015-02-04 07:16:23,670 - u\"Directory['/etc/hadoop/conf.empty']\" {'owner': 'root', 'group': 'hadoop', 'recursive': True}
> 2015-02-04 07:16:23,671 - u\"Link['/etc/hadoop/conf']\" {'not_if': 'ls /etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'}
> 2015-02-04 07:16:23,676 - Skipping u\"Link['/etc/hadoop/conf']\" due to not_if
> 2015-02-04 07:16:23,692 - u\"File['/etc/hadoop/conf/hadoop-env.sh']\" {'content': InlineTemplate(...), 'owner': 'root', 'group': 'hadoop'}
> 2015-02-04 07:16:23,723 - u\"Execute['('setenforce', '0')']\" {'sudo': True, 'only_if': 'test -f /selinux/enforce'}
> 2015-02-04 07:16:23,747 - Skipping u\"Execute['('setenforce', '0')']\" due to only_if
> 2015-02-04 07:16:23,751 - u\"Directory['/usr/hdp/current/hadoop-client/lib/native/Linux-i386-32']\" {'recursive': True}
> 2015-02-04 07:16:23,752 - u\"Directory['/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64']\" {'recursive': True}
> 2015-02-04 07:16:23,753 - u\"Link['/usr/hdp/current/hadoop-client/lib/native/Linux-i386-32/libsnappy.so']\" {'to': '/usr/hdp/current/hadoop-client/lib/libsnappy.so'}
> 2015-02-04 07:16:23,753 - u\"Link['/usr/hdp/current/hadoop-client/lib/native/Linux-i386-32/libsnappy.so']\" replacing old symlink to /grid/0/hdp/2.2.1.0-2340/hadoop/lib/libsnappy.so
> 2015-02-04 07:16:23,765 - Warning: linking to nonexistent location /usr/hdp/current/hadoop-client/lib/libsnappy.so
> 2015-02-04 07:16:23,765 - Creating symbolic u\"Link['/usr/hdp/current/hadoop-client/lib/native/Linux-i386-32/libsnappy.so']\"
> 2015-02-04 07:16:23,778 - u\"Link['/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64/libsnappy.so']\" {'to': '/usr/hdp/current/hadoop-client/lib64/libsnappy.so'}
> 2015-02-04 07:16:23,780 - u\"Link['/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64/libsnappy.so']\" replacing old symlink to /grid/0/hdp/2.2.1.0-2340/hadoop/lib64/libsnappy.so
> 2015-02-04 07:16:23,792 - Warning: linking to nonexistent location /usr/hdp/current/hadoop-client/lib64/libsnappy.so
> 2015-02-04 07:16:23,793 - Creating symbolic u\"Link['/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64/libsnappy.so']\"
> 2015-02-04 07:16:23,807 - u\"Directory['/var/log/hadoop']\" {'owner': 'root', 'group': 'hadoop', 'mode': 0775, 'recursive': True}
> 2015-02-04 07:16:23,808 - u\"Directory['/var/run/hadoop']\" {'owner': 'root', 'group': 'root', 'recursive': True}
> 2015-02-04 07:16:23,809 - u\"Directory['/tmp/hadoop-hdfs']\" {'owner': 'hdfs', 'recursive': True}
> 2015-02-04 07:16:23,815 - u\"File['/etc/hadoop/conf/commons-logging.properties']\" {'content': Template('commons-logging.properties.j2'), 'owner': 'root'}
> 2015-02-04 07:16:23,833 - u\"File['/etc/hadoop/conf/health_check']\" {'content': Template('health_check-v2.j2'), 'owner': 'root'}
> 2015-02-04 07:16:23,848 - u\"File['/etc/hadoop/conf/log4j.properties']\" {'content': '...', 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
> 2015-02-04 07:16:23,879 - u\"File['/etc/hadoop/conf/hadoop-metrics2.properties']\" {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'}
> 2015-02-04 07:16:23,895 - u\"File['/etc/hadoop/conf/task-log4j.properties']\" {'content': StaticFile('task-log4j.properties'), 'mode': 0755}
> 2015-02-04 07:16:24,098 - Error while executing command 'restart':
> Traceback (most recent call last):
>   File \"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py\", line 184, in execute
>     method(env)
>   File \"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py\", line 338, in restart
>     self.stop(env)
>   File \"/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/datanode.py\", line 67, in stop
>     import params
>   File \"/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/params.py\", line 245, in <module>
>     _jn_principal_name = _jn_principal_name.replace('_HOST', hostname.lower())
>   File \"/usr/lib/python2.6/site-packages/resource_management/libraries/script/config_dictionary.py\", line 79, in __getattr__
>     raise Fail(\"Configuration parameter '\" + self.name + \"' was not found in configurations dictionary!\")
> Fail: Configuration parameter 'dfs.journalnode.kerberos.principal' was not found in configurations dictionary!
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)