You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Dmytro Sen (JIRA)" <ji...@apache.org> on 2014/09/25 13:52:33 UTC

[jira] [Created] (AMBARI-7484) Versioned RPMS install fails on preinstall Hook.

Dmytro Sen created AMBARI-7484:
----------------------------------

             Summary: Versioned RPMS install fails on preinstall Hook.
                 Key: AMBARI-7484
                 URL: https://issues.apache.org/jira/browse/AMBARI-7484
             Project: Ambari
          Issue Type: Bug
          Components: stacks
    Affects Versions: 1.7.0
         Environment: HDP2.2
            Reporter: Dmytro Sen
            Assignee: Dmytro Sen
            Priority: Blocker
             Fix For: 1.7.0


Versioned RPMS install fails on preinstall Hook.

{code}
': ['/usr/libexec/hdp/ganglia', '/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
2014-09-24 23:35:34,085 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPFlumeServer -m -o root -g hadoop'] {'path': ['/usr/libexec/hdp/ganglia', '/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
2014-09-24 23:35:34,119 - Execute['/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSlaves -m -o root -g hadoop'] {'path': ['/usr/libexec/hdp/ganglia', '/usr/sbin', '/sbin:/usr/local/bin', '/bin', '/usr/bin']}
2014-09-24 23:35:34,154 - Execute['chkconfig gmond off'] {'path': ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
2014-09-24 23:35:34,165 - Execute['chkconfig gmetad off'] {'path': ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']}
2014-09-24 23:35:34,284 - Directory['/etc/hadoop/conf.empty'] {'owner': 'root', 'group': 'root', 'recursive': True}
2014-09-24 23:35:34,285 - Creating directory Directory['/etc/hadoop/conf.empty']
2014-09-24 23:35:34,287 - Link['/etc/hadoop/conf'] {'not_if': 'ls /etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'}
2014-09-24 23:35:34,297 - Creating symbolic Link['/etc/hadoop/conf']
2014-09-24 23:35:34,309 - File['/etc/hadoop/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs'}
2014-09-24 23:35:34,309 - Writing File['/etc/hadoop/conf/hadoop-env.sh'] because it doesn't exist
2014-09-24 23:35:34,310 - Changing owner for /etc/hadoop/conf/hadoop-env.sh from 0 to hdfs
2014-09-24 23:35:34,310 - Execute['ln -s /usr/hdp/2.9.9.9* /usr/hdp/current'] {'not_if': 'ls /usr/hdp/current'}
2014-09-24 23:35:34,329 - Error while executing command 'install':
Traceback (most recent call last):
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 122, in execute
    method(env)
  File "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py", line 31, in hook
    setup_hadoop_env()
  File "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py", line 44, in setup_hadoop_env
    not_if=format('ls {versioned_hdp_root}')
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, in __init__
    self.env.run()
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 149, in run
    self.run_action(resource, action)
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 115, in run_action
    provider_action()
  File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 237, in action_run
    raise ex
Fail: Execution of 'ln -s /usr/hdp/2.9.9.9* /usr/hdp/current' returned 1. ln: creating symbolic link `/usr/hdp/current': No such file or directory
{code}




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)