You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by sm...@apache.org on 2015/06/06 03:44:02 UTC

ambari git commit: AMBARI-11747. Modify resource pre-upload script to handle recent changes to stack and HdfsResource (smohanty)

Repository: ambari
Updated Branches:
  refs/heads/trunk 89ba4a4bb -> e973a534f


AMBARI-11747. Modify resource pre-upload script to handle recent changes to stack and HdfsResource (smohanty)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/e973a534
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/e973a534
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/e973a534

Branch: refs/heads/trunk
Commit: e973a534f0fcb331dc8d71bdb59843189fdfc2b5
Parents: 89ba4a4
Author: Sumit Mohanty <sm...@hortonworks.com>
Authored: Fri Jun 5 18:43:45 2015 -0700
Committer: Sumit Mohanty <sm...@hortonworks.com>
Committed: Fri Jun 5 18:43:45 2015 -0700

----------------------------------------------------------------------
 .../main/resources/scripts/Ambaripreupload.py   | 37 +++++++++++++++-----
 1 file changed, 28 insertions(+), 9 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/e973a534/ambari-server/src/main/resources/scripts/Ambaripreupload.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/scripts/Ambaripreupload.py b/ambari-server/src/main/resources/scripts/Ambaripreupload.py
index 548983c..fd72f24 100644
--- a/ambari-server/src/main/resources/scripts/Ambaripreupload.py
+++ b/ambari-server/src/main/resources/scripts/Ambaripreupload.py
@@ -63,20 +63,21 @@ def getPropertyValueFromConfigXMLFile(xmlfile, name, defaultValue=None):
         if len(node.childNodes) > 0:
           return node.childNodes[0].nodeValue
         else:
-          return ''
+          return defaultValue
   return defaultValue
 
 def get_fs_root(fsdefaultName=None):
-  return getPropertyValueFromConfigXMLFile("/etc/hadoop/conf/core-site.xml", "fs.defaultFS")
-  if fsdefaultName is None:
-    fsdefaultName = "fake"
+  fsdefaultName = "fake"
    
   while (not fsdefaultName.startswith("wasb://")):
     fsdefaultName =  getPropertyValueFromConfigXMLFile("/etc/hadoop/conf/core-site.xml", "fs.defaultFS")
     if fsdefaultName is None:
       fsdefaultName = "fake"
+    print "Waiting to read appropriate value of fs.defaultFS from /etc/hadoop/conf/core-site.xml ..."
     time.sleep(10)
-  
+    pass
+
+  print "Returning fs.defaultFS -> " + fsdefaultName
   return fsdefaultName
  
 # These values must be the suffix of the properties in cluster-env.xml
@@ -94,9 +95,20 @@ class params:
   execute_path = "/usr/hdp/current/hadoop-client/bin"
   ambari_libs_dir = "/var/lib/ambari-agent/lib"
   hdfs_site = ConfigDictionary({'dfs.webhdfs.enabled':False, 
-                                'dfs.namenode.http-address': getPropertyValueFromConfigXMLFile("/etc/hadoop/conf/hdfs-site.xml", "dfs.namenode.http-address")
   })
   fs_default = get_fs_root()
+  oozie_env_sh_template = \
+'''
+#!/bin/bash
+
+export OOZIE_CONFIG=${OOZIE_CONFIG:-/usr/hdp/current/oozie/conf}
+export OOZIE_DATA=${OOZIE_DATA:-/var/lib/oozie/data}
+export OOZIE_LOG=${OOZIE_LOG:-/var/log/oozie}
+export CATALINA_BASE=${CATALINA_BASE:-/usr/hdp/current/oozie-server/oozie-server}
+export CATALINA_TMPDIR=${CATALINA_TMPDIR:-/var/tmp/oozie}
+export CATALINA_PID=${CATALINA_PID:-/var/run/oozie/oozie.pid}
+export OOZIE_CATALINA_HOME=/usr/lib/bigtop-tomcat
+'''
   
   HdfsResource = functools.partial(
     HdfsResource,
@@ -213,7 +225,12 @@ no_op_test = "ls /var/run/oozie/oozie.pid >/dev/null 2>&1 && ps -p `cat /var/run
 
 with Environment() as env:
   env.set_params(params)
-  
+
+  File("/etc/oozie/conf/oozie-env.sh",
+       owner=params.oozie_user,
+       content=params.oozie_env_sh_template
+  )
+
   hashcode_file = format("{oozie_home}/.hashcode")
   hashcode = hashlib.md5(format('{oozie_home}/oozie-sharelib.tar.gz')).hexdigest()
   skip_recreate_sharelib = format("test -f {hashcode_file} && test -d {oozie_home}/share && [[ `cat {hashcode_file}` == '{hashcode}' ]]")
@@ -250,13 +267,14 @@ with Environment() as env:
     source = oozie_shared_lib,
   )
 
+  print "Copying tarballs..."
   copy_tarballs_to_hdfs("/usr/hdp/current/hadoop-client/mapreduce.tar.gz", hdfs_path_prefix+"/hdp/apps/{{ hdp_stack_version }}/mapreduce/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
   copy_tarballs_to_hdfs("/usr/hdp/current/tez-client/lib/tez.tar.gz", hdfs_path_prefix+"/hdp/apps/{{ hdp_stack_version }}/tez/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
   copy_tarballs_to_hdfs("/usr/hdp/current/hive-client/hive.tar.gz", hdfs_path_prefix+"/hdp/apps/{{ hdp_stack_version }}/hive/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
   copy_tarballs_to_hdfs("/usr/hdp/current/pig-client/pig.tar.gz", hdfs_path_prefix+"/hdp/apps/{{ hdp_stack_version }}/pig/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
   copy_tarballs_to_hdfs("/usr/hdp/current/hadoop-mapreduce-client/hadoop-streaming.jar", hdfs_path_prefix+"/hdp/apps/{{ hdp_stack_version }}/mapreduce/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
   copy_tarballs_to_hdfs("/usr/hdp/current/sqoop-client/sqoop.tar.gz", hdfs_path_prefix+"/hdp/apps/{{ hdp_stack_version }}/sqoop/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
-  
+
   
   # jar shouldn't be used before (read comment below)
   File(format("{ambari_libs_dir}/fast-hdfs-resource.jar"),
@@ -268,4 +286,5 @@ with Environment() as env:
   params.HdfsResource(None, 
                logoutput=True,
                action="execute"
-  )
\ No newline at end of file
+  )
+  print "Completed tarball copy. Ambari preupload script completed."