You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by jl...@apache.org on 2016/03/09 19:21:17 UTC
ambari git commit: AMBARI-15329: Code Cleanup: Remove hdp hardcodings
in functions, variables etc. (jluniya)
Repository: ambari
Updated Branches:
refs/heads/trunk f5bd0585b -> 456b45118
AMBARI-15329: Code Cleanup: Remove hdp hardcodings in functions, variables etc. (jluniya)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/456b4511
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/456b4511
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/456b4511
Branch: refs/heads/trunk
Commit: 456b451180b427e89f52a5dc0a644e27fe346efb
Parents: f5bd058
Author: Jayush Luniya <jl...@hortonworks.com>
Authored: Wed Mar 9 10:09:45 2016 -0800
Committer: Jayush Luniya <jl...@hortonworks.com>
Committed: Wed Mar 9 10:09:45 2016 -0800
----------------------------------------------------------------------
.../src/main/python/amc_service.py | 2 +-
.../src/main/python/main.py | 4 +-
.../0.5.0.2.1/package/scripts/params_windows.py | 2 +-
.../0.5.0.2.1/package/scripts/service_check.py | 2 +-
.../1.4.0.2.0/package/scripts/flume_check.py | 2 +-
.../1.4.0.2.0/package/scripts/params_windows.py | 2 +-
.../package/scripts/params_windows.py | 2 +-
.../0.96.0.2.0/package/scripts/service_check.py | 2 +-
.../package/scripts/setup_ranger_hbase.py | 6 +-
.../HBASE/0.96.0.2.0/package/scripts/upgrade.py | 4 +-
.../package/scripts/setup_ranger_hdfs.py | 10 +--
.../2.1.0.2.0/package/scripts/status_params.py | 12 ++--
.../package/scripts/hcat_service_check.py | 2 +-
.../package/scripts/hive_server_upgrade.py | 6 +-
.../package/scripts/params_windows.py | 4 +-
.../0.12.0.2.0/package/scripts/service_check.py | 2 +-
.../package/scripts/setup_ranger_hive.py | 6 +-
.../package/scripts/webhcat_service_check.py | 2 +-
.../KAFKA/0.8.1.2.2/package/scripts/params.py | 4 +-
.../0.5.0.2.2/package/scripts/params_windows.py | 4 +-
.../package/scripts/setup_ranger_knox.py | 6 +-
.../4.0.0.2.0/package/scripts/params_windows.py | 2 +-
.../4.0.0.2.0/package/scripts/service_check.py | 2 +-
.../package/scripts/params_windows.py | 4 +-
.../0.12.0.2.0/package/scripts/service_check.py | 2 +-
.../RANGER/0.4.0/package/scripts/upgrade.py | 6 +-
.../0.5.0.2.3/package/scripts/kms_server.py | 4 +-
.../0.5.0.2.3/package/scripts/upgrade.py | 6 +-
.../package/scripts/params_windows.py | 4 +-
.../0.60.0.2.2/package/scripts/service_check.py | 2 +-
.../1.4.4.2.0/package/scripts/params_windows.py | 2 +-
.../1.4.4.2.0/package/scripts/service_check.py | 2 +-
.../0.9.1.2.1/package/scripts/params_windows.py | 2 +-
.../0.9.1.2.1/package/scripts/service_check.py | 2 +-
.../package/scripts/setup_ranger_storm.py | 6 +-
.../0.4.0.2.1/package/scripts/params_windows.py | 4 +-
.../0.4.0.2.1/package/scripts/service_check.py | 2 +-
.../3.4.5.2.0/package/scripts/params_windows.py | 4 +-
.../3.4.5.2.0/package/scripts/service_check.py | 2 +-
.../resources/host_scripts/alert_disk_space.py | 20 +++---
.../0.8/hooks/after-INSTALL/scripts/hook.py | 2 +-
.../0.8/hooks/after-INSTALL/scripts/params.py | 2 +-
.../scripts/shared_initialization.py | 2 +-
.../0.8/hooks/before-ANY/scripts/params.py | 2 +-
.../HDFS/package/scripts/status_params.py | 12 ++--
.../2.0.6/hooks/after-INSTALL/scripts/hook.py | 4 +-
.../2.0.6/hooks/after-INSTALL/scripts/params.py | 2 +-
.../scripts/shared_initialization.py | 2 +-
.../2.0.6/hooks/before-ANY/scripts/params.py | 2 +-
.../2.1/hooks/after-INSTALL/scripts/params.py | 4 +-
.../src/test/python/TestVersionSelectUtil.py | 8 +--
.../custom_actions/TestInstallPackages.py | 72 ++++++++++----------
52 files changed, 139 insertions(+), 139 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py
----------------------------------------------------------------------
diff --git a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py
index f616fd0..a19fd11 100644
--- a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py
+++ b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py
@@ -165,7 +165,7 @@ def init_service_debug(options):
sys.frozen = 'windows_exe' # Fake py2exe so we can debug
-def ensure_hdp_service_soft_dependencies():
+def ensure_hadoop_service_soft_dependencies():
if SERVICE_STATUS_RUNNING != WinServiceController.QueryStatus(EMBEDDED_HBASE_MASTER_SERVICE):
err = 'ERROR: Service "{0}" was not started.'.format(EMBEDDED_HBASE_MASTER_SERVICE)
raise FatalException(1, err)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py
----------------------------------------------------------------------
diff --git a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py
index 172861e..b6b4e0b 100644
--- a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py
+++ b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py
@@ -109,8 +109,8 @@ def server_process_main(options, scmStatus=None):
#Ensure the 3 Hadoop services required are started on the local machine
if not options.no_embedded_hbase:
- from amc_service import ensure_hdp_service_soft_dependencies
- ensure_hdp_service_soft_dependencies()
+ from amc_service import ensure_hadoop_service_soft_dependencies
+ ensure_hadoop_service_soft_dependencies()
if scmStatus is not None:
scmStatus.reportStartPending()
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py
index 9136957..b213028 100644
--- a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py
@@ -35,7 +35,7 @@ falcon_home = None
falcon_log_dir = "."
if os.environ.has_key("HADOOP_HOME"):
- hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
+ stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
if os.environ.has_key("FALCON_CONF_DIR"):
falcon_conf_dir = os.environ["FALCON_CONF_DIR"]
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py
index e633dcb..473b7c3 100644
--- a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py
@@ -45,7 +45,7 @@ class FalconServiceCheckWindows(FalconServiceCheck):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "FALCON"
Execute(format("cmd /C {smoke_cmd} {service}"), user=params.falcon_user, logoutput=True, tries = 3, try_sleep = 20)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py
index ae166f8..25d9a08 100644
--- a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py
+++ b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py
@@ -27,7 +27,7 @@ class FlumeServiceCheck(Script):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "FLUME"
Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True, user=params.hdfs_user)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py
index 17177ec..66e9852 100644
--- a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py
@@ -26,7 +26,7 @@ config = Script.get_config()
hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"]
-hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
+stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
flume_home = os.environ['FLUME_HOME']
flume_conf_dir = os.path.join(flume_home, 'conf')
flume_user = hadoop_user
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py
index c132503..7d634cb 100644
--- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py
@@ -27,7 +27,7 @@ config = Script.get_config()
hbase_conf_dir = os.environ["HBASE_CONF_DIR"]
hbase_bin_dir = os.path.join(os.environ["HBASE_HOME"],'bin')
hbase_executable = os.path.join(hbase_bin_dir,"hbase.cmd")
-hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
+stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"]
hbase_user = hadoop_user
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py
index 97cdd32..6ecb58e 100644
--- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py
@@ -34,7 +34,7 @@ class HbaseServiceCheckWindows(HbaseServiceCheck):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd")
service = "HBASE"
Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hbase_user, logoutput=True)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py
index 5c68583..ffd0715 100644
--- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py
+++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py
@@ -29,10 +29,10 @@ def setup_ranger_hbase(upgrade_type=None):
else:
from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin
- hdp_version = None
+ stack_version = None
if upgrade_type is not None:
- hdp_version = params.version
+ stack_version = params.version
if params.retryAble:
Logger.info("HBase: Setup ranger: command retry enables thus retrying if ranger admin is down !")
@@ -80,6 +80,6 @@ def setup_ranger_hbase(upgrade_type=None):
component_list=['hbase-client', 'hbase-master', 'hbase-regionserver'], audit_db_is_enabled=params.xa_audit_db_is_enabled,
credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password,
ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password,
- hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble)
+ stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble)
else:
Logger.info('Ranger admin not installed')
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
index c5ba682..92c0f70 100644
--- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
+++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py
@@ -26,12 +26,12 @@ from resource_management.libraries.functions import stack_select
from resource_management.libraries.functions.version import compare_versions, format_stack_version
from resource_management.libraries.functions.decorator import retry
-def prestart(env, hdp_component):
+def prestart(env, stack_component):
import params
if params.version and compare_versions(format_stack_version(params.version), '2.2.0.0') >= 0:
conf_select.select(params.stack_name, "hbase", params.version)
- stack_select.select(hdp_component, params.version)
+ stack_select.select(stack_component, params.version)
def post_regionserver(env):
import params
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py
index ff93e39..209ac91 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py
@@ -34,10 +34,10 @@ def setup_ranger_hdfs(upgrade_type=None):
else:
from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin
- hdp_version = None
+ stack_version = None
if upgrade_type is not None:
- hdp_version = params.version
+ stack_version = params.version
if params.retryAble:
Logger.info("HDFS: Setup ranger: command retry enables thus retrying if ranger admin is down !")
@@ -58,11 +58,11 @@ def setup_ranger_hdfs(upgrade_type=None):
component_list=['hadoop-client'], audit_db_is_enabled=params.xa_audit_db_is_enabled,
credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password,
ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password,
- hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble)
+ stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble)
- if hdp_version and params.upgrade_direction == Direction.UPGRADE:
+ if stack_version and params.upgrade_direction == Direction.UPGRADE:
# when upgrading to 2.3+, this env file must be removed
- if compare_versions(hdp_version, '2.3', format=True) > 0:
+ if compare_versions(stack_version, '2.3', format=True) > 0:
source_file = os.path.join(params.hadoop_conf_dir, 'set-hdfs-plugin-env.sh')
target_file = source_file + ".bak"
Execute(("mv", source_file, target_file), sudo=True, only_if=format("test -f {source_file}"))
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py
index 388fa59..cdb683b 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py
@@ -36,12 +36,12 @@ if OSCheck.is_windows_family():
else:
hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
hdfs_user = config['configurations']['hadoop-env']['hdfs_user']
- hdp_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}")
- datanode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-datanode.pid")
- namenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-namenode.pid")
- snamenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid")
- journalnode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-journalnode.pid")
- zkfc_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-zkfc.pid")
+ hadoop_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}")
+ datanode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-datanode.pid")
+ namenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-namenode.pid")
+ snamenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid")
+ journalnode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-journalnode.pid")
+ zkfc_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-zkfc.pid")
nfsgateway_pid_file = format("{hadoop_pid_dir_prefix}/root/hadoop_privileged_nfs3.pid")
# Security related/required params
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py
index 27ff29a..4153821 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py
@@ -27,7 +27,7 @@ from ambari_commons import OSConst
@OsFamilyFuncImpl(os_family=OSConst.WINSRV_FAMILY)
def hcat_service_check():
import params
- smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd")
service = "HCatalog"
Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hcat_user, logoutput=True)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py
index 664cafa..c3d15e5 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py
@@ -121,7 +121,7 @@ def _get_current_hiveserver_version():
if formatted_source_version and compare_versions(formatted_source_version, "2.2") >= 0:
version_hive_bin = format('/usr/hdp/{source_version}/hive/bin')
command = format('{version_hive_bin}/hive --version')
- return_code, hdp_output = shell.call(command, user=params.hive_user, path=hive_execute_path)
+ return_code, output = shell.call(command, user=params.hive_user, path=hive_execute_path)
except Exception, e:
Logger.error(str(e))
raise Fail('Unable to execute hive --version command to retrieve the hiveserver2 version.')
@@ -129,12 +129,12 @@ def _get_current_hiveserver_version():
if return_code != 0:
raise Fail('Unable to determine the current HiveServer2 version because of a non-zero return code of {0}'.format(str(return_code)))
- match = re.search('^(Hive) ([0-9]+.[0-9]+.\S+)', hdp_output, re.MULTILINE)
+ match = re.search('^(Hive) ([0-9]+.[0-9]+.\S+)', output, re.MULTILINE)
if match:
current_hive_server_version = match.group(2)
return current_hive_server_version
else:
- raise Fail('The extracted hiveserver2 version "{0}" does not matching any known pattern'.format(hdp_output))
+ raise Fail('The extracted hiveserver2 version "{0}" does not matching any known pattern'.format(output))
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py
index 7c21b5f..5610019 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py
@@ -28,7 +28,7 @@ config = Script.get_config()
stack_version_unformatted = str(config['hostLevelParams']['stack_version'])
stack_version_formatted = format_stack_version(stack_version_unformatted)
-hdp_root = None
+stack_root = None
hive_conf_dir = None
hive_home = None
hive_lib_dir = None
@@ -39,7 +39,7 @@ hcat_config_dir = None
hive_bin = None
try:
- hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
+ stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
hive_conf_dir = os.environ["HIVE_CONF_DIR"]
hive_home = os.environ["HIVE_HOME"]
hive_lib_dir = os.environ["HIVE_LIB_DIR"]
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
index 0c254be..251e71f 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
@@ -37,7 +37,7 @@ class HiveServiceCheckWindows(HiveServiceCheck):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "HIVE"
Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hive_user, logoutput=True)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py
index 8b2e4e4..f51dbab 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py
@@ -29,10 +29,10 @@ def setup_ranger_hive(upgrade_type = None):
else:
from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin
- hdp_version = None
+ stack_version = None
if upgrade_type is not None:
- hdp_version = params.version
+ stack_version = params.version
if params.retryAble:
Logger.info("Hive: Setup ranger: command retry enables thus retrying if ranger admin is down !")
@@ -72,6 +72,6 @@ def setup_ranger_hive(upgrade_type = None):
component_list=['hive-client', 'hive-metastore', 'hive-server2'], audit_db_is_enabled=params.xa_audit_db_is_enabled,
credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password,
ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password,
- hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble)
+ stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble)
else:
Logger.info('Ranger admin not installed')
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py
index 99f52a5..9965bfa9 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py
@@ -32,7 +32,7 @@ def webhcat_service_check():
# AMBARI-11633 [WinTP2] Webhcat service check fails
# Hive doesn't pass the environment variables correctly to child processes, which fails the smoke test.
# Reducing the amount of URLs checked to the minimum required.
- #smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ #smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
#service = "WEBHCAT"
#Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hcat_user, logoutput=True)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py
index ba6857e..4e73730 100644
--- a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py
+++ b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py
@@ -248,8 +248,8 @@ if has_ranger_admin and is_supported_kafka_ranger:
ssl_truststore_password = unicode(config['configurations']['ranger-kafka-policymgr-ssl']['xasecure.policymgr.clientssl.truststore.password']) if xml_configurations_supported else None
credential_file = format('/etc/ranger/{repo_name}/cred.jceks') if xml_configurations_supported else None
- hdp_version = get_stack_version('kafka-broker')
- setup_ranger_env_sh_source = format('/usr/hdp/{hdp_version}/ranger-kafka-plugin/install/conf.templates/enable/kafka-ranger-env.sh')
+ stack_version = get_stack_version('kafka-broker')
+ setup_ranger_env_sh_source = format('/usr/hdp/{stack_version}/ranger-kafka-plugin/install/conf.templates/enable/kafka-ranger-env.sh')
setup_ranger_env_sh_target = format("{conf_dir}/kafka-ranger-env.sh")
#For SQLA explicitly disable audit to DB for Ranger
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py
index e044d9a..e7a3747 100644
--- a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py
@@ -25,7 +25,7 @@ from status_params import *
# server configurations
config = Script.get_config()
-hdp_root = None
+stack_root = None
knox_home = None
knox_conf_dir = None
knox_logs_dir = None
@@ -38,7 +38,7 @@ knox_master_secret_path = None
knox_cert_store_path = None
try:
- hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
+ stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
knox_home = os.environ['KNOX_HOME']
knox_conf_dir = os.environ['KNOX_CONF_DIR']
knox_logs_dir = os.environ['KNOX_LOG_DIR']
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py
index 1a08d54..13987c8 100644
--- a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py
+++ b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py
@@ -29,9 +29,9 @@ def setup_ranger_knox(upgrade_type=None):
else:
from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin
- hdp_version = None
+ stack_version = None
if upgrade_type is not None:
- hdp_version = params.version
+ stack_version = params.version
if params.retryAble:
Logger.info("Knox: Setup ranger: command retry enables thus retrying if ranger admin is down !")
@@ -72,6 +72,6 @@ def setup_ranger_knox(upgrade_type=None):
component_list=['knox-server'], audit_db_is_enabled=params.xa_audit_db_is_enabled,
credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password,
ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password,
- hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble)
+ stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble)
else:
Logger.info('Ranger admin not installed')
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py
index ec404c4..1f939d4 100644
--- a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py
@@ -24,7 +24,7 @@ from status_params import *
config = Script.get_config()
hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"]
-hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
+stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
oozie_root = os.environ['OOZIE_ROOT']
oozie_home = os.environ['OOZIE_HOME']
oozie_conf_dir = os.path.join(oozie_home,'conf')
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py
index f0cdb2c..8d14836 100644
--- a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py
@@ -130,7 +130,7 @@ class OozieServiceCheckWindows(OozieServiceCheck):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd")
service = "OOZIE"
Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py
index 8a0a519..9cb0ce7 100644
--- a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py
@@ -23,11 +23,11 @@ from resource_management import *
# server configurations
config = Script.get_config()
-hdp_root = None
+stack_root = None
pig_home = None
pig_conf_dir = None
try:
- hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
+ stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
pig_home = os.environ['PIG_HOME']
pig_conf_dir = os.path.join(pig_home,'conf')
except:
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py
index 155e63c..ec2946e 100644
--- a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py
@@ -125,7 +125,7 @@ class PigServiceCheckWindows(PigServiceCheck):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "PIG"
Execute(format("cmd /C {smoke_cmd} {service}", smoke_cmd=smoke_cmd, service=service), logoutput=True, user=params.pig_user, timeout=300)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
index ed8b690..aa75949 100644
--- a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
+++ b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py
@@ -23,9 +23,9 @@ from resource_management.libraries.functions import conf_select
from resource_management.libraries.functions import stack_select
from resource_management.libraries.functions.format import format
-def prestart(env, hdp_component):
+def prestart(env, stack_component):
import params
if params.version and params.stack_is_hdp22_or_further:
- conf_select.select(params.stack_name, hdp_component, params.version)
- stack_select.select(hdp_component, params.version)
+ conf_select.select(params.stack_name, stack_component, params.version)
+ stack_select.select(stack_component, params.version)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
index 9450184..42f1cb9 100755
--- a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
+++ b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py
@@ -17,7 +17,7 @@ See the License for the specific language governing permissions and
limitations under the License.
"""
-from resource_management.libraries.functions import hdp_select
+from resource_management.libraries.functions import stack_select
from resource_management.libraries.script import Script
from resource_management.core.resources.system import Execute
from resource_management.core.exceptions import ComponentIsNotRunning
@@ -82,7 +82,7 @@ class KmsServer(Script):
import params
env.set_params(params)
- upgrade_stack = hdp_select._get_upgrade_stack()
+ upgrade_stack = stack_select._get_upgrade_stack()
if upgrade_stack is None:
raise Fail('Unable to determine the stack and stack version')
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
index 798e8f7..315a417 100644
--- a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
+++ b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py
@@ -22,9 +22,9 @@ from resource_management.libraries.functions import conf_select
from resource_management.libraries.functions import stack_select
from resource_management.libraries.functions.format import format
-def prestart(env, hdp_component):
+def prestart(env, stack_component):
import params
if params.version and params.stack_is_hdp23_or_further:
- conf_select.select(params.stack_name, hdp_component, params.version)
- stack_select.select(hdp_component, params.version)
+ conf_select.select(params.stack_name, stack_component, params.version)
+ stack_select.select(stack_component, params.version)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py
index 366a1c9..fdb8160 100644
--- a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py
@@ -24,13 +24,13 @@ import os
# server configurations
config = Script.get_config()
-hdp_root = None
+stack_root = None
slider_home = None
slider_bin_dir = None
slider_conf_dir = None
storm_slider_conf_dir = None
try:
- hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
+ stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
slider_home = os.environ['SLIDER_HOME']
slider_bin_dir = os.path.join(slider_home, 'bin')
slider_conf_dir = os.path.join(slider_home, 'conf')
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py
index b93b0eb..a16d0e7 100644
--- a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py
@@ -29,7 +29,7 @@ class SliderServiceCheck(Script):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd")
service = "SLIDER"
Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True, user=params.hdfs_user)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py
index e4b71f1..f930765 100644
--- a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py
@@ -24,7 +24,7 @@ config = Script.get_config()
sqoop_user = "sqoop"
-hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
+stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
sqoop_env_cmd_template = config['configurations']['sqoop-env']['content']
sqoop_home_dir = os.environ["SQOOP_HOME"]
sqoop_conf_dir = os.path.join(sqoop_home_dir, "conf")
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py
index 426315f..c0d0e8c 100644
--- a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py
@@ -53,7 +53,7 @@ class SqoopServiceCheckWindows(SqoopServiceCheck):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "SQOOP"
Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py
index 90bc76d..88e6246 100644
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py
@@ -27,7 +27,7 @@ config = Script.get_config()
stack_is_hdp23_or_further = Script.is_stack_greater_or_equal("2.3")
-hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
+stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],".."))
conf_dir = os.environ["STORM_CONF_DIR"]
hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"]
storm_user = hadoop_user
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py
index 99a6439..80ea0f5 100644
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py
@@ -38,7 +38,7 @@ class ServiceCheckWindows(ServiceCheck):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "STORM"
Execute(format("cmd /C {smoke_cmd} {service}", smoke_cmd=smoke_cmd, service=service), user=params.storm_user, logoutput=True)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py
index a76457f..bef1f02 100644
--- a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py
+++ b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py
@@ -32,9 +32,9 @@ def setup_ranger_storm(upgrade_type=None):
else:
from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin
- hdp_version = None
+ stack_version = None
if upgrade_type is not None:
- hdp_version = params.version
+ stack_version = params.version
if params.retryAble:
Logger.info("Storm: Setup ranger: command retry enables thus retrying if ranger admin is down !")
@@ -75,6 +75,6 @@ def setup_ranger_storm(upgrade_type=None):
component_list=['storm-client', 'storm-nimbus'], audit_db_is_enabled=params.xa_audit_db_is_enabled,
credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password,
ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password,
- hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble)
+ stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble)
else:
Logger.info('Ranger admin not installed')
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py
index ad80830..dd732f5 100644
--- a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py
@@ -36,9 +36,9 @@ except KeyError:
stack_version_formatted = ""
-hdp_root = None
+stack_root = None
try:
- hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
+ stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], ".."))
except:
pass
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py
index b70256c..c0c66af 100644
--- a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py
@@ -100,7 +100,7 @@ class TezServiceCheckWindows(TezServiceCheck):
def service_check(self, env):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "TEZ"
Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True, user=params.tez_user)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py
index 480fc8b..c36e152 100644
--- a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py
+++ b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py
@@ -26,11 +26,11 @@ import status_params
config = Script.get_config()
config_dir = None
-hdp_root = None
+stack_root = None
try:
# not used zookeeper_home_dir = os.environ["ZOOKEEPER_HOME"]
config_dir = os.environ["ZOOKEEPER_CONF_DIR"]
- hdp_root = os.environ["HADOOP_NODE_INSTALL_ROOT"]
+ stack_root = os.environ["HADOOP_NODE_INSTALL_ROOT"]
except:
pass
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py
index b53e5f5..622a5eb 100644
--- a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py
@@ -65,7 +65,7 @@ class ZookeeperServiceCheckWindows(ZookeeperServiceCheck):
import params
env.set_params(params)
- smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd")
+ smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd")
service = "Zookeeper"
Execute(format("cmd /C {smoke_cmd} {service}"), user=params.zk_user, logoutput=True, tries=3, try_sleep=20)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/host_scripts/alert_disk_space.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/host_scripts/alert_disk_space.py b/ambari-server/src/main/resources/host_scripts/alert_disk_space.py
index 94f19d6..b7a7038 100644
--- a/ambari-server/src/main/resources/host_scripts/alert_disk_space.py
+++ b/ambari-server/src/main/resources/host_scripts/alert_disk_space.py
@@ -37,10 +37,10 @@ PERCENT_USED_WARNING_DEFAULT = 50
PERCENT_USED_CRITICAL_DEFAULT = 80
# the location where HDP installs components when using HDP 2.2+
-HDP_HOME_DIR = "/usr/hdp"
+STACK_HOME_DIR = "/usr/hdp"
# the location where HDP installs components when using HDP 2.0 to 2.1
-HDP_HOME_LEGACY_DIR = "/usr/lib"
+STACK_HOME_LEGACY_DIR = "/usr/lib"
def get_tokens():
"""
@@ -66,14 +66,14 @@ def execute(configurations={}, parameters={}, host_name=None):
"""
# determine the location of HDP home
- hdp_home = None
- if os.path.isdir(HDP_HOME_DIR):
- hdp_home = HDP_HOME_DIR
- elif os.path.isdir(HDP_HOME_LEGACY_DIR):
- hdp_home = HDP_HOME_LEGACY_DIR
-
- # if hdp home was found, use it; otherwise default to None
- path = hdp_home if hdp_home is not None else None
+ stack_home = None
+ if os.path.isdir(STACK_HOME_DIR):
+ stack_home = STACK_HOME_DIR
+ elif os.path.isdir(STACK_HOME_LEGACY_DIR):
+ stack_home = STACK_HOME_LEGACY_DIR
+
+ # if stack home was found, use it; otherwise default to None
+ path = stack_home if stack_home is not None else None
try:
disk_usage = _get_disk_usage(path)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py
index 71ac3df..6c2806d 100644
--- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py
+++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py
@@ -28,7 +28,7 @@ class AfterInstallHook(Hook):
import params
env.set_params(params)
- setup_hdp_install_directory()
+ setup_install_directory()
setup_config()
if __name__ == "__main__":
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py
index f5851aa..dd66d58 100644
--- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py
@@ -35,7 +35,7 @@ else:
hadoop_conf_dir = "/etc/hadoop/conf"
hadoop_conf_empty_dir = "/etc/hadoop/conf.empty"
-versioned_hdp_root = '/usr/bigtop/current'
+versioned_stack_root = '/usr/bigtop/current'
#security params
security_enabled = config['configurations']['cluster-env']['security_enabled']
#java params
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py
index a930f54..8094129 100644
--- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py
+++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py
@@ -19,7 +19,7 @@ limitations under the License.
import os
from resource_management import *
-def setup_hdp_install_directory():
+def setup_install_directory():
import params
if params.rpm_version:
Execute(format('ambari-python-wrap /usr/bin/bigtop-select set all `ambari-python-wrap /usr/bin/bigtop-select versions | grep ^{rpm_version}- | tail -1`'),
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py
index 7a93d41..97e560a 100644
--- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py
@@ -47,7 +47,7 @@ else:
hadoop_conf_dir = "/etc/hadoop/conf"
hadoop_conf_empty_dir = "/etc/hadoop/conf.empty"
-versioned_hdp_root = '/usr/bigtop/current'
+versioned_stack_root = '/usr/bigtop/current'
#security params
security_enabled = config['configurations']['cluster-env']['security_enabled']
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py
index 0027a4c..c2adadb 100644
--- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py
+++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py
@@ -23,9 +23,9 @@ config = Script.get_config()
hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
hdfs_user = config['configurations']['hadoop-env']['hdfs_user']
-hdp_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}")
-datanode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-datanode.pid")
-namenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-namenode.pid")
-snamenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid")
-journalnode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-journalnode.pid")
-zkfc_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-zkfc.pid")
+hadoop_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}")
+datanode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-datanode.pid")
+namenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-namenode.pid")
+snamenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid")
+journalnode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-journalnode.pid")
+zkfc_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-zkfc.pid")
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py
index 73412e0..8a583b3 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py
@@ -20,7 +20,7 @@ limitations under the License.
from resource_management.libraries.script.hook import Hook
from shared_initialization import link_configs
from shared_initialization import setup_config
-from shared_initialization import setup_hdp_symlinks
+from shared_initialization import setup_stack_symlinks
class AfterInstallHook(Hook):
@@ -28,7 +28,7 @@ class AfterInstallHook(Hook):
import params
env.set_params(params)
- setup_hdp_symlinks()
+ setup_stack_symlinks()
setup_config()
link_configs(self.stroutfile)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
index ab7dea5..cd04492 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
@@ -49,7 +49,7 @@ if Script.is_stack_greater_or_equal("2.2"):
# not supported in HDP 2.2+
hadoop_conf_empty_dir = None
-versioned_hdp_root = '/usr/hdp/current'
+versioned_stack_root = '/usr/hdp/current'
#security params
security_enabled = config['configurations']['cluster-env']['security_enabled']
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py
index 8cf75cb..96dc104 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py
@@ -28,7 +28,7 @@ from resource_management.libraries.resources.xml_config import XmlConfig
from resource_management.libraries.script import Script
-def setup_hdp_symlinks():
+def setup_stack_symlinks():
"""
Invokes hdp-select set all against a calculated fully-qualified, "normalized" version based on a
stack version, such as "2.3". This should always be called after a component has been
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
index 7e37873..e4cad39 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
@@ -103,7 +103,7 @@ hadoop_libexec_dir = stack_select.get_hadoop_dir("libexec", force_latest_on_upgr
hadoop_conf_empty_dir = "/etc/hadoop/conf.empty"
hadoop_secure_dn_user = hdfs_user
hadoop_dir = "/etc/hadoop"
-versioned_hdp_root = '/usr/hdp/current'
+versioned_stack_root = '/usr/hdp/current'
hadoop_java_io_tmpdir = os.path.join(tmp_dir, "hadoop_java_io_tmpdir")
datanode_max_locked_memory = config['configurations']['hdfs-site']['dfs.datanode.max.locked.memory']
is_datanode_max_locked_memory_set = not is_empty(config['configurations']['hdfs-site']['dfs.datanode.max.locked.memory'])
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py
index bfc605c..d541a15 100644
--- a/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py
@@ -39,8 +39,8 @@ if os.environ.has_key("HADOOP_NODE_INSTALL_ROOT"):
hadoop_install_root = os.environ["HADOOP_NODE_INSTALL_ROOT"]
-hdp_log_dir = "c:\\hadoop\\logs"
-hdp_data_dir = "c:\\hadoop"
+stack_log_dir = "c:\\hadoop\\logs"
+stack_data_dir = "c:\\hadoop"
db_flavor = "MSSQL"
db_hostname = "localhost"
db_port = "1433"
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/test/python/TestVersionSelectUtil.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/TestVersionSelectUtil.py b/ambari-server/src/test/python/TestVersionSelectUtil.py
index f753023..9fa24cc 100644
--- a/ambari-server/src/test/python/TestVersionSelectUtil.py
+++ b/ambari-server/src/test/python/TestVersionSelectUtil.py
@@ -39,7 +39,7 @@ class TestVersionSelectUtil(TestCase):
@patch('__builtin__.open')
@patch("resource_management.core.shell.call")
def test_get_component_version(self, call_mock, open_mock):
- hdp_expected_version = "2.2.1.0-2175"
+ stack_expected_version = "2.2.1.0-2175"
# Mock classes for reading from a file
class MagicFile(object):
@@ -49,7 +49,7 @@ class TestVersionSelectUtil(TestCase):
"zookeeper-client"
])
def read(self, value):
- return (value + " - " + hdp_expected_version) if value in self.allowed_names else ("ERROR: Invalid package - " + value)
+ return (value + " - " + stack_expected_version) if value in self.allowed_names else ("ERROR: Invalid package - " + value)
def __exit__(self, exc_type, exc_val, exc_tb):
pass
@@ -88,6 +88,6 @@ class TestVersionSelectUtil(TestCase):
# Pass
version = self.module.get_component_version("HDP", "hadoop-hdfs-namenode")
- self.assertEquals(version, hdp_expected_version)
+ self.assertEquals(version, stack_expected_version)
version = self.module.get_component_version("HDP", "hadoop-hdfs-datanode")
- self.assertEquals(version, hdp_expected_version)
+ self.assertEquals(version, stack_expected_version)
http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/test/python/custom_actions/TestInstallPackages.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/custom_actions/TestInstallPackages.py b/ambari-server/src/test/python/custom_actions/TestInstallPackages.py
index 80ba480..f022c80 100644
--- a/ambari-server/src/test/python/custom_actions/TestInstallPackages.py
+++ b/ambari-server/src/test/python/custom_actions/TestInstallPackages.py
@@ -70,9 +70,9 @@ class TestInstallPackages(RMFTestCase):
def test_normal_flow_rhel(self,
write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock):
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[VERSION_STUB]
]
@@ -129,10 +129,10 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
def test_normal_flow_sles(self, write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock, is_suse_family_mock):
+ stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock, is_suse_family_mock):
is_suse_family_mock = True
Script.stack_version_from_distro_select = VERSION_STUB
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[VERSION_STUB]
]
@@ -190,10 +190,10 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
def test_exclude_existing_repo(self, write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
allInstalledPackages_mock, put_structured_out_mock,
is_redhat_family_mock, list_ambari_managed_repos_mock):
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[VERSION_STUB]
]
@@ -324,11 +324,11 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
def test_format_package_name(self, write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
allInstalledPackages_mock, put_structured_out_mock,
package_mock, is_suse_family_mock):
Script.stack_version_from_distro_select = VERSION_STUB
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[VERSION_STUB]
]
@@ -387,9 +387,9 @@ class TestInstallPackages(RMFTestCase):
def test_version_reporting__build_number_defined(self,
write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock):
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[OLD_VERSION_STUB], # before installation attempt
[OLD_VERSION_STUB, VERSION_STUB]
]
@@ -419,12 +419,12 @@ class TestInstallPackages(RMFTestCase):
self.assertTrue(write_actual_version_to_history_file_mock.called)
self.assertEquals(write_actual_version_to_history_file_mock.call_args[0], (VERSION_STUB_WITHOUT_BUILD_NUMBER, VERSION_STUB))
- hdp_versions_mock.reset_mock()
+ stack_versions_mock.reset_mock()
write_actual_version_to_history_file_mock.reset_mock()
put_structured_out_mock.reset_mock()
# Test retrying install again
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[OLD_VERSION_STUB, VERSION_STUB],
[OLD_VERSION_STUB, VERSION_STUB]
]
@@ -464,14 +464,14 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file")
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
@patch("os.path.exists")
- def test_version_reporting__build_number_not_defined__usr_hdp_present__no_components_installed(self,
+ def test_version_reporting__build_number_not_defined_stack_root_present__no_components_installed(self,
exists_mock,
write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock):
exists_mock.return_value = True
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[]
]
@@ -508,7 +508,7 @@ class TestInstallPackages(RMFTestCase):
self.assertFalse(write_actual_version_to_history_file_mock.called)
- hdp_versions_mock.reset_mock()
+ stack_versions_mock.reset_mock()
write_actual_version_to_history_file_mock.reset_mock()
put_structured_out_mock.reset_mock()
@@ -520,14 +520,14 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file")
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
@patch("os.path.exists")
- def test_version_reporting__build_number_not_defined__usr_hdp_absent(self,
+ def test_version_reporting__build_number_not_defined_stack_root_absent(self,
exists_mock,
write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock):
exists_mock.return_value = False
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[]
]
@@ -562,13 +562,13 @@ class TestInstallPackages(RMFTestCase):
self.assertFalse(write_actual_version_to_history_file_mock.called)
- hdp_versions_mock.reset_mock()
+ stack_versions_mock.reset_mock()
write_actual_version_to_history_file_mock.reset_mock()
put_structured_out_mock.reset_mock()
# Test retrying install again (correct build number, provided by other nodes, is now received from server)
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[]
]
@@ -611,12 +611,12 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.stack_select.get_stack_versions")
@patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file")
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
- def test_version_reporting__build_number_not_defined__usr_hdp_present(self,
+ def test_version_reporting__build_number_not_defined_stack_root_present(self,
write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock):
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[OLD_VERSION_STUB], # before installation attempt
[OLD_VERSION_STUB, VERSION_STUB]
]
@@ -646,12 +646,12 @@ class TestInstallPackages(RMFTestCase):
self.assertTrue(write_actual_version_to_history_file_mock.called)
self.assertEquals(write_actual_version_to_history_file_mock.call_args[0], (VERSION_STUB_WITHOUT_BUILD_NUMBER, VERSION_STUB))
- hdp_versions_mock.reset_mock()
+ stack_versions_mock.reset_mock()
write_actual_version_to_history_file_mock.reset_mock()
put_structured_out_mock.reset_mock()
# Test retrying install again
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[OLD_VERSION_STUB, VERSION_STUB],
[OLD_VERSION_STUB, VERSION_STUB]
]
@@ -689,12 +689,12 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.stack_select.get_stack_versions")
@patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file")
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
- def test_version_reporting__wrong_build_number_specified__usr_hdp_present(self,
+ def test_version_reporting__wrong_build_number_specified_stack_root_present(self,
write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock):
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[OLD_VERSION_STUB], # before installation attempt
[OLD_VERSION_STUB, VERSION_STUB]
]
@@ -724,12 +724,12 @@ class TestInstallPackages(RMFTestCase):
self.assertTrue(write_actual_version_to_history_file_mock.called)
self.assertEquals(write_actual_version_to_history_file_mock.call_args[0], ('2.2.0.1', VERSION_STUB))
- hdp_versions_mock.reset_mock()
+ stack_versions_mock.reset_mock()
write_actual_version_to_history_file_mock.reset_mock()
put_structured_out_mock.reset_mock()
# Test retrying install again
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[OLD_VERSION_STUB, VERSION_STUB],
[OLD_VERSION_STUB, VERSION_STUB]
]
@@ -768,14 +768,14 @@ class TestInstallPackages(RMFTestCase):
@patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file")
@patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file")
@patch("os.path.exists")
- def test_version_reporting__wrong_build_number_specified__usr_hdp_absent(self,
+ def test_version_reporting__wrong_build_number_specified_stack_root_absent(self,
exists_mock,
write_actual_version_to_history_file_mock,
read_actual_version_from_history_file_mock,
- hdp_versions_mock,
+ stack_versions_mock,
put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock):
exists_mock.return_value = False
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[]
]
@@ -810,13 +810,13 @@ class TestInstallPackages(RMFTestCase):
self.assertFalse(write_actual_version_to_history_file_mock.called)
- hdp_versions_mock.reset_mock()
+ stack_versions_mock.reset_mock()
write_actual_version_to_history_file_mock.reset_mock()
put_structured_out_mock.reset_mock()
# Test retrying install again (correct build number, provided by other nodes, is now received from server)
- hdp_versions_mock.side_effect = [
+ stack_versions_mock.side_effect = [
[], # before installation attempt
[]
]