You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by ao...@apache.org on 2015/05/20 18:59:14 UTC
ambari git commit: AMBARI-11271. Move hiveserver2 beeline check to
Hive SC and do mupliple other service-level optimizations (aonishuk)
Repository: ambari
Updated Branches:
refs/heads/trunk e70733ae2 -> 461b374ef
AMBARI-11271. Move hiveserver2 beeline check to Hive SC and do mupliple other service-level optimizations (aonishuk)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/461b374e
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/461b374e
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/461b374e
Branch: refs/heads/trunk
Commit: 461b374efee9e1e1a4e2ec2fb6ad763b82e9c62a
Parents: e70733a
Author: Andrew Onishuk <ao...@hortonworks.com>
Authored: Wed May 20 19:59:03 2015 +0300
Committer: Andrew Onishuk <ao...@hortonworks.com>
Committed: Wed May 20 19:59:03 2015 +0300
----------------------------------------------------------------------
.../HDFS/2.1.0.2.0/package/scripts/hdfs.py | 15 ++++
.../2.1.0.2.0/package/scripts/hdfs_datanode.py | 5 --
.../2.1.0.2.0/package/scripts/hdfs_namenode.py | 6 --
.../package/scripts/hdfs_nfsgateway.py | 5 --
.../2.1.0.2.0/package/scripts/hdfs_snamenode.py | 5 --
.../2.1.0.2.0/package/scripts/journalnode.py | 5 --
.../2.1.0.2.0/package/scripts/params_linux.py | 14 +++-
.../HDFS/2.1.0.2.0/package/scripts/utils.py | 33 +++++---
.../HIVE/0.12.0.2.0/metainfo.xml | 2 +-
.../0.12.0.2.0/package/scripts/hive_service.py | 35 --------
.../0.12.0.2.0/package/scripts/service_check.py | 27 ++++++-
.../0.8.1.2.2/package/scripts/kafka_broker.py | 5 +-
.../0.5.0.2.2/package/scripts/knox_gateway.py | 5 +-
.../YARN/2.1.0.2.0/package/scripts/service.py | 7 +-
.../package/scripts/zookeeper_server.py | 1 -
.../scripts/shared_initialization.py | 4 +-
.../scripts/shared_initialization.py | 44 ----------
.../python/stacks/2.0.6/HDFS/test_datanode.py | 85 +++++++-------------
.../stacks/2.0.6/HDFS/test_journalnode.py | 40 +++++----
.../python/stacks/2.0.6/HDFS/test_namenode.py | 24 ++++++
.../python/stacks/2.0.6/HDFS/test_nfsgateway.py | 52 ++++++------
.../python/stacks/2.0.6/HDFS/test_snamenode.py | 50 ++++++------
.../test/python/stacks/2.0.6/HDFS/test_zkfc.py | 84 +++++++++++++++----
.../stacks/2.0.6/HIVE/test_hive_server.py | 40 +--------
.../stacks/2.0.6/YARN/test_historyserver.py | 20 +++--
.../stacks/2.0.6/YARN/test_nodemanager.py | 18 +++--
.../stacks/2.0.6/YARN/test_resourcemanager.py | 19 +++--
.../2.0.6/ZOOKEEPER/test_zookeeper_server.py | 2 -
.../hooks/before-INSTALL/test_before_install.py | 5 +-
.../hooks/before-START/test_before_start.py | 12 ---
.../stacks/2.1/YARN/test_apptimelineserver.py | 10 ++-
31 files changed, 331 insertions(+), 348 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs.py
index 418571a..fa7f399 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs.py
@@ -29,6 +29,9 @@ from ambari_commons import OSConst
def hdfs(name=None):
import params
+ if params.create_lib_snappy_symlinks:
+ install_snappy()
+
# On some OS this folder could be not exists, so we will create it before pushing there files
Directory(params.limits_conf_dir,
recursive=True,
@@ -107,6 +110,18 @@ def hdfs(name=None):
if params.lzo_enabled and len(params.lzo_packages) > 0:
Package(params.lzo_packages)
+
+def install_snappy():
+ import params
+ Directory([params.so_target_dir_x86, params.so_target_dir_x64],
+ recursive=True,
+ )
+ Link(params.so_target_x86,
+ to=params.so_src_x86,
+ )
+ Link(params.so_target_x64,
+ to=params.so_src_x64,
+ )
@OsFamilyFuncImpl(os_family=OSConst.WINSRV_FAMILY)
def hdfs(component=None):
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_datanode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_datanode.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_datanode.py
index d02bb39..df847bd 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_datanode.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_datanode.py
@@ -51,11 +51,6 @@ def datanode(action=None):
handle_dfs_data_dir(create_dirs, params)
elif action == "start" or action == "stop":
import params
- Directory(params.hadoop_pid_dir_prefix,
- mode=0755,
- owner=params.hdfs_user,
- group=params.user_group
- )
service(
action=action, name="datanode",
user=params.hdfs_user,
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py
index 453d824..3900967 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py
@@ -47,12 +47,6 @@ def namenode(action=None, do_format=True, rolling_restart=False, env=None):
group=params.user_group
)
- Directory(params.hadoop_pid_dir_prefix,
- mode=0755,
- owner=params.hdfs_user,
- group=params.user_group
- )
-
if params.dfs_ha_enabled and \
params.dfs_ha_namenode_standby is not None and \
params.hostname == params.dfs_ha_namenode_standby:
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_nfsgateway.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_nfsgateway.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_nfsgateway.py
index ac0e24d..efebfc5 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_nfsgateway.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_nfsgateway.py
@@ -63,11 +63,6 @@ def nfsgateway(action=None, format=False):
if action == "configure":
return
elif action == "start" or action == "stop":
- Directory(params.hadoop_pid_dir_prefix,
- mode=0755,
- owner=params.root_user,
- group=params.root_group
- )
service(
action=action,
name="nfs3",
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_snamenode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_snamenode.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_snamenode.py
index 78ef977..aded211 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_snamenode.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_snamenode.py
@@ -39,11 +39,6 @@ def snamenode(action=None, format=False):
group=params.user_group)
elif action == "start" or action == "stop":
import params
- Directory(params.hadoop_pid_dir_prefix,
- mode=0755,
- owner=params.hdfs_user,
- group=params.user_group
- )
service(
action=action,
name="secondarynamenode",
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
index ada05e1..0cae36f 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/journalnode.py
@@ -59,11 +59,6 @@ class JournalNodeDefault(JournalNode):
env.set_params(params)
self.configure(env)
- Directory(params.hadoop_pid_dir_prefix,
- mode=0755,
- owner=params.hdfs_user,
- group=params.user_group
- )
service(
action="start", name="journalnode", user=params.hdfs_user,
create_pid_dir=True,
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
index 3f1fb0d..6a8a47a 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
@@ -69,6 +69,7 @@ hadoop_bin_dir = conf_select.get_hadoop_dir("bin")
hadoop_home = "/usr/lib/hadoop"
hadoop_secure_dn_user = hdfs_user
hadoop_conf_dir = conf_select.get_hadoop_conf_dir()
+hadoop_lib_home = conf_select.get_hadoop_dir("lib")
# hadoop parameters for 2.2+
if Script.is_hdp_stack_greater_or_equal("2.2"):
@@ -93,10 +94,11 @@ if Script.is_hdp_stack_greater_or_equal("2.2"):
else:
hadoop_secure_dn_user = '""'
-
ambari_libs_dir = "/var/lib/ambari-agent/lib"
limits_conf_dir = "/etc/security/limits.d"
+create_lib_snappy_symlinks = not Script.is_hdp_stack_greater_or_equal("2.2")
+
if Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.1") and not OSCheck.is_suse_family():
# deprecated rhel jsvc_path
jsvc_path = "/usr/libexec/bigtop-utils"
@@ -106,6 +108,16 @@ else:
execute_path = os.environ['PATH'] + os.pathsep + hadoop_bin_dir
ulimit_cmd = "ulimit -c unlimited ; "
+snappy_so = "libsnappy.so"
+so_target_dir_x86 = format("{hadoop_lib_home}/native/Linux-i386-32")
+so_target_dir_x64 = format("{hadoop_lib_home}/native/Linux-amd64-64")
+so_target_x86 = format("{so_target_dir_x86}/{snappy_so}")
+so_target_x64 = format("{so_target_dir_x64}/{snappy_so}")
+so_src_dir_x86 = format("{hadoop_home}/lib")
+so_src_dir_x64 = format("{hadoop_home}/lib64")
+so_src_x86 = format("{so_src_dir_x86}/{snappy_so}")
+so_src_x64 = format("{so_src_dir_x64}/{snappy_so}")
+
#security params
smoke_user_keytab = config['configurations']['cluster-env']['smokeuser_keytab']
hdfs_user_keytab = config['configurations']['hadoop-env']['hdfs_user_keytab']
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
index 217f2f0..21976f4 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
@@ -167,20 +167,35 @@ def service(action=None, name=None, user=None, options="", create_pid_dir=False,
"ls {pid_file} >/dev/null 2>&1 &&"
" ps -p `cat {pid_file}` >/dev/null 2>&1")
- if create_pid_dir:
- Directory(pid_dir,
- owner=user,
- recursive=True)
- if create_log_dir:
+ # on STOP directories shouldn't be created
+ # since during stop still old dirs are used (which were created during previous start)
+ if action != "stop":
if name == "nfs3":
- Directory(log_dir,
- mode=0775,
+ Directory(params.hadoop_pid_dir_prefix,
+ mode=0755,
owner=params.root_user,
- group=params.user_group)
+ group=params.root_group
+ )
else:
- Directory(log_dir,
+ Directory(params.hadoop_pid_dir_prefix,
+ mode=0755,
+ owner=params.hdfs_user,
+ group=params.user_group
+ )
+ if create_pid_dir:
+ Directory(pid_dir,
owner=user,
recursive=True)
+ if create_log_dir:
+ if name == "nfs3":
+ Directory(log_dir,
+ mode=0775,
+ owner=params.root_user,
+ group=params.user_group)
+ else:
+ Directory(log_dir,
+ owner=user,
+ recursive=True)
if params.security_enabled and name == "datanode":
## The directory where pid files are stored in the secure data environment.
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/metainfo.xml b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/metainfo.xml
index 32bc0ba..e12935d 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/metainfo.xml
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/metainfo.xml
@@ -279,7 +279,7 @@
<commandScript>
<script>scripts/service_check.py</script>
<scriptType>PYTHON</scriptType>
- <timeout>120</timeout>
+ <timeout>300</timeout>
</commandScript>
<requiredServices>
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py
index 86312d2..ea70b8a 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py
@@ -91,41 +91,6 @@ def hive_service(name, action='start', rolling_restart=False):
Execute(db_connection_check_command,
path='/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin', tries=5, try_sleep=10)
-
- # AMBARI-5800 - wait for the server to come up instead of just the PID existance
- if name == 'hiveserver2':
- SOCKET_WAIT_SECONDS = 120
-
- start_time = time.time()
- end_time = start_time + SOCKET_WAIT_SECONDS
-
- is_service_socket_valid = False
- print "Waiting for the Hive server to start..."
- if params.security_enabled:
- kinitcmd=format("{kinit_path_local} -kt {smoke_user_keytab} {smokeuser_principal}; ")
- else:
- kinitcmd=None
- while time.time() < end_time:
- try:
- check_thrift_port_sasl(params.hostname, params.hive_server_port, params.hive_server2_authentication,
- params.hive_server_principal, kinitcmd, params.smokeuser,
- transport_mode=params.hive_transport_mode, http_endpoint=params.hive_http_endpoint,
- ssl=params.hive_ssl, ssl_keystore=params.hive_ssl_keystore_path,
- ssl_password=params.hive_ssl_keystore_password)
- is_service_socket_valid = True
- break
- except Exception, e:
- time.sleep(5)
-
- elapsed_time = time.time() - start_time
-
- if not is_service_socket_valid:
- raise Fail("Connection to Hive server %s on port %s failed after %d seconds" %
- (params.hostname, params.hive_server_port, elapsed_time))
-
- print "Successfully connected to Hive at %s on port %s after %d seconds" %\
- (params.hostname, params.hive_server_port, elapsed_time)
-
elif action == 'stop':
daemon_kill_cmd = format("{sudo} kill `cat {pid_file}`")
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
index 04bd3b2..4876fb2 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py
@@ -21,6 +21,7 @@ limitations under the License.
from resource_management import *
import socket
import sys
+import time
from hcat_service_check import hcat_service_check
from webhcat_service_check import webhcat_service_check
from ambari_commons import OSConst
@@ -58,8 +59,17 @@ class HiveServiceCheckDefault(HiveServiceCheck):
else:
kinitcmd=None
+ SOCKET_WAIT_SECONDS = 290
+
+ start_time = time.time()
+ end_time = start_time + SOCKET_WAIT_SECONDS
+
+ print "Waiting for the Hive server to start..."
+
workable_server_available = False
- for address in address_list:
+ i = 0
+ while time.time() < end_time and not workable_server_available:
+ address = address_list[i]
try:
check_thrift_port_sasl(address, port, params.hive_server2_authentication,
params.hive_server_principal, kinitcmd, params.smokeuser,
@@ -70,9 +80,20 @@ class HiveServiceCheckDefault(HiveServiceCheck):
workable_server_available = True
except:
print "Connection to %s on port %s failed" % (address, port)
-
+ time.sleep(5)
+
+ i += 1
+ if i == len(address_list)-1:
+ i = 0
+
+ elapsed_time = time.time() - start_time
+
if not workable_server_available:
- exit(1)
+ raise Fail("Connection to Hive server %s on port %s failed after %d seconds" %
+ (params.hostname, params.hive_server_port, elapsed_time))
+
+ print "Successfully connected to Hive at %s on port %s after %d seconds" %\
+ (params.hostname, params.hive_server_port, elapsed_time)
hcat_service_check()
webhcat_service_check()
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/kafka_broker.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/kafka_broker.py b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/kafka_broker.py
index 31d949a..1b36f28 100644
--- a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/kafka_broker.py
+++ b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/kafka_broker.py
@@ -56,12 +56,13 @@ class KafkaBroker(Script):
def stop(self, env, rolling_restart=False):
import params
env.set_params(params)
- self.configure(env)
daemon_cmd = format('source {params.conf_dir}/kafka-env.sh; {params.kafka_bin} stop')
Execute(daemon_cmd,
user=params.kafka_user,
)
- Execute (format("rm -f {params.kafka_pid_file}"))
+ File (params.kafka_pid_file,
+ action = "delete"
+ )
def status(self, env):
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
index 956d09b..fd7b451 100644
--- a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
+++ b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/knox_gateway.py
@@ -123,13 +123,14 @@ class KnoxGatewayDefault(KnoxGateway):
def stop(self, env, rolling_restart=False):
import params
env.set_params(params)
- self.configure(env)
daemon_cmd = format('{knox_bin} stop')
Execute(daemon_cmd,
environment={'JAVA_HOME': params.java_home},
user=params.knox_user,
)
- Execute (format("rm -f {knox_pid_file}"))
+ File(params.knox_pid_file,
+ action="delete",
+ )
def status(self, env):
import status_params
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/service.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/service.py b/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/service.py
index 3d6b8d1..f368bd4 100644
--- a/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/service.py
+++ b/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/service.py
@@ -64,7 +64,12 @@ def service(componentName, action='start', serviceName='yarn'):
Execute(daemon_cmd, user = usr, not_if = check_process)
# Ensure that the process with the expected PID exists.
- Execute(check_process, user = usr, not_if = check_process, initial_wait = 5)
+ Execute(check_process,
+ user=usr,
+ not_if = check_process,
+ tries=5,
+ try_sleep=1,
+ )
elif action == 'stop':
daemon_cmd = format("{cmd} stop {componentName}")
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/zookeeper_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/zookeeper_server.py b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/zookeeper_server.py
index d685d1d..f3e9d10 100644
--- a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/zookeeper_server.py
+++ b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/zookeeper_server.py
@@ -57,7 +57,6 @@ class ZookeeperServer(Script):
def stop(self, env, rolling_restart=False):
import params
env.set_params(params)
- self.configure(env, rolling_restart=rolling_restart)
zookeeper_service(action = 'stop')
@OsFamilyImpl(os_family=OsFamilyImpl.DEFAULT)
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/shared_initialization.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/shared_initialization.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/shared_initialization.py
index b58959a..e8746d4 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/shared_initialization.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-INSTALL/scripts/shared_initialization.py
@@ -28,7 +28,7 @@ def setup_java():
"""
import params
- jdk_curl_target = format("{artifact_dir}/{jdk_name}")
+ jdk_curl_target = format("{tmp_dir}/{jdk_name}")
java_dir = os.path.dirname(params.java_home)
java_exec = format("{java_home}/bin/java")
tmp_java_dir = format("{tmp_dir}/jdk")
@@ -39,8 +39,10 @@ def setup_java():
Directory(params.artifact_dir,
recursive = True,
)
+
File(jdk_curl_target,
content = DownloadSource(format("{jdk_location}/{jdk_name}")),
+ not_if = format("test -f {jdk_curl_target}")
)
if params.jdk_name.endswith(".bin"):
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/shared_initialization.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/shared_initialization.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/shared_initialization.py
index dd93818..f04c6a0 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/shared_initialization.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/shared_initialization.py
@@ -33,9 +33,6 @@ def setup_hadoop():
sudo=True,
)
- if params.current_service == "HDFS":
- install_snappy()
-
#directories
if params.has_namenode:
Directory(params.hdfs_log_dir_prefix,
@@ -94,23 +91,6 @@ def setup_hadoop():
content=Template("hadoop-metrics2.properties.j2")
)
-def setup_database():
- """
- Load DB
- """
- import params
- db_driver_download_url = None
-
- if params.server_db_name == 'oracle' and params.oracle_driver_url != "":
- db_driver_download_url = params.oracle_driver_symlink_url
- elif params.server_db_name == 'mysql' and params.mysql_driver_url != "":
- db_driver_download_url = params.mysql_driver_symlink_url
-
- if db_driver_download_url:
- File(format("{hadoop_lib_home}/{db_driver_filename}"),
- content = DownloadSource(db_driver_download_url),
- )
-
def setup_configs():
"""
@@ -150,30 +130,6 @@ def generate_include_file():
group=params.user_group
)
-
-def install_snappy():
- import params
-
- snappy_so = "libsnappy.so"
- so_target_dir_x86 = format("{hadoop_lib_home}/native/Linux-i386-32")
- so_target_dir_x64 = format("{hadoop_lib_home}/native/Linux-amd64-64")
- so_target_x86 = format("{so_target_dir_x86}/{snappy_so}")
- so_target_x64 = format("{so_target_dir_x64}/{snappy_so}")
- so_src_dir_x86 = format("{hadoop_home}/lib")
- so_src_dir_x64 = format("{hadoop_home}/lib64")
- so_src_x86 = format("{so_src_dir_x86}/{snappy_so}")
- so_src_x64 = format("{so_src_dir_x64}/{snappy_so}")
- if params.has_namenode and params.create_lib_snappy_symlinks:
- Directory([so_target_dir_x86, so_target_dir_x64],
- recursive=True,
- )
- Link(so_target_x86,
- to=so_src_x86,
- )
- Link(so_target_x64,
- to=so_src_x64,
- )
-
def create_javahome_symlink():
if os.path.exists("/usr/jdk/jdk1.6.0_31") and not os.path.exists("/usr/jdk64/jdk1.6.0_31"):
Directory("/usr/jdk64/",
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
index 4eb8178..d594dcf 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
@@ -80,19 +80,6 @@ class TestDatanode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'hdfs',
- group = 'hadoop',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid` >/dev/null 2>&1',
@@ -163,7 +150,7 @@ class TestDatanode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assert_configure_secured("2.2")
+ self.assert_configure_secured("2.2", snappy_enabled=False)
self.assertResourceCalled('Directory', '/var/run/hadoop',
owner = 'hdfs',
group = 'hadoop',
@@ -204,7 +191,7 @@ class TestDatanode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assert_configure_secured("2.2")
+ self.assert_configure_secured("2.2", snappy_enabled=False)
self.assertResourceCalled('Directory', '/var/run/hadoop',
owner = 'hdfs',
group = 'hadoop',
@@ -237,19 +224,6 @@ class TestDatanode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'hdfs',
- group = 'hadoop',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid` >/dev/null 2>&1',
@@ -279,19 +253,6 @@ class TestDatanode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'hdfs',
- group = 'hadoop',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid` >/dev/null 2>&1',
@@ -323,19 +284,6 @@ class TestDatanode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'hdfs',
- group = 'hadoop',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-datanode.pid` >/dev/null 2>&1',
@@ -350,6 +298,18 @@ class TestDatanode(RMFTestCase):
self.assertNoMoreResources()
def assert_configure_default(self):
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -399,11 +359,24 @@ class TestDatanode(RMFTestCase):
cd_access='a'
)
- def assert_configure_secured(self, stackVersion=STACK_VERSION):
+ def assert_configure_secured(self, stackVersion=STACK_VERSION, snappy_enabled=True):
conf_dir = '/etc/hadoop/conf'
if stackVersion != self.STACK_VERSION:
conf_dir = '/usr/hdp/current/hadoop-client/conf'
-
+
+ if snappy_enabled:
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
index 3fe2acb..469beef 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
@@ -79,14 +79,6 @@ class TestJournalnode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-journalnode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-journalnode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-journalnode.pid` >/dev/null 2>&1',
@@ -151,14 +143,6 @@ class TestJournalnode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-journalnode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-journalnode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-journalnode.pid` >/dev/null 2>&1',
@@ -179,6 +163,18 @@ class TestJournalnode(RMFTestCase):
recursive = True,
cd_access='a'
)
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -221,6 +217,18 @@ class TestJournalnode(RMFTestCase):
recursive = True,
cd_access='a'
)
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
index c753c39..c57287b 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
@@ -813,6 +813,18 @@ class TestNamenode(RMFTestCase):
self.assertNoMoreResources()
def assert_configure_default(self):
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -856,6 +868,18 @@ class TestNamenode(RMFTestCase):
)
def assert_configure_secured(self):
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
index 7255ea5..c84902f 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
@@ -82,20 +82,6 @@ class TestNFSGateway(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'root',
- group = 'root',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/root',
- owner = 'root',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/root',
- owner = 'root',
- group = 'hadoop',
- mode = 0775
- )
self.assertResourceCalled('File', '/var/run/hadoop/root/hadoop_privileged_nfs3.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/root/hadoop_privileged_nfs3.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/root/hadoop_privileged_nfs3.pid` >/dev/null 2>&1',
@@ -163,20 +149,6 @@ class TestNFSGateway(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'root',
- group = 'root',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/root',
- owner = 'root',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/root',
- owner = 'root',
- group = 'hadoop',
- mode = 0775
- )
self.assertResourceCalled('File', '/var/run/hadoop/root/hadoop_privileged_nfs3.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/root/hadoop_privileged_nfs3.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/root/hadoop_privileged_nfs3.pid` >/dev/null 2>&1',
@@ -191,6 +163,18 @@ class TestNFSGateway(RMFTestCase):
self.assertNoMoreResources()
def assert_configure_default(self):
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -227,6 +211,18 @@ class TestNFSGateway(RMFTestCase):
)
def assert_configure_secured(self):
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
index fd665c3..543a0e4 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
@@ -86,19 +86,6 @@ class TestSNamenode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'hdfs',
- group = 'hadoop',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-secondarynamenode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-secondarynamenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-secondarynamenode.pid` >/dev/null 2>&1',
@@ -173,19 +160,6 @@ class TestSNamenode(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop',
- owner = 'hdfs',
- group = 'hadoop',
- mode = 0755
- )
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-secondarynamenode.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-secondarynamenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-secondarynamenode.pid` >/dev/null 2>&1',
@@ -200,6 +174,18 @@ class TestSNamenode(RMFTestCase):
self.assertNoMoreResources()
def assert_configure_default(self):
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -250,6 +236,18 @@ class TestSNamenode(RMFTestCase):
)
def assert_configure_secured(self):
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
index c2fdffa..1de19f3 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
@@ -34,6 +34,18 @@ class TestZkfc(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -73,6 +85,11 @@ class TestZkfc(RMFTestCase):
group = 'hadoop',
mode = 0755
)
+ self.assertResourceCalled('Directory', '/var/run/hadoop',
+ owner = 'hdfs',
+ group = 'hadoop',
+ mode = 0755,
+ )
self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
owner = 'hdfs',
recursive = True,
@@ -100,14 +117,6 @@ class TestZkfc(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-zkfc.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-zkfc.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-zkfc.pid` >/dev/null 2>&1',
@@ -129,6 +138,18 @@ class TestZkfc(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -168,6 +189,11 @@ class TestZkfc(RMFTestCase):
group = 'hadoop',
mode = 0755
)
+ self.assertResourceCalled('Directory', '/var/run/hadoop',
+ owner = 'hdfs',
+ group = 'hadoop',
+ mode = 0755,
+ )
self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
owner = 'hdfs',
recursive = True,
@@ -194,14 +220,6 @@ class TestZkfc(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/var/log/hadoop/hdfs',
- owner = 'hdfs',
- recursive = True,
- )
self.assertResourceCalled('File', '/var/run/hadoop/hdfs/hadoop-hdfs-zkfc.pid',
action = ['delete'],
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-zkfc.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-zkfc.pid` >/dev/null 2>&1',
@@ -223,6 +241,18 @@ class TestZkfc(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -263,6 +293,11 @@ class TestZkfc(RMFTestCase):
mode = 0755
)
# TODO: verify that the znode initialization occurs prior to ZKFC startup
+ self.assertResourceCalled('Directory', '/var/run/hadoop',
+ owner = 'hdfs',
+ group = 'hadoop',
+ mode = 0755,
+ )
self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
owner = 'hdfs',
recursive = True,
@@ -289,6 +324,18 @@ class TestZkfc(RMFTestCase):
hdp_stack_version = self.STACK_VERSION,
target = RMFTestCase.TARGET_COMMON_SERVICES
)
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
+ recursive = True,
+ )
+ self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
+ recursive = True,
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
+ to = '/usr/lib/hadoop/lib/libsnappy.so',
+ )
+ self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
+ to = '/usr/lib/hadoop/lib64/libsnappy.so',
+ )
self.assertResourceCalled('Directory', '/etc/security/limits.d',
owner = 'root',
group = 'root',
@@ -329,6 +376,11 @@ class TestZkfc(RMFTestCase):
mode = 0755
)
# TODO: verify that the znode initialization occurs prior to ZKFC startup
+ self.assertResourceCalled('Directory', '/var/run/hadoop',
+ owner = 'hdfs',
+ group = 'hadoop',
+ mode = 0755,
+ )
self.assertResourceCalled('Directory', '/var/run/hadoop/hdfs',
owner = 'hdfs',
recursive = True,
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
index 90f3e06..3859236 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
@@ -80,11 +80,6 @@ class TestHiveServer(RMFTestCase):
tries=5,
try_sleep=10
)
- self.assertResourceCalled('Execute', "! beeline -u 'jdbc:hive2://c6401.ambari.apache.org:10000/;transportMode=binary;auth=noSasl' -e '' 2>&1| awk '{print}'|grep -i -e 'Connection refused' -e 'Invalid URL'",
- path = ['/bin/', '/usr/bin/', '/usr/lib/hive/bin/', '/usr/sbin/'],
- user = 'ambari-qa',
- timeout = 30,
- )
self.assertNoMoreResources()
@patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
@@ -114,11 +109,6 @@ class TestHiveServer(RMFTestCase):
self.assertResourceCalled('Execute', '/usr/jdk64/jdk1.7.0_45/bin/java -cp /usr/lib/ambari-agent/DBConnectionVerification.jar:/usr/lib/hive/lib//mysql-connector-java.jar org.apache.ambari.server.DBConnectionVerification \'jdbc:mysql://c6402.ambari.apache.org/hive?createDatabaseIfNotExist=true\' hive \'!`"\'"\'"\' 1\' com.mysql.jdbc.Driver',
path=['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'], tries=5, try_sleep=10
)
- self.assertResourceCalled('Execute', "! beeline -u 'jdbc:hive2://c6401.ambari.apache.org:10000/;transportMode=binary;auth=noSasl' -e '' 2>&1| awk '{print}'|grep -i -e 'Connection refused' -e 'Invalid URL'",
- path = ['/bin/', '/usr/bin/', '/usr/lib/hive/bin/', '/usr/sbin/'],
- user = 'ambari-qa',
- timeout = 30,
- )
self.assertNoMoreResources()
@patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
@@ -147,11 +137,6 @@ class TestHiveServer(RMFTestCase):
self.assertResourceCalled('Execute', '/usr/jdk64/jdk1.7.0_45/bin/java -cp /usr/lib/ambari-agent/DBConnectionVerification.jar:/usr/lib/hive/lib//mysql-connector-java.jar org.apache.ambari.server.DBConnectionVerification \'jdbc:mysql://c6402.ambari.apache.org/hive?createDatabaseIfNotExist=true\' hive \'!`"\'"\'"\' 1\' com.mysql.jdbc.Driver',
path=['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'], tries=5, try_sleep=10
)
- self.assertResourceCalled('Execute', "! beeline -u 'jdbc:hive2://c6401.ambari.apache.org:10000/;transportMode=binary;auth=noSasl' -e '' 2>&1| awk '{print}'|grep -i -e 'Connection refused' -e 'Invalid URL'",
- path = ['/bin/', '/usr/bin/', '/usr/lib/hive/bin/', '/usr/sbin/'],
- user = 'ambari-qa',
- timeout = 30,
- )
self.assertNoMoreResources()
@@ -181,11 +166,6 @@ class TestHiveServer(RMFTestCase):
self.assertResourceCalled('Execute', '/usr/jdk64/jdk1.7.0_45/bin/java -cp /usr/lib/ambari-agent/DBConnectionVerification.jar:/usr/lib/hive/lib//mysql-connector-java.jar org.apache.ambari.server.DBConnectionVerification \'jdbc:mysql://c6402.ambari.apache.org/hive?createDatabaseIfNotExist=true\' hive \'!`"\'"\'"\' 1\' com.mysql.jdbc.Driver',
path=['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'], tries=5, try_sleep=10
)
- self.assertResourceCalled('Execute', "! beeline -u 'jdbc:hive2://c6401.ambari.apache.org:10000/;transportMode=binary;auth=noSasl' -e '' 2>&1| awk '{print}'|grep -i -e 'Connection refused' -e 'Invalid URL'",
- path = ['/bin/', '/usr/bin/', '/usr/lib/hive/bin/', '/usr/sbin/'],
- user = 'ambari-qa',
- timeout = 30,
- )
self.assertNoMoreResources()
@patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
@@ -258,16 +238,6 @@ class TestHiveServer(RMFTestCase):
tries=5,
try_sleep=10,
)
- self.assertResourceCalled('Execute',
- '/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa@EXAMPLE.COM; ',
- user='ambari-qa',
- )
- self.assertResourceCalled('Execute',
- "! beeline -u 'jdbc:hive2://c6401.ambari.apache.org:10000/;transportMode=binary;principal=hive/_HOST@EXAMPLE.COM' -e '' 2>&1| awk '{print}'|grep -i -e 'Connection refused' -e 'Invalid URL'",
- path=['/bin/', '/usr/bin/', '/usr/lib/hive/bin/', '/usr/sbin/'],
- user='ambari-qa',
- timeout=30,
- )
self.assertNoMoreResources()
self.assertTrue(check_fs_root_mock.called)
@@ -701,10 +671,7 @@ class TestHiveServer(RMFTestCase):
@patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=True))
- @patch("hive_server.HiveServer.pre_rolling_restart")
- @patch("hive_server.HiveServer.start")
- def test_stop_during_upgrade(self, hive_server_start_mock,
- hive_server_pre_rolling_mock):
+ def test_stop_during_upgrade(self):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer", command = "restart", config_file = "hive-upgrade.json",
@@ -720,9 +687,8 @@ class TestHiveServer(RMFTestCase):
self.assertResourceCalled('Execute', 'hdp-select set hive-server2 2.2.1.0-2065',)
- @patch("hive_server.HiveServer.pre_rolling_restart")
- @patch("hive_server.HiveServer.start")
- def test_stop_during_upgrade_bad_hive_version(self, hive_server_start_mock, hive_server_pre_rolling_mock):
+
+ def test_stop_during_upgrade_bad_hive_version(self):
try:
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer", command = "restart", config_file = "hive-upgrade.json",
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/YARN/test_historyserver.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/YARN/test_historyserver.py b/ambari-server/src/test/python/stacks/2.0.6/YARN/test_historyserver.py
index bc527e7..cd5041b 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/YARN/test_historyserver.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/YARN/test_historyserver.py
@@ -63,10 +63,12 @@ class TestHistoryServer(RMFTestCase):
self.assertResourceCalled('Execute', 'ulimit -c unlimited; export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop-mapreduce/sbin/mr-jobhistory-daemon.sh --config /etc/hadoop/conf start historyserver',
not_if=pid_check_cmd,
user='mapred')
- self.assertResourceCalled('Execute', pid_check_cmd,
- not_if=pid_check_cmd,
- initial_wait=5,
- user='mapred')
+ self.assertResourceCalled('Execute', 'ls /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid` >/dev/null 2>&1',
+ not_if = 'ls /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid` >/dev/null 2>&1',
+ tries = 5,
+ user = 'mapred',
+ try_sleep = 1,
+ )
self.assertNoMoreResources()
def test_stop_default(self):
@@ -115,10 +117,12 @@ class TestHistoryServer(RMFTestCase):
self.assertResourceCalled('Execute', 'ulimit -c unlimited; export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop-mapreduce/sbin/mr-jobhistory-daemon.sh --config /etc/hadoop/conf start historyserver',
not_if=pid_check_cmd,
user='mapred')
- self.assertResourceCalled('Execute', pid_check_cmd,
- user='mapred',
- not_if=pid_check_cmd,
- initial_wait=5)
+ self.assertResourceCalled('Execute', 'ls /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid` >/dev/null 2>&1',
+ not_if = 'ls /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-mapreduce/mapred/mapred-mapred-historyserver.pid` >/dev/null 2>&1',
+ tries = 5,
+ user = 'mapred',
+ try_sleep = 1,
+ )
self.assertNoMoreResources()
def test_stop_secured(self):
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/YARN/test_nodemanager.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/YARN/test_nodemanager.py b/ambari-server/src/test/python/stacks/2.0.6/YARN/test_nodemanager.py
index 005b533..5c517f1 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/YARN/test_nodemanager.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/YARN/test_nodemanager.py
@@ -63,10 +63,12 @@ class TestNodeManager(RMFTestCase):
self.assertResourceCalled('Execute', 'ulimit -c unlimited; export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop-yarn/sbin/yarn-daemon.sh --config /etc/hadoop/conf start nodemanager',
not_if=pid_check_cmd,
user='yarn')
- self.assertResourceCalled('Execute', pid_check_cmd,
- user='yarn',
- not_if=pid_check_cmd,
- initial_wait=5)
+ self.assertResourceCalled('Execute', 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid` >/dev/null 2>&1',
+ not_if = 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid` >/dev/null 2>&1',
+ tries = 5,
+ user = 'yarn',
+ try_sleep = 1,
+ )
self.assertNoMoreResources()
def test_stop_default(self):
@@ -113,9 +115,11 @@ class TestNodeManager(RMFTestCase):
not_if=pid_check_cmd,
user='yarn')
self.assertResourceCalled('Execute', 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid` >/dev/null 2>&1',
- user='yarn',
- not_if=pid_check_cmd,
- initial_wait=5)
+ not_if = 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-nodemanager.pid` >/dev/null 2>&1',
+ tries = 5,
+ user = 'yarn',
+ try_sleep = 1,
+ )
self.assertNoMoreResources()
def test_stop_secured(self):
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/YARN/test_resourcemanager.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/YARN/test_resourcemanager.py b/ambari-server/src/test/python/stacks/2.0.6/YARN/test_resourcemanager.py
index a2ee27f..b775f48 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/YARN/test_resourcemanager.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/YARN/test_resourcemanager.py
@@ -101,10 +101,11 @@ class TestResourceManager(RMFTestCase):
user = 'yarn',
)
self.assertResourceCalled('Execute', 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid` >/dev/null 2>&1',
- initial_wait = 5,
- not_if = 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid` >/dev/null 2>&1',
- user = 'yarn',
- )
+ not_if = 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid` >/dev/null 2>&1',
+ tries = 5,
+ user = 'yarn',
+ try_sleep = 1,
+ )
self.assertNoMoreResources()
def test_stop_default(self):
@@ -149,10 +150,12 @@ class TestResourceManager(RMFTestCase):
self.assertResourceCalled('Execute', 'ulimit -c unlimited; export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop-yarn/sbin/yarn-daemon.sh --config /etc/hadoop/conf start resourcemanager',
not_if=pid_check_cmd,
user='yarn')
- self.assertResourceCalled('Execute', pid_check_cmd,
- user='yarn',
- not_if=pid_check_cmd,
- initial_wait=5)
+ self.assertResourceCalled('Execute', 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid` >/dev/null 2>&1',
+ not_if = 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-resourcemanager.pid` >/dev/null 2>&1',
+ tries = 5,
+ user = 'yarn',
+ try_sleep = 1,
+ )
self.assertNoMoreResources()
def test_stop_secured(self):
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/ZOOKEEPER/test_zookeeper_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/ZOOKEEPER/test_zookeeper_server.py b/ambari-server/src/test/python/stacks/2.0.6/ZOOKEEPER/test_zookeeper_server.py
index ca8cf16..dc2d062 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/ZOOKEEPER/test_zookeeper_server.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/ZOOKEEPER/test_zookeeper_server.py
@@ -64,7 +64,6 @@ class TestZookeeperServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assert_configure_default()
self.assertResourceCalled('Execute', 'source /etc/zookeeper/conf/zookeeper-env.sh ; env ZOOCFGDIR=/etc/zookeeper/conf ZOOCFG=zoo.cfg /usr/lib/zookeeper/bin/zkServer.sh stop',
user = 'zookeeper',
)
@@ -109,7 +108,6 @@ class TestZookeeperServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assert_configure_secured()
self.assertResourceCalled('Execute', 'source /etc/zookeeper/conf/zookeeper-env.sh ; env ZOOCFGDIR=/etc/zookeeper/conf ZOOCFG=zoo.cfg /usr/lib/zookeeper/bin/zkServer.sh stop',
user = 'zookeeper',
)
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/hooks/before-INSTALL/test_before_install.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-INSTALL/test_before_install.py b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-INSTALL/test_before_install.py
index e752427..228192e 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-INSTALL/test_before_install.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-INSTALL/test_before_install.py
@@ -45,15 +45,16 @@ class TestHookBeforeInstall(RMFTestCase):
self.assertResourceCalled('Directory', '/tmp/AMBARI-artifacts/',
recursive = True,
)
- self.assertResourceCalled('File', '/tmp/AMBARI-artifacts//jdk-7u67-linux-x64.tar.gz',
+ self.assertResourceCalled('File', '/tmp/jdk-7u67-linux-x64.tar.gz',
content = DownloadSource('http://c6401.ambari.apache.org:8080/resources//jdk-7u67-linux-x64.tar.gz'),
+ not_if = 'test -f /tmp/jdk-7u67-linux-x64.tar.gz',
)
self.assertResourceCalled('Directory', '/usr/jdk64',)
self.assertResourceCalled('Execute', ('chmod', 'a+x', u'/usr/jdk64'),
not_if = 'test -e /usr/jdk64/jdk1.7.0_45/bin/java',
sudo = True,
)
- self.assertResourceCalled('Execute', 'mkdir -p /tmp/jdk && cd /tmp/jdk && tar -xf /tmp/AMBARI-artifacts//jdk-7u67-linux-x64.tar.gz && ambari-sudo.sh cp -rp /tmp/jdk/* /usr/jdk64',
+ self.assertResourceCalled('Execute', 'mkdir -p /tmp/jdk && cd /tmp/jdk && tar -xf /tmp/jdk-7u67-linux-x64.tar.gz && ambari-sudo.sh cp -rp /tmp/jdk/* /usr/jdk64',
not_if = 'test -e /usr/jdk64/jdk1.7.0_45/bin/java',
)
self.assertResourceCalled('File', '/usr/jdk64/jdk1.7.0_45/bin/java',
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
index fb2699a..0895b7e 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
@@ -176,18 +176,6 @@ class TestHookBeforeStart(RMFTestCase):
not_if = "(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)",
sudo=True,
)
- self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-i386-32',
- recursive = True,
- )
- self.assertResourceCalled('Directory', '/usr/lib/hadoop/lib/native/Linux-amd64-64',
- recursive = True,
- )
- self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so',
- to = '/usr/lib/libsnappy.so',
- )
- self.assertResourceCalled('Link', '/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so',
- to = '/usr/lib64/libsnappy.so',
- )
self.assertResourceCalled('Directory', '/var/log/hadoop',
owner = 'root',
group = 'hadoop',
http://git-wip-us.apache.org/repos/asf/ambari/blob/461b374e/ambari-server/src/test/python/stacks/2.1/YARN/test_apptimelineserver.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.1/YARN/test_apptimelineserver.py b/ambari-server/src/test/python/stacks/2.1/YARN/test_apptimelineserver.py
index 9e8b405..c8a3033 100644
--- a/ambari-server/src/test/python/stacks/2.1/YARN/test_apptimelineserver.py
+++ b/ambari-server/src/test/python/stacks/2.1/YARN/test_apptimelineserver.py
@@ -64,10 +64,12 @@ class TestAppTimelineServer(RMFTestCase):
self.assertResourceCalled('Execute', 'ulimit -c unlimited; export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop-yarn/sbin/yarn-daemon.sh --config /etc/hadoop/conf start timelineserver',
not_if=pid_check_cmd,
user='yarn')
- self.assertResourceCalled('Execute', pid_check_cmd,
- initial_wait=5,
- not_if=pid_check_cmd,
- user='yarn')
+ self.assertResourceCalled('Execute', 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-timelineserver.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-timelineserver.pid` >/dev/null 2>&1',
+ not_if = 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-timelineserver.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop-yarn/yarn/yarn-yarn-timelineserver.pid` >/dev/null 2>&1',
+ tries = 5,
+ user = 'yarn',
+ try_sleep = 1,
+ )
self.assertNoMoreResources()
def test_stop_default(self):