You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by jo...@apache.org on 2015/05/11 21:32:51 UTC
[6/8] ambari git commit: Revert "AMBARI-11034. Implement HdfsResouces
and it's usage (aonishuk)"
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/yarn.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/yarn.py b/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/yarn.py
index 1e75eb0..583e6e9 100644
--- a/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/yarn.py
+++ b/ambari-server/src/main/resources/common-services/YARN/2.1.0.2.0/package/scripts/yarn.py
@@ -61,35 +61,37 @@ def yarn(name = None):
@OsFamilyFuncImpl(os_family=OsFamilyImpl.DEFAULT)
def yarn(name = None):
import params
- if name == "historyserver":
+ if name in ["nodemanager","historyserver"]:
if params.yarn_log_aggregation_enabled:
- params.HdfsResource(params.yarn_nm_app_log_dir,
- action="create_on_execute",
- type="directory",
+ params.HdfsDirectory(params.yarn_nm_app_log_dir,
+ action="create_delayed",
owner=params.yarn_user,
group=params.user_group,
mode=0777,
recursive_chmod=True
)
- params.HdfsResource("/mapred",
- type="directory",
- action="create_on_execute",
+ params.HdfsDirectory("/mapred",
+ action="create_delayed",
owner=params.mapred_user
)
- params.HdfsResource("/mapred/system",
- type="directory",
- action="create_on_execute",
+ params.HdfsDirectory("/mapred/system",
+ action="create_delayed",
owner=params.hdfs_user
)
- params.HdfsResource(params.mapreduce_jobhistory_done_dir,
- type="directory",
- action="create_on_execute",
+ params.HdfsDirectory(params.mapreduce_jobhistory_intermediate_done_dir,
+ action="create_delayed",
owner=params.mapred_user,
group=params.user_group,
- change_permissions_for_parents=True,
mode=0777
)
- params.HdfsResource(None, action="execute")
+
+ params.HdfsDirectory(params.mapreduce_jobhistory_done_dir,
+ action="create_delayed",
+ owner=params.mapred_user,
+ group=params.user_group,
+ mode=01777
+ )
+ params.HdfsDirectory(None, action="create")
if name == "nodemanager":
Directory(params.nm_local_dirs.split(',') + params.nm_log_dirs.split(','),
@@ -176,14 +178,12 @@ def yarn(name = None):
group=params.user_group
)
if not is_empty(params.node_label_enable) and params.node_label_enable or is_empty(params.node_label_enable) and params.node_labels_dir:
- params.HdfsResource(params.node_labels_dir,
- type="directory",
- action="create_on_execute",
+ params.HdfsDirectory(params.node_labels_dir,
+ action="create",
owner=params.yarn_user,
group=params.user_group,
mode=0700
)
- params.HdfsResource(None, action="execute")
elif name == 'apptimelineserver':
Directory(params.ats_leveldb_dir,
owner=params.yarn_user,
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_master.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_master.py b/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_master.py
index cc404eb..b56d15a 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_master.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_master.py
@@ -280,42 +280,36 @@ class TestHBaseMaster(RMFTestCase):
owner='hbase',
content='log4jproperties\nline2'
)
-
- self.assertResourceCalled('HdfsResource', 'hdfs://c6401.ambari.apache.org:8020/apps/hbase/data',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
-
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hbase',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- )
- self.assertResourceCalled('HdfsResource', '/apps/hbase/staging',
- security_enabled = False,
- hadoop_conf_dir = '/etc/hadoop/conf',
- keytab = UnknownConfigurationMock(),
-
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hbase',
- hadoop_bin_dir = '/usr/bin',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0711,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
-
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', 'hdfs://c6401.ambari.apache.org:8020/apps/hbase/data',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/apps/hbase/staging',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ mode = 0711,
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ bin_dir = '/usr/bin',
+ action = ['create'],
+ )
def assert_configure_secured(self):
self.assertResourceCalled('Directory', '/etc/hbase',
@@ -407,41 +401,36 @@ class TestHBaseMaster(RMFTestCase):
owner='hbase',
content='log4jproperties\nline2'
)
- self.assertResourceCalled('HdfsResource', 'hdfs://c6401.ambari.apache.org:8020/apps/hbase/data',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
-
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hbase',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- )
- self.assertResourceCalled('HdfsResource', '/apps/hbase/staging',
- security_enabled = True,
- hadoop_conf_dir = '/etc/hadoop/conf',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
-
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hbase',
- hadoop_bin_dir = '/usr/bin',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0711,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
-
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', 'hdfs://c6401.ambari.apache.org:8020/apps/hbase/data',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/apps/hbase/staging',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0711,
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ bin_dir = '/usr/bin',
+ action = ['create'],
+ )
def test_start_default_22(self):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hbase_master.py",
@@ -539,38 +528,35 @@ class TestHBaseMaster(RMFTestCase):
owner='hbase',
content='log4jproperties\nline2')
- self.assertResourceCalled('HdfsResource', 'hdfs://nn1/apps/hbase/data',
- security_enabled = False,
- hadoop_bin_dir = '/usr/hdp/current/hadoop-client/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hbase',
- hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf',
- type = 'directory',
- action = ['create_on_execute'],
- )
- self.assertResourceCalled('HdfsResource', '/apps/hbase/staging',
- security_enabled = False,
- hadoop_bin_dir = '/usr/hdp/current/hadoop-client/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hbase',
- hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0711,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- hadoop_bin_dir = '/usr/hdp/current/hadoop-client/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', 'hdfs://nn1/apps/hbase/data',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/usr/hdp/current/hadoop-client/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ owner = 'hbase',
+ bin_dir = '/usr/hdp/current/hadoop-client/bin',
+ action = ['create_delayed'])
+
+ self.assertResourceCalled('HdfsDirectory', '/apps/hbase/staging',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/usr/hdp/current/hadoop-client/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ mode = 0711,
+ owner = 'hbase',
+ bin_dir = '/usr/hdp/current/hadoop-client/bin',
+ action = ['create_delayed'])
+
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/usr/hdp/current/hadoop-client/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ bin_dir = '/usr/hdp/current/hadoop-client/bin',
+ action = ['create'])
self.assertResourceCalled('Execute', '/usr/hdp/current/hbase-master/bin/hbase-daemon.sh --config /usr/hdp/current/hbase-master/conf start master',
not_if = 'ls /var/run/hbase/hbase-hbase-master.pid >/dev/null 2>&1 && ps -p `cat /var/run/hbase/hbase-hbase-master.pid` >/dev/null 2>&1',
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_regionserver.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_regionserver.py b/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_regionserver.py
index 3175a01..ea8d9d4 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_regionserver.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HBASE/test_hbase_regionserver.py
@@ -207,6 +207,36 @@ class TestHbaseRegionServer(RMFTestCase):
owner='hbase',
content='log4jproperties\nline2'
)
+ self.assertResourceCalled('HdfsDirectory', 'hdfs://c6401.ambari.apache.org:8020/apps/hbase/data',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/apps/hbase/staging',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0711,
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ bin_dir = '/usr/bin',
+ action = ['create'],
+ )
def assert_configure_secured(self):
self.assertResourceCalled('Directory', '/etc/hbase',
@@ -298,6 +328,36 @@ class TestHbaseRegionServer(RMFTestCase):
owner='hbase',
content='log4jproperties\nline2'
)
+ self.assertResourceCalled('HdfsDirectory', 'hdfs://c6401.ambari.apache.org:8020/apps/hbase/data',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/apps/hbase/staging',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0711,
+ owner = 'hbase',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ bin_dir = '/usr/bin',
+ action = ['create'],
+ )
def test_start_default_22(self):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hbase_regionserver.py",
@@ -393,6 +453,37 @@ class TestHbaseRegionServer(RMFTestCase):
group='hadoop',
owner='hbase',
content='log4jproperties\nline2')
+
+ self.assertResourceCalled('HdfsDirectory', 'hdfs://nn1/apps/hbase/data',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/usr/hdp/current/hadoop-client/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ owner = 'hbase',
+ bin_dir = '/usr/hdp/current/hadoop-client/bin',
+ action = ['create_delayed'])
+
+ self.assertResourceCalled('HdfsDirectory', '/apps/hbase/staging',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/usr/hdp/current/hadoop-client/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ mode = 0711,
+ owner = 'hbase',
+ bin_dir = '/usr/hdp/current/hadoop-client/bin',
+ action = ['create_delayed'])
+
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/usr/hdp/current/hadoop-client/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = "/usr/bin/kinit",
+ bin_dir = '/usr/hdp/current/hadoop-client/bin',
+ action = ['create'])
+
self.assertResourceCalled('Execute', '/usr/hdp/current/hbase-regionserver/bin/hbase-daemon.sh --config /usr/hdp/current/hbase-regionserver/conf start regionserver',
not_if = 'ls /var/run/hbase/hbase-hbase-regionserver.pid >/dev/null 2>&1 && ps -p `cat /var/run/hbase/hbase-hbase-regionserver.pid` >/dev/null 2>&1',
user = 'hbase')
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
index 9b7a91f..8b7846c 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_datanode.py
@@ -380,9 +380,6 @@ class TestDatanode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/var/lib/hadoop-hdfs',
owner = 'hdfs',
group = 'hadoop',
@@ -434,9 +431,6 @@ class TestDatanode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'root',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/var/lib/hadoop-hdfs',
owner = 'hdfs',
group = 'hadoop',
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
index d05e2dc..9975b36 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_journalnode.py
@@ -209,9 +209,6 @@ class TestJournalnode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
def assert_configure_secured(self):
self.assertResourceCalled('Directory', '/grid/0/hdfs/journal',
@@ -250,9 +247,6 @@ class TestJournalnode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'root',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
@patch('time.sleep')
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
index 488e1a2..2e7af32 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py
@@ -100,40 +100,38 @@ class TestNamenode(RMFTestCase):
user = 'hdfs',
try_sleep = 10,
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hdfs',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/ambari-qa',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0770,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- only_if = None,
- keytab = UnknownConfigurationMock(),
- hadoop_bin_dir = '/usr/bin',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', '/tmp',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0777,
+ owner = 'hdfs',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/ambari-qa',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0770,
+ owner = 'ambari-qa',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ action = ['create'],
+ bin_dir = '/usr/bin',
+ only_if = None,
+ )
self.assertNoMoreResources()
pass
@@ -207,40 +205,38 @@ class TestNamenode(RMFTestCase):
user = 'hdfs',
try_sleep = 10,
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hdfs',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/ambari-qa',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0770,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- only_if = None,
- keytab = UnknownConfigurationMock(),
- hadoop_bin_dir = '/usr/bin',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', '/tmp',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0777,
+ owner = 'hdfs',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/ambari-qa',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0770,
+ owner = 'ambari-qa',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ action = ['create'],
+ bin_dir = '/usr/bin',
+ only_if = None,
+ )
self.assertNoMoreResources()
def test_stop_default(self):
@@ -336,40 +332,38 @@ class TestNamenode(RMFTestCase):
user = 'hdfs',
try_sleep = 10,
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hdfs',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/ambari-qa',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0770,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = True,
- only_if = None,
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- hadoop_bin_dir = '/usr/bin',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', '/tmp',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0777,
+ owner = 'hdfs',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/ambari-qa',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0770,
+ owner = 'ambari-qa',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ action = ['create'],
+ bin_dir = '/usr/bin',
+ only_if = None,
+ )
self.assertNoMoreResources()
def test_stop_secured(self):
@@ -435,39 +429,37 @@ class TestNamenode(RMFTestCase):
user = 'hdfs',
try_sleep = 10,
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hdfs',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/ambari-qa',
+ self.assertResourceCalled('HdfsDirectory', '/tmp',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0777,
+ owner = 'hdfs',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/ambari-qa',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0770,
+ owner = 'ambari-qa',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0770,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
+ action = ['create'],
+ bin_dir = '/usr/bin',
only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
- keytab = UnknownConfigurationMock(),
- hadoop_bin_dir = '/usr/bin',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
)
self.assertNoMoreResources()
@@ -516,39 +508,37 @@ class TestNamenode(RMFTestCase):
user = 'hdfs',
try_sleep = 10,
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hdfs',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/ambari-qa',
+ self.assertResourceCalled('HdfsDirectory', '/tmp',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0777,
+ owner = 'hdfs',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/ambari-qa',
+ security_enabled = True,
+ keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0770,
+ owner = 'ambari-qa',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
keytab = '/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0770,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = True,
+ action = ['create'],
+ bin_dir = '/usr/bin',
only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- hadoop_bin_dir = '/usr/bin',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
)
self.assertNoMoreResources()
@@ -606,40 +596,38 @@ class TestNamenode(RMFTestCase):
user = 'hdfs',
try_sleep = 10,
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hdfs',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/ambari-qa',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0770,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
- keytab = UnknownConfigurationMock(),
- hadoop_bin_dir = '/usr/bin',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', '/tmp',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0777,
+ owner = 'hdfs',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/ambari-qa',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0770,
+ owner = 'ambari-qa',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ action = ['create'],
+ bin_dir = '/usr/bin',
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
+ )
self.assertNoMoreResources()
# tests namenode start command when NameNode HA is enabled, and
@@ -693,40 +681,38 @@ class TestNamenode(RMFTestCase):
user = 'hdfs',
try_sleep = 10,
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hdfs',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/ambari-qa',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0770,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn2 | grep active'",
- keytab = UnknownConfigurationMock(),
- hadoop_bin_dir = '/usr/bin',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
+ self.assertResourceCalled('HdfsDirectory', '/tmp',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0777,
+ owner = 'hdfs',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/ambari-qa',
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ mode = 0770,
+ owner = 'ambari-qa',
+ bin_dir = '/usr/bin',
+ action = ['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled = False,
+ keytab = UnknownConfigurationMock(),
+ conf_dir = '/etc/hadoop/conf',
+ hdfs_user = 'hdfs',
+ kinit_path_local = '/usr/bin/kinit',
+ action = ['create'],
+ bin_dir = '/usr/bin',
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn2 | grep active'",
+ )
self.assertNoMoreResources()
def test_decommission_default(self):
@@ -843,9 +829,6 @@ class TestNamenode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode',
owner = 'hdfs',
group = 'hadoop',
@@ -885,9 +868,6 @@ class TestNamenode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'root',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode',
owner = 'hdfs',
group = 'hadoop',
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
index 9d6d165..d683198 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_nfsgateway.py
@@ -221,9 +221,6 @@ class TestNFSGateway(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
def assert_configure_secured(self):
self.assertResourceCalled('Directory', '/etc/security/limits.d',
@@ -256,9 +253,6 @@ class TestNFSGateway(RMFTestCase):
content = Template('slaves.j2'),
owner = 'root',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
@patch("resource_management.libraries.functions.security_commons.build_expectations")
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_service_check.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_service_check.py
index 36abd40..3c098b6 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_service_check.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_service_check.py
@@ -59,45 +59,36 @@ class TestServiceCheck(RMFTestCase):
bin_dir = '/usr/bin',
user = 'hdfs',
)
- self.assertResourceCalled('HdfsResource', '/tmp',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
+ self.assertResourceCalled('ExecuteHadoop', 'fs -mkdir /tmp',
+ conf_dir = '/etc/hadoop/conf',
+ logoutput = True,
+ not_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]/usr/bin/hadoop --config /etc/hadoop/conf fs -test -e /tmp'",
+ try_sleep = 3,
+ tries = 5,
+ bin_dir = '/usr/bin',
user = 'hdfs',
- action = ['create_on_execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- mode = 0777,
)
- self.assertResourceCalled('HdfsResource', '/tmp/',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
+ self.assertResourceCalled('ExecuteHadoop', 'fs -chmod 777 /tmp',
+ conf_dir = '/etc/hadoop/conf',
+ logoutput = True,
+ try_sleep = 3,
+ tries = 5,
+ bin_dir = '/usr/bin',
user = 'hdfs',
- action = ['delete_on_execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'file',
)
- self.assertResourceCalled('HdfsResource', '/tmp/',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- source = '/etc/passwd',
+ self.assertResourceCalled('ExecuteHadoop', 'fs -rm /tmp/; hadoop --config /etc/hadoop/conf fs -put /etc/passwd /tmp/',
+ logoutput = True,
+ tries = 5,
+ conf_dir = '/etc/hadoop/conf',
+ bin_dir = '/usr/bin',
+ try_sleep = 3,
user = 'hdfs',
- action = ['create_on_execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'file',
)
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
+ self.assertResourceCalled('ExecuteHadoop', 'fs -test -e /tmp/',
+ logoutput = True,
+ tries = 5,
+ conf_dir = '/etc/hadoop/conf',
+ bin_dir = '/usr/bin',
+ try_sleep = 3,
user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
)
- self.assertNoMoreResources()
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
index 8630fcf..5bedf5b 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_snamenode.py
@@ -230,9 +230,6 @@ class TestSNamenode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/hadoop/hdfs/namesecondary',
owner = 'hdfs',
group = 'hadoop',
@@ -279,9 +276,6 @@ class TestSNamenode(RMFTestCase):
content = Template('slaves.j2'),
owner = 'root',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/hadoop/hdfs/namesecondary',
owner = 'hdfs',
group = 'hadoop',
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
index e34258b..8e7f34f 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_zkfc.py
@@ -64,9 +64,6 @@ class TestZkfc(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/var/run/hadoop',
owner = 'hdfs',
group = 'hadoop',
@@ -158,9 +155,6 @@ class TestZkfc(RMFTestCase):
content = Template('slaves.j2'),
owner = 'root',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/var/run/hadoop',
owner = 'hdfs',
group = 'hadoop',
@@ -251,9 +245,6 @@ class TestZkfc(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/var/run/hadoop',
owner = 'hdfs',
group = 'hadoop',
@@ -316,9 +307,6 @@ class TestZkfc(RMFTestCase):
content = Template('slaves.j2'),
owner = 'hdfs',
)
- self.assertResourceCalled('File', '/var/lib/ambari-agent/lib/fast-hdfs-resource.jar',
- content = StaticFile('fast-hdfs-resource.jar'),
- )
self.assertResourceCalled('Directory', '/var/run/hadoop',
owner = 'hdfs',
group = 'hadoop',
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
index 9a98a85..670092d 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_server.py
@@ -22,19 +22,16 @@ import socket
import subprocess
from mock.mock import MagicMock, patch
-from resource_management.libraries.functions import version
from resource_management.core import shell
-from resource_management.libraries.script.script import Script
+from resource_management.libraries.functions import dynamic_variable_interpretation
from stacks.utils.RMFTestCase import *
-
-@patch.object(version, "get_hdp_build_version", new = MagicMock(return_value="2.0.0.0-1234"))
@patch("resource_management.libraries.functions.check_thrift_port_sasl", new=MagicMock())
class TestHiveServer(RMFTestCase):
COMMON_SERVICES_PACKAGE_DIR = "HIVE/0.12.0.2.0/package"
STACK_VERSION = "2.0.6"
UPGRADE_STACK_VERSION = "2.2"
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
+
def test_configure_default(self):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer",
@@ -47,7 +44,7 @@ class TestHiveServer(RMFTestCase):
self.assertNoMoreResources()
@patch("socket.socket")
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
+ @patch.object(dynamic_variable_interpretation, "copy_tarballs_to_hdfs", new=MagicMock())
def test_start_default(self, socket_mock):
s = socket_mock.return_value
@@ -86,8 +83,9 @@ class TestHiveServer(RMFTestCase):
)
self.assertNoMoreResources()
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
- def test_start_default_no_copy(self):
+
+ @patch.object(dynamic_variable_interpretation, "_get_tar_source_and_dest_folder")
+ def test_start_default_no_copy(self, get_tar_mock):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer",
@@ -97,6 +95,7 @@ class TestHiveServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES
)
+ get_tar_mock.return_value = ("a", "b")
self.assert_configure_default()
self.assertResourceCalled('Execute', 'hive --config /etc/hive/conf.server --service metatool -updateLocation hdfs://c6401.ambari.apache.org:8020 OK.',
@@ -119,9 +118,10 @@ class TestHiveServer(RMFTestCase):
timeout = 30,
)
self.assertNoMoreResources()
+ self.assertFalse(get_tar_mock.called)
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
- def test_start_default_alt_tmp(self):
+ @patch.object(dynamic_variable_interpretation, "_get_tar_source_and_dest_folder")
+ def test_start_default_alt_tmp(self, get_tar_mock):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer",
command = "start",
@@ -130,6 +130,7 @@ class TestHiveServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES
)
+ get_tar_mock.return_value = ("a", "b")
self.assert_configure_default(no_tmp=True)
self.assertResourceCalled('Execute', 'hive --config /etc/hive/conf.server --service metatool -updateLocation hdfs://c6401.ambari.apache.org:8020 OK.',
@@ -152,10 +153,11 @@ class TestHiveServer(RMFTestCase):
timeout = 30,
)
self.assertNoMoreResources()
+ self.assertFalse(get_tar_mock.called)
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
- def test_start_default_alt_nn_ha_tmp(self):
+ @patch.object(dynamic_variable_interpretation, "_get_tar_source_and_dest_folder")
+ def test_start_default_alt_nn_ha_tmp(self, get_tar_mock):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer",
command = "start",
@@ -164,6 +166,7 @@ class TestHiveServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES
)
+ get_tar_mock.return_value = ("a", "b")
self.assert_configure_default(no_tmp=True)
self.assertResourceCalled('Execute', 'hive --config /etc/hive/conf.server --service metatool -updateLocation hdfs://c6401.ambari.apache.org:8020 OK.',
@@ -186,8 +189,9 @@ class TestHiveServer(RMFTestCase):
timeout = 30,
)
self.assertNoMoreResources()
+ self.assertFalse(get_tar_mock.called)
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
+ @patch.object(dynamic_variable_interpretation, "copy_tarballs_to_hdfs", new=MagicMock())
def test_stop_default(self):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer",
@@ -213,7 +217,7 @@ class TestHiveServer(RMFTestCase):
self.assertNoMoreResources()
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
+
def test_configure_secured(self):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer",
@@ -227,7 +231,6 @@ class TestHiveServer(RMFTestCase):
@patch("hive_service.check_fs_root")
@patch("socket.socket")
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
def test_start_secured(self, socket_mock, check_fs_root_mock):
s = socket_mock.return_value
@@ -273,7 +276,6 @@ class TestHiveServer(RMFTestCase):
@patch("socket.socket")
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=False))
def test_stop_secured(self, socket_mock):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/hive_server.py",
classname = "HiveServer",
@@ -300,89 +302,51 @@ class TestHiveServer(RMFTestCase):
self.assertNoMoreResources()
def assert_configure_default(self, no_tmp = False):
- self.assertResourceCalled('HdfsResource', '/apps/webhcat',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hcat',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0755,
- )
- self.assertResourceCalled('HdfsResource', '/user/hcat',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hcat',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0755,
- )
- self.assertResourceCalled('HdfsResource', '/apps/webhcat/hive.tar.gz',
- security_enabled = False,
- hadoop_conf_dir = '/etc/hadoop/conf',
- keytab = UnknownConfigurationMock(),
- source = '/usr/share/HDP-webhcat/hive.tar.gz',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['create_on_execute'],
- group = 'hadoop',
- hadoop_bin_dir = '/usr/bin',
- type = 'file',
- mode = 0755,
- )
- self.assertResourceCalled('HdfsResource', '/apps/hive/warehouse',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hive',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/hive',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hive',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0700,
+ self.assertResourceCalled('HdfsDirectory', '/apps/hive/warehouse',
+ security_enabled=False,
+ keytab=UnknownConfigurationMock(),
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ mode=0777,
+ owner='hive',
+ bin_dir='/usr/bin',
+ action=['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/hive',
+ security_enabled=False,
+ keytab=UnknownConfigurationMock(),
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ mode=0700,
+ owner='hive',
+ bin_dir='/usr/bin',
+ action=['create_delayed'],
)
+
if not no_tmp:
- self.assertResourceCalled('HdfsResource', '/custompath/tmp/hive',
- security_enabled = False,
- hadoop_conf_dir = '/etc/hadoop/conf',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hive',
- group = 'hdfs',
- hadoop_bin_dir = '/usr/bin',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
+ self.assertResourceCalled('HdfsDirectory', '/custompath/tmp/hive',
+ security_enabled=False,
+ keytab=UnknownConfigurationMock(),
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ mode=0777,
+ owner='hive',
+ group='hdfs',
+ action=['create_delayed'],
+ bin_dir='/usr/bin',
)
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
+
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled=False,
+ keytab=UnknownConfigurationMock(),
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ action=['create'],
+ bin_dir='/usr/bin',
)
self.assertResourceCalled('Directory', '/etc/hive',
mode=0755,
@@ -489,88 +453,49 @@ class TestHiveServer(RMFTestCase):
def assert_configure_secured(self):
- self.assertResourceCalled('HdfsResource', '/apps/webhcat',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hcat',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0755,
- )
- self.assertResourceCalled('HdfsResource', '/user/hcat',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hcat',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0755,
- )
- self.assertResourceCalled('HdfsResource', '/apps/webhcat/hive.tar.gz',
- security_enabled = True,
- hadoop_conf_dir = '/etc/hadoop/conf',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- source = '/usr/share/HDP-webhcat/hive.tar.gz',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['create_on_execute'],
- group = 'hadoop',
- hadoop_bin_dir = '/usr/bin',
- type = 'file',
- mode = 0755,
- )
- self.assertResourceCalled('HdfsResource', '/apps/hive/warehouse',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hive',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', '/user/hive',
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hive',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0700,
- )
- self.assertResourceCalled('HdfsResource', '/custompath/tmp/hive',
- security_enabled = True,
- hadoop_conf_dir = '/etc/hadoop/conf',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- owner = 'hive',
- group = 'hdfs',
- hadoop_bin_dir = '/usr/bin',
- type = 'directory',
- action = ['create_on_execute'],
- mode = 0777,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
+ self.assertResourceCalled('HdfsDirectory', '/apps/hive/warehouse',
+ security_enabled=True,
+ keytab='/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ mode=0777,
+ owner='hive',
+ bin_dir='/usr/bin',
+ action=['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/user/hive',
+ security_enabled=True,
+ keytab='/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ mode=0700,
+ owner='hive',
+ bin_dir='/usr/bin',
+ action=['create_delayed'],
+ )
+ self.assertResourceCalled('HdfsDirectory', '/custompath/tmp/hive',
+ security_enabled=True,
+ keytab='/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ mode=0777,
+ owner='hive',
+ group='hdfs',
+ action=['create_delayed'],
+ bin_dir='/usr/bin',
+ )
+
+ self.assertResourceCalled('HdfsDirectory', None,
+ security_enabled=True,
+ keytab='/etc/security/keytabs/hdfs.headless.keytab',
+ conf_dir='/etc/hadoop/conf',
+ hdfs_user='hdfs',
+ kinit_path_local='/usr/bin/kinit',
+ action=['create'],
+ bin_dir='/usr/bin',
)
self.assertResourceCalled('Directory', '/etc/hive',
mode=0755,
@@ -699,7 +624,6 @@ class TestHiveServer(RMFTestCase):
self.assert_configure_default()
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=True))
@patch("hive_server.HiveServer.pre_rolling_restart")
@patch("hive_server.HiveServer.start")
def test_stop_during_upgrade(self, hive_server_start_mock,
@@ -847,7 +771,6 @@ class TestHiveServer(RMFTestCase):
)
put_structured_out_mock.assert_called_with({"securityState": "UNSECURED"})
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=True))
def test_pre_rolling_restart(self):
config_file = self.get_src_folder()+"/test/python/stacks/2.0.6/configs/default.json"
with open(config_file, "r") as f:
@@ -862,32 +785,9 @@ class TestHiveServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES)
self.assertResourceCalled('Execute',
'hdp-select set hive-server2 %s' % version,)
- self.assertResourceCalled('HdfsResource', 'hdfs:///hdp/apps/2.0.0.0-1234/mapreduce//mapreduce.tar.gz',
- security_enabled = False,
- hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf',
- keytab = UnknownConfigurationMock(),
- source = '/usr/hdp/current/hadoop-client/mapreduce.tar.gz',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['create_on_execute'],
- group = 'hadoop',
- hadoop_bin_dir = '/usr/hdp/current/hadoop-client/bin',
- type = 'file',
- mode = 0444,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- hadoop_bin_dir = '/usr/hdp/current/hadoop-client/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf',
- )
self.assertNoMoreResources()
@patch("resource_management.core.shell.call")
- @patch.object(Script, "is_hdp_stack_greater_or_equal", new = MagicMock(return_value=True))
def test_pre_rolling_restart_23(self, call_mock):
config_file = self.get_src_folder()+"/test/python/stacks/2.0.6/configs/default.json"
with open(config_file, "r") as f:
@@ -907,28 +807,6 @@ class TestHiveServer(RMFTestCase):
self.assertResourceCalled('Execute',
'hdp-select set hive-server2 %s' % version,)
- self.assertResourceCalled('HdfsResource', 'hdfs:///hdp/apps/2.0.0.0-1234/mapreduce//mapreduce.tar.gz',
- security_enabled = False,
- hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf',
- keytab = UnknownConfigurationMock(),
- source = '/usr/hdp/current/hadoop-client/mapreduce.tar.gz',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['create_on_execute'],
- group = 'hadoop',
- hadoop_bin_dir = '/usr/hdp/current/hadoop-client/bin',
- type = 'file',
- mode = 0444,
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- hadoop_bin_dir = '/usr/hdp/current/hadoop-client/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/usr/hdp/current/hadoop-client/conf',
- )
self.assertNoMoreResources()
self.assertEquals(2, mocks_dict['call'].call_count)
http://git-wip-us.apache.org/repos/asf/ambari/blob/e833066e/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
index 86d348f..7bbd0a9 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
@@ -24,13 +24,12 @@ import datetime, sys, socket
import resource_management.libraries.functions
@patch.object(resource_management.libraries.functions, "get_unique_id_and_date", new = MagicMock(return_value=''))
@patch("socket.socket")
-@patch("time.time", new=MagicMock(return_value=1431110511.43))
class TestServiceCheck(RMFTestCase):
COMMON_SERVICES_PACKAGE_DIR = "HIVE/0.12.0.2.0/package"
STACK_VERSION = "2.0.6"
-
- def test_service_check_default(self, socket_mock):
+ @patch("sys.exit")
+ def test_service_check_default(self, sys_exit_mock, socket_mock):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/service_check.py",
classname="HiveServiceCheck",
@@ -83,52 +82,16 @@ class TestServiceCheck(RMFTestCase):
content = StaticFile('templetonSmoke.sh'),
mode = 0755,
)
- self.assertResourceCalled('File', '/tmp/idtest.ambari-qa.1431110511.43.pig',
- content = Template('templeton_smoke.pig.j2', templeton_test_input='/tmp/idtest.ambari-qa.1431110511.43.in', templeton_test_output='/tmp/idtest.ambari-qa.1431110511.43.out'),
- )
- self.assertResourceCalled('HdfsResource', '/tmp/idtest.ambari-qa.1431110511.43.pig',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- source = '/tmp/idtest.ambari-qa.1431110511.43.pig',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'file',
- action = ['create_on_execute'],
- )
- self.assertResourceCalled('HdfsResource', '/tmp/idtest.ambari-qa.1431110511.43.in',
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- source = '/etc/passwd',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'file',
- action = ['create_on_execute'],
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = False,
- hadoop_bin_dir = '/usr/bin',
- keytab = UnknownConfigurationMock(),
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
- self.assertResourceCalled('Execute', '/tmp/templetonSmoke.sh c6402.ambari.apache.org ambari-qa 50111 idtest.ambari-qa.1431110511.43.pig no_keytab false /usr/bin/kinit no_principal',
- logoutput = True,
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
- tries = 3,
- try_sleep = 5,
- )
+ self.assertResourceCalled('Execute', '/tmp/templetonSmoke.sh c6402.ambari.apache.org ambari-qa 50111 no_keytab false /usr/bin/kinit no_principal',
+ logoutput = True,
+ path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
+ tries = 3,
+ try_sleep = 5,
+ )
self.assertNoMoreResources()
-
- def test_service_check_secured(self, socket_mock):
+ @patch("sys.exit")
+ def test_service_check_secured(self, sys_exit_mock, socket_mock):
self.executeScript(self.COMMON_SERVICES_PACKAGE_DIR + "/scripts/service_check.py",
classname="HiveServiceCheck",
@@ -182,47 +145,10 @@ class TestServiceCheck(RMFTestCase):
content = StaticFile('templetonSmoke.sh'),
mode = 0755,
)
-
- self.assertResourceCalled('File', '/tmp/idtest.ambari-qa.1431110511.43.pig',
- content = Template('templeton_smoke.pig.j2', templeton_test_input='/tmp/idtest.ambari-qa.1431110511.43.in', templeton_test_output='/tmp/idtest.ambari-qa.1431110511.43.out'),
- )
- self.assertResourceCalled('HdfsResource', '/tmp/idtest.ambari-qa.1431110511.43.pig',
- action = ['create_on_execute'],
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- source = '/tmp/idtest.ambari-qa.1431110511.43.pig',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'file',
- )
- self.assertResourceCalled('HdfsResource', '/tmp/idtest.ambari-qa.1431110511.43.in',
- action = ['create_on_execute'],
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- source = '/etc/passwd',
- user = 'hdfs',
- owner = 'ambari-qa',
- hadoop_conf_dir = '/etc/hadoop/conf',
- type = 'file',
- )
- self.assertResourceCalled('HdfsResource', None,
- security_enabled = True,
- hadoop_bin_dir = '/usr/bin',
- keytab = '/etc/security/keytabs/hdfs.headless.keytab',
- kinit_path_local = '/usr/bin/kinit',
- user = 'hdfs',
- action = ['execute'],
- hadoop_conf_dir = '/etc/hadoop/conf',
- )
- self.assertResourceCalled('Execute', '/tmp/templetonSmoke.sh c6402.ambari.apache.org ambari-qa 50111 idtest.ambari-qa.1431110511.43.pig /etc/security/keytabs/smokeuser.headless.keytab true /usr/bin/kinit ambari-qa@EXAMPLE.COM',
- logoutput = True,
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
- tries = 3,
- try_sleep = 5,
- )
+ self.assertResourceCalled('Execute', '/tmp/templetonSmoke.sh c6402.ambari.apache.org ambari-qa 50111 /etc/security/keytabs/smokeuser.headless.keytab true /usr/bin/kinit ambari-qa@EXAMPLE.COM',
+ logoutput = True,
+ path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
+ tries = 3,
+ try_sleep = 5,
+ )
self.assertNoMoreResources()
\ No newline at end of file