You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by ao...@apache.org on 2014/01/31 20:50:29 UTC

[04/51] [partial] AMBARI-4491. Move all the supported versions in Baikal for stack to python code (remove dependence on puppet). (aonishuk)

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/GANGLIA/test_ganglia_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/GANGLIA/test_ganglia_server.py b/ambari-server/src/test/python/stacks/1.3.3/GANGLIA/test_ganglia_server.py
deleted file mode 100644
index 3314f10..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/GANGLIA/test_ganglia_server.py
+++ /dev/null
@@ -1,226 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-
-from stacks.utils.RMFTestCase import *
-
-
-class TestGangliaServer(RMFTestCase):
-
-  def test_configure_default(self):
-    self.executeScript("1.3.3/services/GANGLIA/package/scripts/ganglia_server.py",
-                       classname="GangliaServer",
-                       command="configure",
-                       config_file="default.json"
-    )
-    self.assertResourceCalled('Group', 'hadoop',
-                              )
-    self.assertResourceCalled('Group', 'nobody',
-                              )
-    self.assertResourceCalled('Group', 'nobody',
-                              )
-    self.assertResourceCalled('User', 'nobody',
-                              groups = [u'nobody'],
-                              )
-    self.assertResourceCalled('User', 'nobody',
-                              groups = [u'nobody'],
-                              )
-    self.assertResourceCalled('Directory', '/usr/libexec/hdp/ganglia',
-                              owner = 'root',
-                              group = 'root',
-                              recursive = True,
-                              )
-    self.assertResourceCalled('File', '/etc/init.d/hdp-gmetad',
-                              content = StaticFile('gmetad.init'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/etc/init.d/hdp-gmond',
-                              content = StaticFile('gmond.init'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/checkGmond.sh',
-                              content = StaticFile('checkGmond.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/checkRrdcached.sh',
-                              content = StaticFile('checkRrdcached.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/gmetadLib.sh',
-                              content = StaticFile('gmetadLib.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/gmondLib.sh',
-                              content = StaticFile('gmondLib.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/rrdcachedLib.sh',
-                              content = StaticFile('rrdcachedLib.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/setupGanglia.sh',
-                              content = StaticFile('setupGanglia.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/startGmetad.sh',
-                              content = StaticFile('startGmetad.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/startGmond.sh',
-                              content = StaticFile('startGmond.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/startRrdcached.sh',
-                              content = StaticFile('startRrdcached.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/stopGmetad.sh',
-                              content = StaticFile('stopGmetad.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/stopGmond.sh',
-                              content = StaticFile('stopGmond.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/stopRrdcached.sh',
-                              content = StaticFile('stopRrdcached.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/usr/libexec/hdp/ganglia/teardownGanglia.sh',
-                              content = StaticFile('teardownGanglia.sh'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('TemplateConfig', '/usr/libexec/hdp/ganglia/gangliaClusters.conf',
-                              owner = 'root',
-                              template_tag = None,
-                              group = 'root',
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('TemplateConfig', '/usr/libexec/hdp/ganglia/gangliaEnv.sh',
-                              owner = 'root',
-                              template_tag = None,
-                              group = 'root',
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('TemplateConfig', '/usr/libexec/hdp/ganglia/gangliaLib.sh',
-                              owner = 'root',
-                              template_tag = None,
-                              group = 'root',
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPNameNode -m -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPJobTracker -m -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseMaster -m -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHistoryServer -m -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPDataNode -m -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPTaskTracker -m -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseRegionServer -m -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -t -o root -g hadoop',
-                              path = ['/usr/libexec/hdp/ganglia',
-                                      '/usr/sbin',
-                                      '/sbin:/usr/local/bin',
-                                      '/bin',
-                                      '/usr/bin'],
-                              )
-    self.assertResourceCalled('Directory', '/var/lib/ganglia/dwoo',
-                              owner = 'nobody',
-                              recursive = True,
-                              mode = 0777,
-                              )
-    self.assertResourceCalled('Directory', '/srv/www/cgi-bin',
-                              recursive = True,
-                              )
-    self.assertResourceCalled('File', '/srv/www/cgi-bin/rrd.py',
-                              content = StaticFile('rrd.py'),
-                              mode = 0755,
-                              )
-    self.assertResourceCalled('File', '/etc/ganglia/gmetad.conf',
-                              owner = 'root',
-                              group = 'hadoop',
-                              )
-    self.assertNoMoreResources()
-
-  def test_start_default(self):
-    self.executeScript("1.3.3/services/GANGLIA/package/scripts/ganglia_server.py",
-                       classname="GangliaServer",
-                       command="start",
-                       config_file="default.json"
-    )
-    self.assertResourceCalled('Execute', 'service hdp-gmetad start >> /tmp/gmetad.log  2>&1 ; /bin/ps auwx | /bin/grep [g]metad  >> /tmp/gmetad.log  2>&1',
-                              path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                              )
-    self.assertResourceCalled('MonitorWebserver', 'restart',
-                              )
-    self.assertNoMoreResources()
-
-  def test_stop_default(self):
-    self.executeScript("1.3.3/services/GANGLIA/package/scripts/ganglia_server.py",
-                       classname="GangliaServer",
-                       command="stop",
-                       config_file="default.json"
-    )
-    self.assertResourceCalled('Execute', 'service hdp-gmetad stop >> /tmp/gmetad.log  2>&1 ; /bin/ps auwx | /bin/grep [g]metad  >> /tmp/gmetad.log  2>&1',
-                              path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                              )
-    self.assertResourceCalled('MonitorWebserver', 'restart',
-                              )
-    self.assertNoMoreResources()
-

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_client.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_client.py b/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_client.py
deleted file mode 100644
index 7845d97..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_client.py
+++ /dev/null
@@ -1,115 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHBaseClient(RMFTestCase):
-  
-  def test_configure_secured(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_client.py",
-                   classname = "HbaseClient",
-                   command = "configure",
-                   config_file="secured.json"
-    )
-    
-    self.assertResourceCalled('Directory', '/etc/hbase/conf',
-      owner = 'hbase',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hbase-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hbase-site'], # don't hardcode all the properties
-    )
-    self.assertResourceCalled('XmlConfig', 'hdfs-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hdfs-site'], # don't hardcode all the properties
-    )
-    self.assertResourceCalled('File', '/etc/hbase/conf/hbase-policy.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase-env.sh',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hadoop-metrics.properties',
-      owner = 'hbase',
-      template_tag = 'GANGLIA-RS',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/regionservers',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase_client_jaas.conf',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertNoMoreResources()
-    
-  def test_configure_default(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_client.py",
-                   classname = "HbaseClient",
-                   command = "configure",
-                   config_file="default.json"
-    )
-    
-    self.assertResourceCalled('Directory', '/etc/hbase/conf',
-      owner = 'hbase',
-      group = 'hadoop',
-      recursive = True,
-    )    
-    self.assertResourceCalled('XmlConfig', 'hbase-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hbase-site'], # don't hardcode all the properties
-    )    
-    self.assertResourceCalled('XmlConfig', 'hdfs-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hdfs-site'], # don't hardcode all the properties
-    )    
-    self.assertResourceCalled('File', '/etc/hbase/conf/hbase-policy.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-    )    
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase-env.sh',
-      owner = 'hbase',
-      template_tag = None,
-    )    
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hadoop-metrics.properties',
-      owner = 'hbase',
-      template_tag = 'GANGLIA-RS',
-    )    
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/regionservers',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertNoMoreResources()
-    
-
-    
-

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_master.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_master.py b/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_master.py
deleted file mode 100644
index 7e6d67b..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_master.py
+++ /dev/null
@@ -1,224 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHBaseMaster(RMFTestCase):
-  def test_configure_default(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_master.py",
-                   classname = "HbaseMaster",
-                   command = "configure",
-                   config_file="default.json"
-    )
-    
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-    
-  def test_start_default(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_master.py",
-                   classname = "HbaseMaster",
-                   command = "start",
-                   config_file="default.json"
-    )
-    
-    self.assert_configure_default()
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf start master',
-      not_if = 'ls /var/run/hbase/hbase-hbase-master.pid >/dev/null 2>&1 && ps `cat /var/run/hbase/hbase-hbase-master.pid` >/dev/null 2>&1',
-      user = 'hbase'
-    )
-    self.assertNoMoreResources()
-    
-  def test_stop_default(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_master.py",
-                   classname = "HbaseMaster",
-                   command = "stop",
-                   config_file="default.json"
-    )
-    
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf stop master && rm -f /var/run/hbase/hbase-hbase-master.pid',
-      not_if = None,
-      user = 'hbase',
-    )
-    self.assertNoMoreResources()
-
-  def test_decom_default(self):
-    self.executeScript("2.1.1/services/HBASE/package/scripts/hbase_master.py",
-                       classname = "HbaseMaster",
-                       command = "decommission",
-                       config_file="default.json"
-    )
-
-    self.assertResourceCalled('Execute', ' /usr/lib/hbase/bin/hbase --config /etc/hbase/conf org.jruby.Main /usr/lib/hbase/bin/region_mover.rb unload host1',
-                              logoutput = True,
-                              user = 'hbase',
-                              )
-    self.assertResourceCalled('Execute', ' /usr/lib/hbase/bin/hbase --config /etc/hbase/conf org.jruby.Main /usr/lib/hbase/bin/region_mover.rb unload host2',
-                              logoutput = True,
-                              user = 'hbase',
-                              )
-    self.assertNoMoreResources()
-
-  def test_configure_secured(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_master.py",
-                   classname = "HbaseMaster",
-                   command = "configure",
-                   config_file="secured.json"
-    )
-    
-    self.assert_configure_secured()
-    self.assertNoMoreResources()
-    
-  def test_start_secured(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_master.py",
-                   classname = "HbaseMaster",
-                   command = "start",
-                   config_file="secured.json"
-    )
-    
-    self.assert_configure_secured()
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf start master',
-      not_if = 'ls /var/run/hbase/hbase-hbase-master.pid >/dev/null 2>&1 && ps `cat /var/run/hbase/hbase-hbase-master.pid` >/dev/null 2>&1',
-      user = 'hbase',
-    )
-    self.assertNoMoreResources()
-    
-  def test_stop_secured(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_master.py",
-                   classname = "HbaseMaster",
-                   command = "stop",
-                   config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf stop master && rm -f /var/run/hbase/hbase-hbase-master.pid',
-      not_if = None,
-      user = 'hbase',
-    )
-    self.assertNoMoreResources()
-
-  def test_decom_secure(self):
-    self.executeScript("2.1.1/services/HBASE/package/scripts/hbase_master.py",
-                       classname = "HbaseMaster",
-                       command = "decommission",
-                       config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', '/usr/bin/kinit -kt /etc/security/keytabs/hbase.headless.keytab hbase; /usr/lib/hbase/bin/hbase --config /etc/hbase/conf org.jruby.Main /usr/lib/hbase/bin/region_mover.rb unload host1',
-                              logoutput = True,
-                              user = 'hbase',
-                              )
-    self.assertNoMoreResources()
-
-  def assert_configure_default(self):
-    self.assertResourceCalled('Directory', '/etc/hbase/conf',
-      owner = 'hbase',
-      group = 'hadoop',
-      recursive = True,
-    )   
-    self.assertResourceCalled('XmlConfig', 'hbase-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hbase-site'], # don't hardcode all the properties
-    )
-    self.assertResourceCalled('XmlConfig', 'hdfs-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hdfs-site'], # don't hardcode all the properties
-    )  
-    self.assertResourceCalled('File', '/etc/hbase/conf/hbase-policy.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase-env.sh',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hadoop-metrics.properties',
-      owner = 'hbase',
-      template_tag = 'GANGLIA-MASTER',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/regionservers',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )   
-    self.assertResourceCalled('Directory', '/hadoop/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )    
-    self.assertResourceCalled('Directory', '/var/log/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
-  
-  def assert_configure_secured(self):
-    self.assertResourceCalled('Directory', '/etc/hbase/conf',
-      owner = 'hbase',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hbase-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hbase-site'], # don't hardcode all the properties
-    )
-    self.assertResourceCalled('XmlConfig', 'hdfs-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hdfs-site'], # don't hardcode all the properties
-    )
-    self.assertResourceCalled('File', '/etc/hbase/conf/hbase-policy.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase-env.sh',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hadoop-metrics.properties',
-      owner = 'hbase',
-      template_tag = 'GANGLIA-MASTER',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/regionservers',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase_master_jaas.conf',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/hadoop/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_regionserver.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_regionserver.py b/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_regionserver.py
deleted file mode 100644
index 1ad610b..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HBASE/test_hbase_regionserver.py
+++ /dev/null
@@ -1,196 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHbaseRegionServer(RMFTestCase):
-  
-  def test_configure_default(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_regionserver.py",
-                   classname = "HbaseRegionServer",
-                   command = "configure",
-                   config_file="default.json"
-    )
-    
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-    
-  def test_start_default(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_regionserver.py",
-                   classname = "HbaseRegionServer",
-                   command = "start",
-                   config_file="default.json"
-    )
-    
-    self.assert_configure_default()
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf start regionserver',
-      not_if = 'ls /var/run/hbase/hbase-hbase-regionserver.pid >/dev/null 2>&1 && ps `cat /var/run/hbase/hbase-hbase-regionserver.pid` >/dev/null 2>&1',
-      user = 'hbase'
-    )
-    self.assertNoMoreResources()
-    
-  def test_stop_default(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_regionserver.py",
-                   classname = "HbaseRegionServer",
-                   command = "stop",
-                   config_file="default.json"
-    )
-    
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf stop regionserver && rm -f /var/run/hbase/hbase-hbase-regionserver.pid',
-      not_if = None,
-      user = 'hbase',
-    )
-    self.assertNoMoreResources()
-    
-  def test_configure_secured(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_regionserver.py",
-                   classname = "HbaseRegionServer",
-                   command = "configure",
-                   config_file="secured.json"
-    )
-    
-    self.assert_configure_secured()
-    self.assertNoMoreResources()
-    
-  def test_start_secured(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_regionserver.py",
-                   classname = "HbaseRegionServer",
-                   command = "start",
-                   config_file="secured.json"
-    )
-    
-    self.assert_configure_secured()
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf start regionserver',
-      not_if = 'ls /var/run/hbase/hbase-hbase-regionserver.pid >/dev/null 2>&1 && ps `cat /var/run/hbase/hbase-hbase-regionserver.pid` >/dev/null 2>&1',
-      user = 'hbase',
-    )
-    self.assertNoMoreResources()
-    
-  def test_stop_secured(self):
-    self.executeScript("1.3.3/services/HBASE/package/scripts/hbase_regionserver.py",
-                   classname = "HbaseRegionServer",
-                   command = "stop",
-                   config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', '/usr/lib/hbase/bin/hbase-daemon.sh --config /etc/hbase/conf stop regionserver && rm -f /var/run/hbase/hbase-hbase-regionserver.pid',
-      not_if = None,
-      user = 'hbase',
-    )
-    self.assertNoMoreResources()
-
-  def assert_configure_default(self):
-    self.assertResourceCalled('Directory', '/etc/hbase/conf',
-      owner = 'hbase',
-      group = 'hadoop',
-      recursive = True,
-    )   
-    self.assertResourceCalled('XmlConfig', 'hbase-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hbase-site'], # don't hardcode all the properties
-    )
-    self.assertResourceCalled('XmlConfig', 'hdfs-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hdfs-site'], # don't hardcode all the properties
-    )  
-    self.assertResourceCalled('File', '/etc/hbase/conf/hbase-policy.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase-env.sh',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hadoop-metrics.properties',
-      owner = 'hbase',
-      template_tag = 'GANGLIA-RS',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/regionservers',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )   
-    self.assertResourceCalled('Directory', '/hadoop/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )    
-    self.assertResourceCalled('Directory', '/var/log/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
-  
-  
-  def assert_configure_secured(self):
-    self.assertResourceCalled('Directory', '/etc/hbase/conf',
-      owner = 'hbase',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hbase-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hbase-site'], # don't hardcode all the properties
-    )
-    self.assertResourceCalled('XmlConfig', 'hdfs-site.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-      conf_dir = '/etc/hbase/conf',
-      configurations = self.getConfig()['configurations']['hdfs-site'], # don't hardcode all the properties      
-    )
-    self.assertResourceCalled('File', '/etc/hbase/conf/hbase-policy.xml',
-      owner = 'hbase',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase-env.sh',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hadoop-metrics.properties',
-      owner = 'hbase',
-      template_tag = 'GANGLIA-RS',
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/regionservers',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hbase/conf/hbase_regionserver_jaas.conf',
-      owner = 'hbase',
-      template_tag = None,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/hadoop/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hbase',
-      owner = 'hbase',
-      recursive = True,
-    )
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hcat_client.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hcat_client.py b/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hcat_client.py
deleted file mode 100644
index 5a935aa..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hcat_client.py
+++ /dev/null
@@ -1,67 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHcatClient(RMFTestCase):
-
-  def test_configure_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hcat_client.py",
-                       classname = "HCatClient",
-                       command = "configure",
-                       config_file="default.json"
-    )
-
-    self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
-      owner = 'hcat',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('Directory', '/var/run/webhcat',
-      owner = 'hcat',
-      recursive = True,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hcatalog/conf/hcat-env.sh',
-      owner = 'hcat',
-      group = 'hadoop',
-    )
-    self.assertNoMoreResources()
-
-
-
-  def test_configure_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hcat_client.py",
-                         classname = "HCatClient",
-                         command = "configure",
-                         config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
-      owner = 'hcat',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('Directory', '/var/run/webhcat',
-      owner = 'hcat',
-      recursive = True,
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/hcatalog/conf/hcat-env.sh',
-      owner = 'hcat',
-      group = 'hadoop',
-    )
-    self.assertNoMoreResources()
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_client.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_client.py b/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_client.py
deleted file mode 100644
index 56eff70..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_client.py
+++ /dev/null
@@ -1,97 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHiveClient(RMFTestCase):
-
-  def test_configure_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_client.py",
-                       classname = "HiveClient",
-                       command = "configure",
-                       config_file="default.json"
-    )
-    self.assertResourceCalled('Directory', '/etc/hive/conf',
-      owner = 'hive',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hive-site.xml',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 420,
-      conf_dir = '/etc/hive/conf',
-      configurations = self.getConfig()['configurations']['hive-site'],
-    )
-    self.assertResourceCalled('Execute', "/bin/sh -c 'cd /usr/lib/ambari-agent/ && curl -kf --retry 5 http://c6401.ambari.apache.org:8080/resources/DBConnectionVerification.jar -o DBConnectionVerification.jar'",
-      not_if = '[ -f DBConnectionVerification.jar]',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh',
-      content = Template('hive-env.sh.j2', conf_dir="/etc/hive/conf"),
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-default.xml.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertNoMoreResources()
-
-
-
-  def test_configure_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_client.py",
-                       classname = "HiveClient",
-                       command = "configure",
-                       config_file="secured.json"
-    )
-    self.assertResourceCalled('Directory', '/etc/hive/conf',
-      owner = 'hive',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hive-site.xml',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 420,
-      conf_dir = '/etc/hive/conf',
-      configurations = self.getConfig()['configurations']['hive-site'],
-    )
-    self.assertResourceCalled('Execute', "/bin/sh -c 'cd /usr/lib/ambari-agent/ && curl -kf --retry 5 http://c6401.ambari.apache.org:8080/resources/DBConnectionVerification.jar -o DBConnectionVerification.jar'",
-      not_if = '[ -f DBConnectionVerification.jar]',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh',
-      content = Template('hive-env.sh.j2', conf_dir="/etc/hive/conf"),
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-default.xml.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertNoMoreResources()
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_metastore.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_metastore.py b/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_metastore.py
deleted file mode 100644
index d43cf4b..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_metastore.py
+++ /dev/null
@@ -1,214 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHiveMetastore(RMFTestCase):
-
-  def test_configure_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_metastore.py",
-                       classname = "HiveMetastore",
-                       command = "configure",
-                       config_file="default.json"
-    )
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-
-  def test_start_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_metastore.py",
-                       classname = "HiveMetastore",
-                       command = "start",
-                       config_file="default.json"
-    )
-
-    self.assert_configure_default()
-    self.assertResourceCalled('Execute', 'env HADOOP_HOME=/usr JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/start_metastore_script /var/log/hive/hive.out /var/log/hive/hive.log /var/run/hive/hive.pid /etc/hive/conf.server',
-                              not_if = 'ls /var/run/hive/hive.pid >/dev/null 2>&1 && ps `cat /var/run/hive/hive.pid` >/dev/null 2>&1',
-                              user = 'hive'
-    )
-
-    self.assertResourceCalled('Execute', '/usr/jdk64/jdk1.7.0_45/bin/java -cp /usr/lib/ambari-agent/DBConnectionVerification.jar:/usr/share/java/mysql-connector-java.jar org.apache.ambari.server.DBConnectionVerification jdbc:mysql://c6402.ambari.apache.org/hive?createDatabaseIfNotExist=true hive asd com.mysql.jdbc.Driver',
-                              path=['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']
-    )
-
-    self.assertNoMoreResources()
-
-  def test_stop_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_metastore.py",
-                       classname = "HiveMetastore",
-                       command = "stop",
-                       config_file="default.json"
-    )
-
-    self.assertResourceCalled('Execute', 'kill `cat /var/run/hive/hive.pid` >/dev/null 2>&1 && rm -f /var/run/hive/hive.pid')
-    self.assertNoMoreResources()
-
-  def test_configure_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_metastore.py",
-                       classname = "HiveMetastore",
-                       command = "configure",
-                       config_file="secured.json"
-    )
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-
-  def test_start_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_metastore.py",
-                       classname = "HiveMetastore",
-                       command = "start",
-                       config_file="secured.json"
-    )
-
-    self.assert_configure_secured()
-    self.assertResourceCalled('Execute', 'env HADOOP_HOME=/usr JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/start_metastore_script /var/log/hive/hive.out /var/log/hive/hive.log /var/run/hive/hive.pid /etc/hive/conf.server',
-                              not_if = 'ls /var/run/hive/hive.pid >/dev/null 2>&1 && ps `cat /var/run/hive/hive.pid` >/dev/null 2>&1',
-                              user = 'hive'
-    )
-
-    self.assertResourceCalled('Execute', '/usr/jdk64/jdk1.7.0_45/bin/java -cp /usr/lib/ambari-agent/DBConnectionVerification.jar:/usr/share/java/mysql-connector-java.jar org.apache.ambari.server.DBConnectionVerification jdbc:mysql://c6402.ambari.apache.org/hive?createDatabaseIfNotExist=true hive asd com.mysql.jdbc.Driver',
-                              path=['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']
-    )
-
-    self.assertNoMoreResources()
-
-  def test_stop_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_metastore.py",
-                       classname = "HiveMetastore",
-                       command = "stop",
-                       config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', 'kill `cat /var/run/hive/hive.pid` >/dev/null 2>&1 && rm -f /var/run/hive/hive.pid')
-    self.assertNoMoreResources()
-
-  def assert_configure_default(self):
-    self.assertResourceCalled('Execute', 'hive mkdir -p /tmp/HDP-artifacts/ ; cp /usr/share/java/mysql-connector-java.jar /usr/lib/hive/lib//mysql-connector-java.jar',
-      creates = '/usr/lib/hive/lib//mysql-connector-java.jar',
-      path = ['/bin', 'usr/bin/'],
-      not_if = 'test -f /usr/lib/hive/lib//mysql-connector-java.jar',
-    )
-    self.assertResourceCalled('Directory', '/etc/hive/conf.server',
-      owner = 'hive',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hive-site.xml',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 384,
-      conf_dir = '/etc/hive/conf.server',
-      configurations = self.getConfig()['configurations']['hive-site'],
-    )
-    self.assertResourceCalled('Execute', "/bin/sh -c 'cd /usr/lib/ambari-agent/ && curl -kf --retry 5 http://c6401.ambari.apache.org:8080/resources/DBConnectionVerification.jar -o DBConnectionVerification.jar'",
-      not_if = '[ -f DBConnectionVerification.jar]',
-    )
-    self.assertResourceCalled('File', '/tmp/start_metastore_script',
-      content = StaticFile('startMetastore.sh'),
-      mode = 493,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/lib/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf.server/hive-env.sh',
-      content = Template('hive-env.sh.j2', conf_dir="/etc/hive/conf.server"),
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-default.xml.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-
-  def assert_configure_secured(self):
-    self.assertResourceCalled('Execute', 'hive mkdir -p /tmp/HDP-artifacts/ ; cp /usr/share/java/mysql-connector-java.jar /usr/lib/hive/lib//mysql-connector-java.jar',
-      creates = '/usr/lib/hive/lib//mysql-connector-java.jar',
-      path = ['/bin', 'usr/bin/'],
-      not_if = 'test -f /usr/lib/hive/lib//mysql-connector-java.jar',
-    )
-    self.assertResourceCalled('Directory', '/etc/hive/conf.server',
-      owner = 'hive',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hive-site.xml',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 384,
-      conf_dir = '/etc/hive/conf.server',
-      configurations = self.getConfig()['configurations']['hive-site'],
-    )
-    self.assertResourceCalled('Execute', "/bin/sh -c 'cd /usr/lib/ambari-agent/ && curl -kf --retry 5 http://c6401.ambari.apache.org:8080/resources/DBConnectionVerification.jar -o DBConnectionVerification.jar'",
-      not_if = '[ -f DBConnectionVerification.jar]',
-    )
-    self.assertResourceCalled('File', '/tmp/start_metastore_script',
-      content = StaticFile('startMetastore.sh'),
-      mode = 493,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/lib/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf.server/hive-env.sh',
-      content = Template('hive-env.sh.j2', conf_dir="/etc/hive/conf.server"),
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-default.xml.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_server.py b/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_server.py
deleted file mode 100644
index 44effe7..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_server.py
+++ /dev/null
@@ -1,215 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHiveServer(RMFTestCase):
-
-  def test_configure_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_server.py",
-                       classname = "HiveServer",
-                       command = "configure",
-                       config_file="default.json"
-    )
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-  
-  def test_start_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_server.py",
-                         classname = "HiveServer",
-                         command = "start",
-                         config_file="default.json"
-    )
-
-    self.assert_configure_default()
-    self.assertResourceCalled('Execute', 'env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/start_hiveserver2_script /var/log/hive/hive-server2.out /var/log/hive/hive-server2.log /var/run/hive/hive-server.pid /etc/hive/conf.server',
-                              not_if = 'ls /var/run/hive/hive-server.pid >/dev/null 2>&1 && ps `cat /var/run/hive/hive-server.pid` >/dev/null 2>&1',
-                              user = 'hive'
-    )
-
-    self.assertResourceCalled('Execute', '/usr/jdk64/jdk1.7.0_45/bin/java -cp /usr/lib/ambari-agent/DBConnectionVerification.jar:/usr/share/java/mysql-connector-java.jar org.apache.ambari.server.DBConnectionVerification jdbc:mysql://c6402.ambari.apache.org/hive?createDatabaseIfNotExist=true hive asd com.mysql.jdbc.Driver',
-                              path=['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']
-    )
-
-    self.assertNoMoreResources()
-
-  def test_stop_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_server.py",
-                       classname = "HiveServer",
-                       command = "stop",
-                       config_file="default.json"
-    )
-
-    self.assertResourceCalled('Execute', 'kill `cat /var/run/hive/hive-server.pid` >/dev/null 2>&1 && rm -f /var/run/hive/hive-server.pid')
-    self.assertNoMoreResources()
-
-    
-  def test_configure_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_server.py",
-                       classname = "HiveServer",
-                       command = "configure",
-                       config_file="secured.json"
-    )
-    self.assert_configure_secured()
-    self.assertNoMoreResources()
-
-  def test_start_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_server.py",
-                       classname = "HiveServer",
-                       command = "start",
-                       config_file="secured.json"
-    )
-
-    self.assert_configure_secured()
-    self.assertResourceCalled('Execute', 'env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/start_hiveserver2_script /var/log/hive/hive-server2.out /var/log/hive/hive-server2.log /var/run/hive/hive-server.pid /etc/hive/conf.server',
-                              not_if = 'ls /var/run/hive/hive-server.pid >/dev/null 2>&1 && ps `cat /var/run/hive/hive-server.pid` >/dev/null 2>&1',
-                              user = 'hive'
-    )
-
-    self.assertResourceCalled('Execute', '/usr/jdk64/jdk1.7.0_45/bin/java -cp /usr/lib/ambari-agent/DBConnectionVerification.jar:/usr/share/java/mysql-connector-java.jar org.apache.ambari.server.DBConnectionVerification jdbc:mysql://c6402.ambari.apache.org/hive?createDatabaseIfNotExist=true hive asd com.mysql.jdbc.Driver',
-                              path=['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin']
-    )
-
-    self.assertNoMoreResources()
-
-  def test_stop_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/hive_server.py",
-                       classname = "HiveServer",
-                       command = "stop",
-                       config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', 'kill `cat /var/run/hive/hive-server.pid` >/dev/null 2>&1 && rm -f /var/run/hive/hive-server.pid')
-    self.assertNoMoreResources()
-
-  def assert_configure_default(self):
-    self.assertResourceCalled('Execute', 'hive mkdir -p /tmp/HDP-artifacts/ ; cp /usr/share/java/mysql-connector-java.jar /usr/lib/hive/lib//mysql-connector-java.jar',
-      creates = '/usr/lib/hive/lib//mysql-connector-java.jar',
-      path = ['/bin', 'usr/bin/'],
-      not_if = 'test -f /usr/lib/hive/lib//mysql-connector-java.jar',
-    )
-    self.assertResourceCalled('Directory', '/etc/hive/conf.server',
-      owner = 'hive',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hive-site.xml',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 384,
-      conf_dir = '/etc/hive/conf.server',
-      configurations = self.getConfig()['configurations']['hive-site'],
-    )
-    self.assertResourceCalled('Execute', "/bin/sh -c 'cd /usr/lib/ambari-agent/ && curl -kf --retry 5 http://c6401.ambari.apache.org:8080/resources/DBConnectionVerification.jar -o DBConnectionVerification.jar'",
-      not_if = '[ -f DBConnectionVerification.jar]',
-    )
-    self.assertResourceCalled('File', '/tmp/start_hiveserver2_script',
-      content = StaticFile('startHiveserver2.sh'),
-      mode = 493,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/lib/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf.server/hive-env.sh',
-      content = Template('hive-env.sh.j2', conf_dir="/etc/hive/conf.server"),
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-default.xml.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-
-  def assert_configure_secured(self):
-    self.assertResourceCalled('Execute', 'hive mkdir -p /tmp/HDP-artifacts/ ; cp /usr/share/java/mysql-connector-java.jar /usr/lib/hive/lib//mysql-connector-java.jar',
-      creates = '/usr/lib/hive/lib//mysql-connector-java.jar',
-      path = ['/bin', 'usr/bin/'],
-      not_if = 'test -f /usr/lib/hive/lib//mysql-connector-java.jar',
-    )
-    self.assertResourceCalled('Directory', '/etc/hive/conf.server',
-      owner = 'hive',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('XmlConfig', 'hive-site.xml',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 384,
-      conf_dir = '/etc/hive/conf.server',
-      configurations = self.getConfig()['configurations']['hive-site'],
-    )
-    self.assertResourceCalled('Execute', "/bin/sh -c 'cd /usr/lib/ambari-agent/ && curl -kf --retry 5 http://c6401.ambari.apache.org:8080/resources/DBConnectionVerification.jar -o DBConnectionVerification.jar'",
-      not_if = '[ -f DBConnectionVerification.jar]',
-    )
-    self.assertResourceCalled('File', '/tmp/start_hiveserver2_script',
-      content = StaticFile('startHiveserver2.sh'),
-      mode = 493,
-    )
-    self.assertResourceCalled('Directory', '/var/run/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/lib/hive',
-      owner = 'hive',
-      group = 'hadoop',
-      mode = 493,
-      recursive = True,
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf.server/hive-env.sh',
-      content = Template('hive-env.sh.j2', conf_dir="/etc/hive/conf.server"),
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-default.xml.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hive/conf/hive-env.sh.template',
-      owner = 'hive',
-      group = 'hadoop',
-    )
-

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_service_check.py b/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_service_check.py
deleted file mode 100644
index c5af2a6..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_hive_service_check.py
+++ /dev/null
@@ -1,117 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-import datetime
-import  resource_management.libraries.functions
-@patch.object(resource_management.libraries.functions, "get_unique_id_and_date", new = MagicMock(return_value=''))
-class TestServiceCheck(RMFTestCase):
-
-  def test_service_check_default(self):
-
-    self.executeScript("1.3.3/services/HIVE/package/scripts/service_check.py",
-                        classname="HiveServiceCheck",
-                        command="service_check",
-                        config_file="default.json"
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
-                        content = StaticFile('hiveserver2Smoke.sh'),
-                        mode = 493,
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
-                        content = StaticFile('hiveserver2.sql'),
-    )
-    self.assertResourceCalled('Execute', "env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://[u'c6402.ambari.apache.org']:10000 /tmp/hiveserver2.sql",
-                        logoutput = True,
-                        path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
-    self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
-                        content = StaticFile('hcatSmoke.sh'),
-                        mode = 493,
-    )
-    self.assertResourceCalled('Execute', 'sh /tmp/hcatSmoke.sh hcatsmoke prepare',
-                        logoutput = True,
-                        path = ['/usr/sbin', '/usr/local/nin', '/bin', '/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
-    self.assertResourceCalled('ExecuteHadoop', 'fs -test -e /apps/hive/warehouse/hcatsmoke',
-                        logoutput = True,
-                        user = 'hdfs',
-                        conf_dir = '/etc/hadoop/conf',
-    )
-    self.assertResourceCalled('Execute', 'sh /tmp/hcatSmoke.sh hcatsmoke cleanup',
-                        logoutput = True,
-                        path = ['/usr/sbin', '/usr/local/nin', '/bin', '/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
-    self.assertNoMoreResources()
-
-  def test_service_check_secured(self):
-
-    self.executeScript("1.3.3/services/HIVE/package/scripts/service_check.py",
-                        classname="HiveServiceCheck",
-                        command="service_check",
-                        config_file="secured.json"
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
-                        content = StaticFile('hiveserver2Smoke.sh'),
-                        mode = 493,
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
-                        content = StaticFile('hiveserver2.sql'),
-    )
-    self.assertResourceCalled('Execute', "/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa; env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://[u'c6402.ambari.apache.org']:10000/\\;principal=/etc/security/keytabs/hive.service.keytab /tmp/hiveserver2.sql",
-                        logoutput = True,
-                        path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
-    self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
-                        content = StaticFile('hcatSmoke.sh'),
-                        mode = 493,
-    )
-    self.assertResourceCalled('Execute', '/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa; sh /tmp/hcatSmoke.sh hcatsmoke prepare',
-                        logoutput = True,
-                        path = ['/usr/sbin', '/usr/local/nin', '/bin', '/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
-    self.assertResourceCalled('ExecuteHadoop', 'fs -test -e /apps/hive/warehouse/hcatsmoke',
-                        logoutput = True,
-                        user = 'hdfs',
-                        conf_dir = '/etc/hadoop/conf',
-    )
-    self.assertResourceCalled('Execute', '/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa; sh /tmp/hcatSmoke.sh hcatsmoke cleanup',
-                        logoutput = True,
-                        path = ['/usr/sbin', '/usr/local/nin', '/bin', '/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
-    self.assertNoMoreResources()

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_mysql_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_mysql_server.py b/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_mysql_server.py
deleted file mode 100644
index f79b954..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/HIVE/test_mysql_server.py
+++ /dev/null
@@ -1,141 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestMySqlServer(RMFTestCase):
-  def test_configure_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/mysql_server.py",
-                       classname = "MysqlServer",
-                       command = "configure",
-                       config_file="default.json"
-    )
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-
-  def test_start_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/mysql_server.py",
-                       classname = "MysqlServer",
-                       command = "start",
-                       config_file="default.json"
-    )
-
-    self.assertResourceCalled('Execute', 'service mysql start',
-                       logoutput = True,
-                       path = ['/usr/local/bin/:/bin/:/sbin/'],
-                       tries = 1,
-    )
-    self.assertNoMoreResources()
-
-  def test_stop_default(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/mysql_server.py",
-                       classname = "MysqlServer",
-                       command = "stop",
-                       config_file="default.json"
-    )
-
-    self.assertResourceCalled('Execute', 'service mysql stop',
-                              logoutput = True,
-                              path = ['/usr/local/bin/:/bin/:/sbin/'],
-                              tries = 1,
-    )
-    self.assertNoMoreResources()
-
-
-  def test_configure_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/mysql_server.py",
-                       classname = "MysqlServer",
-                       command = "configure",
-                       config_file="secured.json"
-    )
-    self.assert_configure_secured()
-    self.assertNoMoreResources()
-
-  def test_start_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/mysql_server.py",
-                       classname = "MysqlServer",
-                       command = "start",
-                       config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', 'service mysql start',
-                              logoutput = True,
-                              path = ['/usr/local/bin/:/bin/:/sbin/'],
-                              tries = 1,
-                              )
-    self.assertNoMoreResources()
-
-  def test_stop_secured(self):
-    self.executeScript("2.1.1/services/HIVE/package/scripts/mysql_server.py",
-                       classname = "MysqlServer",
-                       command = "stop",
-                       config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', 'service mysql stop',
-                              logoutput = True,
-                              path = ['/usr/local/bin/:/bin/:/sbin/'],
-                              tries = 1,
-                              )
-    self.assertNoMoreResources()
-
-  def assert_configure_default(self):
-    self.assertResourceCalled('Execute', 'service mysql start',
-      logoutput = True,
-      path = ['/usr/local/bin/:/bin/:/sbin/'],
-      tries = 1,
-    )
-    self.assertResourceCalled('File', '/tmp/addMysqlUser.sh',
-      content = StaticFile('addMysqlUser.sh'),
-      mode = 493,
-    )
-    self.assertResourceCalled('Execute', ('bash', '-x', '/tmp/addMysqlUser.sh', 'mysql', u'hive', 'asd', u'c6402.ambari.apache.org'),
-      logoutput = True,
-      path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-      tries = 3,
-      try_sleep = 5,
-    )
-    self.assertResourceCalled('Execute', 'service mysql stop',
-      logoutput = True,
-      path = ['/usr/local/bin/:/bin/:/sbin/'],
-      tries = 1,
-    )
-
-  def assert_configure_secured(self):
-    self.assertResourceCalled('Execute', 'service mysql start',
-      logoutput = True,
-      path = ['/usr/local/bin/:/bin/:/sbin/'],
-      tries = 1,
-    )
-    self.assertResourceCalled('File', '/tmp/addMysqlUser.sh',
-      content = StaticFile('addMysqlUser.sh'),
-      mode = 493,
-    )
-    self.assertResourceCalled('Execute', ('bash', '-x', '/tmp/addMysqlUser.sh', 'mysql', u'hive', 'asd', u'c6402.ambari.apache.org'),
-      logoutput = True,
-      path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-      tries = 3,
-      try_sleep = 5,
-    )
-    self.assertResourceCalled('Execute', 'service mysql stop',
-      logoutput = True,
-      path = ['/usr/local/bin/:/bin/:/sbin/'],
-      tries = 1,
-    )

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_client.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_client.py b/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_client.py
deleted file mode 100644
index 632ae49..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_client.py
+++ /dev/null
@@ -1,87 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestMapreduceClient(RMFTestCase):
-
-  def test_configure_default(self):
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/client.py",
-                       classname = "Client",
-                       command = "configure",
-                       config_file="default.json"
-    )
-    self.assertResourceCalled('Directory', '/var/run/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/hadoop/mapred',
-      owner = 'mapred',
-      recursive = True,
-      mode = 493,
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.exclude',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.include',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
-    self.assertNoMoreResources()
-
-  def test_configure_secured(self):
-
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/client.py",
-      classname = "Client",
-      command = "configure",
-      config_file="secured.json"
-    )
-    self.assertResourceCalled('Directory', '/var/run/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/hadoop/mapred',
-      owner = 'mapred',
-      recursive = True,
-      mode = 493,
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.exclude',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.include',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
-
-    self.assertNoMoreResources()
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/43f14b34/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_historyserver.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_historyserver.py b/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_historyserver.py
deleted file mode 100644
index 51735ce..0000000
--- a/ambari-server/src/test/python/stacks/1.3.3/MAPREDUCE/test_mapreduce_historyserver.py
+++ /dev/null
@@ -1,157 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-from mock.mock import MagicMock, call, patch
-from stacks.utils.RMFTestCase import *
-
-class TestHistoryServer(RMFTestCase):
-
-  def test_configure_default(self):
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/historyserver.py",
-                       classname = "Historyserver",
-                       command = "configure",
-                       config_file="default.json"
-    )
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-
-  def test_start_default(self):
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/historyserver.py",
-                         classname = "Historyserver",
-                         command = "start",
-                         config_file="default.json"
-      )
-
-    self.assert_configure_default()
-    self.assertResourceCalled('Execute', 'export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop/bin/hadoop-daemon.sh --config /etc/hadoop/conf start historyserver',
-                              user = 'mapred',
-                              not_if = 'ls /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid >/dev/null 2>&1 && ps `cat /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid` >/dev/null 2>&1'
-    )
-    self.assertResourceCalled('Execute', 'ls /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid >/dev/null 2>&1 && ps `cat /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid` >/dev/null 2>&1',
-                              user = 'mapred',
-                              initial_wait = 5,
-                              not_if= 'ls /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid >/dev/null 2>&1 && ps `cat /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid` >/dev/null 2>&1'
-    )
-    self.assertNoMoreResources()
-
-  def test_stop_default(self):
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/historyserver.py",
-                       classname = "Historyserver",
-                       command = "stop",
-                       config_file="default.json"
-    )
-
-    self.assertResourceCalled('Execute', 'export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop/bin/hadoop-daemon.sh --config /etc/hadoop/conf stop historyserver',
-                              user = 'mapred'
-    )
-    self.assertResourceCalled('Execute', 'rm -f /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid')
-    self.assertNoMoreResources()
-
-
-  def test_configure_secured(self):
-
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/historyserver.py",
-                       classname = "Historyserver",
-                       command = "configure",
-                       config_file="secured.json"
-    )
-    self.assert_configure_secured()
-    self.assertNoMoreResources()
-
-  def test_start_secured(self):
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/historyserver.py",
-                         classname = "Historyserver",
-                         command = "start",
-                         config_file="secured.json"
-    )
-
-    self.assert_configure_default()
-    self.assertResourceCalled('Execute', 'export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop/bin/hadoop-daemon.sh --config /etc/hadoop/conf start historyserver',
-                              user = 'mapred',
-                              not_if = 'ls /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid >/dev/null 2>&1 && ps `cat /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid` >/dev/null 2>&1'
-    )
-    self.assertResourceCalled('Execute', 'ls /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid >/dev/null 2>&1 && ps `cat /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid` >/dev/null 2>&1',
-                              user = 'mapred',
-                              initial_wait = 5,
-                              not_if= 'ls /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid >/dev/null 2>&1 && ps `cat /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid` >/dev/null 2>&1'
-    )
-    self.assertNoMoreResources()
-
-  def test_stop_secured(self):
-    self.executeScript("1.3.3/services/MAPREDUCE/package/scripts/historyserver.py",
-                       classname = "Historyserver",
-                       command = "stop",
-                       config_file="secured.json"
-    )
-
-    self.assertResourceCalled('Execute', 'export HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && /usr/lib/hadoop/bin/hadoop-daemon.sh --config /etc/hadoop/conf stop historyserver',
-                              user = 'mapred'
-    )
-    self.assertResourceCalled('Execute', 'rm -f /var/run/hadoop/mapred/hadoop-mapred-historyserver.pid')
-    self.assertNoMoreResources()
-
-  def assert_configure_default(self):
-    self.assertResourceCalled('Directory', '/var/run/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/hadoop/mapred',
-      owner = 'mapred',
-      recursive = True,
-      mode = 493,
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.exclude',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.include',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
-
-  def assert_configure_secured(self):
-    self.assertResourceCalled('Directory', '/var/run/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/var/log/hadoop/mapred',
-      owner = 'mapred',
-      group = 'hadoop',
-      recursive = True,
-    )
-    self.assertResourceCalled('Directory', '/hadoop/mapred',
-      owner = 'mapred',
-      recursive = True,
-      mode = 493,
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.exclude',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
-    self.assertResourceCalled('File', '/etc/hadoop/conf/mapred.include',
-      owner = 'mapred',
-      group = 'hadoop',
-    )
\ No newline at end of file