You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by jo...@apache.org on 2014/11/11 21:43:08 UTC

[05/17] ambari git commit: AMBARI-8276 - Alerts: Remove Nagios Service From The Stack (jonathanhurley)

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
index 8095048..ff5539f 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
@@ -20,8 +20,8 @@ package org.apache.ambari.server.metadata;
 
 import static junit.framework.Assert.assertEquals;
 import static junit.framework.Assert.assertFalse;
-import static junit.framework.Assert.assertTrue;
 import static junit.framework.Assert.assertNotNull;
+import static junit.framework.Assert.assertTrue;
 import static org.easymock.EasyMock.createMock;
 import static org.easymock.EasyMock.expect;
 import static org.easymock.EasyMock.replay;
@@ -31,11 +31,11 @@ import java.io.IOException;
 import java.io.InputStream;
 import java.util.Collections;
 import java.util.HashMap;
-import java.util.HashSet;
 import java.util.Map;
 import java.util.Set;
 
 import junit.framework.Assert;
+
 import org.apache.ambari.server.AmbariException;
 import org.apache.ambari.server.Role;
 import org.apache.ambari.server.RoleCommand;
@@ -52,7 +52,6 @@ import org.codehaus.jackson.annotate.JsonAutoDetect;
 import org.codehaus.jackson.annotate.JsonMethod;
 import org.codehaus.jackson.map.ObjectMapper;
 import org.codehaus.jackson.type.TypeReference;
-import org.easymock.IExpectationSetters;
 import org.junit.After;
 import org.junit.Before;
 import org.junit.Test;
@@ -274,16 +273,16 @@ public class RoleCommandOrderTest {
 
     InputStream testJsonIS = getClass().getClassLoader().
             getResourceAsStream(TEST_RCO_DATA_FILE);
-    
+
     ObjectMapper mapper = new ObjectMapper();
     Map<String,Object> testData = mapper.readValue(testJsonIS,
         new TypeReference<Map<String,Object>>() {});
-        
+
     rco.addDependencies(testData);
 
     mapper.setVisibility(JsonMethod.ALL, JsonAutoDetect.Visibility.ANY);
     String dump = mapper.writeValueAsString(rco.getDependencies());
-    
+
     String expected = "{\"RoleCommandPair{role=SECONDARY_NAMENODE, " +
         "cmd=UPGRADE}\":[{\"role\":{\"name\":\"NAMENODE\"},\"cmd\":\"UPGRADE\"}]," +
         "\"RoleCommandPair{role=SECONDARY_NAMENODE, cmd=START}\":[{\"role\":{\"name\":\"NAMENODE\"}," +
@@ -295,8 +294,8 @@ public class RoleCommandOrderTest {
 
     assertEquals(expected, dump);
   }
-  
-  
+
+
   @Test
   public void testInitializeDefault() throws IOException {
     RoleCommandOrder rco = injector.getInstance(RoleCommandOrder.class);
@@ -313,9 +312,9 @@ public class RoleCommandOrderTest {
 
     replay(cluster);
     replay(hdfsService);
-    
+
     rco.initialize(cluster);
-    
+
     verify(cluster);
     verify(hdfsService);
   }
@@ -333,19 +332,20 @@ public class RoleCommandOrderTest {
     Map<String,ServiceComponent> hdfsComponents = Collections.singletonMap("NAMENODE", namenode);
     expect(hdfsService.getServiceComponents()).andReturn(hdfsComponents).anyTimes();
 
-    Service nagiosService = createMock(Service.class);
-    expect(cluster.getService("NAGIOS")).andReturn(nagiosService).atLeastOnce();
-    expect(nagiosService.getCluster()).andReturn(cluster).anyTimes();
+    Service hbaseService = createMock(Service.class);
+    expect(cluster.getService("HBASE")).andReturn(hbaseService).atLeastOnce();
+    expect(hbaseService.getCluster()).andReturn(cluster).anyTimes();
 
-    ServiceComponent nagiosServer = createMock(ServiceComponent.class);
-    expect(nagiosServer.getName()).andReturn("NAGIOS_SERVER").anyTimes();
+    ServiceComponent hbaseMaster = createMock(ServiceComponent.class);
+    expect(hbaseMaster.getName()).andReturn("HBASE_MASTER").anyTimes();
 
-    Map<String,ServiceComponent> nagiosComponents = Collections.singletonMap("NAGIOS_SERVER", nagiosServer);
-    expect(nagiosService.getServiceComponents()).andReturn(nagiosComponents).anyTimes();
+    Map<String, ServiceComponent> hbaseComponents = Collections.singletonMap(
+        "HBASE_MASTER", hbaseMaster);
+    expect(hbaseService.getServiceComponents()).andReturn(hbaseComponents).anyTimes();
 
     Map<String, Service> installedServices = new HashMap<String, Service>();
     installedServices.put("HDFS", hdfsService);
-    installedServices.put("NAGIOS", nagiosService);
+    installedServices.put("HBASE", hbaseService);
     expect(cluster.getServices()).andReturn(installedServices).atLeastOnce();
 
 
@@ -356,14 +356,14 @@ public class RoleCommandOrderTest {
     expect(cluster.getCurrentStackVersion()).andReturn(new StackId("HDP", "2.0.5"));
 
     //replay
-    replay(cluster, hdfsService, nagiosService, nagiosServer, namenode);
+    replay(cluster, hdfsService, hbaseService, hbaseMaster, namenode);
 
     rco.initialize(cluster);
 
-    Set<Service> transitiveServices =
-      rco.getTransitiveServices(cluster.getService("NAGIOS"), RoleCommand.START);
+    Set<Service> transitiveServices = rco.getTransitiveServices(
+        cluster.getService("HBASE"), RoleCommand.START);
 
-    //HDFS should be started before NAGIOS start
+    // HDFS should be started before HBASE start
     Assert.assertNotNull(transitiveServices);
     Assert.assertFalse(transitiveServices.isEmpty());
     Assert.assertEquals(1, transitiveServices.size());

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
index 8d68f9c..8bead43 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
@@ -21,37 +21,35 @@ package org.apache.ambari.server.metadata;
 
 import static org.mockito.Mockito.mock;
 import static org.mockito.Mockito.when;
-
 import junit.framework.Assert;
 
 import org.apache.ambari.server.AmbariException;
 import org.apache.ambari.server.Role;
 import org.apache.ambari.server.RoleCommand;
+import org.apache.ambari.server.orm.GuiceJpaInitializer;
+import org.apache.ambari.server.orm.InMemoryDefaultTestModule;
 import org.apache.ambari.server.stageplanner.RoleGraphNode;
-import org.junit.After;
-import org.junit.Test;
 import org.apache.ambari.server.state.Service;
 import org.apache.ambari.server.state.ServiceComponent;
 import org.apache.ambari.server.state.StackId;
+import org.apache.ambari.server.state.cluster.ClusterImpl;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
 
 import com.google.inject.Guice;
 import com.google.inject.Injector;
 import com.google.inject.persist.PersistService;
 
-import org.apache.ambari.server.state.cluster.ClusterImpl;
-import org.apache.ambari.server.orm.GuiceJpaInitializer;
-import org.apache.ambari.server.orm.InMemoryDefaultTestModule;
-import org.junit.Before;
-
 public class RoleGraphTest {
 
-  
+
   private Injector injector;
-  
+
   @Before
   public void setup() throws Exception {
     injector = Guice.createInjector(new InMemoryDefaultTestModule());
-    injector.getInstance(GuiceJpaInitializer.class);    
+    injector.getInstance(GuiceJpaInitializer.class);
   }
 
   @After
@@ -63,9 +61,9 @@ public class RoleGraphTest {
   public void testValidateOrder() throws AmbariException {
     RoleCommandOrder rco = injector.getInstance(RoleCommandOrder.class);
     ClusterImpl cluster = mock(ClusterImpl.class);
-    
+
     when(cluster.getCurrentStackVersion()).thenReturn(new StackId("HDP-2.0.6"));
-    
+
     rco.initialize(cluster);
 
     RoleGraphNode datanode_upgrade = new RoleGraphNode(Role.DATANODE, RoleCommand.UPGRADE);
@@ -82,25 +80,10 @@ public class RoleGraphTest {
     RoleGraphNode datanode_start = new RoleGraphNode(Role.DATANODE, RoleCommand.START);
     RoleGraphNode datanode_install = new RoleGraphNode(Role.DATANODE, RoleCommand.INSTALL);
     RoleGraphNode jobtracker_start = new RoleGraphNode(Role.JOBTRACKER, RoleCommand.START);
-    RoleGraphNode tasktracker_start = new RoleGraphNode(Role.TASKTRACKER, RoleCommand.START);
     Assert.assertEquals(1, rco.order(datanode_start, datanode_install));
     Assert.assertEquals(1, rco.order(jobtracker_start, datanode_start));
     Assert.assertEquals(0, rco.order(jobtracker_start, jobtracker_start));
 
-    RoleGraphNode hive_client_install = new RoleGraphNode(Role.HIVE_CLIENT,
-      RoleCommand.INSTALL);
-    RoleGraphNode mapred_client_install = new RoleGraphNode(Role.MAPREDUCE_CLIENT,
-      RoleCommand.INSTALL);
-    RoleGraphNode hcat_client_install = new RoleGraphNode(Role.HCAT,
-      RoleCommand.INSTALL);
-    RoleGraphNode nagios_server_install = new RoleGraphNode(Role.NAGIOS_SERVER,
-      RoleCommand.INSTALL);
-    RoleGraphNode oozie_client_install = new RoleGraphNode(Role.OOZIE_CLIENT,
-      RoleCommand.INSTALL);
-    Assert.assertEquals(1, rco.order(nagios_server_install, hive_client_install));
-    Assert.assertEquals(1, rco.order(nagios_server_install, mapred_client_install));
-    Assert.assertEquals(1, rco.order(nagios_server_install, hcat_client_install));
-    Assert.assertEquals(1, rco.order(nagios_server_install, oozie_client_install));
 
     RoleGraphNode pig_service_check = new RoleGraphNode(Role.PIG_SERVICE_CHECK, RoleCommand.SERVICE_CHECK);
     RoleGraphNode resourcemanager_start = new RoleGraphNode(Role.RESOURCEMANAGER, RoleCommand.START);
@@ -109,78 +92,50 @@ public class RoleGraphTest {
     RoleGraphNode hdfs_service_check = new RoleGraphNode(Role.HDFS_SERVICE_CHECK, RoleCommand.SERVICE_CHECK);
     RoleGraphNode snamenode_start = new RoleGraphNode(Role.SECONDARY_NAMENODE, RoleCommand.START);
     Assert.assertEquals(-1, rco.order(snamenode_start, hdfs_service_check));
-    
+
     RoleGraphNode mapred2_service_check = new RoleGraphNode(Role.MAPREDUCE2_SERVICE_CHECK, RoleCommand.SERVICE_CHECK);
     RoleGraphNode rm_start = new RoleGraphNode(Role.RESOURCEMANAGER, RoleCommand.START);
     RoleGraphNode nm_start = new RoleGraphNode(Role.NODEMANAGER, RoleCommand.START);
     RoleGraphNode hs_start = new RoleGraphNode(Role.HISTORYSERVER, RoleCommand.START);
-    RoleGraphNode nagios_start = new RoleGraphNode(Role.NAGIOS_SERVER, RoleCommand.START);
+
     Assert.assertEquals(-1, rco.order(rm_start, mapred2_service_check));
-    Assert.assertEquals(-1, rco.order(nm_start, mapred2_service_check)); 
+    Assert.assertEquals(-1, rco.order(nm_start, mapred2_service_check));
     Assert.assertEquals(-1, rco.order(hs_start, mapred2_service_check));
     Assert.assertEquals(-1, rco.order(hs_start, mapred2_service_check));
     Assert.assertEquals(1, rco.order(nm_start, rm_start));
-    
+
     //Non-HA mode
     RoleGraphNode nn_start = new RoleGraphNode(Role.NAMENODE, RoleCommand.START);
     RoleGraphNode jn_start = new RoleGraphNode(Role.JOURNALNODE, RoleCommand.START);
     RoleGraphNode zk_server_start = new RoleGraphNode(Role.ZOOKEEPER_SERVER, RoleCommand.START);
     RoleGraphNode hbase_master_start = new RoleGraphNode(Role.HBASE_MASTER, RoleCommand.START);
-    RoleGraphNode hbase_reg_srv_start = new RoleGraphNode(Role.HBASE_REGIONSERVER, RoleCommand.START);
-    RoleGraphNode ganglia_server_start = new RoleGraphNode(Role.GANGLIA_SERVER, RoleCommand.START);
-    RoleGraphNode ganglia_monitor_start = new RoleGraphNode(Role.GANGLIA_MONITOR, RoleCommand.START);
-    RoleGraphNode hcat_start = new RoleGraphNode(Role.HCAT, RoleCommand.START);
     RoleGraphNode hive_srv_start = new RoleGraphNode(Role.HIVE_SERVER, RoleCommand.START);
     RoleGraphNode hive_ms_start = new RoleGraphNode(Role.HIVE_METASTORE, RoleCommand.START);
-    RoleGraphNode hue_start = new RoleGraphNode(Role.HUE_SERVER, RoleCommand.START);
     RoleGraphNode mysql_start = new RoleGraphNode(Role.MYSQL_SERVER, RoleCommand.START);
     RoleGraphNode oozie_srv_start = new RoleGraphNode(Role.OOZIE_SERVER, RoleCommand.START);
-    RoleGraphNode pig_start = new RoleGraphNode(Role.PIG, RoleCommand.START);
-    RoleGraphNode sqoop_start = new RoleGraphNode(Role.SQOOP, RoleCommand.START);
     RoleGraphNode webhcat_srv_start = new RoleGraphNode(Role.WEBHCAT_SERVER, RoleCommand.START);
     RoleGraphNode flume_start = new RoleGraphNode(Role.FLUME_HANDLER, RoleCommand.START);
     RoleGraphNode zkfc_start = new RoleGraphNode(Role.ZKFC, RoleCommand.START);
-    
+
     Assert.assertEquals(0, rco.order(nn_start, jn_start));
     Assert.assertEquals(0, rco.order(nn_start, zk_server_start));
     Assert.assertEquals(0, rco.order(zkfc_start, nn_start));
-    // Check that Nagios starts after other components
-    Assert.assertEquals(1, rco.order(nagios_start, nn_start));
-    Assert.assertEquals(1, rco.order(nagios_start, snamenode_start));
-    Assert.assertEquals(1, rco.order(nagios_start, datanode_start));
-    Assert.assertEquals(1, rco.order(nagios_start, resourcemanager_start));
-    Assert.assertEquals(1, rco.order(nagios_start, nm_start));
-    Assert.assertEquals(1, rco.order(nagios_start, hbase_master_start));
-    Assert.assertEquals(1, rco.order(nagios_start, hbase_reg_srv_start));
-    Assert.assertEquals(1, rco.order(nagios_start, ganglia_server_start));
-    Assert.assertEquals(1, rco.order(nagios_start, ganglia_monitor_start));
-    Assert.assertEquals(1, rco.order(nagios_start, hcat_start));
-    Assert.assertEquals(1, rco.order(nagios_start, hs_start));
-    Assert.assertEquals(1, rco.order(nagios_start, hive_srv_start));
-    Assert.assertEquals(1, rco.order(nagios_start, hue_start));
-    Assert.assertEquals(1, rco.order(nagios_start, jobtracker_start));
-    Assert.assertEquals(1, rco.order(nagios_start, tasktracker_start));
-    Assert.assertEquals(1, rco.order(nagios_start, zk_server_start));
-    Assert.assertEquals(1, rco.order(nagios_start, mysql_start));
-    Assert.assertEquals(1, rco.order(nagios_start, oozie_srv_start));
-    Assert.assertEquals(1, rco.order(nagios_start, pig_start));
-    Assert.assertEquals(1, rco.order(nagios_start, sqoop_start));
-    Assert.assertEquals(1, rco.order(nagios_start, webhcat_srv_start));
-    Assert.assertEquals(1, rco.order(nagios_start, flume_start));
-
-
-
-    //Enable HA for cluster
+
+    Assert.assertEquals(1, rco.order(flume_start, oozie_srv_start));
+    Assert.assertEquals(1, rco.order(hbase_master_start, zk_server_start));
+    Assert.assertEquals(1, rco.order(hive_srv_start, mysql_start));
+    Assert.assertEquals(1, rco.order(hive_ms_start, mysql_start));
+    Assert.assertEquals(1, rco.order(webhcat_srv_start, datanode_start));
+
+    // Enable HA for cluster
     Service hdfsServiceMock = mock(Service.class);
     ServiceComponent jnComponentMock = mock(ServiceComponent.class);
     when(cluster.getService("HDFS")).thenReturn(hdfsServiceMock);
     when(hdfsServiceMock.getServiceComponent("JOURNALNODE")).thenReturn(jnComponentMock);
-    
+
     rco.initialize(cluster);
     Assert.assertEquals(1, rco.order(nn_start, jn_start));
     Assert.assertEquals(1, rco.order(nn_start, zk_server_start));
     Assert.assertEquals(1, rco.order(zkfc_start, nn_start));
-    Assert.assertEquals(1, rco.order(nagios_start, zkfc_start));
-    Assert.assertEquals(1, rco.order(nagios_start, jn_start));
   }
 }

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java b/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
index a6cbc6a..0502e1a 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
@@ -18,6 +18,24 @@
 
 package org.apache.ambari.server.stack;
 
+import static org.easymock.EasyMock.createNiceMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertNotNull;
+import static org.junit.Assert.assertNull;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+
+import java.io.File;
+import java.util.Collection;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Map;
+
 import org.apache.ambari.server.AmbariException;
 import org.apache.ambari.server.configuration.Configuration;
 import org.apache.ambari.server.metadata.ActionMetadata;
@@ -33,24 +51,6 @@ import org.apache.ambari.server.state.stack.OsFamily;
 import org.junit.BeforeClass;
 import org.junit.Test;
 
-import java.io.File;
-import java.util.Collection;
-import java.util.HashMap;
-import java.util.HashSet;
-import java.util.List;
-import java.util.Map;
-
-import static org.easymock.EasyMock.createNiceMock;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertNull;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
-
 /**
  * StackManager unit tests.
  */
@@ -197,7 +197,6 @@ public class StackManagerTest {
     expectedServices.add("HDFS");
     expectedServices.add("HIVE");
     expectedServices.add("MAPREDUCE2");
-    expectedServices.add("NAGIOS");
     expectedServices.add("OOZIE");
     expectedServices.add("PIG");
     expectedServices.add("SQOOP");
@@ -205,6 +204,7 @@ public class StackManagerTest {
     expectedServices.add("ZOOKEEPER");
     expectedServices.add("STORM");
     expectedServices.add("FLUME");
+    expectedServices.add("FAKENAGIOS");
 
     ServiceInfo pigService = null;
     for (ServiceInfo service : services) {
@@ -391,16 +391,22 @@ public class StackManagerTest {
 
   @Test
   public void testMonitoringServicePropertyInheritance() throws Exception{
-    StackInfo stack = stackManager.getStack("HDP", "2.0.7");
+    StackInfo stack = stackManager.getStack("HDP", "2.0.8");
     Collection<ServiceInfo> allServices = stack.getServices();
     assertEquals(13, allServices.size());
+
+    boolean monitoringServiceFound = false;
+
     for (ServiceInfo serviceInfo : allServices) {
-      if (serviceInfo.getName().equals("NAGIOS")) {
+      if (serviceInfo.getName().equals("FAKENAGIOS")) {
+        monitoringServiceFound = true;
         assertTrue(serviceInfo.isMonitoringService());
       } else {
         assertNull(serviceInfo.isMonitoringService());
       }
     }
+
+    assertTrue(monitoringServiceFound);
   }
 
   @Test
@@ -408,7 +414,7 @@ public class StackManagerTest {
     StackInfo stack = stackManager.getStack("HDP", "2.0.6");
     Collection<ServiceInfo> allServices = stack.getServices();
 
-    assertEquals(12, allServices.size());
+    assertEquals(11, allServices.size());
     HashSet<String> expectedServices = new HashSet<String>();
     expectedServices.add("GANGLIA");
     expectedServices.add("HBASE");
@@ -416,7 +422,6 @@ public class StackManagerTest {
     expectedServices.add("HDFS");
     expectedServices.add("HIVE");
     expectedServices.add("MAPREDUCE2");
-    expectedServices.add("NAGIOS");
     expectedServices.add("OOZIE");
     expectedServices.add("PIG");
     expectedServices.add("ZOOKEEPER");

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java b/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
index f008980..dd2a519 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
@@ -62,7 +62,7 @@ public class TestStagePlanner {
     RoleCommandOrder rco = injector.getInstance(RoleCommandOrder.class);
     ClusterImpl cluster = mock(ClusterImpl.class);
     when(cluster.getCurrentStackVersion()).thenReturn(new StackId("HDP-2.0.6"));
-    
+
     rco.initialize(cluster);
 
     RoleGraph rg = new RoleGraph(rco);
@@ -140,9 +140,6 @@ public class TestStagePlanner {
     stage.addHostRoleExecutionCommand("host7", Role.WEBHCAT_SERVER,
         RoleCommand.START, new ServiceComponentHostStartEvent("WEBHCAT_SERVER",
             "host7", now), "cluster1", "WEBHCAT");
-    stage.addHostRoleExecutionCommand("host8", Role.NAGIOS_SERVER,
-        RoleCommand.START, new ServiceComponentHostStartEvent("NAGIOS_SERVER",
-            "host8", now), "cluster1", "NAGIOS");
     stage.addHostRoleExecutionCommand("host4", Role.GANGLIA_MONITOR,
         RoleCommand.START, new ServiceComponentHostStartEvent("GANGLIA_MONITOR",
             "host4", now), "cluster1", "GANGLIA");
@@ -156,6 +153,6 @@ public class TestStagePlanner {
     for (Stage s: outStages) {
       System.out.println(s.toString());
     }
-    assertEquals(5, outStages.size());
+    assertEquals(4, outStages.size());
   }
 }

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py b/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py
deleted file mode 100644
index a75e92a..0000000
--- a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py
+++ /dev/null
@@ -1,549 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-import StringIO
-
-import os, sys
-import pprint
-import subprocess
-from unittest import TestCase
-from mock.mock import Mock, MagicMock, patch
-import mm_wrapper
-
-class TestOrWrapper(TestCase):
-
-  dummy_ignore_file = """
-vm-4.vm HIVE HIVE_METASTORE
-vm-5.vm GANGLIA GANGLIA_MONITOR
-vm-4.vm YARN NODEMANAGER
-vm-3.vm YARN NODEMANAGER
-vm-3.vm HBASE HBASE_REGIONSERVER
-vm-4.vm HBASE HBASE_REGIONSERVER
-vm-4.vm STORM STORM_REST_API
-vm-4.vm HDFS DATANODE
-vm-4.vm STORM SUPERVISOR
-vm-4.vm STORM NIMBUS
-vm-4.vm STORM STORM_UI_SERVER
-vm-3.vm STORM SUPERVISOR
-vm-4.vm HDFS SECONDARY_NAMENODE
-vm-3.vm FLUME FLUME_HANDLER
-vm-4.vm GANGLIA GANGLIA_SERVER
-vm-4.vm HIVE HIVE_SERVER
-vm-4.vm ZOOKEEPER ZOOKEEPER_SERVER
-vm-4.vm WEBHCAT WEBHCAT_SERVER
-vm-3.vm HBASE HBASE_MASTER
-vm-4.vm GANGLIA GANGLIA_MONITOR
-vm-3.vm GANGLIA GANGLIA_MONITOR
-vm-3.vm HDFS NAMENODE
-vm-4.vm HIVE MYSQL_SERVER
-vm-4.vm YARN APP_TIMELINE_SERVER
-vm-4.vm FALCON FALCON_SERVER
-vm-3.vm HDFS DATANODE
-vm-4.vm YARN RESOURCEMANAGER
-vm-4.vm OOZIE OOZIE_SERVER
-vm-4.vm MAPREDUCE2 HISTORYSERVER
-vm-4.vm STORM DRPC_SERVER
-vm-4.vm FLUME FLUME_HANDLER
-vm-3.vm ZOOKEEPER ZOOKEEPER_SERVER
-"""
-
-  default_empty_check_result = {
-    'message': 'No checks have been run (no hostnames provided)',
-    'retcode': -1,
-    'real_retcode': None
-  }
-
-
-  @patch("__builtin__.open")
-  def test_ignored_host_list(self, open_mock):
-    # Check with empty file content
-    open_mock.return_value.__enter__.return_value.read.return_value = ""
-    lst = mm_wrapper.ignored_host_list('STORM', 'SUPERVISOR')
-    self.assertEqual(pprint.pformat(lst), '[]')
-    # Check with dummy content
-    open_mock.return_value.__enter__.return_value.read.return_value = self.dummy_ignore_file
-    lst = mm_wrapper.ignored_host_list('STORM', 'SUPERVISOR')
-    self.assertEqual(pprint.pformat(lst), "['vm-4.vm', 'vm-3.vm']")
-    # Check if service name/comp name are not defined
-    open_mock.return_value.__enter__.return_value.read.return_value = self.dummy_ignore_file
-    lst = mm_wrapper.ignored_host_list('', '')
-    self.assertEqual(pprint.pformat(lst), "[]")
-
-
-  @patch("sys.exit")
-  def test_print_usage(self, exit_mock):
-    mm_wrapper.print_usage()
-    self.assertTrue(exit_mock.called)
-    self.assertEqual(exit_mock.call_args_list[0][0][0], 1)
-
-
-  def test_get_real_component(self):
-    with patch.dict(os.environ, {'NAGIOS__SERVICEHOST_COMPONENT': 'SUPERVISOR'}, clear=True):
-      component = mm_wrapper.get_real_component()
-      self.assertEqual(component, 'SUPERVISOR')
-    with patch.dict(os.environ, {'NAGIOS__SERVICEHOST_COMPONENT': 'JOBHISTORY'}, clear=True):
-      component = mm_wrapper.get_real_component()
-      self.assertEqual(component, 'MAPREDUCE2')
-
-
-  @patch("mm_wrapper.print_usage")
-  def test_parse_args(self, print_usage_mock):
-    args = ['or', 'h1', 'h2', '--', 'prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    mode, hostnames, command_line = mm_wrapper.parse_args(args)
-    self.assertEquals(mode, mm_wrapper.OR)
-    self.assertEquals(hostnames, ['h1', 'h2'])
-    self.assertEquals(command_line, ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt'])
-
-    args = ['and', 'h1', 'h2', '--', 'prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    mode, hostnames, command_line = mm_wrapper.parse_args(args)
-    self.assertEquals(mode, mm_wrapper.AND)
-    self.assertEquals(hostnames, ['h1', 'h2'])
-    self.assertEquals(command_line, ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt'])
-
-    args = ['env_only', 'h1', 'h2', '--', 'prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    mode, hostnames, command_line = mm_wrapper.parse_args(args)
-    self.assertEquals(mode, mm_wrapper.ENV_ONLY)
-    self.assertEquals(hostnames, ['h1', 'h2'])
-    self.assertEquals(command_line, ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt'])
-
-    # Check wrong usage
-    args = []
-    mm_wrapper.parse_args(args)
-    self.assertTrue(print_usage_mock.called)
-
-
-  @patch("mm_wrapper.ignored_host_list")
-  @patch("mm_wrapper.work_in_or_mode")
-  @patch("mm_wrapper.work_in_and_mode")
-  @patch("mm_wrapper.work_in_env_only_mode")
-  @patch("mm_wrapper.work_in_filter_mm_mode")
-  @patch("mm_wrapper.work_in_legacy_check_wrapper_mode")
-  def test_do_work(self, work_in_legacy_check_wrapper_mode, work_in_filter_mm_mode_mock,
-                   work_in_env_only_mode_mock, work_in_and_mode_mock,
-                   work_in_or_mode_mock,
-                   ignored_host_list_mock):
-    hostnames = ['h1', 'h2', 'h3', 'h4']
-    ignored_host_list_mock.return_value = ['h2', 'h3']
-    command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    work_in_or_mode_mock.return_value = {
-      'message': "or_mode mode result",
-      'retcode': 0,
-      'real_retcode': None
-    }
-    work_in_and_mode_mock.return_value = {
-      'message': "and_mode mode result",
-      'retcode': 0,
-      'real_retcode': None
-    }
-    work_in_env_only_mode_mock.return_value = {
-      'message': "env_only mode result",
-      'retcode': 0,
-      'real_retcode': None
-    }
-    work_in_filter_mm_mode_mock.return_value = {
-      'message': "filter_mm mode result",
-      'retcode': 0,
-      'real_retcode': None
-    }
-    work_in_legacy_check_wrapper_mode.return_value = {
-      'message': "legacy_check_wrapper mode result",
-      'retcode': 0,
-      'real_retcode': None
-    }
-    result = mm_wrapper.do_work(mm_wrapper.OR, hostnames, command_line)
-    self.assertEquals(str(result), "(['or_mode mode result'], 0)")
-
-    result = mm_wrapper.do_work(mm_wrapper.AND, hostnames, command_line)
-    self.assertEquals(str(result), "(['and_mode mode result'], 0)")
-
-    result = mm_wrapper.do_work(mm_wrapper.ENV_ONLY, hostnames, command_line)
-    self.assertEquals(str(result), "(['env_only mode result'], 0)")
-
-    result = mm_wrapper.do_work(mm_wrapper.FILTER_MM, hostnames, command_line)
-    self.assertEquals(str(result), "(['filter_mm mode result'], 0)")
-
-    result = mm_wrapper.do_work(mm_wrapper.LEGACY_CHECK_WRAPPER, hostnames, command_line)
-    self.assertEquals(str(result), "(['legacy_check_wrapper mode result'], 0)")
-
-    # Check behaviour when real_retcode is defined
-    work_in_or_mode_mock.return_value = {
-      'message': "or_mode mode result",
-      'retcode': 0,
-      'real_retcode': 1
-    }
-    result = mm_wrapper.do_work(mm_wrapper.OR, hostnames, command_line)
-    self.assertEquals(str(result), "(['or_mode mode result', 'AMBARIPASSIVE=1'], 0)")
-
-
-  @patch("mm_wrapper.check_output")
-  def test_work_in_or_mode(self, check_output_mock):
-    hostnames = ['h1', 'h2', 'h3', 'h4']
-    ignored_hosts = ['h2', 'h3']
-    command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    custom_env = {'MM_HOSTS': ignored_hosts}
-
-    # Normal usage
-    check_output_mock.return_value = 'Dummy message'
-    result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(str(result),
-                      "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
-    self.assertEquals(check_output_mock.call_count, 1)  # Exited on first success
-    self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
-    for check_tupple in zip(check_output_mock.call_args_list, hostnames):
-      self.assertEquals(check_tupple[0][0][0], ['prog', '-h', check_tupple[1], '-opt', 'yet', 'another', 'opt'])
-
-    check_output_mock.reset_mock()
-
-    # Failed all checks
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(3, 'dummy cmd')
-    error.output = 'dummy output2'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output3'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(2, 'dummy cmd')
-    error.output = 'dummy output4'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-
-    result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 4)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 2}")
-
-    check_output_mock.reset_mock()
-
-    # Failed all but MM host component checks
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(0, 'dummy cmd')
-    error.output = 'dummy output2'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(2, 'dummy cmd')
-    error.output = 'dummy output3'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(3, 'dummy cmd')
-    error.output = 'dummy output4'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-
-    result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 4)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 3}")
-
-    check_output_mock.reset_mock()
-
-    # Components check only for one check is successful
-    ignored_hosts = []
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(0, 'dummy cmd')
-    error.output = 'dummy output2'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(2, 'dummy cmd')
-    error.output = 'dummy output3'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(3, 'dummy cmd')
-    error.output = 'dummy output4'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-
-    result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 2)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output2', 'real_retcode': None, 'retcode': 0}")
-
-
-  @patch("mm_wrapper.check_output")
-  def test_work_in_and_mode(self, check_output_mock):
-    hostnames = ['h1', 'h2', 'h3', 'h4']
-    ignored_hosts = ['h2', 'h3']
-    command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    custom_env = {'MM_HOSTS': ignored_hosts}
-
-    # Normal usage
-    check_output_mock.return_value = 'Dummy message'
-    result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(str(result),
-                      "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
-    self.assertEquals(check_output_mock.call_count, 4)
-    self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
-    for check_tupple in zip(check_output_mock.call_args_list, hostnames):
-      self.assertEquals(check_tupple[0][0][0], ['prog', '-h', check_tupple[1], '-opt', 'yet', 'another', 'opt'])
-
-    check_output_mock.reset_mock()
-
-    # Failed all checks
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(3, 'dummy cmd')
-    error.output = 'dummy output2'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output3'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(2, 'dummy cmd')
-    error.output = 'dummy output4'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 4)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 2}")
-
-    check_output_mock.reset_mock()
-
-    # Failed all but MM host component checks
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(0, 'dummy cmd')
-    error.output = 'dummy output2'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(2, 'dummy cmd')
-    error.output = 'dummy output3'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(3, 'dummy cmd')
-    error.output = 'dummy output4'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 4)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 3}")
-
-    check_output_mock.reset_mock()
-
-    # Components check only for one check is successful
-    ignored_hosts = []
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(0, 'dummy cmd')
-    error.output = 'dummy output2'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(2, 'dummy cmd')
-    error.output = 'dummy output3'
-    check_output_side_effects.append(error)
-
-    error = subprocess.CalledProcessError(3, 'dummy cmd')
-    error.output = 'dummy output4'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
-                                        custom_env,
-                                        self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 4)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 3}")
-
-
-  @patch("mm_wrapper.check_output")
-  def test_work_in_env_only_mode(self, check_output_mock):
-    hostnames = ['h1', 'h2', 'h3', 'h4']
-    ignored_hosts = ['h2', 'h3']
-    command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    custom_env = {'MM_HOSTS' : ignored_hosts}
-
-    # Normal usage
-    check_output_mock.return_value = 'Dummy message'
-    result = mm_wrapper.work_in_env_only_mode(hostnames, command_line, custom_env)
-    self.assertEquals(str(result),
-                      "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
-    self.assertEquals(check_output_mock.call_count, 1)
-    self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
-    self.assertEquals(check_output_mock.call_args[0][0],
-                      ['prog', '-h', 'h1', 'h2', 'h3', 'h4', '-opt', 'yet', 'another', 'opt'])
-
-    check_output_mock.reset_mock()
-
-    # Failed all checks
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_env_only_mode(hostnames, command_line, custom_env)
-    self.assertEquals(check_output_mock.call_count, 1)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output1', 'real_retcode': None, 'retcode': 1}")
-
-    check_output_mock.reset_mock()
-
-
-  @patch("mm_wrapper.check_output")
-  def test_work_in_filter_mm_mode(self, check_output_mock):
-    hostnames = ['h1', 'h2', 'h3', 'h4']
-    ignored_hosts = ['h2', 'h3']
-    command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
-    custom_env = {'MM_HOSTS' : ignored_hosts}
-
-    # Normal usage
-    check_output_mock.return_value = 'Dummy message'
-    result = mm_wrapper.work_in_filter_mm_mode(hostnames, ignored_hosts, command_line,
-                                               custom_env,
-                                               self.default_empty_check_result)
-    self.assertEquals(str(result),
-                      "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
-    self.assertEquals(check_output_mock.call_count, 1)
-    self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
-    self.assertEquals(check_output_mock.call_args[0][0],
-                      ['prog', '-h', 'h1', 'h4', '-opt', 'yet', 'another', 'opt'])
-
-    check_output_mock.reset_mock()
-
-    # Failed all checks
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_filter_mm_mode(hostnames, ignored_hosts, command_line,
-                                               custom_env,
-                                               self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 1)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output1', 'real_retcode': None, 'retcode': 1}")
-
-    check_output_mock.reset_mock()
-
-    # All host components are in MM
-    ignored_hosts = hostnames
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_filter_mm_mode(hostnames, ignored_hosts, command_line,
-                                               custom_env,
-                                               self.default_empty_check_result)
-    self.assertEquals(check_output_mock.call_count, 0)
-    self.assertEquals(str(result),
-                      "{'message': 'No checks have been run (no hostnames provided)', "
-                      "'real_retcode': None, 'retcode': -1}")
-
-    check_output_mock.reset_mock()
-
-
-  @patch("mm_wrapper.check_output")
-  @patch.dict(os.environ, {'NAGIOS_HOSTNAME': 'h2'}, clear=True)
-  def test_work_in_legacy_check_wrapper_mode(self, check_output_mock):
-    command_line = ['prog', '-opt', 'yet', 'another', 'opt']
-    ignored_hosts = []
-    custom_env = {'MM_HOSTS': ignored_hosts}
-
-    # Normal usage
-    ignored_hosts = []
-    check_output_mock.return_value = 'Dummy message'
-    result = mm_wrapper.work_in_legacy_check_wrapper_mode(ignored_hosts, command_line,
-                                               custom_env)
-    self.assertEquals(str(result),
-                      "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
-    self.assertEquals(check_output_mock.call_count, 1)
-    self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
-    self.assertEquals(check_output_mock.call_args[0][0],
-                      ['prog', '-opt', 'yet', 'another', 'opt'])
-
-    check_output_mock.reset_mock()
-
-    # Failed check on host that is not in MM state
-    ignored_hosts = ['h3']
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_legacy_check_wrapper_mode(ignored_hosts, command_line,
-                                               custom_env)
-    self.assertEquals(check_output_mock.call_count, 1)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output1', 'real_retcode': None, 'retcode': 1}")
-
-    check_output_mock.reset_mock()
-
-    # Failed check on host that is in MM state
-    ignored_hosts = ['h2']
-    check_output_side_effects = []
-    error = subprocess.CalledProcessError(1, 'dummy cmd')
-    error.output = 'dummy output1'
-    check_output_side_effects.append(error)
-
-    check_output_mock.side_effect = check_output_side_effects
-    result = mm_wrapper.work_in_legacy_check_wrapper_mode(ignored_hosts, command_line,
-                                               custom_env)
-    self.assertEquals(check_output_mock.call_count, 1)
-    self.assertEquals(str(result),
-                      "{'message': 'dummy output1', 'real_retcode': 1, 'retcode': 0}")
-
-    check_output_mock.reset_mock()
-

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py b/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py
deleted file mode 100644
index f1e4650..0000000
--- a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py
+++ /dev/null
@@ -1,282 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-
-from mock.mock import MagicMock, patch
-from stacks.utils.RMFTestCase import *
-
-
-class TestNagiosServer(RMFTestCase):
-  def test_configure_default(self):
-    self.executeScript("1.3.2/services/NAGIOS/package/scripts/nagios_server.py",
-                       classname="NagiosServer",
-                       command="configure",
-                       config_file="default.json"
-    )
-    self.assert_configure_default()
-    self.assertNoMoreResources()
-
-
-  def test_start_default(self):
-    self.executeScript(
-      "1.3.2/services/NAGIOS/package/scripts/nagios_service.py",
-      classname="NagiosServer",
-      command="start",
-      config_file="default.json"
-    )
-    self.assert_configure_default()
-    self.assertResourceCalled('Execute', 'service nagios start',
-                              path=['/usr/local/bin/:/bin/:/sbin/']
-    )
-    self.assertResourceCalled('MonitorWebserver', 'restart',
-    )
-    self.assertNoMoreResources()
-
-
-  @patch('os.path.isfile')
-  def test_stop_default(self, os_path_isfile_mock):
-    src_dir = RMFTestCase._getSrcFolder()    
-    os_path_isfile_mock.side_effect = [False, True]
-       
-    self.executeScript(
-      "1.3.2/services/NAGIOS/package/scripts/nagios_service.py",
-      classname="NagiosServer",
-      command="stop",
-      config_file="default.json"
-    )
-    
-    self.assertResourceCalled('Execute','service nagios stop', path=['/usr/local/bin/:/bin/:/sbin/'])
-    self.assertResourceCalled('Execute','rm -f /var/run/nagios/nagios.pid')
-    self.assertResourceCalled('MonitorWebserver', 'restart')
-    
-    self.assertNoMoreResources()
-
-
-  def assert_configure_default(self):
-    self.assertResourceCalled('File', '/etc/apache2/conf.d/nagios.conf',
-                              owner='nagios',
-                              group='nagios',
-                              content=Template("nagios.conf.j2"),
-                              mode=0644
-    )
-    self.assertResourceCalled('Directory', '/etc/nagios',
-                              owner='nagios',
-                              group='nagios',
-    )
-    self.assertResourceCalled('Directory', '/usr/lib64/nagios/plugins'
-    )
-    self.assertResourceCalled('Directory', '/etc/nagios/objects'
-    )
-    self.assertResourceCalled('Directory', '/var/run/nagios',
-                              owner='nagios',
-                              group='nagios',
-                              mode=0755,
-                              recursive=True
-    )
-    self.assertResourceCalled('Directory', '/var/nagios',
-                              owner='nagios',
-                              group='nagios',
-                              recursive=True
-    )
-    self.assertResourceCalled('Directory', '/var/nagios/spool/checkresults',
-                              owner='nagios',
-                              group='nagios',
-                              recursive=True
-    )
-    self.assertResourceCalled('Directory', '/var/nagios/rw',
-                              owner='nagios',
-                              group='nagios',
-                              recursive=True
-    )
-    self.assertResourceCalled('Directory', '/var/log/nagios',
-                              owner='nagios',
-                              group='nagios',
-                              mode=0755
-    )
-    self.assertResourceCalled('Directory', '/var/log/nagios/archives',
-                              owner='nagios',
-                              group='nagios',
-                              mode=0755
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/nagios/nagios.cfg',
-                              owner='nagios',
-                              group='nagios',
-                              mode=None
-    )
-    self.assertResourceCalled('TemplateConfig', '/etc/nagios/resource.cfg',
-                              owner='nagios',
-                              group='nagios',
-                              mode=None
-    )
-    self.assertResourceCalled('TemplateConfig',
-                              '/etc/nagios/objects/hadoop-hosts.cfg',
-                              owner='nagios',
-                              group='hadoop',
-                              mode=None
-    )
-    self.assertResourceCalled('TemplateConfig',
-                              '/etc/nagios/objects/hadoop-hostgroups.cfg',
-                              owner='nagios',
-                              group='hadoop',
-                              mode=None
-    )
-    self.assertResourceCalled('TemplateConfig',
-                              '/etc/nagios/objects/hadoop-servicegroups.cfg',
-                              owner='nagios',
-                              group='hadoop',
-                              mode=None
-    )
-    self.assertResourceCalled('TemplateConfig',
-                              '/etc/nagios/objects/hadoop-services.cfg',
-                              owner='nagios',
-                              group='hadoop',
-                              mode=None
-    )
-    self.assertResourceCalled('TemplateConfig',
-                              '/etc/nagios/objects/hadoop-commands.cfg',
-                              owner='nagios',
-                              group='hadoop',
-                              mode=None
-    )
-    self.assertResourceCalled('TemplateConfig',
-                              '/etc/nagios/objects/contacts.cfg',
-                              owner='nagios',
-                              group='hadoop',
-                              mode=None
-    )
-    self.assertResourceCalled('File', '/usr/lib64/nagios/plugins/check_cpu.pl',
-                              content=StaticFile('check_cpu.pl'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File', '/usr/lib64/nagios/plugins/check_cpu.php',
-                              content=StaticFile('check_cpu.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File', '/usr/lib64/nagios/plugins/check_cpu_ha.php',
-                              content=StaticFile('check_cpu_ha.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_datanode_storage.php',
-                              content=StaticFile('check_datanode_storage.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_aggregate.php',
-                              content=StaticFile('check_aggregate.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_hdfs_blocks.php',
-                              content=StaticFile('check_hdfs_blocks.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_hdfs_capacity.php',
-                              content=StaticFile('check_hdfs_capacity.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_rpcq_latency.php',
-                              content=StaticFile('check_rpcq_latency.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_webui.sh',
-                              content=StaticFile('check_webui.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_name_dir_status.php',
-                              content=StaticFile('check_name_dir_status.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_oozie_status.sh',
-                              content=StaticFile('check_oozie_status.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_templeton_status.sh',
-                              content=StaticFile('check_templeton_status.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_hive_metastore_status.sh',
-                              content=StaticFile(
-                                'check_hive_metastore_status.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_hue_status.sh',
-                              content=StaticFile('check_hue_status.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_mapred_local_dir_used.sh',
-                              content=StaticFile(
-                                'check_mapred_local_dir_used.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_nodemanager_health.sh',
-                              content=StaticFile('check_nodemanager_health.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_namenodes_ha.sh',
-                              content=StaticFile('check_namenodes_ha.sh'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/hdp_nagios_init.php',
-                              content=StaticFile('hdp_nagios_init.php'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/mm_wrapper.py',
-                              content=StaticFile('mm_wrapper.py'),
-                              mode=0755
-    )
-    self.assertResourceCalled('File',
-                              '/usr/lib64/nagios/plugins/check_hive_thrift_port.py',
-                              content=StaticFile('check_hive_thrift_port.py'),
-                              mode=0755
-    )
-    self.assertResourceCalled('Execute',
-                              'htpasswd2 -c -b  /etc/nagios/htpasswd.users nagiosadmin \'!`"\'"\'"\' 1\''
-    )
-
-    self.assertResourceCalled('File', '/etc/nagios/htpasswd.users',
-                              owner='nagios',
-                              group='nagios',
-                              mode=0640
-    )
-    self.assertResourceCalled('Execute', 'usermod -G nagios wwwrun'
-    )
-
-    self.assertResourceCalled('File', '/etc/nagios/command.cfg',
-                              owner='nagios',
-                              group='nagios'
-    )
-    self.assertResourceCalled('File', '/var/nagios/ignore.dat',
-                              owner='nagios',
-                              group='nagios',
-                              mode=0664
-    )

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
index 31bc5d4..9ba4073 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
@@ -326,14 +326,6 @@
             "hadoop_heapsize": "1024", 
             "jtnode_opt_maxnewsize": "200m"
         }, 
-        "nagios-env": {
-            "hive_metastore_user_passwd": "password", 
-            "nagios_web_password": "!`\"' 1",
-            "nagios_user": "nagios", 
-            "nagios_group": "nagios", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "user@com.ua"
-        }, 
         "hive-env": {
             "hive_metastore_user_passwd": "password", 
             "hcat_pid_dir": "/var/run/webhcat", 
@@ -493,9 +485,6 @@
             "c6401.ambari.apache.org", 
             "c6402.ambari.apache.org"
         ], 
-        "nagios_server_host": [
-            "c6402.ambari.apache.org"
-        ], 
         "hive_metastore_hosts": [
             "c6402.ambari.apache.org"
         ], 

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
index d9f9f71..3448b83 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
@@ -18,8 +18,8 @@
         "ambari_db_rca_username": "mapred", 
         "java_home": "/usr/jdk64/jdk1.7.0_45", 
         "mysql_jdbc_url": "http://c6401.ambari.apache.org:8080/resources//mysql-connector-java.jar",
-        "group_list": "[\"hadoop\",\"nobody\",\"users\",\"nagios\"]",
-        "user_list": "[\"hive\",\"oozie\",\"nobody\",\"nagios\",\"ambari-qa\",\"flume\",\"hdfs\",\"storm\",\"mapred\",\"hbase\",\"tez\",\"zookeeper\",\"falcon\",\"sqoop\",\"yarn\",\"hcat\"]"
+        "group_list": "[\"hadoop\",\"nobody\",\"users\"]",
+        "user_list": "[\"hive\",\"oozie\",\"nobody\",\"ambari-qa\",\"flume\",\"hdfs\",\"storm\",\"mapred\",\"hbase\",\"tez\",\"zookeeper\",\"falcon\",\"sqoop\",\"yarn\",\"hcat\"]"
     }, 
     "commandType": "EXECUTION_COMMAND", 
     "roleParams": {}, 
@@ -335,14 +335,6 @@
             "jtnode_opt_maxnewsize": "200m",
             "rca_properties": "\nambari.jobhistory.database={ambari_db_rca_url}\nambari.jobhistory.driver={ambari_db_rca_driver}\nambari.jobhistory.user={ambari_db_rca_username}\nambari.jobhistory.password={ambari_db_rca_password}\nambari.jobhistory.logger=${{hadoop.root.logger}}\n\nlog4j.appender.JHA=org.apache.ambari.log4j.hadoop.mapreduce.jobhistory.JobHistoryAppender\nlog4j.appender.JHA.database={ambari_db_rca_url}\nlog4j.appender.JHA.driver={ambari_db_rca_driver}\nlog4j.appender.JHA.user={ambari_db_rca_username}\nlog4j.appender.JHA.password={ambari_db_rca_password}\n\nlog4j.logger.org.apache.hadoop.mapred.JobHistory$JobHistoryLogger=DEBUG,JHA\nlog4j.additivity.org.apache.hadoop.mapred.JobHistory$JobHistoryLogger=true\n\n"
         }, 
-        "nagios-env": {
-            "hive_metastore_user_passwd": "password", 
-            "nagios_web_password": "!`\"' 1",
-            "nagios_user": "nagios", 
-            "nagios_group": "nagios", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "user@com.ua"
-        }, 
         "hive-env": {
             "hive_metastore_user_passwd": "password", 
             "hcat_pid_dir": "/var/run/webhcat", 
@@ -559,9 +551,6 @@
             "c6401.ambari.apache.org", 
             "c6402.ambari.apache.org"
         ], 
-        "nagios_server_host": [
-            "c6402.ambari.apache.org"
-        ], 
         "hive_metastore_hosts": [
             "c6402.ambari.apache.org"
         ], 

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
index 2ec3a22..fa8d183 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
@@ -326,14 +326,6 @@
             "hadoop_heapsize": "1024", 
             "jtnode_opt_maxnewsize": "200m"
         }, 
-        "nagios-env": {
-            "hive_metastore_user_passwd": "password", 
-            "nagios_web_password": "!`\"' 1",
-            "nagios_user": "nagios", 
-            "nagios_group": "nagios", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "user@com.ua"
-        }, 
         "hive-env": {
             "hive_metastore_user_passwd": "password", 
             "hcat_pid_dir": "/var/run/webhcat", 
@@ -541,9 +533,6 @@
             "c6401.ambari.apache.org", 
             "c6402.ambari.apache.org"
         ], 
-        "nagios_server_host": [
-            "c6402.ambari.apache.org"
-        ], 
         "hive_metastore_hosts": [
             "c6402.ambari.apache.org"
         ], 

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
index da7a878..77a06f2 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
@@ -326,14 +326,6 @@
             "hadoop_heapsize": "1024", 
             "jtnode_opt_maxnewsize": "200m"
         }, 
-        "nagios-env": {
-            "hive_metastore_user_passwd": "password", 
-            "nagios_web_password": "!`\"' 1",
-            "nagios_user": "nagios", 
-            "nagios_group": "nagios", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "user@com.ua"
-        }, 
         "hive-env": {
             "hive_metastore_user_passwd": "password", 
             "hcat_pid_dir": "/var/run/webhcat", 
@@ -544,9 +536,6 @@
             "c6401.ambari.apache.org", 
             "c6402.ambari.apache.org"
         ], 
-        "nagios_server_host": [
-            "c6402.ambari.apache.org"
-        ], 
         "hive_metastore_hosts": [
             "c6402.ambari.apache.org"
         ], 

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json b/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
index 3c08702..1278350 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
@@ -170,15 +170,13 @@
             "tasktracker_task_controller": "org.apache.hadoop.mapred.LinuxTaskController", 
             "oozie_keytab": "/etc/security/keytabs/oozie.service.keytab", 
             "hadoop_http_principal_name": "HTTP/_HOST", 
-            "kinit_path_local": "/usr/bin", 
-            "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab", 
+            "kinit_path_local": "/usr/bin",  
             "hbase_regionserver_heapsize": "1024m",
             "hbase_regionserver_xmn_max": "512",
             "hbase_regionserver_xmn_ratio": "0.2",
             "datanode_primary_name": "dn", 
             "namenode_principal_name": "nn/_HOST", 
-            "namenode_keytab": "/etc/security/keytabs/nn.service.keytab", 
-            "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM", 
+            "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",  
             "dfs_datanode_http_address": "1022",
             "hbase_user_keytab": "/etc/security/keytabs/hbase.headless.keytab", 
             "jobtracker_primary_name": "jt", 
@@ -188,8 +186,7 @@
             "clientPort": "2181", 
             "oozie_jdbc_driver": "org.apache.derby.jdbc.EmbeddedDriver", 
             "hive_metastore_primary_name": "hive", 
-            "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab", 
-            "nagios_primary_name": "nagios", 
+            "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab",  
             "jobtracker_principal_name": "jt/_HOST", 
             "hive_database": "New MySQL Database", 
             "hcat_pid_dir": "/var/run/webhcat",
@@ -198,7 +195,6 @@
             "oozie_pid_dir": "/var/run/oozie", 
             "datanode_principal_name": "dn/_HOST", 
             "hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab", 
-            "nagios_group": "nagios", 
             "hcat_user": "hcat", 
             "hadoop_heapsize": "1024", 
             "hbase_regionserver_primary_name": "hbase", 
@@ -219,7 +215,6 @@
             "smokeuser_keytab": "/etc/security/keytabs/smokeuser.headless.keytab", 
             "kerberos_domain": "EXAMPLE.COM", 
             "snamenode_keytab": "/etc/security/keytabs/nn.service.keytab", 
-            "nagios_server": "c6402.ambari.apache.org", 
             "ganglia_runtime_dir": "/var/run/ganglia/hdp", 
             "lzo_enabled": "true", 
             "oozie_principal_name": "oozie/c6402.ambari.apache.org", 
@@ -241,7 +236,6 @@
             "gmetad_user": "nobody", 
             "oozie_http_keytab": "/etc/security/keytabs/spnego.service.keytab", 
             "hive_metastore": "c6402.ambari.apache.org", 
-            "nagios_user": "nagios", 
             "security_enabled": "true", 
             "proxyuser_group": "users", 
             "namenode_formatted_mark_dir": "/var/run/hadoop/hdfs/namenode/formatted/", 
@@ -261,8 +255,6 @@
             "jtnode_heapsize": "1024m", 
             "yarn_user": "yarn", 
             "gmond_user": "nobody", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "q@q.q", 
             "snamenode_primary_name": "nn", 
             "hdfs_user": "hdfs", 
             "oozie_database_type": "derby", 
@@ -290,7 +282,6 @@
             "user_group": "hadoop", 
             "hive_user": "hive", 
             "webHCat_http_primary_name": "HTTP", 
-            "nagios_web_password": "!`\"' 1", 
             "smokeuser": "ambari-qa", 
             "ganglia_conf_dir": "/etc/ganglia/hdp", 
             "hbase_master_heapsize": "1024m", 
@@ -508,16 +499,6 @@
             "hadoop_heapsize": "1024", 
             "jtnode_opt_maxnewsize": "200m"
         }, 
-        "nagios-env": {
-            "hive_metastore_user_passwd": "password", 
-            "nagios_web_password": "password", 
-            "nagios_user": "nagios", 
-            "nagios_group": "nagios", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "user@com.ua",
-            "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM",
-            "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab"
-        }, 
         "hive-env": {
             "hive_metastore_user_passwd": "password", 
             "hcat_pid_dir": "/var/run/webhcat", 
@@ -740,9 +721,6 @@
             "c6401.ambari.apache.org", 
             "c6402.ambari.apache.org"
         ], 
-        "nagios_server_host": [
-            "c6402.ambari.apache.org"
-        ], 
         "hive_metastore_hosts": [
             "c6402.ambari.apache.org"
         ], 

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
index 74ea204..b41af13 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
@@ -170,15 +170,13 @@
             "tasktracker_task_controller": "org.apache.hadoop.mapred.LinuxTaskController", 
             "oozie_keytab": "/etc/security/keytabs/oozie.service.keytab", 
             "hadoop_http_principal_name": "HTTP/_HOST", 
-            "kinit_path_local": "/usr/bin", 
-            "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab", 
+            "kinit_path_local": "/usr/bin",  
             "hbase_regionserver_heapsize": "1024m",
             "hbase_regionserver_xmn_max": "512",
             "hbase_regionserver_xmn_ratio": "0.2",
             "datanode_primary_name": "dn", 
             "namenode_principal_name": "nn/_HOST", 
-            "namenode_keytab": "/etc/security/keytabs/nn.service.keytab", 
-            "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM", 
+            "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",  
             "dfs_datanode_http_address": "1022", 
             "hbase_user_keytab": "/etc/security/keytabs/hbase.headless.keytab", 
             "jobtracker_primary_name": "jt", 
@@ -188,8 +186,7 @@
             "clientPort": "2181", 
             "oozie_jdbc_driver": "org.apache.derby.jdbc.EmbeddedDriver", 
             "hive_metastore_primary_name": "hive", 
-            "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab", 
-            "nagios_primary_name": "nagios", 
+            "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab",  
             "jobtracker_principal_name": "jt/_HOST", 
             "hive_database": "New MySQL Database", 
             "hcat_pid_dir": "/var/run/webhcat",
@@ -197,8 +194,7 @@
             "snappy_enabled": "true", 
             "oozie_pid_dir": "/var/run/oozie", 
             "datanode_principal_name": "dn/_HOST", 
-            "hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab", 
-            "nagios_group": "nagios", 
+            "hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab",  
             "hcat_user": "hcat", 
             "hadoop_heapsize": "1024", 
             "hbase_regionserver_primary_name": "hbase", 
@@ -218,8 +214,7 @@
             "namenode_heapsize": "1024m", 
             "smokeuser_keytab": "/etc/security/keytabs/smokeuser.headless.keytab", 
             "kerberos_domain": "EXAMPLE.COM", 
-            "snamenode_keytab": "/etc/security/keytabs/nn.service.keytab", 
-            "nagios_server": "c6402.ambari.apache.org", 
+            "snamenode_keytab": "/etc/security/keytabs/nn.service.keytab",  
             "ganglia_runtime_dir": "/var/run/ganglia/hdp", 
             "lzo_enabled": "true", 
             "oozie_principal_name": "oozie/c6402.ambari.apache.org", 
@@ -240,8 +235,7 @@
             "hadoop_http_keytab": "/etc/security/keytabs/spnego.service.keytab", 
             "gmetad_user": "nobody", 
             "oozie_http_keytab": "/etc/security/keytabs/spnego.service.keytab", 
-            "hive_metastore": "c6402.ambari.apache.org", 
-            "nagios_user": "nagios", 
+            "hive_metastore": "c6402.ambari.apache.org",  
             "security_enabled": "true", 
             "proxyuser_group": "users", 
             "namenode_formatted_mark_dir": "/var/run/hadoop/hdfs/namenode/formatted/", 
@@ -261,8 +255,6 @@
             "jtnode_heapsize": "1024m", 
             "yarn_user": "yarn", 
             "gmond_user": "nobody", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "q@q.q", 
             "snamenode_primary_name": "nn", 
             "hdfs_user": "hdfs", 
             "oozie_database_type": "derby", 
@@ -289,8 +281,7 @@
             "hbase_log_dir": "/var/log/hbase", 
             "user_group": "hadoop", 
             "hive_user": "hive", 
-            "webHCat_http_primary_name": "HTTP", 
-            "nagios_web_password": "!`\"' 1", 
+            "webHCat_http_primary_name": "HTTP",  
             "smokeuser": "ambari-qa", 
             "ganglia_conf_dir": "/etc/ganglia/hdp", 
             "hbase_master_heapsize": "1024m", 
@@ -506,16 +497,6 @@
             "hadoop_heapsize": "1024", 
             "jtnode_opt_maxnewsize": "200m"
         }, 
-        "nagios-env": {
-            "hive_metastore_user_passwd": "password", 
-            "nagios_web_password": "password", 
-            "nagios_user": "nagios", 
-            "nagios_group": "nagios", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "user@com.ua",
-            "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM",
-            "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab"
-        }, 
         "hive-env": {
             "hive_metastore_user_passwd": "password", 
             "hcat_pid_dir": "/var/run/webhcat", 
@@ -733,9 +714,6 @@
             "c6401.ambari.apache.org", 
             "c6402.ambari.apache.org"
         ], 
-        "nagios_server_host": [
-            "c6402.ambari.apache.org"
-        ], 
         "hive_metastore_hosts": [
             "c6402.ambari.apache.org"
         ], 

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
index 0412d50..124db7c 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
@@ -169,15 +169,13 @@
             "tasktracker_task_controller": "org.apache.hadoop.mapred.LinuxTaskController", 
             "oozie_keytab": "/etc/security/keytabs/oozie.service.keytab", 
             "hadoop_http_principal_name": "HTTP/_HOST", 
-            "kinit_path_local": "/usr/bin", 
-            "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab", 
+            "kinit_path_local": "/usr/bin",  
             "hbase_regionserver_heapsize": "1024m",
             "hbase_regionserver_xmn_max": "512",
             "hbase_regionserver_xmn_ratio": "0.2",
             "datanode_primary_name": "dn", 
             "namenode_principal_name": "nn/_HOST", 
-            "namenode_keytab": "/etc/security/keytabs/nn.service.keytab", 
-            "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM", 
+            "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",  
             "dfs_datanode_http_address": "1022", 
             "hbase_user_keytab": "/etc/security/keytabs/hbase.headless.keytab", 
             "jobtracker_primary_name": "jt", 
@@ -188,7 +186,6 @@
             "oozie_jdbc_driver": "org.apache.derby.jdbc.EmbeddedDriver", 
             "hive_metastore_primary_name": "hive", 
             "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab", 
-            "nagios_primary_name": "nagios", 
             "jobtracker_principal_name": "jt/_HOST", 
             "hive_database": "New MySQL Database", 
             "hcat_pid_dir": "/etc/run/webhcat", 
@@ -197,7 +194,6 @@
             "oozie_pid_dir": "/var/run/oozie", 
             "datanode_principal_name": "dn/_HOST", 
             "hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab", 
-            "nagios_group": "nagios", 
             "hcat_user": "hcat", 
             "hadoop_heapsize": "1024", 
             "hbase_regionserver_primary_name": "hbase", 
@@ -217,7 +213,6 @@
             "smokeuser_keytab": "/etc/security/keytabs/smokeuser.headless.keytab", 
             "kerberos_domain": "EXAMPLE.COM", 
             "snamenode_keytab": "/etc/security/keytabs/nn.service.keytab", 
-            "nagios_server": "c6402.ambari.apache.org", 
             "ganglia_runtime_dir": "/var/run/ganglia/hdp", 
             "lzo_enabled": "true", 
             "oozie_principal_name": "oozie/c6402.ambari.apache.org", 
@@ -238,8 +233,7 @@
             "hadoop_http_keytab": "/etc/security/keytabs/spnego.service.keytab", 
             "gmetad_user": "nobody", 
             "oozie_http_keytab": "/etc/security/keytabs/spnego.service.keytab", 
-            "hive_metastore": "c6402.ambari.apache.org", 
-            "nagios_user": "nagios", 
+            "hive_metastore": "c6402.ambari.apache.org",  
             "security_enabled": "true", 
             "proxyuser_group": "users", 
             "namenode_formatted_mark_dir": "/var/run/hadoop/hdfs/namenode/formatted/", 
@@ -259,8 +253,6 @@
             "jtnode_heapsize": "1024m", 
             "yarn_user": "yarn", 
             "gmond_user": "nobody", 
-            "nagios_web_login": "nagiosadmin", 
-            "nagios_contact": "q@q.q", 
             "snamenode_primary_name": "nn", 
             "hdfs_user": "hdfs", 
             "oozie_database_type": "derby", 
@@ -288,7 +280,6 @@
             "user_group": "hadoop", 
             "hive_user": "hive", 
             "webHCat_http_primary_name": "HTTP", 
-            "nagios_web_password": "!`\"' 1", 
             "smokeuser": "ambari-qa", 
             "ganglia_conf_dir": "/etc/ganglia/hdp", 
             "hbase_master_heapsize": "1024m", 
@@ -591,9 +582,6 @@
             "c6401.ambari.apache.org", 
             "c6402.ambari.apache.org"
         ], 
-        "nagios_server_host": [
-            "c6402.ambari.apache.org"
-        ], 
         "hive_metastore_hosts": [
             "c6402.ambari.apache.org"
         ], 

http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py b/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
index 836274a..16aa939 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
+++ b/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
@@ -45,9 +45,6 @@ class TestHookBeforeInstall(RMFTestCase):
     self.assertResourceCalled('Group', 'users',
         ignore_failures = False,
     )
-    self.assertResourceCalled('Group', 'nagios',
-        ignore_failures = False,
-    )
     self.assertResourceCalled('User', 'hive',
         gid = 'hadoop',
         ignore_failures = False,
@@ -63,11 +60,6 @@ class TestHookBeforeInstall(RMFTestCase):
         ignore_failures = False,
         groups = [u'nobody'],
     )
-    self.assertResourceCalled('User', 'nagios',
-        gid = 'nagios',
-        ignore_failures = False,
-        groups = [u'hadoop'],
-    )
     self.assertResourceCalled('User', 'ambari-qa',
         gid = 'hadoop',
         ignore_failures = False,