You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by jo...@apache.org on 2014/11/11 21:43:08 UTC
[05/17] ambari git commit: AMBARI-8276 - Alerts: Remove Nagios
Service From The Stack (jonathanhurley)
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
index 8095048..ff5539f 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleCommandOrderTest.java
@@ -20,8 +20,8 @@ package org.apache.ambari.server.metadata;
import static junit.framework.Assert.assertEquals;
import static junit.framework.Assert.assertFalse;
-import static junit.framework.Assert.assertTrue;
import static junit.framework.Assert.assertNotNull;
+import static junit.framework.Assert.assertTrue;
import static org.easymock.EasyMock.createMock;
import static org.easymock.EasyMock.expect;
import static org.easymock.EasyMock.replay;
@@ -31,11 +31,11 @@ import java.io.IOException;
import java.io.InputStream;
import java.util.Collections;
import java.util.HashMap;
-import java.util.HashSet;
import java.util.Map;
import java.util.Set;
import junit.framework.Assert;
+
import org.apache.ambari.server.AmbariException;
import org.apache.ambari.server.Role;
import org.apache.ambari.server.RoleCommand;
@@ -52,7 +52,6 @@ import org.codehaus.jackson.annotate.JsonAutoDetect;
import org.codehaus.jackson.annotate.JsonMethod;
import org.codehaus.jackson.map.ObjectMapper;
import org.codehaus.jackson.type.TypeReference;
-import org.easymock.IExpectationSetters;
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
@@ -274,16 +273,16 @@ public class RoleCommandOrderTest {
InputStream testJsonIS = getClass().getClassLoader().
getResourceAsStream(TEST_RCO_DATA_FILE);
-
+
ObjectMapper mapper = new ObjectMapper();
Map<String,Object> testData = mapper.readValue(testJsonIS,
new TypeReference<Map<String,Object>>() {});
-
+
rco.addDependencies(testData);
mapper.setVisibility(JsonMethod.ALL, JsonAutoDetect.Visibility.ANY);
String dump = mapper.writeValueAsString(rco.getDependencies());
-
+
String expected = "{\"RoleCommandPair{role=SECONDARY_NAMENODE, " +
"cmd=UPGRADE}\":[{\"role\":{\"name\":\"NAMENODE\"},\"cmd\":\"UPGRADE\"}]," +
"\"RoleCommandPair{role=SECONDARY_NAMENODE, cmd=START}\":[{\"role\":{\"name\":\"NAMENODE\"}," +
@@ -295,8 +294,8 @@ public class RoleCommandOrderTest {
assertEquals(expected, dump);
}
-
-
+
+
@Test
public void testInitializeDefault() throws IOException {
RoleCommandOrder rco = injector.getInstance(RoleCommandOrder.class);
@@ -313,9 +312,9 @@ public class RoleCommandOrderTest {
replay(cluster);
replay(hdfsService);
-
+
rco.initialize(cluster);
-
+
verify(cluster);
verify(hdfsService);
}
@@ -333,19 +332,20 @@ public class RoleCommandOrderTest {
Map<String,ServiceComponent> hdfsComponents = Collections.singletonMap("NAMENODE", namenode);
expect(hdfsService.getServiceComponents()).andReturn(hdfsComponents).anyTimes();
- Service nagiosService = createMock(Service.class);
- expect(cluster.getService("NAGIOS")).andReturn(nagiosService).atLeastOnce();
- expect(nagiosService.getCluster()).andReturn(cluster).anyTimes();
+ Service hbaseService = createMock(Service.class);
+ expect(cluster.getService("HBASE")).andReturn(hbaseService).atLeastOnce();
+ expect(hbaseService.getCluster()).andReturn(cluster).anyTimes();
- ServiceComponent nagiosServer = createMock(ServiceComponent.class);
- expect(nagiosServer.getName()).andReturn("NAGIOS_SERVER").anyTimes();
+ ServiceComponent hbaseMaster = createMock(ServiceComponent.class);
+ expect(hbaseMaster.getName()).andReturn("HBASE_MASTER").anyTimes();
- Map<String,ServiceComponent> nagiosComponents = Collections.singletonMap("NAGIOS_SERVER", nagiosServer);
- expect(nagiosService.getServiceComponents()).andReturn(nagiosComponents).anyTimes();
+ Map<String, ServiceComponent> hbaseComponents = Collections.singletonMap(
+ "HBASE_MASTER", hbaseMaster);
+ expect(hbaseService.getServiceComponents()).andReturn(hbaseComponents).anyTimes();
Map<String, Service> installedServices = new HashMap<String, Service>();
installedServices.put("HDFS", hdfsService);
- installedServices.put("NAGIOS", nagiosService);
+ installedServices.put("HBASE", hbaseService);
expect(cluster.getServices()).andReturn(installedServices).atLeastOnce();
@@ -356,14 +356,14 @@ public class RoleCommandOrderTest {
expect(cluster.getCurrentStackVersion()).andReturn(new StackId("HDP", "2.0.5"));
//replay
- replay(cluster, hdfsService, nagiosService, nagiosServer, namenode);
+ replay(cluster, hdfsService, hbaseService, hbaseMaster, namenode);
rco.initialize(cluster);
- Set<Service> transitiveServices =
- rco.getTransitiveServices(cluster.getService("NAGIOS"), RoleCommand.START);
+ Set<Service> transitiveServices = rco.getTransitiveServices(
+ cluster.getService("HBASE"), RoleCommand.START);
- //HDFS should be started before NAGIOS start
+ // HDFS should be started before HBASE start
Assert.assertNotNull(transitiveServices);
Assert.assertFalse(transitiveServices.isEmpty());
Assert.assertEquals(1, transitiveServices.size());
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
index 8d68f9c..8bead43 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/metadata/RoleGraphTest.java
@@ -21,37 +21,35 @@ package org.apache.ambari.server.metadata;
import static org.mockito.Mockito.mock;
import static org.mockito.Mockito.when;
-
import junit.framework.Assert;
import org.apache.ambari.server.AmbariException;
import org.apache.ambari.server.Role;
import org.apache.ambari.server.RoleCommand;
+import org.apache.ambari.server.orm.GuiceJpaInitializer;
+import org.apache.ambari.server.orm.InMemoryDefaultTestModule;
import org.apache.ambari.server.stageplanner.RoleGraphNode;
-import org.junit.After;
-import org.junit.Test;
import org.apache.ambari.server.state.Service;
import org.apache.ambari.server.state.ServiceComponent;
import org.apache.ambari.server.state.StackId;
+import org.apache.ambari.server.state.cluster.ClusterImpl;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
import com.google.inject.Guice;
import com.google.inject.Injector;
import com.google.inject.persist.PersistService;
-import org.apache.ambari.server.state.cluster.ClusterImpl;
-import org.apache.ambari.server.orm.GuiceJpaInitializer;
-import org.apache.ambari.server.orm.InMemoryDefaultTestModule;
-import org.junit.Before;
-
public class RoleGraphTest {
-
+
private Injector injector;
-
+
@Before
public void setup() throws Exception {
injector = Guice.createInjector(new InMemoryDefaultTestModule());
- injector.getInstance(GuiceJpaInitializer.class);
+ injector.getInstance(GuiceJpaInitializer.class);
}
@After
@@ -63,9 +61,9 @@ public class RoleGraphTest {
public void testValidateOrder() throws AmbariException {
RoleCommandOrder rco = injector.getInstance(RoleCommandOrder.class);
ClusterImpl cluster = mock(ClusterImpl.class);
-
+
when(cluster.getCurrentStackVersion()).thenReturn(new StackId("HDP-2.0.6"));
-
+
rco.initialize(cluster);
RoleGraphNode datanode_upgrade = new RoleGraphNode(Role.DATANODE, RoleCommand.UPGRADE);
@@ -82,25 +80,10 @@ public class RoleGraphTest {
RoleGraphNode datanode_start = new RoleGraphNode(Role.DATANODE, RoleCommand.START);
RoleGraphNode datanode_install = new RoleGraphNode(Role.DATANODE, RoleCommand.INSTALL);
RoleGraphNode jobtracker_start = new RoleGraphNode(Role.JOBTRACKER, RoleCommand.START);
- RoleGraphNode tasktracker_start = new RoleGraphNode(Role.TASKTRACKER, RoleCommand.START);
Assert.assertEquals(1, rco.order(datanode_start, datanode_install));
Assert.assertEquals(1, rco.order(jobtracker_start, datanode_start));
Assert.assertEquals(0, rco.order(jobtracker_start, jobtracker_start));
- RoleGraphNode hive_client_install = new RoleGraphNode(Role.HIVE_CLIENT,
- RoleCommand.INSTALL);
- RoleGraphNode mapred_client_install = new RoleGraphNode(Role.MAPREDUCE_CLIENT,
- RoleCommand.INSTALL);
- RoleGraphNode hcat_client_install = new RoleGraphNode(Role.HCAT,
- RoleCommand.INSTALL);
- RoleGraphNode nagios_server_install = new RoleGraphNode(Role.NAGIOS_SERVER,
- RoleCommand.INSTALL);
- RoleGraphNode oozie_client_install = new RoleGraphNode(Role.OOZIE_CLIENT,
- RoleCommand.INSTALL);
- Assert.assertEquals(1, rco.order(nagios_server_install, hive_client_install));
- Assert.assertEquals(1, rco.order(nagios_server_install, mapred_client_install));
- Assert.assertEquals(1, rco.order(nagios_server_install, hcat_client_install));
- Assert.assertEquals(1, rco.order(nagios_server_install, oozie_client_install));
RoleGraphNode pig_service_check = new RoleGraphNode(Role.PIG_SERVICE_CHECK, RoleCommand.SERVICE_CHECK);
RoleGraphNode resourcemanager_start = new RoleGraphNode(Role.RESOURCEMANAGER, RoleCommand.START);
@@ -109,78 +92,50 @@ public class RoleGraphTest {
RoleGraphNode hdfs_service_check = new RoleGraphNode(Role.HDFS_SERVICE_CHECK, RoleCommand.SERVICE_CHECK);
RoleGraphNode snamenode_start = new RoleGraphNode(Role.SECONDARY_NAMENODE, RoleCommand.START);
Assert.assertEquals(-1, rco.order(snamenode_start, hdfs_service_check));
-
+
RoleGraphNode mapred2_service_check = new RoleGraphNode(Role.MAPREDUCE2_SERVICE_CHECK, RoleCommand.SERVICE_CHECK);
RoleGraphNode rm_start = new RoleGraphNode(Role.RESOURCEMANAGER, RoleCommand.START);
RoleGraphNode nm_start = new RoleGraphNode(Role.NODEMANAGER, RoleCommand.START);
RoleGraphNode hs_start = new RoleGraphNode(Role.HISTORYSERVER, RoleCommand.START);
- RoleGraphNode nagios_start = new RoleGraphNode(Role.NAGIOS_SERVER, RoleCommand.START);
+
Assert.assertEquals(-1, rco.order(rm_start, mapred2_service_check));
- Assert.assertEquals(-1, rco.order(nm_start, mapred2_service_check));
+ Assert.assertEquals(-1, rco.order(nm_start, mapred2_service_check));
Assert.assertEquals(-1, rco.order(hs_start, mapred2_service_check));
Assert.assertEquals(-1, rco.order(hs_start, mapred2_service_check));
Assert.assertEquals(1, rco.order(nm_start, rm_start));
-
+
//Non-HA mode
RoleGraphNode nn_start = new RoleGraphNode(Role.NAMENODE, RoleCommand.START);
RoleGraphNode jn_start = new RoleGraphNode(Role.JOURNALNODE, RoleCommand.START);
RoleGraphNode zk_server_start = new RoleGraphNode(Role.ZOOKEEPER_SERVER, RoleCommand.START);
RoleGraphNode hbase_master_start = new RoleGraphNode(Role.HBASE_MASTER, RoleCommand.START);
- RoleGraphNode hbase_reg_srv_start = new RoleGraphNode(Role.HBASE_REGIONSERVER, RoleCommand.START);
- RoleGraphNode ganglia_server_start = new RoleGraphNode(Role.GANGLIA_SERVER, RoleCommand.START);
- RoleGraphNode ganglia_monitor_start = new RoleGraphNode(Role.GANGLIA_MONITOR, RoleCommand.START);
- RoleGraphNode hcat_start = new RoleGraphNode(Role.HCAT, RoleCommand.START);
RoleGraphNode hive_srv_start = new RoleGraphNode(Role.HIVE_SERVER, RoleCommand.START);
RoleGraphNode hive_ms_start = new RoleGraphNode(Role.HIVE_METASTORE, RoleCommand.START);
- RoleGraphNode hue_start = new RoleGraphNode(Role.HUE_SERVER, RoleCommand.START);
RoleGraphNode mysql_start = new RoleGraphNode(Role.MYSQL_SERVER, RoleCommand.START);
RoleGraphNode oozie_srv_start = new RoleGraphNode(Role.OOZIE_SERVER, RoleCommand.START);
- RoleGraphNode pig_start = new RoleGraphNode(Role.PIG, RoleCommand.START);
- RoleGraphNode sqoop_start = new RoleGraphNode(Role.SQOOP, RoleCommand.START);
RoleGraphNode webhcat_srv_start = new RoleGraphNode(Role.WEBHCAT_SERVER, RoleCommand.START);
RoleGraphNode flume_start = new RoleGraphNode(Role.FLUME_HANDLER, RoleCommand.START);
RoleGraphNode zkfc_start = new RoleGraphNode(Role.ZKFC, RoleCommand.START);
-
+
Assert.assertEquals(0, rco.order(nn_start, jn_start));
Assert.assertEquals(0, rco.order(nn_start, zk_server_start));
Assert.assertEquals(0, rco.order(zkfc_start, nn_start));
- // Check that Nagios starts after other components
- Assert.assertEquals(1, rco.order(nagios_start, nn_start));
- Assert.assertEquals(1, rco.order(nagios_start, snamenode_start));
- Assert.assertEquals(1, rco.order(nagios_start, datanode_start));
- Assert.assertEquals(1, rco.order(nagios_start, resourcemanager_start));
- Assert.assertEquals(1, rco.order(nagios_start, nm_start));
- Assert.assertEquals(1, rco.order(nagios_start, hbase_master_start));
- Assert.assertEquals(1, rco.order(nagios_start, hbase_reg_srv_start));
- Assert.assertEquals(1, rco.order(nagios_start, ganglia_server_start));
- Assert.assertEquals(1, rco.order(nagios_start, ganglia_monitor_start));
- Assert.assertEquals(1, rco.order(nagios_start, hcat_start));
- Assert.assertEquals(1, rco.order(nagios_start, hs_start));
- Assert.assertEquals(1, rco.order(nagios_start, hive_srv_start));
- Assert.assertEquals(1, rco.order(nagios_start, hue_start));
- Assert.assertEquals(1, rco.order(nagios_start, jobtracker_start));
- Assert.assertEquals(1, rco.order(nagios_start, tasktracker_start));
- Assert.assertEquals(1, rco.order(nagios_start, zk_server_start));
- Assert.assertEquals(1, rco.order(nagios_start, mysql_start));
- Assert.assertEquals(1, rco.order(nagios_start, oozie_srv_start));
- Assert.assertEquals(1, rco.order(nagios_start, pig_start));
- Assert.assertEquals(1, rco.order(nagios_start, sqoop_start));
- Assert.assertEquals(1, rco.order(nagios_start, webhcat_srv_start));
- Assert.assertEquals(1, rco.order(nagios_start, flume_start));
-
-
-
- //Enable HA for cluster
+
+ Assert.assertEquals(1, rco.order(flume_start, oozie_srv_start));
+ Assert.assertEquals(1, rco.order(hbase_master_start, zk_server_start));
+ Assert.assertEquals(1, rco.order(hive_srv_start, mysql_start));
+ Assert.assertEquals(1, rco.order(hive_ms_start, mysql_start));
+ Assert.assertEquals(1, rco.order(webhcat_srv_start, datanode_start));
+
+ // Enable HA for cluster
Service hdfsServiceMock = mock(Service.class);
ServiceComponent jnComponentMock = mock(ServiceComponent.class);
when(cluster.getService("HDFS")).thenReturn(hdfsServiceMock);
when(hdfsServiceMock.getServiceComponent("JOURNALNODE")).thenReturn(jnComponentMock);
-
+
rco.initialize(cluster);
Assert.assertEquals(1, rco.order(nn_start, jn_start));
Assert.assertEquals(1, rco.order(nn_start, zk_server_start));
Assert.assertEquals(1, rco.order(zkfc_start, nn_start));
- Assert.assertEquals(1, rco.order(nagios_start, zkfc_start));
- Assert.assertEquals(1, rco.order(nagios_start, jn_start));
}
}
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java b/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
index a6cbc6a..0502e1a 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/stack/StackManagerTest.java
@@ -18,6 +18,24 @@
package org.apache.ambari.server.stack;
+import static org.easymock.EasyMock.createNiceMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertNotNull;
+import static org.junit.Assert.assertNull;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+
+import java.io.File;
+import java.util.Collection;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Map;
+
import org.apache.ambari.server.AmbariException;
import org.apache.ambari.server.configuration.Configuration;
import org.apache.ambari.server.metadata.ActionMetadata;
@@ -33,24 +51,6 @@ import org.apache.ambari.server.state.stack.OsFamily;
import org.junit.BeforeClass;
import org.junit.Test;
-import java.io.File;
-import java.util.Collection;
-import java.util.HashMap;
-import java.util.HashSet;
-import java.util.List;
-import java.util.Map;
-
-import static org.easymock.EasyMock.createNiceMock;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertNull;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
-
/**
* StackManager unit tests.
*/
@@ -197,7 +197,6 @@ public class StackManagerTest {
expectedServices.add("HDFS");
expectedServices.add("HIVE");
expectedServices.add("MAPREDUCE2");
- expectedServices.add("NAGIOS");
expectedServices.add("OOZIE");
expectedServices.add("PIG");
expectedServices.add("SQOOP");
@@ -205,6 +204,7 @@ public class StackManagerTest {
expectedServices.add("ZOOKEEPER");
expectedServices.add("STORM");
expectedServices.add("FLUME");
+ expectedServices.add("FAKENAGIOS");
ServiceInfo pigService = null;
for (ServiceInfo service : services) {
@@ -391,16 +391,22 @@ public class StackManagerTest {
@Test
public void testMonitoringServicePropertyInheritance() throws Exception{
- StackInfo stack = stackManager.getStack("HDP", "2.0.7");
+ StackInfo stack = stackManager.getStack("HDP", "2.0.8");
Collection<ServiceInfo> allServices = stack.getServices();
assertEquals(13, allServices.size());
+
+ boolean monitoringServiceFound = false;
+
for (ServiceInfo serviceInfo : allServices) {
- if (serviceInfo.getName().equals("NAGIOS")) {
+ if (serviceInfo.getName().equals("FAKENAGIOS")) {
+ monitoringServiceFound = true;
assertTrue(serviceInfo.isMonitoringService());
} else {
assertNull(serviceInfo.isMonitoringService());
}
}
+
+ assertTrue(monitoringServiceFound);
}
@Test
@@ -408,7 +414,7 @@ public class StackManagerTest {
StackInfo stack = stackManager.getStack("HDP", "2.0.6");
Collection<ServiceInfo> allServices = stack.getServices();
- assertEquals(12, allServices.size());
+ assertEquals(11, allServices.size());
HashSet<String> expectedServices = new HashSet<String>();
expectedServices.add("GANGLIA");
expectedServices.add("HBASE");
@@ -416,7 +422,6 @@ public class StackManagerTest {
expectedServices.add("HDFS");
expectedServices.add("HIVE");
expectedServices.add("MAPREDUCE2");
- expectedServices.add("NAGIOS");
expectedServices.add("OOZIE");
expectedServices.add("PIG");
expectedServices.add("ZOOKEEPER");
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java b/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
index f008980..dd2a519 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/stageplanner/TestStagePlanner.java
@@ -62,7 +62,7 @@ public class TestStagePlanner {
RoleCommandOrder rco = injector.getInstance(RoleCommandOrder.class);
ClusterImpl cluster = mock(ClusterImpl.class);
when(cluster.getCurrentStackVersion()).thenReturn(new StackId("HDP-2.0.6"));
-
+
rco.initialize(cluster);
RoleGraph rg = new RoleGraph(rco);
@@ -140,9 +140,6 @@ public class TestStagePlanner {
stage.addHostRoleExecutionCommand("host7", Role.WEBHCAT_SERVER,
RoleCommand.START, new ServiceComponentHostStartEvent("WEBHCAT_SERVER",
"host7", now), "cluster1", "WEBHCAT");
- stage.addHostRoleExecutionCommand("host8", Role.NAGIOS_SERVER,
- RoleCommand.START, new ServiceComponentHostStartEvent("NAGIOS_SERVER",
- "host8", now), "cluster1", "NAGIOS");
stage.addHostRoleExecutionCommand("host4", Role.GANGLIA_MONITOR,
RoleCommand.START, new ServiceComponentHostStartEvent("GANGLIA_MONITOR",
"host4", now), "cluster1", "GANGLIA");
@@ -156,6 +153,6 @@ public class TestStagePlanner {
for (Stage s: outStages) {
System.out.println(s.toString());
}
- assertEquals(5, outStages.size());
+ assertEquals(4, outStages.size());
}
}
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py b/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py
deleted file mode 100644
index a75e92a..0000000
--- a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_mm_wrapper.py
+++ /dev/null
@@ -1,549 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-import StringIO
-
-import os, sys
-import pprint
-import subprocess
-from unittest import TestCase
-from mock.mock import Mock, MagicMock, patch
-import mm_wrapper
-
-class TestOrWrapper(TestCase):
-
- dummy_ignore_file = """
-vm-4.vm HIVE HIVE_METASTORE
-vm-5.vm GANGLIA GANGLIA_MONITOR
-vm-4.vm YARN NODEMANAGER
-vm-3.vm YARN NODEMANAGER
-vm-3.vm HBASE HBASE_REGIONSERVER
-vm-4.vm HBASE HBASE_REGIONSERVER
-vm-4.vm STORM STORM_REST_API
-vm-4.vm HDFS DATANODE
-vm-4.vm STORM SUPERVISOR
-vm-4.vm STORM NIMBUS
-vm-4.vm STORM STORM_UI_SERVER
-vm-3.vm STORM SUPERVISOR
-vm-4.vm HDFS SECONDARY_NAMENODE
-vm-3.vm FLUME FLUME_HANDLER
-vm-4.vm GANGLIA GANGLIA_SERVER
-vm-4.vm HIVE HIVE_SERVER
-vm-4.vm ZOOKEEPER ZOOKEEPER_SERVER
-vm-4.vm WEBHCAT WEBHCAT_SERVER
-vm-3.vm HBASE HBASE_MASTER
-vm-4.vm GANGLIA GANGLIA_MONITOR
-vm-3.vm GANGLIA GANGLIA_MONITOR
-vm-3.vm HDFS NAMENODE
-vm-4.vm HIVE MYSQL_SERVER
-vm-4.vm YARN APP_TIMELINE_SERVER
-vm-4.vm FALCON FALCON_SERVER
-vm-3.vm HDFS DATANODE
-vm-4.vm YARN RESOURCEMANAGER
-vm-4.vm OOZIE OOZIE_SERVER
-vm-4.vm MAPREDUCE2 HISTORYSERVER
-vm-4.vm STORM DRPC_SERVER
-vm-4.vm FLUME FLUME_HANDLER
-vm-3.vm ZOOKEEPER ZOOKEEPER_SERVER
-"""
-
- default_empty_check_result = {
- 'message': 'No checks have been run (no hostnames provided)',
- 'retcode': -1,
- 'real_retcode': None
- }
-
-
- @patch("__builtin__.open")
- def test_ignored_host_list(self, open_mock):
- # Check with empty file content
- open_mock.return_value.__enter__.return_value.read.return_value = ""
- lst = mm_wrapper.ignored_host_list('STORM', 'SUPERVISOR')
- self.assertEqual(pprint.pformat(lst), '[]')
- # Check with dummy content
- open_mock.return_value.__enter__.return_value.read.return_value = self.dummy_ignore_file
- lst = mm_wrapper.ignored_host_list('STORM', 'SUPERVISOR')
- self.assertEqual(pprint.pformat(lst), "['vm-4.vm', 'vm-3.vm']")
- # Check if service name/comp name are not defined
- open_mock.return_value.__enter__.return_value.read.return_value = self.dummy_ignore_file
- lst = mm_wrapper.ignored_host_list('', '')
- self.assertEqual(pprint.pformat(lst), "[]")
-
-
- @patch("sys.exit")
- def test_print_usage(self, exit_mock):
- mm_wrapper.print_usage()
- self.assertTrue(exit_mock.called)
- self.assertEqual(exit_mock.call_args_list[0][0][0], 1)
-
-
- def test_get_real_component(self):
- with patch.dict(os.environ, {'NAGIOS__SERVICEHOST_COMPONENT': 'SUPERVISOR'}, clear=True):
- component = mm_wrapper.get_real_component()
- self.assertEqual(component, 'SUPERVISOR')
- with patch.dict(os.environ, {'NAGIOS__SERVICEHOST_COMPONENT': 'JOBHISTORY'}, clear=True):
- component = mm_wrapper.get_real_component()
- self.assertEqual(component, 'MAPREDUCE2')
-
-
- @patch("mm_wrapper.print_usage")
- def test_parse_args(self, print_usage_mock):
- args = ['or', 'h1', 'h2', '--', 'prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- mode, hostnames, command_line = mm_wrapper.parse_args(args)
- self.assertEquals(mode, mm_wrapper.OR)
- self.assertEquals(hostnames, ['h1', 'h2'])
- self.assertEquals(command_line, ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt'])
-
- args = ['and', 'h1', 'h2', '--', 'prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- mode, hostnames, command_line = mm_wrapper.parse_args(args)
- self.assertEquals(mode, mm_wrapper.AND)
- self.assertEquals(hostnames, ['h1', 'h2'])
- self.assertEquals(command_line, ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt'])
-
- args = ['env_only', 'h1', 'h2', '--', 'prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- mode, hostnames, command_line = mm_wrapper.parse_args(args)
- self.assertEquals(mode, mm_wrapper.ENV_ONLY)
- self.assertEquals(hostnames, ['h1', 'h2'])
- self.assertEquals(command_line, ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt'])
-
- # Check wrong usage
- args = []
- mm_wrapper.parse_args(args)
- self.assertTrue(print_usage_mock.called)
-
-
- @patch("mm_wrapper.ignored_host_list")
- @patch("mm_wrapper.work_in_or_mode")
- @patch("mm_wrapper.work_in_and_mode")
- @patch("mm_wrapper.work_in_env_only_mode")
- @patch("mm_wrapper.work_in_filter_mm_mode")
- @patch("mm_wrapper.work_in_legacy_check_wrapper_mode")
- def test_do_work(self, work_in_legacy_check_wrapper_mode, work_in_filter_mm_mode_mock,
- work_in_env_only_mode_mock, work_in_and_mode_mock,
- work_in_or_mode_mock,
- ignored_host_list_mock):
- hostnames = ['h1', 'h2', 'h3', 'h4']
- ignored_host_list_mock.return_value = ['h2', 'h3']
- command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- work_in_or_mode_mock.return_value = {
- 'message': "or_mode mode result",
- 'retcode': 0,
- 'real_retcode': None
- }
- work_in_and_mode_mock.return_value = {
- 'message': "and_mode mode result",
- 'retcode': 0,
- 'real_retcode': None
- }
- work_in_env_only_mode_mock.return_value = {
- 'message': "env_only mode result",
- 'retcode': 0,
- 'real_retcode': None
- }
- work_in_filter_mm_mode_mock.return_value = {
- 'message': "filter_mm mode result",
- 'retcode': 0,
- 'real_retcode': None
- }
- work_in_legacy_check_wrapper_mode.return_value = {
- 'message': "legacy_check_wrapper mode result",
- 'retcode': 0,
- 'real_retcode': None
- }
- result = mm_wrapper.do_work(mm_wrapper.OR, hostnames, command_line)
- self.assertEquals(str(result), "(['or_mode mode result'], 0)")
-
- result = mm_wrapper.do_work(mm_wrapper.AND, hostnames, command_line)
- self.assertEquals(str(result), "(['and_mode mode result'], 0)")
-
- result = mm_wrapper.do_work(mm_wrapper.ENV_ONLY, hostnames, command_line)
- self.assertEquals(str(result), "(['env_only mode result'], 0)")
-
- result = mm_wrapper.do_work(mm_wrapper.FILTER_MM, hostnames, command_line)
- self.assertEquals(str(result), "(['filter_mm mode result'], 0)")
-
- result = mm_wrapper.do_work(mm_wrapper.LEGACY_CHECK_WRAPPER, hostnames, command_line)
- self.assertEquals(str(result), "(['legacy_check_wrapper mode result'], 0)")
-
- # Check behaviour when real_retcode is defined
- work_in_or_mode_mock.return_value = {
- 'message': "or_mode mode result",
- 'retcode': 0,
- 'real_retcode': 1
- }
- result = mm_wrapper.do_work(mm_wrapper.OR, hostnames, command_line)
- self.assertEquals(str(result), "(['or_mode mode result', 'AMBARIPASSIVE=1'], 0)")
-
-
- @patch("mm_wrapper.check_output")
- def test_work_in_or_mode(self, check_output_mock):
- hostnames = ['h1', 'h2', 'h3', 'h4']
- ignored_hosts = ['h2', 'h3']
- command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- custom_env = {'MM_HOSTS': ignored_hosts}
-
- # Normal usage
- check_output_mock.return_value = 'Dummy message'
- result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(str(result),
- "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
- self.assertEquals(check_output_mock.call_count, 1) # Exited on first success
- self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
- for check_tupple in zip(check_output_mock.call_args_list, hostnames):
- self.assertEquals(check_tupple[0][0][0], ['prog', '-h', check_tupple[1], '-opt', 'yet', 'another', 'opt'])
-
- check_output_mock.reset_mock()
-
- # Failed all checks
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(3, 'dummy cmd')
- error.output = 'dummy output2'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output3'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(2, 'dummy cmd')
- error.output = 'dummy output4'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
-
- result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 4)
- self.assertEquals(str(result),
- "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 2}")
-
- check_output_mock.reset_mock()
-
- # Failed all but MM host component checks
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(0, 'dummy cmd')
- error.output = 'dummy output2'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(2, 'dummy cmd')
- error.output = 'dummy output3'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(3, 'dummy cmd')
- error.output = 'dummy output4'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
-
- result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 4)
- self.assertEquals(str(result),
- "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 3}")
-
- check_output_mock.reset_mock()
-
- # Components check only for one check is successful
- ignored_hosts = []
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(0, 'dummy cmd')
- error.output = 'dummy output2'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(2, 'dummy cmd')
- error.output = 'dummy output3'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(3, 'dummy cmd')
- error.output = 'dummy output4'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
-
- result = mm_wrapper.work_in_or_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 2)
- self.assertEquals(str(result),
- "{'message': 'dummy output2', 'real_retcode': None, 'retcode': 0}")
-
-
- @patch("mm_wrapper.check_output")
- def test_work_in_and_mode(self, check_output_mock):
- hostnames = ['h1', 'h2', 'h3', 'h4']
- ignored_hosts = ['h2', 'h3']
- command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- custom_env = {'MM_HOSTS': ignored_hosts}
-
- # Normal usage
- check_output_mock.return_value = 'Dummy message'
- result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(str(result),
- "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
- self.assertEquals(check_output_mock.call_count, 4)
- self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
- for check_tupple in zip(check_output_mock.call_args_list, hostnames):
- self.assertEquals(check_tupple[0][0][0], ['prog', '-h', check_tupple[1], '-opt', 'yet', 'another', 'opt'])
-
- check_output_mock.reset_mock()
-
- # Failed all checks
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(3, 'dummy cmd')
- error.output = 'dummy output2'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output3'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(2, 'dummy cmd')
- error.output = 'dummy output4'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 4)
- self.assertEquals(str(result),
- "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 2}")
-
- check_output_mock.reset_mock()
-
- # Failed all but MM host component checks
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(0, 'dummy cmd')
- error.output = 'dummy output2'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(2, 'dummy cmd')
- error.output = 'dummy output3'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(3, 'dummy cmd')
- error.output = 'dummy output4'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 4)
- self.assertEquals(str(result),
- "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 3}")
-
- check_output_mock.reset_mock()
-
- # Components check only for one check is successful
- ignored_hosts = []
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(0, 'dummy cmd')
- error.output = 'dummy output2'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(2, 'dummy cmd')
- error.output = 'dummy output3'
- check_output_side_effects.append(error)
-
- error = subprocess.CalledProcessError(3, 'dummy cmd')
- error.output = 'dummy output4'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_and_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 4)
- self.assertEquals(str(result),
- "{'message': 'dummy output4', 'real_retcode': None, 'retcode': 3}")
-
-
- @patch("mm_wrapper.check_output")
- def test_work_in_env_only_mode(self, check_output_mock):
- hostnames = ['h1', 'h2', 'h3', 'h4']
- ignored_hosts = ['h2', 'h3']
- command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- custom_env = {'MM_HOSTS' : ignored_hosts}
-
- # Normal usage
- check_output_mock.return_value = 'Dummy message'
- result = mm_wrapper.work_in_env_only_mode(hostnames, command_line, custom_env)
- self.assertEquals(str(result),
- "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
- self.assertEquals(check_output_mock.call_count, 1)
- self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
- self.assertEquals(check_output_mock.call_args[0][0],
- ['prog', '-h', 'h1', 'h2', 'h3', 'h4', '-opt', 'yet', 'another', 'opt'])
-
- check_output_mock.reset_mock()
-
- # Failed all checks
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_env_only_mode(hostnames, command_line, custom_env)
- self.assertEquals(check_output_mock.call_count, 1)
- self.assertEquals(str(result),
- "{'message': 'dummy output1', 'real_retcode': None, 'retcode': 1}")
-
- check_output_mock.reset_mock()
-
-
- @patch("mm_wrapper.check_output")
- def test_work_in_filter_mm_mode(self, check_output_mock):
- hostnames = ['h1', 'h2', 'h3', 'h4']
- ignored_hosts = ['h2', 'h3']
- command_line = ['prog', '-h', '^^', '-opt', 'yet', 'another', 'opt']
- custom_env = {'MM_HOSTS' : ignored_hosts}
-
- # Normal usage
- check_output_mock.return_value = 'Dummy message'
- result = mm_wrapper.work_in_filter_mm_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(str(result),
- "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
- self.assertEquals(check_output_mock.call_count, 1)
- self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
- self.assertEquals(check_output_mock.call_args[0][0],
- ['prog', '-h', 'h1', 'h4', '-opt', 'yet', 'another', 'opt'])
-
- check_output_mock.reset_mock()
-
- # Failed all checks
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_filter_mm_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 1)
- self.assertEquals(str(result),
- "{'message': 'dummy output1', 'real_retcode': None, 'retcode': 1}")
-
- check_output_mock.reset_mock()
-
- # All host components are in MM
- ignored_hosts = hostnames
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_filter_mm_mode(hostnames, ignored_hosts, command_line,
- custom_env,
- self.default_empty_check_result)
- self.assertEquals(check_output_mock.call_count, 0)
- self.assertEquals(str(result),
- "{'message': 'No checks have been run (no hostnames provided)', "
- "'real_retcode': None, 'retcode': -1}")
-
- check_output_mock.reset_mock()
-
-
- @patch("mm_wrapper.check_output")
- @patch.dict(os.environ, {'NAGIOS_HOSTNAME': 'h2'}, clear=True)
- def test_work_in_legacy_check_wrapper_mode(self, check_output_mock):
- command_line = ['prog', '-opt', 'yet', 'another', 'opt']
- ignored_hosts = []
- custom_env = {'MM_HOSTS': ignored_hosts}
-
- # Normal usage
- ignored_hosts = []
- check_output_mock.return_value = 'Dummy message'
- result = mm_wrapper.work_in_legacy_check_wrapper_mode(ignored_hosts, command_line,
- custom_env)
- self.assertEquals(str(result),
- "{'message': 'Dummy message', 'real_retcode': None, 'retcode': 0}")
- self.assertEquals(check_output_mock.call_count, 1)
- self.assertEquals(check_output_mock.call_args[1]['env']['MM_HOSTS'], ignored_hosts)
- self.assertEquals(check_output_mock.call_args[0][0],
- ['prog', '-opt', 'yet', 'another', 'opt'])
-
- check_output_mock.reset_mock()
-
- # Failed check on host that is not in MM state
- ignored_hosts = ['h3']
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_legacy_check_wrapper_mode(ignored_hosts, command_line,
- custom_env)
- self.assertEquals(check_output_mock.call_count, 1)
- self.assertEquals(str(result),
- "{'message': 'dummy output1', 'real_retcode': None, 'retcode': 1}")
-
- check_output_mock.reset_mock()
-
- # Failed check on host that is in MM state
- ignored_hosts = ['h2']
- check_output_side_effects = []
- error = subprocess.CalledProcessError(1, 'dummy cmd')
- error.output = 'dummy output1'
- check_output_side_effects.append(error)
-
- check_output_mock.side_effect = check_output_side_effects
- result = mm_wrapper.work_in_legacy_check_wrapper_mode(ignored_hosts, command_line,
- custom_env)
- self.assertEquals(check_output_mock.call_count, 1)
- self.assertEquals(str(result),
- "{'message': 'dummy output1', 'real_retcode': 1, 'retcode': 0}")
-
- check_output_mock.reset_mock()
-
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py b/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py
deleted file mode 100644
index f1e4650..0000000
--- a/ambari-server/src/test/python/stacks/1.3.2/NAGIOS/test_nagios_server.py
+++ /dev/null
@@ -1,282 +0,0 @@
-#!/usr/bin/env python
-
-'''
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-'''
-
-from mock.mock import MagicMock, patch
-from stacks.utils.RMFTestCase import *
-
-
-class TestNagiosServer(RMFTestCase):
- def test_configure_default(self):
- self.executeScript("1.3.2/services/NAGIOS/package/scripts/nagios_server.py",
- classname="NagiosServer",
- command="configure",
- config_file="default.json"
- )
- self.assert_configure_default()
- self.assertNoMoreResources()
-
-
- def test_start_default(self):
- self.executeScript(
- "1.3.2/services/NAGIOS/package/scripts/nagios_service.py",
- classname="NagiosServer",
- command="start",
- config_file="default.json"
- )
- self.assert_configure_default()
- self.assertResourceCalled('Execute', 'service nagios start',
- path=['/usr/local/bin/:/bin/:/sbin/']
- )
- self.assertResourceCalled('MonitorWebserver', 'restart',
- )
- self.assertNoMoreResources()
-
-
- @patch('os.path.isfile')
- def test_stop_default(self, os_path_isfile_mock):
- src_dir = RMFTestCase._getSrcFolder()
- os_path_isfile_mock.side_effect = [False, True]
-
- self.executeScript(
- "1.3.2/services/NAGIOS/package/scripts/nagios_service.py",
- classname="NagiosServer",
- command="stop",
- config_file="default.json"
- )
-
- self.assertResourceCalled('Execute','service nagios stop', path=['/usr/local/bin/:/bin/:/sbin/'])
- self.assertResourceCalled('Execute','rm -f /var/run/nagios/nagios.pid')
- self.assertResourceCalled('MonitorWebserver', 'restart')
-
- self.assertNoMoreResources()
-
-
- def assert_configure_default(self):
- self.assertResourceCalled('File', '/etc/apache2/conf.d/nagios.conf',
- owner='nagios',
- group='nagios',
- content=Template("nagios.conf.j2"),
- mode=0644
- )
- self.assertResourceCalled('Directory', '/etc/nagios',
- owner='nagios',
- group='nagios',
- )
- self.assertResourceCalled('Directory', '/usr/lib64/nagios/plugins'
- )
- self.assertResourceCalled('Directory', '/etc/nagios/objects'
- )
- self.assertResourceCalled('Directory', '/var/run/nagios',
- owner='nagios',
- group='nagios',
- mode=0755,
- recursive=True
- )
- self.assertResourceCalled('Directory', '/var/nagios',
- owner='nagios',
- group='nagios',
- recursive=True
- )
- self.assertResourceCalled('Directory', '/var/nagios/spool/checkresults',
- owner='nagios',
- group='nagios',
- recursive=True
- )
- self.assertResourceCalled('Directory', '/var/nagios/rw',
- owner='nagios',
- group='nagios',
- recursive=True
- )
- self.assertResourceCalled('Directory', '/var/log/nagios',
- owner='nagios',
- group='nagios',
- mode=0755
- )
- self.assertResourceCalled('Directory', '/var/log/nagios/archives',
- owner='nagios',
- group='nagios',
- mode=0755
- )
- self.assertResourceCalled('TemplateConfig', '/etc/nagios/nagios.cfg',
- owner='nagios',
- group='nagios',
- mode=None
- )
- self.assertResourceCalled('TemplateConfig', '/etc/nagios/resource.cfg',
- owner='nagios',
- group='nagios',
- mode=None
- )
- self.assertResourceCalled('TemplateConfig',
- '/etc/nagios/objects/hadoop-hosts.cfg',
- owner='nagios',
- group='hadoop',
- mode=None
- )
- self.assertResourceCalled('TemplateConfig',
- '/etc/nagios/objects/hadoop-hostgroups.cfg',
- owner='nagios',
- group='hadoop',
- mode=None
- )
- self.assertResourceCalled('TemplateConfig',
- '/etc/nagios/objects/hadoop-servicegroups.cfg',
- owner='nagios',
- group='hadoop',
- mode=None
- )
- self.assertResourceCalled('TemplateConfig',
- '/etc/nagios/objects/hadoop-services.cfg',
- owner='nagios',
- group='hadoop',
- mode=None
- )
- self.assertResourceCalled('TemplateConfig',
- '/etc/nagios/objects/hadoop-commands.cfg',
- owner='nagios',
- group='hadoop',
- mode=None
- )
- self.assertResourceCalled('TemplateConfig',
- '/etc/nagios/objects/contacts.cfg',
- owner='nagios',
- group='hadoop',
- mode=None
- )
- self.assertResourceCalled('File', '/usr/lib64/nagios/plugins/check_cpu.pl',
- content=StaticFile('check_cpu.pl'),
- mode=0755
- )
- self.assertResourceCalled('File', '/usr/lib64/nagios/plugins/check_cpu.php',
- content=StaticFile('check_cpu.php'),
- mode=0755
- )
- self.assertResourceCalled('File', '/usr/lib64/nagios/plugins/check_cpu_ha.php',
- content=StaticFile('check_cpu_ha.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_datanode_storage.php',
- content=StaticFile('check_datanode_storage.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_aggregate.php',
- content=StaticFile('check_aggregate.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_hdfs_blocks.php',
- content=StaticFile('check_hdfs_blocks.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_hdfs_capacity.php',
- content=StaticFile('check_hdfs_capacity.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_rpcq_latency.php',
- content=StaticFile('check_rpcq_latency.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_webui.sh',
- content=StaticFile('check_webui.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_name_dir_status.php',
- content=StaticFile('check_name_dir_status.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_oozie_status.sh',
- content=StaticFile('check_oozie_status.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_templeton_status.sh',
- content=StaticFile('check_templeton_status.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_hive_metastore_status.sh',
- content=StaticFile(
- 'check_hive_metastore_status.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_hue_status.sh',
- content=StaticFile('check_hue_status.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_mapred_local_dir_used.sh',
- content=StaticFile(
- 'check_mapred_local_dir_used.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_nodemanager_health.sh',
- content=StaticFile('check_nodemanager_health.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_namenodes_ha.sh',
- content=StaticFile('check_namenodes_ha.sh'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/hdp_nagios_init.php',
- content=StaticFile('hdp_nagios_init.php'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/mm_wrapper.py',
- content=StaticFile('mm_wrapper.py'),
- mode=0755
- )
- self.assertResourceCalled('File',
- '/usr/lib64/nagios/plugins/check_hive_thrift_port.py',
- content=StaticFile('check_hive_thrift_port.py'),
- mode=0755
- )
- self.assertResourceCalled('Execute',
- 'htpasswd2 -c -b /etc/nagios/htpasswd.users nagiosadmin \'!`"\'"\'"\' 1\''
- )
-
- self.assertResourceCalled('File', '/etc/nagios/htpasswd.users',
- owner='nagios',
- group='nagios',
- mode=0640
- )
- self.assertResourceCalled('Execute', 'usermod -G nagios wwwrun'
- )
-
- self.assertResourceCalled('File', '/etc/nagios/command.cfg',
- owner='nagios',
- group='nagios'
- )
- self.assertResourceCalled('File', '/var/nagios/ignore.dat',
- owner='nagios',
- group='nagios',
- mode=0664
- )
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
index 31bc5d4..9ba4073 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default.hbasedecom.json
@@ -326,14 +326,6 @@
"hadoop_heapsize": "1024",
"jtnode_opt_maxnewsize": "200m"
},
- "nagios-env": {
- "hive_metastore_user_passwd": "password",
- "nagios_web_password": "!`\"' 1",
- "nagios_user": "nagios",
- "nagios_group": "nagios",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "user@com.ua"
- },
"hive-env": {
"hive_metastore_user_passwd": "password",
"hcat_pid_dir": "/var/run/webhcat",
@@ -493,9 +485,6 @@
"c6401.ambari.apache.org",
"c6402.ambari.apache.org"
],
- "nagios_server_host": [
- "c6402.ambari.apache.org"
- ],
"hive_metastore_hosts": [
"c6402.ambari.apache.org"
],
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
index d9f9f71..3448b83 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default.json
@@ -18,8 +18,8 @@
"ambari_db_rca_username": "mapred",
"java_home": "/usr/jdk64/jdk1.7.0_45",
"mysql_jdbc_url": "http://c6401.ambari.apache.org:8080/resources//mysql-connector-java.jar",
- "group_list": "[\"hadoop\",\"nobody\",\"users\",\"nagios\"]",
- "user_list": "[\"hive\",\"oozie\",\"nobody\",\"nagios\",\"ambari-qa\",\"flume\",\"hdfs\",\"storm\",\"mapred\",\"hbase\",\"tez\",\"zookeeper\",\"falcon\",\"sqoop\",\"yarn\",\"hcat\"]"
+ "group_list": "[\"hadoop\",\"nobody\",\"users\"]",
+ "user_list": "[\"hive\",\"oozie\",\"nobody\",\"ambari-qa\",\"flume\",\"hdfs\",\"storm\",\"mapred\",\"hbase\",\"tez\",\"zookeeper\",\"falcon\",\"sqoop\",\"yarn\",\"hcat\"]"
},
"commandType": "EXECUTION_COMMAND",
"roleParams": {},
@@ -335,14 +335,6 @@
"jtnode_opt_maxnewsize": "200m",
"rca_properties": "\nambari.jobhistory.database={ambari_db_rca_url}\nambari.jobhistory.driver={ambari_db_rca_driver}\nambari.jobhistory.user={ambari_db_rca_username}\nambari.jobhistory.password={ambari_db_rca_password}\nambari.jobhistory.logger=${{hadoop.root.logger}}\n\nlog4j.appender.JHA=org.apache.ambari.log4j.hadoop.mapreduce.jobhistory.JobHistoryAppender\nlog4j.appender.JHA.database={ambari_db_rca_url}\nlog4j.appender.JHA.driver={ambari_db_rca_driver}\nlog4j.appender.JHA.user={ambari_db_rca_username}\nlog4j.appender.JHA.password={ambari_db_rca_password}\n\nlog4j.logger.org.apache.hadoop.mapred.JobHistory$JobHistoryLogger=DEBUG,JHA\nlog4j.additivity.org.apache.hadoop.mapred.JobHistory$JobHistoryLogger=true\n\n"
},
- "nagios-env": {
- "hive_metastore_user_passwd": "password",
- "nagios_web_password": "!`\"' 1",
- "nagios_user": "nagios",
- "nagios_group": "nagios",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "user@com.ua"
- },
"hive-env": {
"hive_metastore_user_passwd": "password",
"hcat_pid_dir": "/var/run/webhcat",
@@ -559,9 +551,6 @@
"c6401.ambari.apache.org",
"c6402.ambari.apache.org"
],
- "nagios_server_host": [
- "c6402.ambari.apache.org"
- ],
"hive_metastore_hosts": [
"c6402.ambari.apache.org"
],
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
index 2ec3a22..fa8d183 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default.non_gmetad_host.json
@@ -326,14 +326,6 @@
"hadoop_heapsize": "1024",
"jtnode_opt_maxnewsize": "200m"
},
- "nagios-env": {
- "hive_metastore_user_passwd": "password",
- "nagios_web_password": "!`\"' 1",
- "nagios_user": "nagios",
- "nagios_group": "nagios",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "user@com.ua"
- },
"hive-env": {
"hive_metastore_user_passwd": "password",
"hcat_pid_dir": "/var/run/webhcat",
@@ -541,9 +533,6 @@
"c6401.ambari.apache.org",
"c6402.ambari.apache.org"
],
- "nagios_server_host": [
- "c6402.ambari.apache.org"
- ],
"hive_metastore_hosts": [
"c6402.ambari.apache.org"
],
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json b/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
index da7a878..77a06f2 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/default_client.json
@@ -326,14 +326,6 @@
"hadoop_heapsize": "1024",
"jtnode_opt_maxnewsize": "200m"
},
- "nagios-env": {
- "hive_metastore_user_passwd": "password",
- "nagios_web_password": "!`\"' 1",
- "nagios_user": "nagios",
- "nagios_group": "nagios",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "user@com.ua"
- },
"hive-env": {
"hive_metastore_user_passwd": "password",
"hcat_pid_dir": "/var/run/webhcat",
@@ -544,9 +536,6 @@
"c6401.ambari.apache.org",
"c6402.ambari.apache.org"
],
- "nagios_server_host": [
- "c6402.ambari.apache.org"
- ],
"hive_metastore_hosts": [
"c6402.ambari.apache.org"
],
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json b/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
index 3c08702..1278350 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/secured.json
@@ -170,15 +170,13 @@
"tasktracker_task_controller": "org.apache.hadoop.mapred.LinuxTaskController",
"oozie_keytab": "/etc/security/keytabs/oozie.service.keytab",
"hadoop_http_principal_name": "HTTP/_HOST",
- "kinit_path_local": "/usr/bin",
- "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab",
+ "kinit_path_local": "/usr/bin",
"hbase_regionserver_heapsize": "1024m",
"hbase_regionserver_xmn_max": "512",
"hbase_regionserver_xmn_ratio": "0.2",
"datanode_primary_name": "dn",
"namenode_principal_name": "nn/_HOST",
- "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",
- "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM",
+ "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",
"dfs_datanode_http_address": "1022",
"hbase_user_keytab": "/etc/security/keytabs/hbase.headless.keytab",
"jobtracker_primary_name": "jt",
@@ -188,8 +186,7 @@
"clientPort": "2181",
"oozie_jdbc_driver": "org.apache.derby.jdbc.EmbeddedDriver",
"hive_metastore_primary_name": "hive",
- "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab",
- "nagios_primary_name": "nagios",
+ "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab",
"jobtracker_principal_name": "jt/_HOST",
"hive_database": "New MySQL Database",
"hcat_pid_dir": "/var/run/webhcat",
@@ -198,7 +195,6 @@
"oozie_pid_dir": "/var/run/oozie",
"datanode_principal_name": "dn/_HOST",
"hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab",
- "nagios_group": "nagios",
"hcat_user": "hcat",
"hadoop_heapsize": "1024",
"hbase_regionserver_primary_name": "hbase",
@@ -219,7 +215,6 @@
"smokeuser_keytab": "/etc/security/keytabs/smokeuser.headless.keytab",
"kerberos_domain": "EXAMPLE.COM",
"snamenode_keytab": "/etc/security/keytabs/nn.service.keytab",
- "nagios_server": "c6402.ambari.apache.org",
"ganglia_runtime_dir": "/var/run/ganglia/hdp",
"lzo_enabled": "true",
"oozie_principal_name": "oozie/c6402.ambari.apache.org",
@@ -241,7 +236,6 @@
"gmetad_user": "nobody",
"oozie_http_keytab": "/etc/security/keytabs/spnego.service.keytab",
"hive_metastore": "c6402.ambari.apache.org",
- "nagios_user": "nagios",
"security_enabled": "true",
"proxyuser_group": "users",
"namenode_formatted_mark_dir": "/var/run/hadoop/hdfs/namenode/formatted/",
@@ -261,8 +255,6 @@
"jtnode_heapsize": "1024m",
"yarn_user": "yarn",
"gmond_user": "nobody",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "q@q.q",
"snamenode_primary_name": "nn",
"hdfs_user": "hdfs",
"oozie_database_type": "derby",
@@ -290,7 +282,6 @@
"user_group": "hadoop",
"hive_user": "hive",
"webHCat_http_primary_name": "HTTP",
- "nagios_web_password": "!`\"' 1",
"smokeuser": "ambari-qa",
"ganglia_conf_dir": "/etc/ganglia/hdp",
"hbase_master_heapsize": "1024m",
@@ -508,16 +499,6 @@
"hadoop_heapsize": "1024",
"jtnode_opt_maxnewsize": "200m"
},
- "nagios-env": {
- "hive_metastore_user_passwd": "password",
- "nagios_web_password": "password",
- "nagios_user": "nagios",
- "nagios_group": "nagios",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "user@com.ua",
- "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM",
- "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab"
- },
"hive-env": {
"hive_metastore_user_passwd": "password",
"hcat_pid_dir": "/var/run/webhcat",
@@ -740,9 +721,6 @@
"c6401.ambari.apache.org",
"c6402.ambari.apache.org"
],
- "nagios_server_host": [
- "c6402.ambari.apache.org"
- ],
"hive_metastore_hosts": [
"c6402.ambari.apache.org"
],
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
index 74ea204..b41af13 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_client.json
@@ -170,15 +170,13 @@
"tasktracker_task_controller": "org.apache.hadoop.mapred.LinuxTaskController",
"oozie_keytab": "/etc/security/keytabs/oozie.service.keytab",
"hadoop_http_principal_name": "HTTP/_HOST",
- "kinit_path_local": "/usr/bin",
- "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab",
+ "kinit_path_local": "/usr/bin",
"hbase_regionserver_heapsize": "1024m",
"hbase_regionserver_xmn_max": "512",
"hbase_regionserver_xmn_ratio": "0.2",
"datanode_primary_name": "dn",
"namenode_principal_name": "nn/_HOST",
- "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",
- "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM",
+ "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",
"dfs_datanode_http_address": "1022",
"hbase_user_keytab": "/etc/security/keytabs/hbase.headless.keytab",
"jobtracker_primary_name": "jt",
@@ -188,8 +186,7 @@
"clientPort": "2181",
"oozie_jdbc_driver": "org.apache.derby.jdbc.EmbeddedDriver",
"hive_metastore_primary_name": "hive",
- "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab",
- "nagios_primary_name": "nagios",
+ "hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab",
"jobtracker_principal_name": "jt/_HOST",
"hive_database": "New MySQL Database",
"hcat_pid_dir": "/var/run/webhcat",
@@ -197,8 +194,7 @@
"snappy_enabled": "true",
"oozie_pid_dir": "/var/run/oozie",
"datanode_principal_name": "dn/_HOST",
- "hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab",
- "nagios_group": "nagios",
+ "hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab",
"hcat_user": "hcat",
"hadoop_heapsize": "1024",
"hbase_regionserver_primary_name": "hbase",
@@ -218,8 +214,7 @@
"namenode_heapsize": "1024m",
"smokeuser_keytab": "/etc/security/keytabs/smokeuser.headless.keytab",
"kerberos_domain": "EXAMPLE.COM",
- "snamenode_keytab": "/etc/security/keytabs/nn.service.keytab",
- "nagios_server": "c6402.ambari.apache.org",
+ "snamenode_keytab": "/etc/security/keytabs/nn.service.keytab",
"ganglia_runtime_dir": "/var/run/ganglia/hdp",
"lzo_enabled": "true",
"oozie_principal_name": "oozie/c6402.ambari.apache.org",
@@ -240,8 +235,7 @@
"hadoop_http_keytab": "/etc/security/keytabs/spnego.service.keytab",
"gmetad_user": "nobody",
"oozie_http_keytab": "/etc/security/keytabs/spnego.service.keytab",
- "hive_metastore": "c6402.ambari.apache.org",
- "nagios_user": "nagios",
+ "hive_metastore": "c6402.ambari.apache.org",
"security_enabled": "true",
"proxyuser_group": "users",
"namenode_formatted_mark_dir": "/var/run/hadoop/hdfs/namenode/formatted/",
@@ -261,8 +255,6 @@
"jtnode_heapsize": "1024m",
"yarn_user": "yarn",
"gmond_user": "nobody",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "q@q.q",
"snamenode_primary_name": "nn",
"hdfs_user": "hdfs",
"oozie_database_type": "derby",
@@ -289,8 +281,7 @@
"hbase_log_dir": "/var/log/hbase",
"user_group": "hadoop",
"hive_user": "hive",
- "webHCat_http_primary_name": "HTTP",
- "nagios_web_password": "!`\"' 1",
+ "webHCat_http_primary_name": "HTTP",
"smokeuser": "ambari-qa",
"ganglia_conf_dir": "/etc/ganglia/hdp",
"hbase_master_heapsize": "1024m",
@@ -506,16 +497,6 @@
"hadoop_heapsize": "1024",
"jtnode_opt_maxnewsize": "200m"
},
- "nagios-env": {
- "hive_metastore_user_passwd": "password",
- "nagios_web_password": "password",
- "nagios_user": "nagios",
- "nagios_group": "nagios",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "user@com.ua",
- "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM",
- "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab"
- },
"hive-env": {
"hive_metastore_user_passwd": "password",
"hcat_pid_dir": "/var/run/webhcat",
@@ -733,9 +714,6 @@
"c6401.ambari.apache.org",
"c6402.ambari.apache.org"
],
- "nagios_server_host": [
- "c6402.ambari.apache.org"
- ],
"hive_metastore_hosts": [
"c6402.ambari.apache.org"
],
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
index 0412d50..124db7c 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
+++ b/ambari-server/src/test/python/stacks/1.3.2/configs/secured_no_jce_name.json
@@ -169,15 +169,13 @@
"tasktracker_task_controller": "org.apache.hadoop.mapred.LinuxTaskController",
"oozie_keytab": "/etc/security/keytabs/oozie.service.keytab",
"hadoop_http_principal_name": "HTTP/_HOST",
- "kinit_path_local": "/usr/bin",
- "nagios_keytab_path": "/etc/security/keytabs/nagios.service.keytab",
+ "kinit_path_local": "/usr/bin",
"hbase_regionserver_heapsize": "1024m",
"hbase_regionserver_xmn_max": "512",
"hbase_regionserver_xmn_ratio": "0.2",
"datanode_primary_name": "dn",
"namenode_principal_name": "nn/_HOST",
- "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",
- "nagios_principal_name": "nagios/c6402.ambari.apache.org@EXAMPLE.COM",
+ "namenode_keytab": "/etc/security/keytabs/nn.service.keytab",
"dfs_datanode_http_address": "1022",
"hbase_user_keytab": "/etc/security/keytabs/hbase.headless.keytab",
"jobtracker_primary_name": "jt",
@@ -188,7 +186,6 @@
"oozie_jdbc_driver": "org.apache.derby.jdbc.EmbeddedDriver",
"hive_metastore_primary_name": "hive",
"hbase_master_keytab": "/etc/security/keytabs/hbase.service.keytab",
- "nagios_primary_name": "nagios",
"jobtracker_principal_name": "jt/_HOST",
"hive_database": "New MySQL Database",
"hcat_pid_dir": "/etc/run/webhcat",
@@ -197,7 +194,6 @@
"oozie_pid_dir": "/var/run/oozie",
"datanode_principal_name": "dn/_HOST",
"hive_metastore_keytab": "/etc/security/keytabs/hive.service.keytab",
- "nagios_group": "nagios",
"hcat_user": "hcat",
"hadoop_heapsize": "1024",
"hbase_regionserver_primary_name": "hbase",
@@ -217,7 +213,6 @@
"smokeuser_keytab": "/etc/security/keytabs/smokeuser.headless.keytab",
"kerberos_domain": "EXAMPLE.COM",
"snamenode_keytab": "/etc/security/keytabs/nn.service.keytab",
- "nagios_server": "c6402.ambari.apache.org",
"ganglia_runtime_dir": "/var/run/ganglia/hdp",
"lzo_enabled": "true",
"oozie_principal_name": "oozie/c6402.ambari.apache.org",
@@ -238,8 +233,7 @@
"hadoop_http_keytab": "/etc/security/keytabs/spnego.service.keytab",
"gmetad_user": "nobody",
"oozie_http_keytab": "/etc/security/keytabs/spnego.service.keytab",
- "hive_metastore": "c6402.ambari.apache.org",
- "nagios_user": "nagios",
+ "hive_metastore": "c6402.ambari.apache.org",
"security_enabled": "true",
"proxyuser_group": "users",
"namenode_formatted_mark_dir": "/var/run/hadoop/hdfs/namenode/formatted/",
@@ -259,8 +253,6 @@
"jtnode_heapsize": "1024m",
"yarn_user": "yarn",
"gmond_user": "nobody",
- "nagios_web_login": "nagiosadmin",
- "nagios_contact": "q@q.q",
"snamenode_primary_name": "nn",
"hdfs_user": "hdfs",
"oozie_database_type": "derby",
@@ -288,7 +280,6 @@
"user_group": "hadoop",
"hive_user": "hive",
"webHCat_http_primary_name": "HTTP",
- "nagios_web_password": "!`\"' 1",
"smokeuser": "ambari-qa",
"ganglia_conf_dir": "/etc/ganglia/hdp",
"hbase_master_heapsize": "1024m",
@@ -591,9 +582,6 @@
"c6401.ambari.apache.org",
"c6402.ambari.apache.org"
],
- "nagios_server_host": [
- "c6402.ambari.apache.org"
- ],
"hive_metastore_hosts": [
"c6402.ambari.apache.org"
],
http://git-wip-us.apache.org/repos/asf/ambari/blob/e4ededeb/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py b/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
index 836274a..16aa939 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
+++ b/ambari-server/src/test/python/stacks/1.3.2/hooks/before-ANY/test_before_any.py
@@ -45,9 +45,6 @@ class TestHookBeforeInstall(RMFTestCase):
self.assertResourceCalled('Group', 'users',
ignore_failures = False,
)
- self.assertResourceCalled('Group', 'nagios',
- ignore_failures = False,
- )
self.assertResourceCalled('User', 'hive',
gid = 'hadoop',
ignore_failures = False,
@@ -63,11 +60,6 @@ class TestHookBeforeInstall(RMFTestCase):
ignore_failures = False,
groups = [u'nobody'],
)
- self.assertResourceCalled('User', 'nagios',
- gid = 'nagios',
- ignore_failures = False,
- groups = [u'hadoop'],
- )
self.assertResourceCalled('User', 'ambari-qa',
gid = 'hadoop',
ignore_failures = False,