You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by dm...@apache.org on 2014/02/27 14:31:25 UTC

[1/2] git commit: Revert "AMBARI-4849. Expose jmx metrics in Storm REST API server via Ambari Server API's for Storm. (Arsen Babych via mahadev)"

Repository: ambari
Updated Branches:
  refs/heads/trunk 81c4a9d97 -> e26ca98c0


Revert "AMBARI-4849. Expose jmx metrics in Storm REST API server via Ambari Server API's for Storm. (Arsen Babych via mahadev)"

This reverts commit 87ee281089847b8516f02a5988e6a8b09de48af3.


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/e26ca98c
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/e26ca98c
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/e26ca98c

Branch: refs/heads/trunk
Commit: e26ca98c01590f37f588786489d2825e6d18fd9c
Parents: 7646ec2
Author: Lisnichenko Dmitro <dl...@hortonworks.com>
Authored: Thu Feb 27 15:11:26 2014 +0200
Committer: Lisnichenko Dmitro <dl...@hortonworks.com>
Committed: Thu Feb 27 15:31:04 2014 +0200

----------------------------------------------------------------------
 .../controller/jmx/JMXPropertyProvider.java     | 205 +++++++------------
 .../HDP/2.1.1/services/STORM/metrics.json       |  97 ---------
 .../AmbariManagementControllerTest.java         |   2 +-
 .../StackDefinedPropertyProviderTest.java       |  49 +----
 .../controller/jmx/JMXPropertyProviderTest.java |   2 -
 .../controller/jmx/TestStreamProvider.java      |   1 -
 6 files changed, 77 insertions(+), 279 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/e26ca98c/ambari-server/src/main/java/org/apache/ambari/server/controller/jmx/JMXPropertyProvider.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/java/org/apache/ambari/server/controller/jmx/JMXPropertyProvider.java b/ambari-server/src/main/java/org/apache/ambari/server/controller/jmx/JMXPropertyProvider.java
index 95c7cbc..4de36ca 100644
--- a/ambari-server/src/main/java/org/apache/ambari/server/controller/jmx/JMXPropertyProvider.java
+++ b/ambari-server/src/main/java/org/apache/ambari/server/controller/jmx/JMXPropertyProvider.java
@@ -40,17 +40,16 @@ import java.util.regex.Pattern;
 
 import org.apache.ambari.server.controller.internal.AbstractPropertyProvider;
 import org.apache.ambari.server.controller.internal.PropertyInfo;
+import org.apache.ambari.server.controller.spi.NoSuchParentResourceException;
 import org.apache.ambari.server.controller.spi.Predicate;
 import org.apache.ambari.server.controller.spi.Request;
 import org.apache.ambari.server.controller.spi.Resource;
 import org.apache.ambari.server.controller.spi.SystemException;
-import org.apache.ambari.server.controller.utilities.PropertyHelper;
+import org.apache.ambari.server.controller.spi.UnsupportedPropertyException;
 import org.apache.ambari.server.controller.utilities.StreamProvider;
-import org.codehaus.jackson.JsonFactory;
 import org.codehaus.jackson.map.DeserializationConfig;
 import org.codehaus.jackson.map.ObjectMapper;
 import org.codehaus.jackson.map.ObjectReader;
-import org.codehaus.jackson.type.TypeReference;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
@@ -59,17 +58,12 @@ import org.slf4j.LoggerFactory;
  */
 public class JMXPropertyProvider extends AbstractPropertyProvider {
 
-
-  // TODO: HACK, value is set at populateResource()
-  private String componentName = null;
-
   private static final String NAME_KEY = "name";
   private static final String PORT_KEY = "tag.port";
   private static final String DOT_REPLACEMENT_CHAR = "#";
   private static final long DEFAULT_POPULATE_TIMEOUT_MILLIS = 10000L;
 
   public static final String TIMED_OUT_MSG = "Timed out waiting for JMX metrics.";
-  public static final String STORM_REST_API = "STORM_REST_API";
 
   /**
    * Thread pool
@@ -95,8 +89,7 @@ public class JMXPropertyProvider extends AbstractPropertyProvider {
     EXECUTOR_SERVICE = threadPoolExecutor;
   }
 
-  private final static ObjectReader jmxObjectReader;
-  private final static ObjectReader stormObjectReader;
+  private final static ObjectReader objectReader;
 
   private static final Map<String, String> DEFAULT_JMX_PORTS = new HashMap<String, String>();
 
@@ -111,19 +104,10 @@ public class JMXPropertyProvider extends AbstractPropertyProvider {
     DEFAULT_JMX_PORTS.put("HISTORYSERVER",      "19888");
     DEFAULT_JMX_PORTS.put("NODEMANAGER",         "8042");
     DEFAULT_JMX_PORTS.put("JOURNALNODE",         "8480");
-    DEFAULT_JMX_PORTS.put("STORM_REST_API",      "8745");
-
-    ObjectMapper jmxObjectMapper = new ObjectMapper();
-    jmxObjectMapper.configure(DeserializationConfig.Feature.USE_ANNOTATIONS, false);
-    jmxObjectReader = jmxObjectMapper.reader(JMXMetricHolder.class);
-
-    JsonFactory factory = new JsonFactory();
-    ObjectMapper stormObjectMapper = new ObjectMapper(factory);
-    TypeReference<HashMap<String,Object>> typeRef
-            = new TypeReference<
-            HashMap<String,Object>
-            >() {};
-    stormObjectReader = jmxObjectMapper.reader(typeRef);
+
+    ObjectMapper objectMapper = new ObjectMapper();
+    objectMapper.configure(DeserializationConfig.Feature.USE_ANNOTATIONS, false);
+    objectReader = objectMapper.reader(JMXMetricHolder.class);
   }
 
   protected final static Logger LOG =
@@ -249,10 +233,7 @@ public class JMXPropertyProvider extends AbstractPropertyProvider {
    * @return the spec
    */
   protected String getSpec(String protocol, String hostName, String port) {
-      if (null == componentName || !componentName.equals(STORM_REST_API))
-        return protocol + "://" + hostName + ":" + port + "/jmx";
-      else
-        return protocol + "://" + hostName + ":" + port + "/api/cluster/summary";
+      return protocol + "://" + hostName + ":" + port + "/jmx";
   }
 
   /**
@@ -320,7 +301,7 @@ public class JMXPropertyProvider extends AbstractPropertyProvider {
       }
     }
 
-    componentName = (String) resource.getPropertyValue(componentNamePropertyId);
+    String componentName = (String) resource.getPropertyValue(componentNamePropertyId);
 
     if (getComponentMetrics().get(componentName) == null) {
       // If there are no metrics defined for the given component then there is nothing to do.
@@ -346,127 +327,91 @@ public class JMXPropertyProvider extends AbstractPropertyProvider {
       InputStream in = streamProvider.readFrom(getSpec(protocol, hostName, port));
 
       try {
+        JMXMetricHolder metricHolder = objectReader.readValue(in);
 
-        if (null == componentName || !componentName.equals(STORM_REST_API)) {
-          getHadoopMetricValue(in, ids, resource, request);
-        } else {
-          getStormMetricValue(in, ids, resource, request);
-        }
+        Map<String, Map<String, Object>> categories = new HashMap<String, Map<String, Object>>();
 
-      } finally {
-        in.close();
-      }
-    } catch (IOException e) {
-      logException(e);
-    }
-    return resource;
-  }
-
-  /**
-   * Hadoop-specific metrics fetching
-   */
-  private void getHadoopMetricValue(InputStream in, Set<String> ids,
-                       Resource resource, Request request) throws IOException {
-    JMXMetricHolder metricHolder = jmxObjectReader.readValue(in);
-
-    Map<String, Map<String, Object>> categories = new HashMap<String, Map<String, Object>>();
-
-    for (Map<String, Object> bean : metricHolder.getBeans()) {
-      String category = getCategory(bean);
-      if (category != null) {
-        categories.put(category, bean);
-      }
-    }
-
-    for (String propertyId : ids) {
-      Map<String, PropertyInfo> propertyInfoMap = getPropertyInfoMap(componentName, propertyId);
-
-      String requestedPropertyId = propertyId;
+        for (Map<String, Object> bean : metricHolder.getBeans()) {
+          String category = getCategory(bean);
+          if (category != null) {
+            categories.put(category, bean);
+          }
+        }
 
-      for (Map.Entry<String, PropertyInfo> entry : propertyInfoMap.entrySet()) {
+        for (String propertyId : ids) {
+          Map<String, PropertyInfo> propertyInfoMap = getPropertyInfoMap(componentName, propertyId);
 
-        PropertyInfo propertyInfo = entry.getValue();
-        propertyId = entry.getKey();
+          String requestedPropertyId = propertyId;
 
-        if (propertyInfo.isPointInTime()) {
+          for (Map.Entry<String, PropertyInfo> entry : propertyInfoMap.entrySet()) {
 
-          String property = propertyInfo.getPropertyId();
-          String category = "";
+            PropertyInfo propertyInfo = entry.getValue();
+            propertyId = entry.getKey();
 
+            if (propertyInfo.isPointInTime()) {
 
-          List<String> keyList = new LinkedList<String>();
+              String property = propertyInfo.getPropertyId();
+              String category = "";
 
-          int keyStartIndex = property.indexOf('[');
-          if (-1 != keyStartIndex) {
-            int keyEndIndex = property.indexOf(']', keyStartIndex);
-            if (-1 != keyEndIndex && keyEndIndex > keyStartIndex) {
-              keyList.add(property.substring(keyStartIndex+1, keyEndIndex));
-            }
-          }
-
-          if (!containsArguments(propertyId)) {
-            int dotIndex = property.indexOf('.', property.indexOf('='));
-            if (-1 != dotIndex) {
-              category = property.substring(0, dotIndex);
-              property = (-1 == keyStartIndex) ?
+              
+              List<String> keyList = new LinkedList<String>();
+              
+              int keyStartIndex = property.indexOf('[');
+              if (-1 != keyStartIndex) {
+                int keyEndIndex = property.indexOf(']', keyStartIndex);
+                if (-1 != keyEndIndex && keyEndIndex > keyStartIndex) {
+                  keyList.add(property.substring(keyStartIndex+1, keyEndIndex));
+                }
+              }
+              
+              if (!containsArguments(propertyId)) {
+                int dotIndex = property.indexOf('.', property.indexOf('='));
+                if (-1 != dotIndex) {
+                  category = property.substring(0, dotIndex);
+                  property = (-1 == keyStartIndex) ?
                       property.substring(dotIndex+1) :
-                      property.substring(dotIndex+1, keyStartIndex);
-            }
-          } else {
-            int firstKeyIndex = keyStartIndex > -1 ? keyStartIndex : property.length();
-            int dotIndex = property.lastIndexOf('.', firstKeyIndex);
-
-            if (dotIndex != -1) {
-              category = property.substring(0, dotIndex);
-              property = property.substring(dotIndex + 1, firstKeyIndex);
-            }
-          }
-
-          if (containsArguments(propertyId)) {
-            Pattern pattern = Pattern.compile(category);
+                        property.substring(dotIndex+1, keyStartIndex);
+                }
+              } else {
+                int firstKeyIndex = keyStartIndex > -1 ? keyStartIndex : property.length();
+                int dotIndex = property.lastIndexOf('.', firstKeyIndex);
 
-            // find all jmx categories that match the regex
-            for (String jmxCat : categories.keySet()) {
-              Matcher matcher = pattern.matcher(jmxCat);
-              if (matcher.matches()) {
-                String newPropertyId = propertyId;
-                for (int i = 0; i < matcher.groupCount(); i++) {
-                  newPropertyId = substituteArgument(newPropertyId, "$" + (i + 1), matcher.group(i + 1));
+                if (dotIndex != -1) {
+                  category = property.substring(0, dotIndex);
+                  property = property.substring(dotIndex + 1, firstKeyIndex);
                 }
-                // We need to do the final filtering here, after the argument substitution
-                if (isRequestedPropertyId(newPropertyId, requestedPropertyId, request)) {
-                  setResourceValue(resource, categories, newPropertyId, jmxCat, property, keyList);
+              }
+
+              if (containsArguments(propertyId)) {
+                Pattern pattern = Pattern.compile(category);
+                
+                // find all jmx categories that match the regex
+                for (String jmxCat : categories.keySet()) {
+                  Matcher matcher = pattern.matcher(jmxCat);
+                  if (matcher.matches()) {
+                    String newPropertyId = propertyId;
+                    for (int i = 0; i < matcher.groupCount(); i++) {
+                      newPropertyId = substituteArgument(newPropertyId, "$" + (i + 1), matcher.group(i + 1));
+                    }
+                    // We need to do the final filtering here, after the argument substitution
+                    if (isRequestedPropertyId(newPropertyId, requestedPropertyId, request)) {
+                      setResourceValue(resource, categories, newPropertyId, jmxCat, property, keyList);
+                    }
+                  }
                 }
+              } else {
+                setResourceValue(resource, categories, propertyId, category, property, keyList);
               }
             }
-          } else {
-            setResourceValue(resource, categories, propertyId, category, property, keyList);
           }
         }
+      } finally {
+        in.close();
       }
+    } catch (IOException e) {
+      logException(e);
     }
-  }
-
-  /**
-   * TODO: Refactor
-   * Storm-specific metrics fetching
-   */
-  private void getStormMetricValue(InputStream in, Set<String> ids,
-                                   Resource resource, Request request) throws IOException {
-    HashMap<String, Object> metricHolder = stormObjectReader.readValue(in);
-    for (String category : ids) {
-      Map<String, PropertyInfo> defProps = getComponentMetrics().get(STORM_REST_API);
-      for (String depProp : defProps.keySet()) {
-        if (depProp.startsWith(category)) {
-          PropertyInfo propInfo = defProps.get(depProp);
-          String propName = propInfo.getPropertyId();
-          Object propertyValue = metricHolder.get(propName);
-          String absId = PropertyHelper.getPropertyId(category, propName);
-          // TODO: Maybe cast to int
-          resource.setProperty(absId, propertyValue);
-        }
-      }
-    }
+    return resource;
   }
 
   private void setResourceValue(Resource resource, Map<String, Map<String, Object>> categories, String propertyId,

http://git-wip-us.apache.org/repos/asf/ambari/blob/e26ca98c/ambari-server/src/main/resources/stacks/HDP/2.1.1/services/STORM/metrics.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.1.1/services/STORM/metrics.json b/ambari-server/src/main/resources/stacks/HDP/2.1.1/services/STORM/metrics.json
index ef3cde7..da75ef8 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.1.1/services/STORM/metrics.json
+++ b/ambari-server/src/main/resources/stacks/HDP/2.1.1/services/STORM/metrics.json
@@ -1,101 +1,4 @@
 {
-  "STORM_REST_API": {
-    "Component": [
-      {
-        "type": "jmx",
-        "metrics": {
-          "metrics/api/cluster/summary/tasks.total": {
-            "metric": "tasks.total",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/slots.total": {
-            "metric": "slots.total",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/slots.free": {
-            "metric": "slots.free",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/supervisors": {
-            "metric": "supervisors",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/executors.total": {
-            "metric": "executors.total",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/slots.used": {
-            "metric": "slots.used",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/topologies": {
-            "metric": "topologies",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/nimbus.uptime": {
-            "metric": "nimbus.uptime",
-            "pointInTime": true,
-            "temporal": false
-          }
-        }
-      }
-    ],
-    "HostComponent": [
-      {
-        "type": "jmx",
-        "metrics": {
-          "metrics/api/cluster/summary/tasks.total": {
-            "metric": "tasks.total",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/slots.total": {
-            "metric": "slots.total",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/slots.free": {
-            "metric": "slots.free",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/supervisors": {
-            "metric": "supervisors",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/executors.total": {
-            "metric": "executors.total",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/slots.used": {
-            "metric": "slots.used",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/topologies": {
-            "metric": "topologies",
-            "pointInTime": true,
-            "temporal": false
-          },
-          "metrics/api/cluster/summary/nimbus.uptime": {
-            "metric": "nimbus.uptime",
-            "pointInTime": true,
-            "temporal": false
-          }
-        }
-      }
-
-    ]
-  },
   "NIMBUS": {
     "Component": [
       {

http://git-wip-us.apache.org/repos/asf/ambari/blob/e26ca98c/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java
index 6e32ac8..dc4c5a5 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java
@@ -147,7 +147,7 @@ public class AmbariManagementControllerTest {
   private static final String REPO_ID = "HDP-1.1.1.16";
   private static final String PROPERTY_NAME = "hbase.regionserver.msginterval";
   private static final String SERVICE_NAME = "HDFS";
-  private static final int STACK_VERSIONS_CNT = 12;
+  private static final int STACK_VERSIONS_CNT = 11;
   private static final int REPOS_CNT = 3;
   private static final int STACKS_CNT = 1;
   private static final int STACK_PROPERTIES_CNT = 81;

http://git-wip-us.apache.org/repos/asf/ambari/blob/e26ca98c/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/StackDefinedPropertyProviderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/StackDefinedPropertyProviderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/StackDefinedPropertyProviderTest.java
index b9b2962..6ce0022 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/StackDefinedPropertyProviderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/StackDefinedPropertyProviderTest.java
@@ -656,54 +656,7 @@ public class StackDefinedPropertyProviderTest {
     Assert.assertEquals(2, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/dfs/journalnode/cluster/mycluster", "lastWriterEpoch")));
     Assert.assertEquals(0, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/dfs/journalnode/cluster/mycluster", "currentLagTxns")));
     Assert.assertEquals(8444, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/dfs/journalnode/cluster/mycluster", "lastWrittenTxId")));
-  }
-
-
-
-  @Test
-  public void testPopulateResources_jmx_Storm() throws Exception {
-    // Adjust stack version for cluster
-    Cluster cluster = clusters.getCluster("c1");
-    cluster.setDesiredStackVersion(new StackId("HDP-2.1.1"));
-
-    TestStreamProvider  streamProvider = new TestStreamProvider();
-    TestJMXHostProvider hostProvider = new TestJMXHostProvider(false);
-    TestGangliaHostProvider gangliaHostProvider = new TestGangliaHostProvider();
-
-    StackDefinedPropertyProvider propertyProvider = new StackDefinedPropertyProvider(
-            Resource.Type.HostComponent,
-            hostProvider,
-            gangliaHostProvider,
-            streamProvider,
-            PropertyHelper.getPropertyId("HostRoles", "cluster_name"),
-            PropertyHelper.getPropertyId("HostRoles", "host_name"),
-            PropertyHelper.getPropertyId("HostRoles", "component_name"),
-            PropertyHelper.getPropertyId("HostRoles", "state"),
-            new EmptyPropertyProvider(),
-            new EmptyPropertyProvider());
-
-    Resource resource = new ResourceImpl(Resource.Type.HostComponent);
-
-    resource.setProperty("HostRoles/cluster_name", "c1");
-    resource.setProperty(HOST_COMPONENT_HOST_NAME_PROPERTY_ID, "domu-12-31-39-0e-34-e1.compute-1.internal");
-    resource.setProperty(HOST_COMPONENT_COMPONENT_NAME_PROPERTY_ID, "STORM_REST_API");
-    resource.setProperty(HOST_COMPONENT_STATE_PROPERTY_ID, "STARTED");
-
-    // request with an empty set should get all supported properties
-    Request request = PropertyHelper.getReadRequest(Collections.<String>emptySet());
-
-    Assert.assertEquals(1, propertyProvider.populateResources(Collections.singleton(resource), request, null).size());
-
-    // see test/resources/storm_rest_api_jmx.json for values
-    Assert.assertEquals(28, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/tasks.total", "tasks.total")));
-    Assert.assertEquals(8, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/slots.total", "slots.total")));
-    Assert.assertEquals(5, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/slots.free", "slots.free")));
-    Assert.assertEquals(2, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/supervisors", "supervisors")));
-    Assert.assertEquals(28, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/executors.total", "executors.total")));
-    Assert.assertEquals(3, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/slots.used", "slots.used")));
-    Assert.assertEquals(1, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/topologies", "topologies")));
-    Assert.assertEquals(4637, resource.getPropertyValue(PropertyHelper.getPropertyId("metrics/api/cluster/summary/nimbus.uptime", "nimbus.uptime")));
-  }
+  }  
 
   @Test
   public void testPopulateResources_NoRegionServer() throws Exception {

http://git-wip-us.apache.org/repos/asf/ambari/blob/e26ca98c/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/JMXPropertyProviderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/JMXPropertyProviderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/JMXPropertyProviderTest.java
index ac7314e..c2910db 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/JMXPropertyProviderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/JMXPropertyProviderTest.java
@@ -450,8 +450,6 @@ public class JMXPropertyProviderTest {
         return null == clusterName ? "60010" : "60011";
       else  if (componentName.equals("JOURNALNODE"))
         return "8480";
-      else  if (componentName.equals("STORM_REST_API"))
-        return "8745";
       else
         return null;
     }

http://git-wip-us.apache.org/repos/asf/ambari/blob/e26ca98c/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/TestStreamProvider.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/TestStreamProvider.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/TestStreamProvider.java
index a7b7553..0c793dc 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/TestStreamProvider.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/jmx/TestStreamProvider.java
@@ -38,7 +38,6 @@ public class TestStreamProvider implements StreamProvider {
     FILE_MAPPING.put("60011", "hbase_hbasemaster_jmx_2.json");
     FILE_MAPPING.put("8088",  "resourcemanager_jmx.json");
     FILE_MAPPING.put("8480",  "hdfs_journalnode_jmx.json");
-    FILE_MAPPING.put("8745",  "storm_rest_api_jmx.json");
   }
 
   /**


[2/2] git commit: Revert "AMBARI-4849. Expose jmx metrics in Storm REST API server via Ambari Server API's for Storm. (Arsen Babych via mahadev)"

Posted by dm...@apache.org.
Revert "AMBARI-4849. Expose jmx metrics in Storm REST API server via Ambari Server API's for Storm. (Arsen Babych via mahadev)"

This reverts commit 3842cd54841a363a70ed2e9debc77206411d70c2.


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/7646ec2f
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/7646ec2f
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/7646ec2f

Branch: refs/heads/trunk
Commit: 7646ec2f16bb059e677db70805b39b63fdbf1437
Parents: 81c4a9d
Author: Lisnichenko Dmitro <dl...@hortonworks.com>
Authored: Thu Feb 27 14:46:29 2014 +0200
Committer: Lisnichenko Dmitro <dl...@hortonworks.com>
Committed: Thu Feb 27 15:31:04 2014 +0200

----------------------------------------------------------------------
 .../resources/stacks/HDP/2.1.1/metainfo.xml     |  24 ----
 .../stacks/HDP/2.1.1/repos/repoinfo.xml         |  61 ----------
 .../stacks/HDP/2.1.1/role_command_order.json    | 100 -----------------
 .../STORM/configuration/placeholder.txt         |  17 ---
 .../HDP/2.1.1/services/STORM/metainfo.xml       | 112 -------------------
 .../HDP/2.1.1/services/STORM/metrics.json       |  99 ----------------
 .../services/STORM/package/placeholder.txt      |  17 ---
 .../src/test/resources/storm_rest_api_jmx.json  |  10 --
 8 files changed, 440 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/stacks/HDP/2.1.1/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/stacks/HDP/2.1.1/metainfo.xml b/ambari-server/src/test/resources/stacks/HDP/2.1.1/metainfo.xml
deleted file mode 100644
index 04416b0..0000000
--- a/ambari-server/src/test/resources/stacks/HDP/2.1.1/metainfo.xml
+++ /dev/null
@@ -1,24 +0,0 @@
-<?xml version="1.0"?>
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-<metainfo>
-    <versions>
-      <active>true</active>
-    </versions>
-    <extends>2.0.8</extends>
-</metainfo>
-

http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/stacks/HDP/2.1.1/repos/repoinfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/stacks/HDP/2.1.1/repos/repoinfo.xml b/ambari-server/src/test/resources/stacks/HDP/2.1.1/repos/repoinfo.xml
deleted file mode 100644
index 1025240..0000000
--- a/ambari-server/src/test/resources/stacks/HDP/2.1.1/repos/repoinfo.xml
+++ /dev/null
@@ -1,61 +0,0 @@
-<?xml version="1.0"?>
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-<reposinfo>
-  <os type="centos6">
-    <repo>
-      <baseurl>http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.0.6.0</baseurl>
-      <repoid>HDP-2.0.8</repoid>
-      <reponame>HDP</reponame>
-    </repo>
-  </os>
-  <os type="centos5">
-    <repo>
-      <baseurl>http://public-repo-1.hortonworks.com/HDP/centos5/2.x/updates/2.0.6.0</baseurl>
-      <repoid>HDP-2.0.8</repoid>
-      <reponame>HDP</reponame>
-    </repo>
-  </os>
-  <os type="redhat6">
-    <repo>
-      <baseurl>http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.0.6.0</baseurl>
-      <repoid>HDP-2.0.8</repoid>
-      <reponame>HDP</reponame>
-    </repo>
-  </os>
-  <os type="redhat5">
-    <repo>
-      <baseurl>http://public-repo-1.hortonworks.com/HDP/centos5/2.x/updates/2.0.6.0</baseurl>
-      <repoid>HDP-2.0.8</repoid>
-      <reponame>HDP</reponame>
-    </repo>
-  </os>
-  <os type="suse11">
-    <repo>
-      <baseurl>http://public-repo-1.hortonworks.com/HDP/suse11/2.x/updates/2.0.6.0</baseurl>
-      <repoid>HDP-2.0.8</repoid>
-      <reponame>HDP</reponame>
-    </repo>
-  </os>
-  <os type="sles11">
-    <repo>
-      <baseurl>http://public-repo-1.hortonworks.com/HDP/suse11/2.x/updates/2.0.6.0</baseurl>
-      <repoid>HDP-2.0.8</repoid>
-      <reponame>HDP</reponame>
-    </repo>
-  </os>
-</reposinfo>

http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/stacks/HDP/2.1.1/role_command_order.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/stacks/HDP/2.1.1/role_command_order.json b/ambari-server/src/test/resources/stacks/HDP/2.1.1/role_command_order.json
deleted file mode 100644
index 84610ca..0000000
--- a/ambari-server/src/test/resources/stacks/HDP/2.1.1/role_command_order.json
+++ /dev/null
@@ -1,100 +0,0 @@
-{
-  "_comment" : "Record format:",
-  "_comment" : "blockedRole-blockedCommand: [blockerRole1-blockerCommand1, blockerRole2-blockerCommand2, ...]",
-  "general_deps" : {
-    "_comment" : "dependencies for all cases",
-    "NAGIOS_SERVER-INSTALL" : ["HIVE_CLIENT-INSTALL", "HCAT-INSTALL",
-        "MAPREDUCE_CLIENT-INSTALL", "OOZIE_CLIENT-INSTALL"],
-    "HBASE_MASTER-START": ["ZOOKEEPER_SERVER-START"],
-    "HBASE_REGIONSERVER-START": ["HBASE_MASTER-START"],
-    "OOZIE_SERVER-START": ["JOBTRACKER-START", "TASKTRACKER-START"],
-    "WEBHCAT_SERVER-START": ["TASKTRACKER-START", "HIVE_SERVER-START"],
-    "HIVE_METASTORE-START": ["MYSQL_SERVER-START"],
-    "HIVE_SERVER-START": ["TASKTRACKER-START", "MYSQL_SERVER-START"],
-    "HUE_SERVER-START": ["HIVE_SERVER-START", "HCAT-START", "OOZIE_SERVER-START"],
-    "FLUME_SERVER-START": ["OOZIE_SERVER-START"],
-    "NAGIOS_SERVER-START": ["HBASE_MASTER-START", "HBASE_REGIONSERVER-START",
-        "GANGLIA_SERVER-START", "GANGLIA_MONITOR-START", "HCAT-START",
-        "HIVE_SERVER-START", "HIVE_METASTORE-START", "HUE_SERVER-START",
-        "JOBTRACKER-START", "TASKTRACKER-START", "ZOOKEEPER_SERVER-START",
-        "MYSQL_SERVER-START", "OOZIE_SERVER-START", "PIG-START", "SQOOP-START",
-        "WEBHCAT_SERVER-START", "FLUME_SERVER-START"],
-    "MAPREDUCE_SERVICE_CHECK-SERVICE_CHECK": ["JOBTRACKER-START", "TASKTRACKER-START"],
-    "OOZIE_SERVICE_CHECK-SERVICE_CHECK": ["OOZIE_SERVER-START"],
-    "WEBHCAT_SERVICE_CHECK-SERVICE_CHECK": ["WEBHCAT_SERVER-START"],
-    "HBASE_SERVICE_CHECK-SERVICE_CHECK": ["HBASE_MASTER-START", "HBASE_REGIONSERVER-START"],
-    "HIVE_SERVICE_CHECK-SERVICE_CHECK": ["HIVE_SERVER-START", "HIVE_METASTORE-START"],
-    "HCAT_SERVICE_CHECK-SERVICE_CHECK": ["HIVE_SERVER-START"],
-    "PIG_SERVICE_CHECK-SERVICE_CHECK": ["JOBTRACKER-START", "TASKTRACKER-START"],
-    "SQOOP_SERVICE_CHECK-SERVICE_CHECK": ["JOBTRACKER-START", "TASKTRACKER-START"],
-    "ZOOKEEPER_SERVICE_CHECK-SERVICE_CHECK": ["ZOOKEEPER_SERVER-START"],
-    "ZOOKEEPER_QUORUM_SERVICE_CHECK-SERVICE_CHECK": ["ZOOKEEPER_SERVER-START"],
-    "ZOOKEEPER_SERVER-STOP" : ["HBASE_MASTER-STOP", "HBASE_REGIONSERVER-STOP"],
-    "HBASE_MASTER-STOP": ["HBASE_REGIONSERVER-STOP"],
-    "TASKTRACKER-UPGRADE": ["JOBTRACKER-UPGRADE"],
-    "MAPREDUCE_CLIENT-UPGRADE": ["TASKTRACKER-UPGRADE", "JOBTRACKER-UPGRADE"],
-    "ZOOKEEPER_SERVER-UPGRADE": ["MAPREDUCE_CLIENT-UPGRADE"],
-    "ZOOKEEPER_CLIENT-UPGRADE": ["ZOOKEEPER_SERVER-UPGRADE"],
-    "HBASE_MASTER-UPGRADE": ["ZOOKEEPER_CLIENT-UPGRADE"],
-    "HBASE_REGIONSERVER-UPGRADE": ["HBASE_MASTER-UPGRADE"],
-    "HBASE_CLIENT-UPGRADE": ["HBASE_REGIONSERVER-UPGRADE"],
-    "HIVE_SERVER-UPGRADE" : ["HBASE_CLIENT-UPGRADE"],
-    "HIVE_METASTORE-UPGRADE" : ["HIVE_SERVER-UPGRADE"],
-    "MYSQL_SERVER-UPGRADE": ["HIVE_METASTORE-UPGRADE"],
-    "HIVE_CLIENT-UPGRADE": ["MYSQL_SERVER-UPGRADE"],
-    "HCAT-UPGRADE": ["HIVE_CLIENT-UPGRADE"],
-    "OOZIE_SERVER-UPGRADE" : ["HCAT-UPGRADE"],
-    "OOZIE_CLIENT-UPGRADE" : ["OOZIE_SERVER-UPGRADE"],
-    "WEBHCAT_SERVER-UPGRADE" : ["OOZIE_CLIENT-UPGRADE"],
-    "PIG-UPGRADE" : ["WEBHCAT_SERVER-UPGRADE"],
-    "SQOOP-UPGRADE" : ["PIG-UPGRADE"],
-    "NAGIOS_SERVER-UPGRADE" : ["SQOOP-UPGRADE"],
-    "GANGLIA_SERVER-UPGRADE" : ["NAGIOS_SERVER-UPGRADE"],
-    "GANGLIA_MONITOR-UPGRADE" : ["GANGLIA_SERVER-UPGRADE"]
-  },
-  "_comment" : "GLUSTERFS-specific dependencies",
-  "optional_glusterfs": {
-    "HBASE_MASTER-START": ["PEERSTATUS-START"],
-    "JOBTRACKER-START": ["PEERSTATUS-START"],
-    "TASKTRACKER-START": ["PEERSTATUS-START"],
-    "GLUSTERFS_SERVICE_CHECK-SERVICE_CHECK": ["PEERSTATUS-START"],
-    "JOBTRACKER-UPGRADE": ["GLUSTERFS_CLIENT-UPGRADE"]
-  },
-  "_comment" : "Dependencies that are used when GLUSTERFS is not present in cluster",
-  "optional_no_glusterfs": {
-    "SECONDARY_NAMENODE-START": ["NAMENODE-START"],
-    "RESOURCEMANAGER-START": ["NAMENODE-START", "DATANODE-START"],
-    "NODEMANAGER-START": ["NAMENODE-START", "DATANODE-START", "RESOURCEMANAGER-START"],
-    "HISTORYSERVER-START": ["NAMENODE-START", "DATANODE-START"],
-    "HBASE_MASTER-START": ["NAMENODE-START", "DATANODE-START"],
-    "JOBTRACKER-START": ["NAMENODE-START", "DATANODE-START"],
-    "TASKTRACKER-START": ["NAMENODE-START", "DATANODE-START"],
-    "HIVE_SERVER-START": ["DATANODE-START"],
-    "WEBHCAT_SERVER-START": ["DATANODE-START"],
-    "NAGIOS_SERVER-START": ["NAMENODE-START", "SECONDARY_NAMENODE-START",
-        "DATANODE-START", "RESOURCEMANAGER-START", "NODEMANAGER-START", "HISTORYSERVER-START"],
-    "HDFS_SERVICE_CHECK-SERVICE_CHECK": ["NAMENODE-START", "DATANODE-START",
-        "SECONDARY_NAMENODE-START"],
-    "MAPREDUCE2_SERVICE_CHECK-SERVICE_CHECK": ["NODEMANAGER-START",
-        "RESOURCEMANAGER-START", "HISTORYSERVER-START", "YARN_SERVICE_CHECK-SERVICE_CHECK"],
-    "YARN_SERVICE_CHECK-SERVICE_CHECK": ["NODEMANAGER-START", "RESOURCEMANAGER-START"],
-    "RESOURCEMANAGER_SERVICE_CHECK-SERVICE_CHECK": ["RESOURCEMANAGER-START"],
-    "PIG_SERVICE_CHECK-SERVICE_CHECK": ["RESOURCEMANAGER-START", "NODEMANAGER-START"],
-    "NAMENODE-STOP": ["JOBTRACKER-STOP", "TASKTRACKER-STOP", "RESOURCEMANAGER-STOP",
-        "NODEMANAGER-STOP", "HISTORYSERVER-STOP", "HBASE_MASTER-STOP"],
-    "DATANODE-STOP": ["JOBTRACKER-STOP", "TASKTRACKER-STOP", "RESOURCEMANAGER-STOP",
-        "NODEMANAGER-STOP", "HISTORYSERVER-STOP", "HBASE_MASTER-STOP"],
-    "SECONDARY_NAMENODE-UPGRADE": ["NAMENODE-UPGRADE"],
-    "DATANODE-UPGRADE": ["SECONDARY_NAMENODE-UPGRADE"],
-    "HDFS_CLIENT-UPGRADE": ["DATANODE-UPGRADE"],
-    "JOBTRACKER-UPGRADE": ["HDFS_CLIENT-UPGRADE"]
-  },
-  "_comment" : "Dependencies that are used in HA NameNode cluster",
-  "optional_ha": {
-    "NAMENODE-START": ["JOURNALNODE-START", "ZOOKEEPER_SERVER-START"],
-    "ZKFC-START": ["NAMENODE-START"],
-    "NAGIOS_SERVER-START": ["ZKFC-START", "JOURNALNODE-START"],
-    "HDFS_SERVICE_CHECK-SERVICE_CHECK": ["ZKFC-START"]
-  }
-}
-

http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/configuration/placeholder.txt
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/configuration/placeholder.txt b/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/configuration/placeholder.txt
deleted file mode 100644
index e2479f3..0000000
--- a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/configuration/placeholder.txt
+++ /dev/null
@@ -1,17 +0,0 @@
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent

http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metainfo.xml b/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metainfo.xml
deleted file mode 100644
index f2bceca..0000000
--- a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metainfo.xml
+++ /dev/null
@@ -1,112 +0,0 @@
-<?xml version="1.0"?>
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-
-<metainfo>
-  <schemaVersion>2.0</schemaVersion>
-  <services>
-    <service>
-      <name>STORM</name>
-      <comment>Apache Hadoop Stream processing framework</comment>
-      <version>0.9.0.1</version>
-      <components>
-
-        <component>
-          <name>NIMBUS</name>
-          <category>MASTER</category>
-          <commandScript>
-            <script>scripts/nimbus.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-
-        <component>
-          <name>STORM_REST_API</name>
-          <category>MASTER</category>
-          <commandScript>
-            <script>scripts/rest_api.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-
-        <component>
-          <name>SUPERVISOR</name>
-          <category>SLAVE</category>
-          <commandScript>
-            <script>scripts/supervisor.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-
-        <component>
-          <name>STORM_UI_SERVER</name>
-          <category>MASTER</category>
-          <commandScript>
-            <script>scripts/ui_server.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-
-        <component>
-          <name>DRPC_SERVER</name>
-          <category>MASTER</category>
-          <commandScript>
-            <script>scripts/drpc_server.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-
-        <component>
-          <name>LOGVIEWER_SERVER</name>
-          <category>MASTER</category>
-          <commandScript>
-            <script>scripts/logviewer_server.py</script>
-            <scriptType>PYTHON</scriptType>
-            <timeout>600</timeout>
-          </commandScript>
-        </component>
-      </components>
-
-
-      <osSpecifics>
-        <osSpecific>
-          <osType>any</osType>
-          <packages>
-            <package>
-              <type>rpm</type>
-              <name>storm</name>
-            </package>
-          </packages>
-        </osSpecific>
-      </osSpecifics>
-
-      <commandScript>
-        <script>scripts/service_check.py</script>
-        <scriptType>PYTHON</scriptType>
-        <timeout>300</timeout>
-      </commandScript>
-
-      <configuration-dependencies>
-      </configuration-dependencies>
-    </service>
-  </services>
-</metainfo>

http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metrics.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metrics.json b/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metrics.json
deleted file mode 100644
index 59bec39..0000000
--- a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/metrics.json
+++ /dev/null
@@ -1,99 +0,0 @@
-{
-    "STORM_REST_API": {
-        "Component": [
-            {
-                "type": "jmx",
-                "metrics": {
-                    "metrics/api/cluster/summary/tasks.total": {
-                        "metric": "tasks.total",
-                        "pointInTime": true,
-                        "temporal": false
-                    },
-                    "metrics/api/cluster/summary/slots.total": {
-                      "metric": "slots.total",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/slots.free": {
-                      "metric": "slots.free",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/supervisors": {
-                      "metric": "supervisors",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/executors.total": {
-                      "metric": "executors.total",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/slots.used": {
-                      "metric": "slots.used",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/topologies": {
-                      "metric": "topologies",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/nimbus.uptime": {
-                      "metric": "nimbus.uptime",
-                      "pointInTime": true,
-                      "temporal": false
-                    }
-                }
-            }
-        ],
-        "HostComponent": [
-            {
-                "type": "jmx",
-                "metrics": {
-                    "metrics/api/cluster/summary/tasks.total": {
-                      "metric": "tasks.total",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/slots.total": {
-                      "metric": "slots.total",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/slots.free": {
-                      "metric": "slots.free",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/supervisors": {
-                      "metric": "supervisors",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/executors.total": {
-                      "metric": "executors.total",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/slots.used": {
-                      "metric": "slots.used",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/topologies": {
-                      "metric": "topologies",
-                      "pointInTime": true,
-                      "temporal": false
-                    },
-                    "metrics/api/cluster/summary/nimbus.uptime": {
-                      "metric": "nimbus.uptime",
-                      "pointInTime": true,
-                      "temporal": false
-                    }
-                }
-            }
-
-        ]
-    }
-}

http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/package/placeholder.txt
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/package/placeholder.txt b/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/package/placeholder.txt
deleted file mode 100644
index e2479f3..0000000
--- a/ambari-server/src/test/resources/stacks/HDP/2.1.1/services/STORM/package/placeholder.txt
+++ /dev/null
@@ -1,17 +0,0 @@
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements.  See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership.  The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License.  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-Ambari Agent

http://git-wip-us.apache.org/repos/asf/ambari/blob/7646ec2f/ambari-server/src/test/resources/storm_rest_api_jmx.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/resources/storm_rest_api_jmx.json b/ambari-server/src/test/resources/storm_rest_api_jmx.json
deleted file mode 100644
index d099d13..0000000
--- a/ambari-server/src/test/resources/storm_rest_api_jmx.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
-  "tasks.total" : 28,
-  "slots.total" : 8,
-  "slots.free" : 5,
-  "supervisors" : 2,
-  "executors.total" : 28,
-  "slots.used" : 3,
-  "topologies" : 1,
-  "nimbus.uptime" : 4637
-}