You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by vb...@apache.org on 2014/04/04 16:30:04 UTC

[1/2] git commit: (vbrodetskyi)

Repository: ambari
Updated Branches:
  refs/heads/trunk 994631c3b -> f96d22e0c


(vbrodetskyi)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/ae08e109
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/ae08e109
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/ae08e109

Branch: refs/heads/trunk
Commit: ae08e109b0edb0c11292b048288e856c1235874e
Parents: 994631c
Author: Vitaly Brodetskyi <vb...@hortonworks.com>
Authored: Fri Apr 4 17:17:52 2014 +0300
Committer: Vitaly Brodetskyi <vb...@hortonworks.com>
Committed: Fri Apr 4 17:21:24 2014 +0300

----------------------------------------------------------------------
 .../ganglia/GangliaHostPropertyProvider.java    |   2 +-
 .../ganglia/GangliaPropertyProvider.java        |  12 +-
 .../GANGLIA/package/scripts/ganglia_monitor.py  | 125 +------------------
 .../package/templates/gangliaClusters.conf.j2   |   2 +-
 .../package/templates/hadoop-services.cfg.j2    |   8 +-
 .../GANGLIA/package/scripts/ganglia_monitor.py  |  87 +------------
 .../package/templates/gangliaClusters.conf.j2   |   2 +-
 ...doop-metrics2-hbase.properties-GANGLIA-RS.j2 |   8 +-
 .../package/templates/hadoop-services.cfg.j2    |   8 +-
 .../ganglia/GangliaPropertyProviderTest.java    |  22 ++--
 .../1.3.2/GANGLIA/test_ganglia_monitor.py       |  81 +-----------
 .../2.0.6/GANGLIA/test_ganglia_monitor.py       |  25 +---
 ambari-web/app/views/main/dashboard.js          |   2 +-
 13 files changed, 41 insertions(+), 343 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
index c87cd3a..74ff7a2 100644
--- a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
+++ b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
@@ -39,7 +39,7 @@ public class GangliaHostPropertyProvider extends GangliaPropertyProvider{
 
   static {
     GANGLIA_CLUSTER_NAMES.add("HDPNameNode");
-    GANGLIA_CLUSTER_NAMES.add("HDPSlaves");
+    GANGLIA_CLUSTER_NAMES.add("HostMetrics");
     GANGLIA_CLUSTER_NAMES.add("HDPJobTracker");
     GANGLIA_CLUSTER_NAMES.add("HDPResourceManager");
     GANGLIA_CLUSTER_NAMES.add("HDPHBaseMaster");

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
index b81c024..39db257 100644
--- a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
+++ b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
@@ -59,16 +59,16 @@ public abstract class GangliaPropertyProvider extends AbstractPropertyProvider {
   
   static {
     GANGLIA_CLUSTER_NAME_MAP.put("NAMENODE",           Arrays.asList("HDPNameNode"));
-    GANGLIA_CLUSTER_NAME_MAP.put("DATANODE",           Arrays.asList("HDPDataNode", "HDPSlaves"));
+    GANGLIA_CLUSTER_NAME_MAP.put("DATANODE",           Arrays.asList("HDPDataNode", "HostMetrics"));
     GANGLIA_CLUSTER_NAME_MAP.put("JOBTRACKER",         Arrays.asList("HDPJobTracker"));
-    GANGLIA_CLUSTER_NAME_MAP.put("TASKTRACKER",        Arrays.asList("HDPTaskTracker", "HDPSlaves"));
+    GANGLIA_CLUSTER_NAME_MAP.put("TASKTRACKER",        Arrays.asList("HDPTaskTracker", "HostMetrics"));
     GANGLIA_CLUSTER_NAME_MAP.put("RESOURCEMANAGER",    Arrays.asList("HDPResourceManager"));
-    GANGLIA_CLUSTER_NAME_MAP.put("NODEMANAGER",        Arrays.asList("HDPNodeManager", "HDPSlaves"));
+    GANGLIA_CLUSTER_NAME_MAP.put("NODEMANAGER",        Arrays.asList("HDPNodeManager", "HostMetrics"));
     GANGLIA_CLUSTER_NAME_MAP.put("HISTORYSERVER",      Arrays.asList("HDPHistoryServer"));
     GANGLIA_CLUSTER_NAME_MAP.put("HBASE_MASTER",       Arrays.asList("HDPHBaseMaster"));
-    GANGLIA_CLUSTER_NAME_MAP.put("HBASE_REGIONSERVER", Arrays.asList("HDPHBaseRegionServer", "HDPSlaves"));
-    GANGLIA_CLUSTER_NAME_MAP.put("FLUME_SERVER",       Arrays.asList("HDPFlumeServer", "HDPSlaves"));
-    GANGLIA_CLUSTER_NAME_MAP.put("JOURNALNODE",        Arrays.asList("HDPJournalNode", "HDPSlaves"));
+    GANGLIA_CLUSTER_NAME_MAP.put("HBASE_REGIONSERVER", Arrays.asList("HDPHBaseRegionServer", "HostMetrics"));
+    GANGLIA_CLUSTER_NAME_MAP.put("FLUME_SERVER",       Arrays.asList("HDPFlumeServer", "HostMetrics"));
+    GANGLIA_CLUSTER_NAME_MAP.put("JOURNALNODE",        Arrays.asList("HDPJournalNode", "HostMetrics"));
     GANGLIA_CLUSTER_NAME_MAP.put("NIMBUS",             Arrays.asList("HDPNimbus"));
     GANGLIA_CLUSTER_NAME_MAP.put("SUPERVISOR",         Arrays.asList("HDPSupervisor"));
   }

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
index 2ab3a3c..9f0aa3e 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
@@ -100,131 +100,18 @@ class GangliaMonitor(Script):
   def generate_slave_configs(self):
     import params
 
-    if params.is_namenode_master:
-      generate_daemon("gmond",
-                      name = "HDPNameNode",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_jtnode_master:
-      generate_daemon("gmond",
-                      name = "HDPJobTracker",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_hsnode_master:
-      generate_daemon("gmond",
-                      name = "HDPHistoryServer",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_hbase_master:
-      generate_daemon("gmond",
-                      name = "HDPHBaseMaster",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_slave:
-      generate_daemon("gmond",
-                      name = "HDPDataNode",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_tasktracker:
-      generate_daemon("gmond",
-                      name = "HDPTaskTracker",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_hbase_rs:
-      generate_daemon("gmond",
-                      name = "HDPHBaseRegionServer",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_flume:
-      generate_daemon("gmond",
-                      name = "HDPFlumeServer",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.pure_slave:
-      generate_daemon("gmond",
-                      name = "HDPSlaves",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
+    generate_daemon("gmond",
+                    name = "HostMetrics",
+                    role = "monitor",
+                    owner = "root",
+                    group = params.user_group)
 
 
   def generate_master_configs(self):
     import params
 
-    if params.has_namenodes:
-      generate_daemon("gmond",
-                      name = "HDPNameNode",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.has_jobtracker:
-      generate_daemon("gmond",
-                      name = "HDPJobTracker",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.has_hbase_masters:
-      generate_daemon("gmond",
-                      name = "HDPHBaseMaster",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.has_historyserver:
-      generate_daemon("gmond",
-                      name = "HDPHistoryServer",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.has_slaves:
-      generate_daemon("gmond",
-                      name = "HDPDataNode",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.has_tasktracker:
-      generate_daemon("gmond",
-                      name = "HDPTaskTracker",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.has_hbase_rs:
-      generate_daemon("gmond",
-                      name = "HDPHBaseRegionServer",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.has_flume:
-      generate_daemon("gmond",
-                      name = "HDPFlumeServer",
-                      role = "server",
-                      owner = "root",
-                      group = params.user_group)
-
     generate_daemon("gmond",
-                    name = "HDPSlaves",
+                    name = "HostMetrics",
                     role = "server",
                     owner = "root",
                     group = params.user_group)

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2 b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
index b60a15c..a577bb5 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
@@ -45,7 +45,7 @@
     HDPNodeManager     	{{ganglia_server_host}}   8657
     HDPTaskTracker     	{{ganglia_server_host}}   8658
     HDPDataNode       	{{ganglia_server_host}}   8659
-    HDPSlaves       	{{ganglia_server_host}}   8660
+    HostMetrics       	{{ganglia_server_host}}   8660
     HDPNameNode         {{ganglia_server_host}}   8661
     HDPJobTracker     	{{ganglia_server_host}}  8662
     HDPHBaseMaster      {{ganglia_server_host}}   8663

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2 b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
index b45d110..2ec136f 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
@@ -189,9 +189,8 @@ define service {
 {% endif %}
 
 {% if hostgroup_defs['resourcemanager'] %}
-{% for hostname in hostgroup_defs['resourcemanager'] %}
 define service {
-        host_name	        {{ hostname }}
+        hostgroup_name          ganglia-server
         use                     hadoop-service
         service_description     GANGLIA::Ganglia Monitor process for ResourceManager
         servicegroups           GANGLIA
@@ -200,13 +199,11 @@ define service {
         retry_check_interval    0.25
         max_check_attempts      4
 }
-{% endfor %}
 {% endif %}
 
 {% if hostgroup_defs['historyserver2'] %}
-{% for hostname in hostgroup_defs['historyserver2'] %}
 define service {
-        host_name	        {{ hostname }}
+        hostgroup_name          ganglia-server
         use                     hadoop-service
         service_description     GANGLIA::Ganglia Monitor process for HistoryServer
         servicegroups           GANGLIA
@@ -215,7 +212,6 @@ define service {
         retry_check_interval    0.25
         max_check_attempts      4
 }
-{% endfor %}
 {% endif %}
 
 {% endif %}

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
index d3f097f..94d06fb 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
@@ -100,95 +100,12 @@ class GangliaMonitor(Script):
   def generate_slave_configs(self):
     import params
 
-    if params.is_namenode_master:
-      generate_daemon("gmond",
-                      name = "HDPNameNode",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_jtnode_master:
-      generate_daemon("gmond",
-                      name = "HDPJobTracker",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_rmnode_master:
-      generate_daemon("gmond",
-                      name = "HDPResourceManager",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_hsnode_master:
-      generate_daemon("gmond",
-                      name = "HDPHistoryServer",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_nimbus_host:
-      generate_daemon("gmond",
-                      name = "HDPNimbus",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-    if params.is_supervisor_host:
-      generate_daemon("gmond",
-                      name = "HDPSupervisor",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_hbase_master:
-      generate_daemon("gmond",
-                      name = "HDPHBaseMaster",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
     generate_daemon("gmond",
-                    name = "HDPSlaves",
+                    name = "HostMetrics",
                     role = "monitor",
                     owner = "root",
                     group = params.user_group)
 
-    if params.is_slave:
-      generate_daemon("gmond",
-                      name = "HDPDataNode",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_tasktracker:
-      generate_daemon("gmond",
-                      name = "HDPTaskTracker",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_hbase_rs:
-      generate_daemon("gmond",
-                      name = "HDPHBaseRegionServer",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_flume:
-      generate_daemon("gmond",
-                      name = "HDPFlumeServer",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
-    if params.is_jn_host:
-      generate_daemon("gmond",
-                      name = "HDPJournalNode",
-                      role = "monitor",
-                      owner = "root",
-                      group = params.user_group)
-
 
   def generate_master_configs(self):
     import params
@@ -285,7 +202,7 @@ class GangliaMonitor(Script):
                       group = params.user_group)
 
     generate_daemon("gmond",
-                    name = "HDPSlaves",
+                    name = "HostMetrics",
                     role = "server",
                     owner = "root",
                     group = params.user_group)

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
index 1672d92..a8fae84 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
@@ -44,7 +44,7 @@
     HDPNodeManager     	{{ganglia_server_host}}   8657
     HDPTaskTracker     	{{ganglia_server_host}}   8658
     HDPDataNode       	{{ganglia_server_host}}   8659
-    HDPSlaves       	{{ganglia_server_host}}   8660
+    HostMetrics       	{{ganglia_server_host}}   8660
     HDPNameNode         {{ganglia_server_host}}   8661
     HDPJobTracker     	{{ganglia_server_host}}   8662
     HDPHBaseMaster      {{ganglia_server_host}}   8663

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
index 2c5ee5e..55e8461 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
@@ -51,21 +51,21 @@ hbase.extendedperiod = 3600
 # hbase.class=org.apache.hadoop.metrics.ganglia.GangliaContext
 hbase.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
 hbase.period=10
-hbase.servers={{ganglia_server_host}}:8660
+hbase.servers={{ganglia_server_host}}:8656
 
 # Configuration of the "jvm" context for ganglia
 # Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)
 # jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext
 jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
 jvm.period=10
-jvm.servers={{ganglia_server_host}}:8660
+jvm.servers={{ganglia_server_host}}:8656
 
 # Configuration of the "rpc" context for ganglia
 # Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)
 # rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext
 rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
 rpc.period=10
-rpc.servers={{ganglia_server_host}}:8660
+rpc.servers={{ganglia_server_host}}:8656
 
 #Ganglia following hadoop example
 hbase.sink.ganglia.class=org.apache.hadoop.metrics2.sink.ganglia.GangliaSink31
@@ -77,4 +77,4 @@ hbase.sink.ganglia.period=10
 .sink.ganglia.slope=jvm.metrics.gcCount=zero,jvm.metrics.memHeapUsedM=both
 .sink.ganglia.dmax=jvm.metrics.threadsBlocked=70,jvm.metrics.memHeapUsedM=40
 
-hbase.sink.ganglia.servers={{ganglia_server_host}}:8660
+hbase.sink.ganglia.servers={{ganglia_server_host}}:8656

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
index b77e77b..b4be13a 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
@@ -204,9 +204,8 @@ define service {
 {% endif %}
 
 {% if hostgroup_defs['resourcemanager'] %}
-{% for hostname in hostgroup_defs['resourcemanager'] %}
 define service {
-        host_name	        {{ hostname }}
+        hostgroup_name          ganglia-server
         use                     hadoop-service
         service_description     GANGLIA::Ganglia Monitor process for ResourceManager
         servicegroups           GANGLIA
@@ -215,13 +214,11 @@ define service {
         retry_check_interval    0.25
         max_check_attempts      4
 }
-{% endfor %}
 {% endif %}
 
 {% if hostgroup_defs['historyserver2'] %}
-{% for hostname in hostgroup_defs['historyserver2'] %}
 define service {
-        host_name	        {{ hostname }}
+        hostgroup_name          ganglia-server
         use                     hadoop-service
         service_description     GANGLIA::Ganglia Monitor process for HistoryServer
         servicegroups           GANGLIA
@@ -230,7 +227,6 @@ define service {
         retry_check_interval    0.25
         max_check_attempts      4
 }
-{% endfor %}
 {% endif %}
 
 {% endif %}

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
index 4e827ab..07b7d6b 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
@@ -125,7 +125,7 @@ public class GangliaPropertyProviderTest {
 
 
     String expected = (configuration.isGangliaSSL() ? "https" : "http") +
-        "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPDataNode%2CHDPSlaves&h=domU-12-31-39-0E-34-E1.compute-1.internal&m=jvm.metrics.gcCount&s=10&e=20&r=1";
+        "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPDataNode%2CHostMetrics&h=domU-12-31-39-0E-34-E1.compute-1.internal&m=jvm.metrics.gcCount&s=10&e=20&r=1";
     Assert.assertEquals(expected, streamProvider.getLastSpec());
 
     Assert.assertEquals(3, PropertyHelper.getProperties(resource).size());
@@ -175,7 +175,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPTaskTracker,HDPSlaves");
+    expectedUri.setParameter("c", "HDPTaskTracker,HostMetrics");
     expectedUri.setParameter("h", "domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setParameter("m", metricsList);
     expectedUri.setParameter("s", "10");
@@ -323,7 +323,7 @@ public class GangliaPropertyProviderTest {
     uriBuilder.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     uriBuilder.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     uriBuilder.setPath("/cgi-bin/rrd.py");
-    uriBuilder.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HDPSlaves,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
+    uriBuilder.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HostMetrics,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
     uriBuilder.setParameter("h", "domU-12-31-39-0E-34-E3.compute-1.internal,domU-12-31-39-0E-34-E1.compute-1.internal,domU-12-31-39-0E-34-E2.compute-1.internal");
     uriBuilder.setParameter("m", "jvm.metrics.gcCount");
     uriBuilder.setParameter("s", "10");
@@ -382,7 +382,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HDPSlaves,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
+    expectedUri.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HostMetrics,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
    
     expectedUri.setParameter("h", hostsList.toString());
     expectedUri.setParameter("m", "jvm.metrics.gcCount");
@@ -438,7 +438,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+    expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
     expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
     expectedUri.setParameter("m", metricsList);
     expectedUri.setParameter("s", "10");
@@ -501,7 +501,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+    expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
     expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
     expectedUri.setParameter("m", metricsList);
     expectedUri.setParameter("e", "now");
@@ -546,7 +546,7 @@ public class GangliaPropertyProviderTest {
     Assert.assertEquals(1, propertyProvider.populateResources(Collections.singleton(resource), request, null).size());
 
     String expected = (configuration.isGangliaSSL() ? "https" : "http") +
-        "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPFlumeServer%2CHDPSlaves&h=ip-10-39-113-33.ec2.internal&m=";
+        "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPFlumeServer%2CHostMetrics&h=ip-10-39-113-33.ec2.internal&m=";
     
     Assert.assertTrue(streamProvider.getLastSpec().startsWith(expected));
 
@@ -593,7 +593,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+    expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
     expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
     expectedUri.setParameter("m", metricsList);
     expectedUri.setParameter("s", "10");
@@ -651,7 +651,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+    expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
     expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
     expectedUri.setParameter("m", metricsList);
     expectedUri.setParameter("s", "10");
@@ -710,7 +710,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+    expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
     expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
     expectedUri.setParameter("m", metricsList);
     expectedUri.setParameter("s", "10");
@@ -769,7 +769,7 @@ public class GangliaPropertyProviderTest {
     expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
     expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
     expectedUri.setPath("/cgi-bin/rrd.py");
-    expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+    expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
     expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
     expectedUri.setParameter("m", metricsList);
     expectedUri.setParameter("s", "10");

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py b/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
index d969263..b646871 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
+++ b/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
@@ -188,35 +188,7 @@ class TestGangliaMonitor(RMFTestCase):
         group = 'root',
         mode = 0755,
     )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPJobTracker -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHistoryServer -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPDataNode -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPTaskTracker -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseRegionServer -o root -g hadoop',
+    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -o root -g hadoop',
         path = ['/usr/libexec/hdp/ganglia',
            '/usr/sbin',
            '/sbin:/usr/local/bin',
@@ -242,56 +214,7 @@ class TestGangliaMonitor(RMFTestCase):
 
 
   def assert_gmond_master_conf_generated(self):
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPNameNode -m -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPJobTracker -m -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseMaster -m -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHistoryServer -m -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPDataNode -m -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPTaskTracker -m -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseRegionServer -m -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSlaves -m -o root -g hadoop',
+    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -m -o root -g hadoop',
         path = ['/usr/libexec/hdp/ganglia',
            '/usr/sbin',
            '/sbin:/usr/local/bin',

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py b/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
index b2f7700..e5acbd1 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
@@ -176,28 +176,7 @@ class TestGangliaMonitor(RMFTestCase):
         group = 'root',
         mode = 0755,
     )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPNameNode -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPNimbus -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSupervisor -o root -g hadoop',
-        path = ['/usr/libexec/hdp/ganglia',
-           '/usr/sbin',
-           '/sbin:/usr/local/bin',
-           '/bin',
-           '/usr/bin'],
-    )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSlaves -o root -g hadoop',
+    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -o root -g hadoop',
         path = ['/usr/libexec/hdp/ganglia',
            '/usr/sbin',
            '/sbin:/usr/local/bin',
@@ -286,7 +265,7 @@ class TestGangliaMonitor(RMFTestCase):
            '/bin',
            '/usr/bin'],
     )
-    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSlaves -m -o root -g hadoop',
+    self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -m -o root -g hadoop',
         path = ['/usr/libexec/hdp/ganglia',
            '/usr/sbin',
            '/sbin:/usr/local/bin',

http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-web/app/views/main/dashboard.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/views/main/dashboard.js b/ambari-web/app/views/main/dashboard.js
index ad49d27..de14dcb 100644
--- a/ambari-web/app/views/main/dashboard.js
+++ b/ambari-web/app/views/main/dashboard.js
@@ -526,7 +526,7 @@ App.MainDashboardView = Em.View.extend(App.UserPref, {
   },
 
   gangliaUrl: function () {
-    return App.router.get('clusterController.gangliaUrl') + "/?r=hour&cs=&ce=&m=&s=by+name&c=HDPSlaves&tab=m&vn=";
+    return App.router.get('clusterController.gangliaUrl') + "/?r=hour&cs=&ce=&m=&s=by+name&c=HostMetrics&tab=m&vn=";
   }.property('App.router.clusterController.gangliaUrl'),
 
   showAlertsPopup: function (event) {


[2/2] git commit: AMBARI-5343. Ambari hiveserver2 smoke test wrong for hiveserver 2 that uses LDAP authentication.(vbrodetskyi)

Posted by vb...@apache.org.
AMBARI-5343. Ambari hiveserver2 smoke test wrong for hiveserver 2 that uses LDAP authentication.(vbrodetskyi)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/f96d22e0
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/f96d22e0
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/f96d22e0

Branch: refs/heads/trunk
Commit: f96d22e0c27d1547f485b5b52998436d44ac8a75
Parents: ae08e10
Author: Vitaly Brodetskyi <vb...@hortonworks.com>
Authored: Fri Apr 4 17:29:19 2014 +0300
Committer: Vitaly Brodetskyi <vb...@hortonworks.com>
Committed: Fri Apr 4 17:29:19 2014 +0300

----------------------------------------------------------------------
 .../services/HIVE/package/scripts/params.py     |  3 +-
 .../HIVE/package/scripts/service_check.py       | 37 ++++++++------------
 .../services/HIVE/package/scripts/params.py     |  3 +-
 .../HIVE/package/scripts/service_check.py       | 37 ++++++++------------
 .../1.3.2/HIVE/test_hive_service_check.py       | 34 +++---------------
 .../2.0.6/HIVE/test_hive_service_check.py       | 36 +++----------------
 6 files changed, 41 insertions(+), 109 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
index 968ade0..e2751a9 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
@@ -47,7 +47,8 @@ check_db_connection_jar = format("/usr/lib/ambari-agent/{check_db_connection_jar
 hive_metastore_port = config['configurations']['global']['hive_metastore_port']
 hive_var_lib = '/var/lib/hive'
 hive_server_host = config['clusterHostInfo']['hive_server_host'][0]
-hive_url = format("jdbc:hive2://{hive_server_host}:10000")
+hive_server_port = 10000
+hive_url = format("jdbc:hive2://{hive_server_host}:{hive_server_port}")
 
 smokeuser = config['configurations']['global']['smokeuser']
 smoke_test_sql = "/tmp/hiveserver2.sql"

http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
index 111e8a1..09ba1bf 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
@@ -19,6 +19,8 @@ limitations under the License.
 """
 
 from resource_management import *
+import socket
+import sys
 
 from hcat_service_check import hcat_service_check
 
@@ -26,29 +28,18 @@ class HiveServiceCheck(Script):
   def service_check(self, env):
     import params
     env.set_params(params)
-    if params.security_enabled:
-      kinit_cmd = format("{kinit_path_local} -kt {smoke_user_keytab} {smokeuser};")
-      hive_principal_ext = format("principal={hive_metastore_keytab_path}")
-      hive_url_ext = format("{hive_url}/\\;{hive_principal_ext}")
-      smoke_cmd = format("{kinit_cmd} env JAVA_HOME={java64_home} {smoke_test_path} {hive_url_ext} {smoke_test_sql}")
-    else:
-      smoke_cmd = format("env JAVA_HOME={java64_home} {smoke_test_path} {hive_url} {smoke_test_sql}")
-
-    File(params.smoke_test_path,
-         content=StaticFile('hiveserver2Smoke.sh'),
-         mode=0755
-    )
-
-    File(params.smoke_test_sql,
-         content=StaticFile('hiveserver2.sql')
-    )
-
-    Execute(smoke_cmd,
-            tries=3,
-            try_sleep=5,
-            path='/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin',
-            logoutput=True,
-            user=params.smokeuser)
+
+    address=format("{hive_server_host}")
+    port=int(format("{hive_server_port}"))
+    s = socket.socket()
+    print "Test connectivity to hive server"
+    try:
+      s.connect((address, port))
+      print "Successfully connected to %s on port %s" % (address, port)
+      s.close()
+    except socket.error, e:
+      print "Connection to %s on port %s failed: %s" % (address, port, e)
+      sys.exit(1)
 
     hcat_service_check()
 

http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
index f8f5231..520343b 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
@@ -50,7 +50,8 @@ hive_metastore_port = config['configurations']['global']['hive_metastore_port']
 hive_var_lib = '/var/lib/hive'
 hive_bin = '/usr/lib/hive/bin'
 hive_server_host = config['clusterHostInfo']['hive_server_host'][0]
-hive_url = format("jdbc:hive2://{hive_server_host}:10000")
+hive_server_port = 10000
+hive_url = format("jdbc:hive2://{hive_server_host}:{hive_server_port}")
 
 smokeuser = config['configurations']['global']['smokeuser']
 smoke_test_sql = "/tmp/hiveserver2.sql"

http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
index 111e8a1..09ba1bf 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
@@ -19,6 +19,8 @@ limitations under the License.
 """
 
 from resource_management import *
+import socket
+import sys
 
 from hcat_service_check import hcat_service_check
 
@@ -26,29 +28,18 @@ class HiveServiceCheck(Script):
   def service_check(self, env):
     import params
     env.set_params(params)
-    if params.security_enabled:
-      kinit_cmd = format("{kinit_path_local} -kt {smoke_user_keytab} {smokeuser};")
-      hive_principal_ext = format("principal={hive_metastore_keytab_path}")
-      hive_url_ext = format("{hive_url}/\\;{hive_principal_ext}")
-      smoke_cmd = format("{kinit_cmd} env JAVA_HOME={java64_home} {smoke_test_path} {hive_url_ext} {smoke_test_sql}")
-    else:
-      smoke_cmd = format("env JAVA_HOME={java64_home} {smoke_test_path} {hive_url} {smoke_test_sql}")
-
-    File(params.smoke_test_path,
-         content=StaticFile('hiveserver2Smoke.sh'),
-         mode=0755
-    )
-
-    File(params.smoke_test_sql,
-         content=StaticFile('hiveserver2.sql')
-    )
-
-    Execute(smoke_cmd,
-            tries=3,
-            try_sleep=5,
-            path='/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin',
-            logoutput=True,
-            user=params.smokeuser)
+
+    address=format("{hive_server_host}")
+    port=int(format("{hive_server_port}"))
+    s = socket.socket()
+    print "Test connectivity to hive server"
+    try:
+      s.connect((address, port))
+      print "Successfully connected to %s on port %s" % (address, port)
+      s.close()
+    except socket.error, e:
+      print "Connection to %s on port %s failed: %s" % (address, port, e)
+      sys.exit(1)
 
     hcat_service_check()
 

http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py b/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
index f1deec7..237be7a 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
+++ b/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
@@ -24,27 +24,14 @@ import  resource_management.libraries.functions
 @patch.object(resource_management.libraries.functions, "get_unique_id_and_date", new = MagicMock(return_value=''))
 class TestServiceCheck(RMFTestCase):
 
-  def test_service_check_default(self):
+  @patch("sys.exit")
+  def test_service_check_default(self, sys_exit_mock):
 
     self.executeScript("1.3.2/services/HIVE/package/scripts/service_check.py",
                         classname="HiveServiceCheck",
                         command="service_check",
                         config_file="default.json"
     )
-    self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
-                        content = StaticFile('hiveserver2Smoke.sh'),
-                        mode = 0755,
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
-                        content = StaticFile('hiveserver2.sql'),
-    )
-    self.assertResourceCalled('Execute', "env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000 /tmp/hiveserver2.sql",
-                        logoutput = True,
-                        path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
     self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
                         content = StaticFile('hcatSmoke.sh'),
                         mode = 0755,
@@ -73,27 +60,14 @@ class TestServiceCheck(RMFTestCase):
     )
     self.assertNoMoreResources()
 
-  def test_service_check_secured(self):
+  @patch("sys.exit")
+  def test_service_check_secured(self, sys_exit_mock):
 
     self.executeScript("1.3.2/services/HIVE/package/scripts/service_check.py",
                         classname="HiveServiceCheck",
                         command="service_check",
                         config_file="secured.json"
     )
-    self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
-                        content = StaticFile('hiveserver2Smoke.sh'),
-                        mode = 0755,
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
-                        content = StaticFile('hiveserver2.sql'),
-    )
-    self.assertResourceCalled('Execute', "/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa; env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000/\\;principal=/etc/security/keytabs/hive.service.keytab /tmp/hiveserver2.sql",
-                        logoutput = True,
-                        path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
     self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
                         content = StaticFile('hcatSmoke.sh'),
                         mode = 0755,

http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
index 6390a70..7eb5810 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
@@ -19,32 +19,19 @@ limitations under the License.
 '''
 from mock.mock import MagicMock, call, patch
 from stacks.utils.RMFTestCase import *
-import datetime
+import datetime, sys
 import  resource_management.libraries.functions
 @patch.object(resource_management.libraries.functions, "get_unique_id_and_date", new = MagicMock(return_value=''))
 class TestServiceCheck(RMFTestCase):
 
-  def test_service_check_default(self):
+  @patch("sys.exit")
+  def test_service_check_default(self, sys_exit_mock):
 
     self.executeScript("2.0.6/services/HIVE/package/scripts/service_check.py",
                         classname="HiveServiceCheck",
                         command="service_check",
                         config_file="default.json"
     )
-    self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
-                        content = StaticFile('hiveserver2Smoke.sh'),
-                        mode = 0755,
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
-                        content = StaticFile('hiveserver2.sql'),
-    )
-    self.assertResourceCalled('Execute', "env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000 /tmp/hiveserver2.sql",
-                        logoutput = True,
-                        path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
     self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
                         content = StaticFile('hcatSmoke.sh'),
                         mode = 0755,
@@ -73,27 +60,14 @@ class TestServiceCheck(RMFTestCase):
     )
     self.assertNoMoreResources()
 
-  def test_service_check_secured(self):
+  @patch("sys.exit")
+  def test_service_check_secured(self, sys_exit_mock):
 
     self.executeScript("2.0.6/services/HIVE/package/scripts/service_check.py",
                         classname="HiveServiceCheck",
                         command="service_check",
                         config_file="secured.json"
     )
-    self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
-                        content = StaticFile('hiveserver2Smoke.sh'),
-                        mode = 0755,
-    )
-    self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
-                        content = StaticFile('hiveserver2.sql'),
-    )
-    self.assertResourceCalled('Execute', "/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa; env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000/\\;principal=/etc/security/keytabs/hive.service.keytab /tmp/hiveserver2.sql",
-                        logoutput = True,
-                        path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
-                        tries = 3,
-                        user = 'ambari-qa',
-                        try_sleep = 5,
-    )
     self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
                         content = StaticFile('hcatSmoke.sh'),
                         mode = 0755,