You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by vb...@apache.org on 2014/04/04 16:30:04 UTC
[1/2] git commit: (vbrodetskyi)
Repository: ambari
Updated Branches:
refs/heads/trunk 994631c3b -> f96d22e0c
(vbrodetskyi)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/ae08e109
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/ae08e109
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/ae08e109
Branch: refs/heads/trunk
Commit: ae08e109b0edb0c11292b048288e856c1235874e
Parents: 994631c
Author: Vitaly Brodetskyi <vb...@hortonworks.com>
Authored: Fri Apr 4 17:17:52 2014 +0300
Committer: Vitaly Brodetskyi <vb...@hortonworks.com>
Committed: Fri Apr 4 17:21:24 2014 +0300
----------------------------------------------------------------------
.../ganglia/GangliaHostPropertyProvider.java | 2 +-
.../ganglia/GangliaPropertyProvider.java | 12 +-
.../GANGLIA/package/scripts/ganglia_monitor.py | 125 +------------------
.../package/templates/gangliaClusters.conf.j2 | 2 +-
.../package/templates/hadoop-services.cfg.j2 | 8 +-
.../GANGLIA/package/scripts/ganglia_monitor.py | 87 +------------
.../package/templates/gangliaClusters.conf.j2 | 2 +-
...doop-metrics2-hbase.properties-GANGLIA-RS.j2 | 8 +-
.../package/templates/hadoop-services.cfg.j2 | 8 +-
.../ganglia/GangliaPropertyProviderTest.java | 22 ++--
.../1.3.2/GANGLIA/test_ganglia_monitor.py | 81 +-----------
.../2.0.6/GANGLIA/test_ganglia_monitor.py | 25 +---
ambari-web/app/views/main/dashboard.js | 2 +-
13 files changed, 41 insertions(+), 343 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
index c87cd3a..74ff7a2 100644
--- a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
+++ b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaHostPropertyProvider.java
@@ -39,7 +39,7 @@ public class GangliaHostPropertyProvider extends GangliaPropertyProvider{
static {
GANGLIA_CLUSTER_NAMES.add("HDPNameNode");
- GANGLIA_CLUSTER_NAMES.add("HDPSlaves");
+ GANGLIA_CLUSTER_NAMES.add("HostMetrics");
GANGLIA_CLUSTER_NAMES.add("HDPJobTracker");
GANGLIA_CLUSTER_NAMES.add("HDPResourceManager");
GANGLIA_CLUSTER_NAMES.add("HDPHBaseMaster");
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
index b81c024..39db257 100644
--- a/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
+++ b/ambari-server/src/main/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProvider.java
@@ -59,16 +59,16 @@ public abstract class GangliaPropertyProvider extends AbstractPropertyProvider {
static {
GANGLIA_CLUSTER_NAME_MAP.put("NAMENODE", Arrays.asList("HDPNameNode"));
- GANGLIA_CLUSTER_NAME_MAP.put("DATANODE", Arrays.asList("HDPDataNode", "HDPSlaves"));
+ GANGLIA_CLUSTER_NAME_MAP.put("DATANODE", Arrays.asList("HDPDataNode", "HostMetrics"));
GANGLIA_CLUSTER_NAME_MAP.put("JOBTRACKER", Arrays.asList("HDPJobTracker"));
- GANGLIA_CLUSTER_NAME_MAP.put("TASKTRACKER", Arrays.asList("HDPTaskTracker", "HDPSlaves"));
+ GANGLIA_CLUSTER_NAME_MAP.put("TASKTRACKER", Arrays.asList("HDPTaskTracker", "HostMetrics"));
GANGLIA_CLUSTER_NAME_MAP.put("RESOURCEMANAGER", Arrays.asList("HDPResourceManager"));
- GANGLIA_CLUSTER_NAME_MAP.put("NODEMANAGER", Arrays.asList("HDPNodeManager", "HDPSlaves"));
+ GANGLIA_CLUSTER_NAME_MAP.put("NODEMANAGER", Arrays.asList("HDPNodeManager", "HostMetrics"));
GANGLIA_CLUSTER_NAME_MAP.put("HISTORYSERVER", Arrays.asList("HDPHistoryServer"));
GANGLIA_CLUSTER_NAME_MAP.put("HBASE_MASTER", Arrays.asList("HDPHBaseMaster"));
- GANGLIA_CLUSTER_NAME_MAP.put("HBASE_REGIONSERVER", Arrays.asList("HDPHBaseRegionServer", "HDPSlaves"));
- GANGLIA_CLUSTER_NAME_MAP.put("FLUME_SERVER", Arrays.asList("HDPFlumeServer", "HDPSlaves"));
- GANGLIA_CLUSTER_NAME_MAP.put("JOURNALNODE", Arrays.asList("HDPJournalNode", "HDPSlaves"));
+ GANGLIA_CLUSTER_NAME_MAP.put("HBASE_REGIONSERVER", Arrays.asList("HDPHBaseRegionServer", "HostMetrics"));
+ GANGLIA_CLUSTER_NAME_MAP.put("FLUME_SERVER", Arrays.asList("HDPFlumeServer", "HostMetrics"));
+ GANGLIA_CLUSTER_NAME_MAP.put("JOURNALNODE", Arrays.asList("HDPJournalNode", "HostMetrics"));
GANGLIA_CLUSTER_NAME_MAP.put("NIMBUS", Arrays.asList("HDPNimbus"));
GANGLIA_CLUSTER_NAME_MAP.put("SUPERVISOR", Arrays.asList("HDPSupervisor"));
}
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
index 2ab3a3c..9f0aa3e 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/scripts/ganglia_monitor.py
@@ -100,131 +100,18 @@ class GangliaMonitor(Script):
def generate_slave_configs(self):
import params
- if params.is_namenode_master:
- generate_daemon("gmond",
- name = "HDPNameNode",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_jtnode_master:
- generate_daemon("gmond",
- name = "HDPJobTracker",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_hsnode_master:
- generate_daemon("gmond",
- name = "HDPHistoryServer",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_hbase_master:
- generate_daemon("gmond",
- name = "HDPHBaseMaster",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_slave:
- generate_daemon("gmond",
- name = "HDPDataNode",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_tasktracker:
- generate_daemon("gmond",
- name = "HDPTaskTracker",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_hbase_rs:
- generate_daemon("gmond",
- name = "HDPHBaseRegionServer",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_flume:
- generate_daemon("gmond",
- name = "HDPFlumeServer",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.pure_slave:
- generate_daemon("gmond",
- name = "HDPSlaves",
- role = "monitor",
- owner = "root",
- group = params.user_group)
+ generate_daemon("gmond",
+ name = "HostMetrics",
+ role = "monitor",
+ owner = "root",
+ group = params.user_group)
def generate_master_configs(self):
import params
- if params.has_namenodes:
- generate_daemon("gmond",
- name = "HDPNameNode",
- role = "server",
- owner = "root",
- group = params.user_group)
-
- if params.has_jobtracker:
- generate_daemon("gmond",
- name = "HDPJobTracker",
- role = "server",
- owner = "root",
- group = params.user_group)
-
- if params.has_hbase_masters:
- generate_daemon("gmond",
- name = "HDPHBaseMaster",
- role = "server",
- owner = "root",
- group = params.user_group)
-
- if params.has_historyserver:
- generate_daemon("gmond",
- name = "HDPHistoryServer",
- role = "server",
- owner = "root",
- group = params.user_group)
-
- if params.has_slaves:
- generate_daemon("gmond",
- name = "HDPDataNode",
- role = "server",
- owner = "root",
- group = params.user_group)
-
- if params.has_tasktracker:
- generate_daemon("gmond",
- name = "HDPTaskTracker",
- role = "server",
- owner = "root",
- group = params.user_group)
-
- if params.has_hbase_rs:
- generate_daemon("gmond",
- name = "HDPHBaseRegionServer",
- role = "server",
- owner = "root",
- group = params.user_group)
-
- if params.has_flume:
- generate_daemon("gmond",
- name = "HDPFlumeServer",
- role = "server",
- owner = "root",
- group = params.user_group)
-
generate_daemon("gmond",
- name = "HDPSlaves",
+ name = "HostMetrics",
role = "server",
owner = "root",
group = params.user_group)
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2 b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
index b60a15c..a577bb5 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/GANGLIA/package/templates/gangliaClusters.conf.j2
@@ -45,7 +45,7 @@
HDPNodeManager {{ganglia_server_host}} 8657
HDPTaskTracker {{ganglia_server_host}} 8658
HDPDataNode {{ganglia_server_host}} 8659
- HDPSlaves {{ganglia_server_host}} 8660
+ HostMetrics {{ganglia_server_host}} 8660
HDPNameNode {{ganglia_server_host}} 8661
HDPJobTracker {{ganglia_server_host}} 8662
HDPHBaseMaster {{ganglia_server_host}} 8663
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2 b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
index b45d110..2ec136f 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/NAGIOS/package/templates/hadoop-services.cfg.j2
@@ -189,9 +189,8 @@ define service {
{% endif %}
{% if hostgroup_defs['resourcemanager'] %}
-{% for hostname in hostgroup_defs['resourcemanager'] %}
define service {
- host_name {{ hostname }}
+ hostgroup_name ganglia-server
use hadoop-service
service_description GANGLIA::Ganglia Monitor process for ResourceManager
servicegroups GANGLIA
@@ -200,13 +199,11 @@ define service {
retry_check_interval 0.25
max_check_attempts 4
}
-{% endfor %}
{% endif %}
{% if hostgroup_defs['historyserver2'] %}
-{% for hostname in hostgroup_defs['historyserver2'] %}
define service {
- host_name {{ hostname }}
+ hostgroup_name ganglia-server
use hadoop-service
service_description GANGLIA::Ganglia Monitor process for HistoryServer
servicegroups GANGLIA
@@ -215,7 +212,6 @@ define service {
retry_check_interval 0.25
max_check_attempts 4
}
-{% endfor %}
{% endif %}
{% endif %}
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
index d3f097f..94d06fb 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/scripts/ganglia_monitor.py
@@ -100,95 +100,12 @@ class GangliaMonitor(Script):
def generate_slave_configs(self):
import params
- if params.is_namenode_master:
- generate_daemon("gmond",
- name = "HDPNameNode",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_jtnode_master:
- generate_daemon("gmond",
- name = "HDPJobTracker",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_rmnode_master:
- generate_daemon("gmond",
- name = "HDPResourceManager",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_hsnode_master:
- generate_daemon("gmond",
- name = "HDPHistoryServer",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_nimbus_host:
- generate_daemon("gmond",
- name = "HDPNimbus",
- role = "monitor",
- owner = "root",
- group = params.user_group)
- if params.is_supervisor_host:
- generate_daemon("gmond",
- name = "HDPSupervisor",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_hbase_master:
- generate_daemon("gmond",
- name = "HDPHBaseMaster",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
generate_daemon("gmond",
- name = "HDPSlaves",
+ name = "HostMetrics",
role = "monitor",
owner = "root",
group = params.user_group)
- if params.is_slave:
- generate_daemon("gmond",
- name = "HDPDataNode",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_tasktracker:
- generate_daemon("gmond",
- name = "HDPTaskTracker",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_hbase_rs:
- generate_daemon("gmond",
- name = "HDPHBaseRegionServer",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_flume:
- generate_daemon("gmond",
- name = "HDPFlumeServer",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
- if params.is_jn_host:
- generate_daemon("gmond",
- name = "HDPJournalNode",
- role = "monitor",
- owner = "root",
- group = params.user_group)
-
def generate_master_configs(self):
import params
@@ -285,7 +202,7 @@ class GangliaMonitor(Script):
group = params.user_group)
generate_daemon("gmond",
- name = "HDPSlaves",
+ name = "HostMetrics",
role = "server",
owner = "root",
group = params.user_group)
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
index 1672d92..a8fae84 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/templates/gangliaClusters.conf.j2
@@ -44,7 +44,7 @@
HDPNodeManager {{ganglia_server_host}} 8657
HDPTaskTracker {{ganglia_server_host}} 8658
HDPDataNode {{ganglia_server_host}} 8659
- HDPSlaves {{ganglia_server_host}} 8660
+ HostMetrics {{ganglia_server_host}} 8660
HDPNameNode {{ganglia_server_host}} 8661
HDPJobTracker {{ganglia_server_host}} 8662
HDPHBaseMaster {{ganglia_server_host}} 8663
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
index 2c5ee5e..55e8461 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/package/templates/hadoop-metrics2-hbase.properties-GANGLIA-RS.j2
@@ -51,21 +51,21 @@ hbase.extendedperiod = 3600
# hbase.class=org.apache.hadoop.metrics.ganglia.GangliaContext
hbase.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
hbase.period=10
-hbase.servers={{ganglia_server_host}}:8660
+hbase.servers={{ganglia_server_host}}:8656
# Configuration of the "jvm" context for ganglia
# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)
# jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext
jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
jvm.period=10
-jvm.servers={{ganglia_server_host}}:8660
+jvm.servers={{ganglia_server_host}}:8656
# Configuration of the "rpc" context for ganglia
# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)
# rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext
rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
rpc.period=10
-rpc.servers={{ganglia_server_host}}:8660
+rpc.servers={{ganglia_server_host}}:8656
#Ganglia following hadoop example
hbase.sink.ganglia.class=org.apache.hadoop.metrics2.sink.ganglia.GangliaSink31
@@ -77,4 +77,4 @@ hbase.sink.ganglia.period=10
.sink.ganglia.slope=jvm.metrics.gcCount=zero,jvm.metrics.memHeapUsedM=both
.sink.ganglia.dmax=jvm.metrics.threadsBlocked=70,jvm.metrics.memHeapUsedM=40
-hbase.sink.ganglia.servers={{ganglia_server_host}}:8660
+hbase.sink.ganglia.servers={{ganglia_server_host}}:8656
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
index b77e77b..b4be13a 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/NAGIOS/package/templates/hadoop-services.cfg.j2
@@ -204,9 +204,8 @@ define service {
{% endif %}
{% if hostgroup_defs['resourcemanager'] %}
-{% for hostname in hostgroup_defs['resourcemanager'] %}
define service {
- host_name {{ hostname }}
+ hostgroup_name ganglia-server
use hadoop-service
service_description GANGLIA::Ganglia Monitor process for ResourceManager
servicegroups GANGLIA
@@ -215,13 +214,11 @@ define service {
retry_check_interval 0.25
max_check_attempts 4
}
-{% endfor %}
{% endif %}
{% if hostgroup_defs['historyserver2'] %}
-{% for hostname in hostgroup_defs['historyserver2'] %}
define service {
- host_name {{ hostname }}
+ hostgroup_name ganglia-server
use hadoop-service
service_description GANGLIA::Ganglia Monitor process for HistoryServer
servicegroups GANGLIA
@@ -230,7 +227,6 @@ define service {
retry_check_interval 0.25
max_check_attempts 4
}
-{% endfor %}
{% endif %}
{% endif %}
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
index 4e827ab..07b7d6b 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/ganglia/GangliaPropertyProviderTest.java
@@ -125,7 +125,7 @@ public class GangliaPropertyProviderTest {
String expected = (configuration.isGangliaSSL() ? "https" : "http") +
- "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPDataNode%2CHDPSlaves&h=domU-12-31-39-0E-34-E1.compute-1.internal&m=jvm.metrics.gcCount&s=10&e=20&r=1";
+ "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPDataNode%2CHostMetrics&h=domU-12-31-39-0E-34-E1.compute-1.internal&m=jvm.metrics.gcCount&s=10&e=20&r=1";
Assert.assertEquals(expected, streamProvider.getLastSpec());
Assert.assertEquals(3, PropertyHelper.getProperties(resource).size());
@@ -175,7 +175,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPTaskTracker,HDPSlaves");
+ expectedUri.setParameter("c", "HDPTaskTracker,HostMetrics");
expectedUri.setParameter("h", "domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setParameter("m", metricsList);
expectedUri.setParameter("s", "10");
@@ -323,7 +323,7 @@ public class GangliaPropertyProviderTest {
uriBuilder.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
uriBuilder.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
uriBuilder.setPath("/cgi-bin/rrd.py");
- uriBuilder.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HDPSlaves,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
+ uriBuilder.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HostMetrics,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
uriBuilder.setParameter("h", "domU-12-31-39-0E-34-E3.compute-1.internal,domU-12-31-39-0E-34-E1.compute-1.internal,domU-12-31-39-0E-34-E2.compute-1.internal");
uriBuilder.setParameter("m", "jvm.metrics.gcCount");
uriBuilder.setParameter("s", "10");
@@ -382,7 +382,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HDPSlaves,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
+ expectedUri.setParameter("c", "HDPJobTracker,HDPHBaseMaster,HDPResourceManager,HDPFlumeServer,HostMetrics,HDPHistoryServer,HDPJournalNode,HDPTaskTracker,HDPHBaseRegionServer,HDPNameNode");
expectedUri.setParameter("h", hostsList.toString());
expectedUri.setParameter("m", "jvm.metrics.gcCount");
@@ -438,7 +438,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+ expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
expectedUri.setParameter("m", metricsList);
expectedUri.setParameter("s", "10");
@@ -501,7 +501,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+ expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
expectedUri.setParameter("m", metricsList);
expectedUri.setParameter("e", "now");
@@ -546,7 +546,7 @@ public class GangliaPropertyProviderTest {
Assert.assertEquals(1, propertyProvider.populateResources(Collections.singleton(resource), request, null).size());
String expected = (configuration.isGangliaSSL() ? "https" : "http") +
- "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPFlumeServer%2CHDPSlaves&h=ip-10-39-113-33.ec2.internal&m=";
+ "://domU-12-31-39-0E-34-E1.compute-1.internal/cgi-bin/rrd.py?c=HDPFlumeServer%2CHostMetrics&h=ip-10-39-113-33.ec2.internal&m=";
Assert.assertTrue(streamProvider.getLastSpec().startsWith(expected));
@@ -593,7 +593,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+ expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
expectedUri.setParameter("m", metricsList);
expectedUri.setParameter("s", "10");
@@ -651,7 +651,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+ expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
expectedUri.setParameter("m", metricsList);
expectedUri.setParameter("s", "10");
@@ -710,7 +710,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+ expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
expectedUri.setParameter("m", metricsList);
expectedUri.setParameter("s", "10");
@@ -769,7 +769,7 @@ public class GangliaPropertyProviderTest {
expectedUri.setScheme((configuration.isGangliaSSL() ? "https" : "http"));
expectedUri.setHost("domU-12-31-39-0E-34-E1.compute-1.internal");
expectedUri.setPath("/cgi-bin/rrd.py");
- expectedUri.setParameter("c", "HDPFlumeServer,HDPSlaves");
+ expectedUri.setParameter("c", "HDPFlumeServer,HostMetrics");
expectedUri.setParameter("h", "ip-10-39-113-33.ec2.internal");
expectedUri.setParameter("m", metricsList);
expectedUri.setParameter("s", "10");
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py b/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
index d969263..b646871 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
+++ b/ambari-server/src/test/python/stacks/1.3.2/GANGLIA/test_ganglia_monitor.py
@@ -188,35 +188,7 @@ class TestGangliaMonitor(RMFTestCase):
group = 'root',
mode = 0755,
)
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPJobTracker -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHistoryServer -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPDataNode -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPTaskTracker -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseRegionServer -o root -g hadoop',
+ self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -o root -g hadoop',
path = ['/usr/libexec/hdp/ganglia',
'/usr/sbin',
'/sbin:/usr/local/bin',
@@ -242,56 +214,7 @@ class TestGangliaMonitor(RMFTestCase):
def assert_gmond_master_conf_generated(self):
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPNameNode -m -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPJobTracker -m -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseMaster -m -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHistoryServer -m -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPDataNode -m -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPTaskTracker -m -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPHBaseRegionServer -m -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSlaves -m -o root -g hadoop',
+ self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -m -o root -g hadoop',
path = ['/usr/libexec/hdp/ganglia',
'/usr/sbin',
'/sbin:/usr/local/bin',
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py b/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
index b2f7700..e5acbd1 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/GANGLIA/test_ganglia_monitor.py
@@ -176,28 +176,7 @@ class TestGangliaMonitor(RMFTestCase):
group = 'root',
mode = 0755,
)
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPNameNode -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPNimbus -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSupervisor -o root -g hadoop',
- path = ['/usr/libexec/hdp/ganglia',
- '/usr/sbin',
- '/sbin:/usr/local/bin',
- '/bin',
- '/usr/bin'],
- )
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSlaves -o root -g hadoop',
+ self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -o root -g hadoop',
path = ['/usr/libexec/hdp/ganglia',
'/usr/sbin',
'/sbin:/usr/local/bin',
@@ -286,7 +265,7 @@ class TestGangliaMonitor(RMFTestCase):
'/bin',
'/usr/bin'],
)
- self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HDPSlaves -m -o root -g hadoop',
+ self.assertResourceCalled('Execute', '/usr/libexec/hdp/ganglia/setupGanglia.sh -c HostMetrics -m -o root -g hadoop',
path = ['/usr/libexec/hdp/ganglia',
'/usr/sbin',
'/sbin:/usr/local/bin',
http://git-wip-us.apache.org/repos/asf/ambari/blob/ae08e109/ambari-web/app/views/main/dashboard.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/views/main/dashboard.js b/ambari-web/app/views/main/dashboard.js
index ad49d27..de14dcb 100644
--- a/ambari-web/app/views/main/dashboard.js
+++ b/ambari-web/app/views/main/dashboard.js
@@ -526,7 +526,7 @@ App.MainDashboardView = Em.View.extend(App.UserPref, {
},
gangliaUrl: function () {
- return App.router.get('clusterController.gangliaUrl') + "/?r=hour&cs=&ce=&m=&s=by+name&c=HDPSlaves&tab=m&vn=";
+ return App.router.get('clusterController.gangliaUrl') + "/?r=hour&cs=&ce=&m=&s=by+name&c=HostMetrics&tab=m&vn=";
}.property('App.router.clusterController.gangliaUrl'),
showAlertsPopup: function (event) {
[2/2] git commit: AMBARI-5343. Ambari hiveserver2 smoke test wrong
for hiveserver 2 that uses LDAP authentication.(vbrodetskyi)
Posted by vb...@apache.org.
AMBARI-5343. Ambari hiveserver2 smoke test wrong for hiveserver 2 that uses LDAP authentication.(vbrodetskyi)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/f96d22e0
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/f96d22e0
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/f96d22e0
Branch: refs/heads/trunk
Commit: f96d22e0c27d1547f485b5b52998436d44ac8a75
Parents: ae08e10
Author: Vitaly Brodetskyi <vb...@hortonworks.com>
Authored: Fri Apr 4 17:29:19 2014 +0300
Committer: Vitaly Brodetskyi <vb...@hortonworks.com>
Committed: Fri Apr 4 17:29:19 2014 +0300
----------------------------------------------------------------------
.../services/HIVE/package/scripts/params.py | 3 +-
.../HIVE/package/scripts/service_check.py | 37 ++++++++------------
.../services/HIVE/package/scripts/params.py | 3 +-
.../HIVE/package/scripts/service_check.py | 37 ++++++++------------
.../1.3.2/HIVE/test_hive_service_check.py | 34 +++---------------
.../2.0.6/HIVE/test_hive_service_check.py | 36 +++----------------
6 files changed, 41 insertions(+), 109 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
index 968ade0..e2751a9 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/params.py
@@ -47,7 +47,8 @@ check_db_connection_jar = format("/usr/lib/ambari-agent/{check_db_connection_jar
hive_metastore_port = config['configurations']['global']['hive_metastore_port']
hive_var_lib = '/var/lib/hive'
hive_server_host = config['clusterHostInfo']['hive_server_host'][0]
-hive_url = format("jdbc:hive2://{hive_server_host}:10000")
+hive_server_port = 10000
+hive_url = format("jdbc:hive2://{hive_server_host}:{hive_server_port}")
smokeuser = config['configurations']['global']['smokeuser']
smoke_test_sql = "/tmp/hiveserver2.sql"
http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
index 111e8a1..09ba1bf 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/package/scripts/service_check.py
@@ -19,6 +19,8 @@ limitations under the License.
"""
from resource_management import *
+import socket
+import sys
from hcat_service_check import hcat_service_check
@@ -26,29 +28,18 @@ class HiveServiceCheck(Script):
def service_check(self, env):
import params
env.set_params(params)
- if params.security_enabled:
- kinit_cmd = format("{kinit_path_local} -kt {smoke_user_keytab} {smokeuser};")
- hive_principal_ext = format("principal={hive_metastore_keytab_path}")
- hive_url_ext = format("{hive_url}/\\;{hive_principal_ext}")
- smoke_cmd = format("{kinit_cmd} env JAVA_HOME={java64_home} {smoke_test_path} {hive_url_ext} {smoke_test_sql}")
- else:
- smoke_cmd = format("env JAVA_HOME={java64_home} {smoke_test_path} {hive_url} {smoke_test_sql}")
-
- File(params.smoke_test_path,
- content=StaticFile('hiveserver2Smoke.sh'),
- mode=0755
- )
-
- File(params.smoke_test_sql,
- content=StaticFile('hiveserver2.sql')
- )
-
- Execute(smoke_cmd,
- tries=3,
- try_sleep=5,
- path='/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin',
- logoutput=True,
- user=params.smokeuser)
+
+ address=format("{hive_server_host}")
+ port=int(format("{hive_server_port}"))
+ s = socket.socket()
+ print "Test connectivity to hive server"
+ try:
+ s.connect((address, port))
+ print "Successfully connected to %s on port %s" % (address, port)
+ s.close()
+ except socket.error, e:
+ print "Connection to %s on port %s failed: %s" % (address, port, e)
+ sys.exit(1)
hcat_service_check()
http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
index f8f5231..520343b 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/params.py
@@ -50,7 +50,8 @@ hive_metastore_port = config['configurations']['global']['hive_metastore_port']
hive_var_lib = '/var/lib/hive'
hive_bin = '/usr/lib/hive/bin'
hive_server_host = config['clusterHostInfo']['hive_server_host'][0]
-hive_url = format("jdbc:hive2://{hive_server_host}:10000")
+hive_server_port = 10000
+hive_url = format("jdbc:hive2://{hive_server_host}:{hive_server_port}")
smokeuser = config['configurations']['global']['smokeuser']
smoke_test_sql = "/tmp/hiveserver2.sql"
http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
index 111e8a1..09ba1bf 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/package/scripts/service_check.py
@@ -19,6 +19,8 @@ limitations under the License.
"""
from resource_management import *
+import socket
+import sys
from hcat_service_check import hcat_service_check
@@ -26,29 +28,18 @@ class HiveServiceCheck(Script):
def service_check(self, env):
import params
env.set_params(params)
- if params.security_enabled:
- kinit_cmd = format("{kinit_path_local} -kt {smoke_user_keytab} {smokeuser};")
- hive_principal_ext = format("principal={hive_metastore_keytab_path}")
- hive_url_ext = format("{hive_url}/\\;{hive_principal_ext}")
- smoke_cmd = format("{kinit_cmd} env JAVA_HOME={java64_home} {smoke_test_path} {hive_url_ext} {smoke_test_sql}")
- else:
- smoke_cmd = format("env JAVA_HOME={java64_home} {smoke_test_path} {hive_url} {smoke_test_sql}")
-
- File(params.smoke_test_path,
- content=StaticFile('hiveserver2Smoke.sh'),
- mode=0755
- )
-
- File(params.smoke_test_sql,
- content=StaticFile('hiveserver2.sql')
- )
-
- Execute(smoke_cmd,
- tries=3,
- try_sleep=5,
- path='/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin',
- logoutput=True,
- user=params.smokeuser)
+
+ address=format("{hive_server_host}")
+ port=int(format("{hive_server_port}"))
+ s = socket.socket()
+ print "Test connectivity to hive server"
+ try:
+ s.connect((address, port))
+ print "Successfully connected to %s on port %s" % (address, port)
+ s.close()
+ except socket.error, e:
+ print "Connection to %s on port %s failed: %s" % (address, port, e)
+ sys.exit(1)
hcat_service_check()
http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py b/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
index f1deec7..237be7a 100644
--- a/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
+++ b/ambari-server/src/test/python/stacks/1.3.2/HIVE/test_hive_service_check.py
@@ -24,27 +24,14 @@ import resource_management.libraries.functions
@patch.object(resource_management.libraries.functions, "get_unique_id_and_date", new = MagicMock(return_value=''))
class TestServiceCheck(RMFTestCase):
- def test_service_check_default(self):
+ @patch("sys.exit")
+ def test_service_check_default(self, sys_exit_mock):
self.executeScript("1.3.2/services/HIVE/package/scripts/service_check.py",
classname="HiveServiceCheck",
command="service_check",
config_file="default.json"
)
- self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
- content = StaticFile('hiveserver2Smoke.sh'),
- mode = 0755,
- )
- self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
- content = StaticFile('hiveserver2.sql'),
- )
- self.assertResourceCalled('Execute', "env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000 /tmp/hiveserver2.sql",
- logoutput = True,
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
- tries = 3,
- user = 'ambari-qa',
- try_sleep = 5,
- )
self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
content = StaticFile('hcatSmoke.sh'),
mode = 0755,
@@ -73,27 +60,14 @@ class TestServiceCheck(RMFTestCase):
)
self.assertNoMoreResources()
- def test_service_check_secured(self):
+ @patch("sys.exit")
+ def test_service_check_secured(self, sys_exit_mock):
self.executeScript("1.3.2/services/HIVE/package/scripts/service_check.py",
classname="HiveServiceCheck",
command="service_check",
config_file="secured.json"
)
- self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
- content = StaticFile('hiveserver2Smoke.sh'),
- mode = 0755,
- )
- self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
- content = StaticFile('hiveserver2.sql'),
- )
- self.assertResourceCalled('Execute', "/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa; env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000/\\;principal=/etc/security/keytabs/hive.service.keytab /tmp/hiveserver2.sql",
- logoutput = True,
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
- tries = 3,
- user = 'ambari-qa',
- try_sleep = 5,
- )
self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
content = StaticFile('hcatSmoke.sh'),
mode = 0755,
http://git-wip-us.apache.org/repos/asf/ambari/blob/f96d22e0/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
index 6390a70..7eb5810 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_service_check.py
@@ -19,32 +19,19 @@ limitations under the License.
'''
from mock.mock import MagicMock, call, patch
from stacks.utils.RMFTestCase import *
-import datetime
+import datetime, sys
import resource_management.libraries.functions
@patch.object(resource_management.libraries.functions, "get_unique_id_and_date", new = MagicMock(return_value=''))
class TestServiceCheck(RMFTestCase):
- def test_service_check_default(self):
+ @patch("sys.exit")
+ def test_service_check_default(self, sys_exit_mock):
self.executeScript("2.0.6/services/HIVE/package/scripts/service_check.py",
classname="HiveServiceCheck",
command="service_check",
config_file="default.json"
)
- self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
- content = StaticFile('hiveserver2Smoke.sh'),
- mode = 0755,
- )
- self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
- content = StaticFile('hiveserver2.sql'),
- )
- self.assertResourceCalled('Execute', "env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000 /tmp/hiveserver2.sql",
- logoutput = True,
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
- tries = 3,
- user = 'ambari-qa',
- try_sleep = 5,
- )
self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
content = StaticFile('hcatSmoke.sh'),
mode = 0755,
@@ -73,27 +60,14 @@ class TestServiceCheck(RMFTestCase):
)
self.assertNoMoreResources()
- def test_service_check_secured(self):
+ @patch("sys.exit")
+ def test_service_check_secured(self, sys_exit_mock):
self.executeScript("2.0.6/services/HIVE/package/scripts/service_check.py",
classname="HiveServiceCheck",
command="service_check",
config_file="secured.json"
)
- self.assertResourceCalled('File', '/tmp/hiveserver2Smoke.sh',
- content = StaticFile('hiveserver2Smoke.sh'),
- mode = 0755,
- )
- self.assertResourceCalled('File', '/tmp/hiveserver2.sql',
- content = StaticFile('hiveserver2.sql'),
- )
- self.assertResourceCalled('Execute', "/usr/bin/kinit -kt /etc/security/keytabs/smokeuser.headless.keytab ambari-qa; env JAVA_HOME=/usr/jdk64/jdk1.7.0_45 /tmp/hiveserver2Smoke.sh jdbc:hive2://c6402.ambari.apache.org:10000/\\;principal=/etc/security/keytabs/hive.service.keytab /tmp/hiveserver2.sql",
- logoutput = True,
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
- tries = 3,
- user = 'ambari-qa',
- try_sleep = 5,
- )
self.assertResourceCalled('File', '/tmp/hcatSmoke.sh',
content = StaticFile('hcatSmoke.sh'),
mode = 0755,