You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by jl...@apache.org on 2014/12/17 21:05:51 UTC
[19/37] ambari git commit: AMBARI-8745: Common Services: Refactor
HDP2.0.6 FLUME, GANGLIA, HBASE services (Jayush Luniya)
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json
deleted file mode 100644
index 2114c12..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/metrics.json
+++ /dev/null
@@ -1,716 +0,0 @@
-{
- "FLUME_HANDLER": {
- "Component": [
- {
- "type": "ganglia",
- "metrics": {
- "metrics/boottime":{
- "metric":"boottime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_aidle":{
- "metric":"cpu_aidle",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_idle":{
- "metric":"cpu_idle",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_nice":{
- "metric":"cpu_nice",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_num":{
- "metric":"cpu_num",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_speed":{
- "metric":"cpu_speed",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_system":{
- "metric":"cpu_system",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_user":{
- "metric":"cpu_user",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_wio":{
- "metric":"cpu_wio",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/disk/disk_free":{
- "metric":"disk_free",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/disk/disk_total":{
- "metric":"disk_total",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/disk/part_max_used":{
- "metric":"part_max_used",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/gcCount":{
- "metric":"jvm.metrics.gcCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/gcTimeMillis":{
- "metric":"jvm.metrics.gcTimeMillis",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logError":{
- "metric":"jvm.metrics.logError",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logFatal":{
- "metric":"jvm.metrics.logFatal",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logInfo":{
- "metric":"jvm.metrics.logInfo",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logWarn":{
- "metric":"jvm.metrics.logWarn",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/maxMemoryM":{
- "metric":"jvm.metrics.maxMemoryM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memHeapCommittedM":{
- "metric":"jvm.metrics.memHeapCommittedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memHeapUsedM":{
- "metric":"jvm.metrics.memHeapUsedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memNonHeapCommittedM":{
- "metric":"jvm.metrics.memNonHeapCommittedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memNonHeapUsedM":{
- "metric":"jvm.metrics.memNonHeapUsedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsBlocked":{
- "metric":"jvm.metrics.threadsBlocked",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsNew":{
- "metric":"jvm.metrics.threadsNew",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsRunnable":{
- "metric":"jvm.metrics.threadsRunnable",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsTerminated":{
- "metric":"jvm.metrics.threadsTerminated",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsTimedWaiting":{
- "metric":"jvm.metrics.threadsTimedWaiting",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsWaiting":{
- "metric":"jvm.metrics.threadsWaiting",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/ChannelCapacity":{
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelCapacity",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/StartTime":{
- "metric":"(\\w+).CHANNEL.(\\w+).StartTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventTakeAttemptCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeAttemptCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventTakeSuccessCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventPutAttemptCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutAttemptCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/StopTime":{
- "metric":"(\\w+).CHANNEL.(\\w+).StopTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/ChannelFillPercentage":{
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelFillPercentage",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/ChannelSize":{
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventPutSuccessCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/ConnectionCreatedCount":{
- "metric":"(\\w+).SINK.(\\w+).ConnectionCreatedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/BatchCompleteCount":{
- "metric":"(\\w+).SINK.(\\w+).BatchCompleteCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/EventDrainSuccessCount":{
- "metric":"(\\w+).SINK.(\\w+).EventDrainSuccessCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/StartTime":{
- "metric":"(\\w+).SINK.(\\w+).StartTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/EventDrainAttemptCount":{
- "metric":"(\\w+).SINK.(\\w+).EventDrainAttemptCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/ConnectionFailedCount":{
- "metric":"(\\w+).SINK.(\\w+).ConnectionFailedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/BatchUnderflowCount":{
- "metric":"(\\w+).SINK.(\\w+).BatchUnderflowCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/ConnectionClosedCount":{
- "metric":"(\\w+).SINK.(\\w+).ConnectionClosedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/StopTime":{
- "metric":"(\\w+).SINK.(\\w+).StopTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/BatchEmptyCount":{
- "metric":"(\\w+).SINK.(\\w+).BatchEmptyCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendBatchReceivedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendBatchReceivedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendAcceptedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendAcceptedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/StartTime":{
- "metric":"(\\w+).SOURCE.(\\w+).StartTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/OpenConnectionCount":{
- "metric":"(\\w+).SOURCE.(\\w+).OpenConnectionCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendBatchAcceptedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendBatchAcceptedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendReceivedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendReceivedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/EventReceivedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).EventReceivedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/StopTime":{
- "metric":"(\\w+).SOURCE.(\\w+).StopTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/EventAcceptedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).EventAcceptedCount",
- "pointInTime":true,
- "temporal":true
- },
-
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/min": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._min",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/max": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._max",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/avg": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._avg",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/sum": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._sum",
- "pointInTime":false,
- "temporal":true
- },
-
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/avg": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._avg",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/max": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._max",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/min": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._min",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/sum": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._sum",
- "pointInTime":false,
- "temporal":true
- },
-
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/avg": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._avg",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/max": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._max",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/min": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._min",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/sum": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._sum",
- "pointInTime":false,
- "temporal":true
- }
-
- }
- }
- ],
- "HostComponent": [
- {
- "type": "ganglia",
- "metrics": {
- "metrics/boottime":{
- "metric":"boottime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_aidle":{
- "metric":"cpu_aidle",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_idle":{
- "metric":"cpu_idle",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_nice":{
- "metric":"cpu_nice",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_num":{
- "metric":"cpu_num",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_speed":{
- "metric":"cpu_speed",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_system":{
- "metric":"cpu_system",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_user":{
- "metric":"cpu_user",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/cpu/cpu_wio":{
- "metric":"cpu_wio",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/disk/disk_free":{
- "metric":"disk_free",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/disk/disk_total":{
- "metric":"disk_total",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/disk/part_max_used":{
- "metric":"part_max_used",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/gcCount":{
- "metric":"jvm.metrics.gcCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/gcTimeMillis":{
- "metric":"jvm.metrics.gcTimeMillis",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logError":{
- "metric":"jvm.metrics.logError",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logFatal":{
- "metric":"jvm.metrics.logFatal",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logInfo":{
- "metric":"jvm.metrics.logInfo",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/logWarn":{
- "metric":"jvm.metrics.logWarn",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/maxMemoryM":{
- "metric":"jvm.metrics.maxMemoryM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memHeapCommittedM":{
- "metric":"jvm.metrics.memHeapCommittedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memHeapUsedM":{
- "metric":"jvm.metrics.memHeapUsedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memNonHeapCommittedM":{
- "metric":"jvm.metrics.memNonHeapCommittedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/memNonHeapUsedM":{
- "metric":"jvm.metrics.memNonHeapUsedM",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsBlocked":{
- "metric":"jvm.metrics.threadsBlocked",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsNew":{
- "metric":"jvm.metrics.threadsNew",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsRunnable":{
- "metric":"jvm.metrics.threadsRunnable",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsTerminated":{
- "metric":"jvm.metrics.threadsTerminated",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsTimedWaiting":{
- "metric":"jvm.metrics.threadsTimedWaiting",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/jvm/threadsWaiting":{
- "metric":"jvm.metrics.threadsWaiting",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/ChannelCapacity":{
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelCapacity",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/StartTime":{
- "metric":"(\\w+).CHANNEL.(\\w+).StartTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventTakeAttemptCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeAttemptCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventTakeSuccessCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventPutAttemptCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutAttemptCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/StopTime":{
- "metric":"(\\w+).CHANNEL.(\\w+).StopTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/ChannelFillPercentage":{
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelFillPercentage",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/ChannelSize":{
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/CHANNEL/$2/EventPutSuccessCount":{
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/ConnectionCreatedCount":{
- "metric":"(\\w+).SINK.(\\w+).ConnectionCreatedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/BatchCompleteCount":{
- "metric":"(\\w+).SINK.(\\w+).BatchCompleteCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/EventDrainSuccessCount":{
- "metric":"(\\w+).SINK.(\\w+).EventDrainSuccessCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/StartTime":{
- "metric":"(\\w+).SINK.(\\w+).StartTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/EventDrainAttemptCount":{
- "metric":"(\\w+).SINK.(\\w+).EventDrainAttemptCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/ConnectionFailedCount":{
- "metric":"(\\w+).SINK.(\\w+).ConnectionFailedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/BatchUnderflowCount":{
- "metric":"(\\w+).SINK.(\\w+).BatchUnderflowCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/ConnectionClosedCount":{
- "metric":"(\\w+).SINK.(\\w+).ConnectionClosedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/StopTime":{
- "metric":"(\\w+).SINK.(\\w+).StopTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SINK/$2/BatchEmptyCount":{
- "metric":"(\\w+).SINK.(\\w+).BatchEmptyCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendBatchReceivedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendBatchReceivedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendAcceptedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendAcceptedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/StartTime":{
- "metric":"(\\w+).SOURCE.(\\w+).StartTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/OpenConnectionCount":{
- "metric":"(\\w+).SOURCE.(\\w+).OpenConnectionCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendBatchAcceptedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendBatchAcceptedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/AppendReceivedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).AppendReceivedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/EventReceivedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).EventReceivedCount",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/StopTime":{
- "metric":"(\\w+).SOURCE.(\\w+).StopTime",
- "pointInTime":true,
- "temporal":true
- },
- "metrics/flume/$1/SOURCE/$2/EventAcceptedCount":{
- "metric":"(\\w+).SOURCE.(\\w+).EventAcceptedCount",
- "pointInTime":true,
- "temporal":true
- },
-
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/avg": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._avg",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/max": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._max",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/min": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._min",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventTakeSuccessCount/rate/sum": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventTakeSuccessCount._rate._sum",
- "pointInTime":false,
- "temporal":true
- },
-
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/avg": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._avg",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/max": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._max",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/min": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._min",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")EventPutSuccessCount/rate/sum": {
- "metric":"(\\w+).CHANNEL.(\\w+).EventPutSuccessCount._rate._sum",
- "pointInTime":false,
- "temporal":true
- },
-
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/avg": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._avg",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/max": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._max",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/min": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._min",
- "pointInTime":false,
- "temporal":true
- },
- "metrics/flume/$1.substring(0)/CHANNEL/$2.replaceAll(\"[^-]+\",\"\")ChannelSize/sum": {
- "metric":"(\\w+).CHANNEL.(\\w+).ChannelSize._sum",
- "pointInTime":false,
- "temporal":true
- }
-
- }
- }
- ]
- }
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py
deleted file mode 100644
index b183bbc..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/files/alert_flume_agent_status.py
+++ /dev/null
@@ -1,99 +0,0 @@
-#!/usr/bin/env python
-
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-"""
-
-import socket
-
-from resource_management.libraries.functions.flume_agent_helper import find_expected_agent_names
-from resource_management.libraries.functions.flume_agent_helper import get_flume_status
-
-RESULT_CODE_OK = 'OK'
-RESULT_CODE_CRITICAL = 'CRITICAL'
-RESULT_CODE_UNKNOWN = 'UNKNOWN'
-
-FLUME_CONF_DIR_KEY = '{{flume-env/flume_conf_dir}}'
-
-FLUME_RUN_DIR = '/var/run/flume'
-
-def get_tokens():
- """
- Returns a tuple of tokens in the format {{site/property}} that will be used
- to build the dictionary passed into execute
- """
- return (FLUME_CONF_DIR_KEY,)
-
-
-def execute(parameters=None, host_name=None):
- """
- Returns a tuple containing the result code and a pre-formatted result label
-
- Keyword arguments:
- parameters (dictionary): a mapping of parameter key to value
- host_name (string): the name of this host where the alert is running
- """
-
- if parameters is None:
- return (RESULT_CODE_UNKNOWN, ['There were no parameters supplied to the script.'])
-
- flume_conf_directory = None
- if FLUME_CONF_DIR_KEY in parameters:
- flume_conf_directory = parameters[FLUME_CONF_DIR_KEY]
-
- if flume_conf_directory is None:
- return (RESULT_CODE_UNKNOWN, ['The Flume configuration directory is a required parameter.'])
-
- if host_name is None:
- host_name = socket.getfqdn()
-
- processes = get_flume_status(flume_conf_directory, FLUME_RUN_DIR)
- expected_agents = find_expected_agent_names(flume_conf_directory)
-
- alert_label = ''
- alert_state = RESULT_CODE_OK
-
- if len(processes) == 0 and len(expected_agents) == 0:
- alert_label = 'No agents defined on {0}'.format(host_name)
- else:
- ok = []
- critical = []
- text_arr = []
-
- for process in processes:
- if not process.has_key('status') or process['status'] == 'NOT_RUNNING':
- critical.append(process['name'])
- else:
- ok.append(process['name'])
-
- if len(critical) > 0:
- text_arr.append("{0} {1} NOT running".format(", ".join(critical),
- "is" if len(critical) == 1 else "are"))
-
- if len(ok) > 0:
- text_arr.append("{0} {1} running".format(", ".join(ok),
- "is" if len(ok) == 1 else "are"))
-
- plural = len(critical) > 1 or len(ok) > 1
- alert_label = "Agent{0} {1} {2}".format(
- "s" if plural else "",
- " and ".join(text_arr),
- "on " + host_name)
-
- alert_state = RESULT_CODE_CRITICAL if len(critical) > 0 else RESULT_CODE_OK
-
- return (alert_state, [alert_label])
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py
deleted file mode 100644
index f4870ad..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume.py
+++ /dev/null
@@ -1,206 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-import glob
-import json
-import os
-from resource_management import *
-from resource_management.libraries.functions.flume_agent_helper import is_flume_process_live
-from resource_management.libraries.functions.flume_agent_helper import find_expected_agent_names
-
-def flume(action = None):
- import params
-
- if action == 'config':
- # remove previously defined meta's
- for n in find_expected_agent_names(params.flume_conf_dir):
- os.unlink(os.path.join(params.flume_conf_dir, n, 'ambari-meta.json'))
-
- Directory(params.flume_conf_dir, recursive=True)
- Directory(params.flume_log_dir, owner=params.flume_user)
-
- flume_agents = {}
- if params.flume_conf_content is not None:
- flume_agents = build_flume_topology(params.flume_conf_content)
-
- for agent in flume_agents.keys():
- flume_agent_conf_dir = os.path.join(params.flume_conf_dir, agent)
- flume_agent_conf_file = os.path.join(flume_agent_conf_dir, 'flume.conf')
- flume_agent_meta_file = os.path.join(flume_agent_conf_dir, 'ambari-meta.json')
- flume_agent_log4j_file = os.path.join(flume_agent_conf_dir, 'log4j.properties')
- flume_agent_env_file = os.path.join(flume_agent_conf_dir, 'flume-env.sh')
-
- Directory(flume_agent_conf_dir)
-
- PropertiesFile(flume_agent_conf_file,
- properties=flume_agents[agent],
- mode = 0644)
-
- File(flume_agent_log4j_file,
- content=Template('log4j.properties.j2', agent_name = agent),
- mode = 0644)
-
- File(flume_agent_meta_file,
- content = json.dumps(ambari_meta(agent, flume_agents[agent])),
- mode = 0644)
-
- File(flume_agent_env_file,
- owner=params.flume_user,
- content=InlineTemplate(params.flume_env_sh_template)
- )
-
- if params.has_metric_collector:
- File(os.path.join(flume_agent_conf_dir, "flume-metrics2.properties"),
- owner=params.flume_user,
- content=Template("flume-metrics2.properties.j2")
- )
-
- elif action == 'start':
- # desired state for service should be STARTED
- if len(params.flume_command_targets) == 0:
- _set_desired_state('STARTED')
-
- # It is important to run this command as a background process.
-
-
- flume_base = as_user(format("{flume_bin} agent --name {{0}} --conf {{1}} --conf-file {{2}} {{3}}"), params.flume_user, env={'JAVA_HOME': params.java_home}) + " &"
-
- for agent in cmd_target_names():
- flume_agent_conf_dir = params.flume_conf_dir + os.sep + agent
- flume_agent_conf_file = flume_agent_conf_dir + os.sep + "flume.conf"
- flume_agent_pid_file = params.flume_run_dir + os.sep + agent + ".pid"
-
- if not os.path.isfile(flume_agent_conf_file):
- continue
-
- if not is_flume_process_live(flume_agent_pid_file):
- # TODO someday make the ganglia ports configurable
- extra_args = ''
- if params.ganglia_server_host is not None:
- extra_args = '-Dflume.monitoring.type=ganglia -Dflume.monitoring.hosts={0}:{1}'
- extra_args = extra_args.format(params.ganglia_server_host, '8655')
- if params.has_metric_collector:
- extra_args = '-Dflume.monitoring.type=org.apache.hadoop.metrics2.sink.flume.FlumeTimelineMetricsSink'
-
- flume_cmd = flume_base.format(agent, flume_agent_conf_dir,
- flume_agent_conf_file, extra_args)
-
- Execute(flume_cmd,
- wait_for_finish=False,
- environment={'JAVA_HOME': params.java_home}
- )
-
- # sometimes startup spawns a couple of threads - so only the first line may count
- pid_cmd = format('pgrep -o -u {flume_user} -f ^{java_home}.*{agent}.* > {flume_agent_pid_file}')
- Execute(pid_cmd, logoutput=True, tries=20, try_sleep=6)
-
- pass
- elif action == 'stop':
- # desired state for service should be INSTALLED
- if len(params.flume_command_targets) == 0:
- _set_desired_state('INSTALLED')
-
- pid_files = glob.glob(params.flume_run_dir + os.sep + "*.pid")
-
- if 0 == len(pid_files):
- return
-
- agent_names = cmd_target_names()
-
-
- for agent in agent_names:
- pid_file = params.flume_run_dir + os.sep + agent + '.pid'
- pid = format('`cat {pid_file}` > /dev/null 2>&1')
- Execute(format('kill {pid}'), ignore_failures=True)
- File(pid_file, action = 'delete')
-
-
-def ambari_meta(agent_name, agent_conf):
- res = {}
-
- sources = agent_conf[agent_name + '.sources'].split(' ')
- res['sources_count'] = len(sources)
-
- sinks = agent_conf[agent_name + '.sinks'].split(' ')
- res['sinks_count'] = len(sinks)
-
- channels = agent_conf[agent_name + '.channels'].split(' ')
- res['channels_count'] = len(channels)
-
- return res
-
-
-# define a map of dictionaries, where the key is agent name
-# and the dictionary is the name/value pair
-def build_flume_topology(content):
-
- result = {}
- agent_names = []
-
- for line in content.split('\n'):
- rline = line.strip()
- if 0 != len(rline) and not rline.startswith('#'):
- pair = rline.split('=')
- lhs = pair[0].strip()
- rhs = pair[1].strip()
-
- part0 = lhs.split('.')[0]
-
- if lhs.endswith(".sources"):
- agent_names.append(part0)
-
- if not result.has_key(part0):
- result[part0] = {}
-
- result[part0][lhs] = rhs
-
- # trim out non-agents
- for k in result.keys():
- if not k in agent_names:
- del result[k]
-
- return result
-
-
-def cmd_target_names():
- import params
-
- if len(params.flume_command_targets) > 0:
- return params.flume_command_targets
- else:
- return find_expected_agent_names(params.flume_conf_dir)
-
-
-def _set_desired_state(state):
- import params
- filename = os.path.join(params.flume_run_dir, 'ambari-state.txt')
- File(filename,
- content = state,
- )
-
-
-def get_desired_state():
- import params
-
- try:
- with open(os.path.join(params.flume_run_dir, 'ambari-state.txt'), 'r') as fp:
- return fp.read()
- except:
- return 'INSTALLED'
-
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py
deleted file mode 100644
index b93b8e8..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_check.py
+++ /dev/null
@@ -1,40 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-from resource_management import *
-
-
-class FlumeServiceCheck(Script):
-
- def service_check(self, env):
- import params
-
- env.set_params(params)
- if params.security_enabled:
- principal_replaced = params.http_principal.replace("_HOST", params.hostname)
- Execute(format("{kinit_path_local} -kt {http_keytab} {principal_replaced}"),
- user=params.smoke_user)
-
- Execute(format('env JAVA_HOME={java_home} {flume_bin} version'),
- logoutput=True,
- tries = 3,
- try_sleep = 20)
-
-if __name__ == "__main__":
- FlumeServiceCheck().execute()
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py
deleted file mode 100644
index a13f507..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/flume_handler.py
+++ /dev/null
@@ -1,79 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-from flume import flume
-from flume import get_desired_state
-
-from resource_management import *
-from resource_management.libraries.functions.flume_agent_helper import find_expected_agent_names
-from resource_management.libraries.functions.flume_agent_helper import get_flume_status
-
-class FlumeHandler(Script):
- def install(self, env):
- import params
-
- self.install_packages(env)
- env.set_params(params)
-
- def start(self, env, rolling_restart=False):
- import params
-
- env.set_params(params)
- self.configure(env)
-
- flume(action='start')
-
- def stop(self, env, rolling_restart=False):
- import params
-
- env.set_params(params)
-
- flume(action='stop')
-
- def configure(self, env):
- import params
-
- env.set_params(params)
-
- flume(action='config')
-
- def status(self, env):
- import params
- env.set_params(params)
-
-
- processes = get_flume_status(params.flume_conf_dir, params.flume_run_dir)
- expected_agents = find_expected_agent_names(params.flume_conf_dir)
-
- json = {}
- json['processes'] = processes
- self.put_structured_out(json)
-
- # only throw an exception if there are agents defined and there is a
- # problem with the processes; if there are no agents defined, then
- # the service should report STARTED (green) ONLY if the desired state is started. otherwise, INSTALLED (red)
- if len(expected_agents) > 0:
- for proc in processes:
- if not proc.has_key('status') or proc['status'] == 'NOT_RUNNING':
- raise ComponentIsNotRunning()
- elif len(expected_agents) == 0 and 'INSTALLED' == get_desired_state():
- raise ComponentIsNotRunning()
-
-if __name__ == "__main__":
- FlumeHandler().execute()
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py
deleted file mode 100644
index 2997867..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/scripts/params.py
+++ /dev/null
@@ -1,80 +0,0 @@
-"""
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-
-"""
-
-from resource_management.libraries.functions.version import format_hdp_stack_version, compare_versions
-from resource_management import *
-
-config = Script.get_config()
-
-user_group = config['configurations']['cluster-env']['user_group']
-proxyuser_group = config['configurations']['hadoop-env']['proxyuser_group']
-
-security_enabled = False
-
-hdp_stack_version = str(config['hostLevelParams']['stack_version'])
-hdp_stack_version = format_hdp_stack_version(hdp_stack_version)
-
-#hadoop params
-if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.2') >= 0:
- flume_bin = '/usr/hdp/current/flume-server/bin/flume-ng'
- flume_hive_home = '/usr/hdp/current/hive-metastore'
- flume_hcat_home = '/usr/hdp/current/hive-webhcat'
-else:
- flume_bin = '/usr/bin/flume-ng'
- flume_hive_home = '/usr/lib/hive'
- flume_hcat_home = '/usr/lib/hive-hcatalog'
-
-flume_conf_dir = '/etc/flume/conf'
-java_home = config['hostLevelParams']['java_home']
-flume_log_dir = '/var/log/flume'
-flume_run_dir = '/var/run/flume'
-flume_user = 'flume'
-flume_group = 'flume'
-
-if 'flume-env' in config['configurations'] and 'flume_user' in config['configurations']['flume-env']:
- flume_user = config['configurations']['flume-env']['flume_user']
-
-if (('flume-conf' in config['configurations']) and('content' in config['configurations']['flume-conf'])):
- flume_conf_content = config['configurations']['flume-conf']['content']
-else:
- flume_conf_content = None
-
-if (('flume-log4j' in config['configurations']) and ('content' in config['configurations']['flume-log4j'])):
- flume_log4j_content = config['configurations']['flume-log4j']['content']
-else:
- flume_log4j_content = None
-
-targets = default('/commandParams/flume_handler', None)
-flume_command_targets = [] if targets is None else targets.split(',')
-
-flume_env_sh_template = config['configurations']['flume-env']['content']
-
-ganglia_server_hosts = default('/clusterHostInfo/ganglia_server_host', [])
-ganglia_server_host = None
-if 0 != len(ganglia_server_hosts):
- ganglia_server_host = ganglia_server_hosts[0]
-
-hostname = None
-if config.has_key('hostname'):
- hostname = config['hostname']
-
-ams_collector_hosts = default("/clusterHostInfo/metric_collector_hosts", [])
-has_metric_collector = not len(ams_collector_hosts) == 0
-if has_metric_collector:
- metric_collector_host = ams_collector_hosts[0]
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2
deleted file mode 100644
index 7458bf8..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume-metrics2.properties.j2
+++ /dev/null
@@ -1,22 +0,0 @@
-{#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#}
-
-collector={{metric_collector_host}}:8188
-collectionFrequency=60000
-maxRowCacheSize=10000
-sendInterval=59000
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2
deleted file mode 100644
index 4dee67f..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/flume.conf.j2
+++ /dev/null
@@ -1,24 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-
-# flume.conf: Add your flume configuration here and start flume
-# Note if you are using the Windows srvice or Unix service
-# provided by the HDP distribution, they will assume the
-# agent's name in this file to be 'a1'
-#
-{{flume_agent_conf_content}}
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2 b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2
deleted file mode 100644
index 3b34db8..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/FLUME/package/templates/log4j.properties.j2
+++ /dev/null
@@ -1,67 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-#
-
-# Define some default values that can be overridden by system properties.
-#
-# For testing, it may also be convenient to specify
-# -Dflume.root.logger=DEBUG,console when launching flume.
-
-#flume.root.logger=DEBUG,console
-flume.root.logger=INFO,LOGFILE
-flume.log.dir={{flume_log_dir}}
-flume.log.file=flume-{{agent_name}}.log
-
-log4j.logger.org.apache.flume.lifecycle = INFO
-log4j.logger.org.jboss = WARN
-log4j.logger.org.mortbay = INFO
-log4j.logger.org.apache.avro.ipc.NettyTransceiver = WARN
-log4j.logger.org.apache.hadoop = INFO
-
-# Define the root logger to the system property "flume.root.logger".
-log4j.rootLogger=${flume.root.logger}
-
-
-# Stock log4j rolling file appender
-# Default log rotation configuration
-log4j.appender.LOGFILE=org.apache.log4j.RollingFileAppender
-log4j.appender.LOGFILE.MaxFileSize=100MB
-log4j.appender.LOGFILE.MaxBackupIndex=10
-log4j.appender.LOGFILE.File=${flume.log.dir}/${flume.log.file}
-log4j.appender.LOGFILE.layout=org.apache.log4j.PatternLayout
-log4j.appender.LOGFILE.layout.ConversionPattern=%d{dd MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
-
-
-# Warning: If you enable the following appender it will fill up your disk if you don't have a cleanup job!
-# This uses the updated rolling file appender from log4j-extras that supports a reliable time-based rolling policy.
-# See http://logging.apache.org/log4j/companions/extras/apidocs/org/apache/log4j/rolling/TimeBasedRollingPolicy.html
-# Add "DAILY" to flume.root.logger above if you want to use this
-log4j.appender.DAILY=org.apache.log4j.rolling.RollingFileAppender
-log4j.appender.DAILY.rollingPolicy=org.apache.log4j.rolling.TimeBasedRollingPolicy
-log4j.appender.DAILY.rollingPolicy.ActiveFileName=${flume.log.dir}/${flume.log.file}
-log4j.appender.DAILY.rollingPolicy.FileNamePattern=${flume.log.dir}/${flume.log.file}.%d{yyyy-MM-dd}
-log4j.appender.DAILY.layout=org.apache.log4j.PatternLayout
-log4j.appender.DAILY.layout.ConversionPattern=%d{dd MMM yyyy HH:mm:ss,SSS} %-5p [%t] (%C.%M:%L) %x - %m%n
-
-
-# console
-# Add "console" to flume.root.logger above if you want to use this
-log4j.appender.console=org.apache.log4j.ConsoleAppender
-log4j.appender.console.target=System.err
-log4j.appender.console.layout=org.apache.log4j.PatternLayout
-log4j.appender.console.layout.ConversionPattern=%d (%t) [%p - %l] %m%n
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json
deleted file mode 100644
index 7605787..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/alerts.json
+++ /dev/null
@@ -1,137 +0,0 @@
-{
- "GANGLIA": {
- "service": [],
- "GANGLIA_SERVER": [
- {
- "name": "ganglia_server_process",
- "label": "Ganglia Server Process",
- "description": "This host-level alert is triggered if the Ganglia server process cannot be established to be up and listening on the network.",
- "interval": 1,
- "scope": "ANY",
- "enabled": true,
- "source": {
- "type": "PORT",
- "uri": "8651",
- "default_port": 8651,
- "reporting": {
- "ok": {
- "text": "TCP OK - {0:.3f}s response on port {1}"
- },
- "warning": {
- "text": "TCP OK - {0:.3f}s response on port {1}",
- "value": 1.5
- },
- "critical": {
- "text": "Connection failed: {0} to {1}:{2}",
- "value": 5.0
- }
- }
- }
- },
- {
- "name": "ganglia_monitor_hdfs_namenode",
- "label": "Ganglia NameNode Process Monitor",
- "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for HDFS NameNode is not up and listening.",
- "interval": 1,
- "scope": "ANY",
- "enabled": true,
- "source": {
- "type": "PORT",
- "uri": "8661",
- "default_port": 8661,
- "reporting": {
- "ok": {
- "text": "TCP OK - {0:.3f}s response on port {1}"
- },
- "warning": {
- "text": "TCP OK - {0:.3f}s response on port {1}",
- "value": 1.5
- },
- "critical": {
- "text": "Connection failed: {0} to {1}:{2}",
- "value": 5.0
- }
- }
- }
- },
- {
- "name": "ganglia_monitor_hbase_master",
- "label": "Ganglia HBase Master Process Monitor",
- "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for the HBase Master process is not up and listening.",
- "interval": 1,
- "scope": "ANY",
- "enabled": true,
- "source": {
- "type": "PORT",
- "uri": "8663",
- "default_port": 8663,
- "reporting": {
- "ok": {
- "text": "TCP OK - {0:.3f}s response on port {1}"
- },
- "warning": {
- "text": "TCP OK - {0:.3f}s response on port {1}",
- "value": 1.5
- },
- "critical": {
- "text": "Connection failed: {0} to {1}:{2}",
- "value": 5.0
- }
- }
- }
- },
- {
- "name": "ganglia_monitor_yarn_resourcemanager",
- "label": "Ganglia ResourceManager Process Monitor",
- "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for the YARN ResourceManager process is not up and listening.",
- "interval": 1,
- "scope": "ANY",
- "enabled": true,
- "source": {
- "type": "PORT",
- "uri": "8664",
- "default_port": 8664,
- "reporting": {
- "ok": {
- "text": "TCP OK - {0:.3f}s response on port {1}"
- },
- "warning": {
- "text": "TCP OK - {0:.3f}s response on port {1}",
- "value": 1.5
- },
- "critical": {
- "text": "Connection failed: {0} to {1}:{2}",
- "value": 5.0
- }
- }
- }
- },
- {
- "name": "ganglia_monitor_mapreduce_history_server",
- "label": "Ganglia History Server Process Monitor",
- "description": "This host-level alert is triggered if the Ganglia gmond process which handles receiving metrics for the MapReduce History Server process is not up and listening.",
- "interval": 1,
- "scope": "ANY",
- "enabled": true,
- "source": {
- "type": "PORT",
- "uri": "8666",
- "default_port": 8666,
- "reporting": {
- "ok": {
- "text": "TCP OK - {0:.3f}s response on port {1}"
- },
- "warning": {
- "text": "TCP OK - {0:.3f}s response on port {1}",
- "value": 1.5
- },
- "critical": {
- "text": "Connection failed: {0} to {1}:{2}",
- "value": 5.0
- }
- }
- }
- }
- ]
- }
-}
\ No newline at end of file
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml
deleted file mode 100644
index 3328acf..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/configuration/ganglia-env.xml
+++ /dev/null
@@ -1,77 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
--->
-
-<configuration>
- <property>
- <name>ganglia_conf_dir</name>
- <value>/etc/ganglia/hdp</value>
- <description>Config directory for Ganglia</description>
- </property>
- <property>
- <name>ganglia_runtime_dir</name>
- <value>/var/run/ganglia/hdp</value>
- <description>Run directories for Ganglia</description>
- </property>
- <property>
- <name>gmetad_user</name>
- <value>nobody</value>
- <property-type>USER GROUP</property-type>
- <description>User </description>
- </property>
- <property>
- <name>gmond_user</name>
- <value>nobody</value>
- <property-type>USER GROUP</property-type>
- <description>User </description>
- </property>
- <property>
- <name>rrdcached_base_dir</name>
- <value>/var/lib/ganglia/rrds</value>
- <description>Default directory for saving the rrd files on ganglia server</description>
- </property>
- <property>
- <name>rrdcached_timeout</name>
- <value>3600</value>
- <description>(-w) Data is written to disk every timeout seconds. If this option is not specified the default interval of 300 seconds will be used.</description>
- </property>
- <property>
- <name>rrdcached_flush_timeout</name>
- <value>7200</value>
- <description>(-f) Every timeout seconds the entire cache is searched for old values which are written to disk. This only concerns files to which updates have stopped, so setting this to a high value, such as 3600 seconds, is acceptable in most cases. This timeout defaults to 3600 seconds.</description>
- </property>
- <property>
- <name>rrdcached_delay</name>
- <value>1800</value>
- <description>(-z) If specified, rrdcached will delay writing of each RRD for a random number of seconds in the range [0,delay). This will avoid too many writes being queued simultaneously. This value should be no greater than the value specified in -w. By default, there is no delay.</description>
- </property>
- <property>
- <name>rrdcached_write_threads</name>
- <value>4</value>
- <description>(-t) Specifies the number of threads used for writing RRD files. The default is 4. Increasing this number will allow rrdcached to have more simultaneous I/O requests into the kernel. This may allow the kernel to re-order disk writes, resulting in better disk throughput.</description>
- </property>
- <property>
- <name>additional_clusters</name>
- <value> </value>
- <description>Add additional desired Ganglia metrics cluster in the form "name1:port1,name2:port2". Ensure that the names and ports are unique across all cluster and ports are available on ganglia server host. Ambari has reserved ports 8667-8669 within its own pool.</description>
- </property>
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
index 4e96ade..33bc837 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/metainfo.xml
@@ -20,108 +20,7 @@
<services>
<service>
<name>GANGLIA</name>
- <displayName>Ganglia</displayName>
- <comment>Ganglia Metrics Collection system (<a href="http://oss.oetiker.ch/rrdtool/" target="_blank">RRDTool</a> will be installed too)</comment>
- <version>3.5.0</version>
- <components>
- <component>
- <name>GANGLIA_SERVER</name>
- <displayName>Ganglia Server</displayName>
- <category>MASTER</category>
- <cardinality>1</cardinality>
- <commandScript>
- <script>scripts/ganglia_server.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>600</timeout>
- </commandScript>
- </component>
-
- <component>
- <name>GANGLIA_MONITOR</name>
- <displayName>Ganglia Monitor</displayName>
- <category>SLAVE</category>
- <cardinality>ALL</cardinality>
- <auto-deploy>
- <enabled>true</enabled>
- </auto-deploy>
- <commandScript>
- <script>scripts/ganglia_monitor.py</script>
- <scriptType>PYTHON</scriptType>
- <timeout>600</timeout>
- </commandScript>
- </component>
- </components>
- <osSpecifics>
- <osSpecific>
- <osFamily>redhat5,redhat6,suse11</osFamily>
- <packages>
- <package>
- <name>python-rrdtool-1.4.5</name>
- </package>
- <package>
- <name>libganglia-3.5.0-99</name>
- </package>
- <package>
- <name>ganglia-devel-3.5.0-99</name>
- </package>
- <package>
- <name>ganglia-gmetad-3.5.0-99</name>
- </package>
- <package>
- <name>ganglia-web-3.5.7-99.noarch</name>
- </package>
- <package>
- <name>ganglia-gmond-3.5.0-99</name>
- </package>
- <package>
- <name>ganglia-gmond-modules-python-3.5.0-99</name>
- </package>
- </packages>
- </osSpecific>
- <osSpecific>
- <osFamily>ubuntu12</osFamily>
- <packages>
- <package>
- <name>python-rrdtool</name>
- </package>
- <package>
- <name>gmetad</name>
- </package>
- <package>
- <name>ganglia-webfrontend</name>
- </package>
- <package>
- <name>ganglia-monitor-python</name>
- </package>
- <package>
- <name>rrdcached</name>
- </package>
- </packages>
- </osSpecific>
- <osSpecific>
- <osFamily>suse11</osFamily>
- <packages>
- <package>
- <name>apache2</name>
- </package>
- <package>
- <name>apache2?mod_php*</name>
- </package>
- </packages>
- </osSpecific>
- <osSpecific>
- <osFamily>redhat5,redhat6</osFamily>
- <packages>
- <package>
- <name>httpd</name>
- </package>
- </packages>
- </osSpecific>
- </osSpecifics>
- <configuration-dependencies>
- <config-type>ganglia-env</config-type>
- </configuration-dependencies>
- <monitoringService>true</monitoringService>
+ <extends>common-services/GANGLIA/3.5.0</extends>
</service>
</services>
</metainfo>
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh
deleted file mode 100644
index f45e371..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmetad.sh
+++ /dev/null
@@ -1,37 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements. See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership. The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License. You may obtain a copy of the License at
-# *
-# * http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Get all our common constants etc. set up.
-source ./gmetadLib.sh;
-
-# Before checking gmetad, check rrdcached.
-./checkRrdcached.sh;
-
-gmetadRunningPid=`getGmetadRunningPid`;
-
-if [ -n "${gmetadRunningPid}" ]
-then
- echo "${GMETAD_BIN} running with PID ${gmetadRunningPid}";
-else
- echo "Failed to find running ${GMETAD_BIN}";
- exit 1;
-fi
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh
deleted file mode 100644
index fbf524a..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkGmond.sh
+++ /dev/null
@@ -1,62 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements. See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership. The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License. You may obtain a copy of the License at
-# *
-# * http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Get all our common constants etc. set up.
-# Pulls in gangliaLib.sh as well, so we can skip pulling it in again.
-source ./gmondLib.sh;
-
-function checkGmondForCluster()
-{
- gmondClusterName=${1};
-
- gmondCoreConfFileName=`getGmondCoreConfFileName ${gmondClusterName}`;
-
- # Skip over (purported) Clusters that don't have their core conf file present.
- if [ -e "${gmondCoreConfFileName}" ]
- then
- gmondRunningPid=`getGmondRunningPid ${gmondClusterName}`;
-
- if [ -n "${gmondRunningPid}" ]
- then
- echo "${GMOND_BIN} for cluster ${gmondClusterName} running with PID ${gmondRunningPid}";
- else
- echo "Failed to find running ${GMOND_BIN} for cluster ${gmondClusterName}";
- exit 1;
- fi
- fi
-}
-
-# main()
-gmondClusterName=${1};
-
-if [ "x" == "x${gmondClusterName}" ]
-then
- # No ${gmondClusterName} passed in as command-line arg, so check
- # all the gmonds we know about.
- for gmondClusterName in `getConfiguredGangliaClusterNames`
- do
- checkGmondForCluster ${gmondClusterName};
- done
-else
- # Just check the one ${gmondClusterName} that was asked for.
- checkGmondForCluster ${gmondClusterName};
-fi
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh
deleted file mode 100644
index 1e0c2e2..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/checkRrdcached.sh
+++ /dev/null
@@ -1,34 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements. See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership. The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License. You may obtain a copy of the License at
-# *
-# * http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Get all our common constants etc. set up.
-source ./rrdcachedLib.sh;
-
-rrdcachedRunningPid=`getRrdcachedRunningPid`;
-
-if [ -n "${rrdcachedRunningPid}" ]
-then
- echo "${RRDCACHED_BIN} running with PID ${rrdcachedRunningPid}";
-else
- echo "Failed to find running ${RRDCACHED_BIN}";
- exit 1;
-fi
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init
deleted file mode 100644
index 20b388e..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetad.init
+++ /dev/null
@@ -1,73 +0,0 @@
-#!/bin/sh
-# chkconfig: 2345 70 40
-# description: hdp-gmetad startup script
-# processname: hdp-gmetad
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements. See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership. The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License. You may obtain a copy of the License at
-# *
-# * http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-# Remember to keep this in-sync with the definition of
-# GANGLIA_RUNTIME_COMPONENTS_UNPACK_DIR in monrpmInstaller.sh.
-HDP_GANGLIA_RUNTIME_COMPONENTS_DIR=/usr/libexec/hdp/ganglia
-HDP_GANLIA_GMETAD_STARTER=${HDP_GANGLIA_RUNTIME_COMPONENTS_DIR}/startGmetad.sh
-HDP_GANLIA_GMETAD_STOPPER=${HDP_GANGLIA_RUNTIME_COMPONENTS_DIR}/stopGmetad.sh
-HDP_GANLIA_GMETAD_CHECKER=${HDP_GANGLIA_RUNTIME_COMPONENTS_DIR}/checkGmetad.sh
-
-RETVAL=0
-
-case "$1" in
- start)
- echo "============================="
- echo "Starting hdp-gmetad..."
- echo "============================="
- [ -f ${HDP_GANLIA_GMETAD_STARTER} ] || exit 1
- eval "${HDP_GANLIA_GMETAD_STARTER}"
- RETVAL=$?
- echo
- [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hdp-gmetad
- ;;
-
- stop)
- echo "=================================="
- echo "Shutting down hdp-gmetad..."
- echo "=================================="
- [ -f ${HDP_GANLIA_GMETAD_STOPPER} ] || exit 1
- eval "${HDP_GANLIA_GMETAD_STOPPER}"
- RETVAL=$?
- echo
- [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hdp-gmetad
- ;;
-
- restart|reload)
- $0 stop
- $0 start
- RETVAL=$?
- ;;
- status)
- echo "======================================="
- echo "Checking status of hdp-gmetad..."
- echo "======================================="
- [ -f ${HDP_GANLIA_GMETAD_CHECKER} ] || exit 1
- eval "${HDP_GANLIA_GMETAD_CHECKER}"
- RETVAL=$?
- ;;
- *)
- echo "Usage: $0 {start|stop|restart|status}"
- exit 1
-esac
-
-exit $RETVAL
http://git-wip-us.apache.org/repos/asf/ambari/blob/b4cd4cb7/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh
deleted file mode 100644
index 6a24bed..0000000
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/GANGLIA/package/files/gmetadLib.sh
+++ /dev/null
@@ -1,204 +0,0 @@
-#!/usr/bin/env bash
-
-#/*
-# * Licensed to the Apache Software Foundation (ASF) under one
-# * or more contributor license agreements. See the NOTICE file
-# * distributed with this work for additional information
-# * regarding copyright ownership. The ASF licenses this file
-# * to you under the Apache License, Version 2.0 (the
-# * "License"); you may not use this file except in compliance
-# * with the License. You may obtain a copy of the License at
-# *
-# * http://www.apache.org/licenses/LICENSE-2.0
-# *
-# * Unless required by applicable law or agreed to in writing, software
-# * distributed under the License is distributed on an "AS IS" BASIS,
-# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# * See the License for the specific language governing permissions and
-# * limitations under the License.
-# */
-
-cd `dirname ${0}`;
-
-# Slurp in all our user-customizable settings.
-source ./gangliaEnv.sh;
-
-# Get access to Ganglia-wide constants etc.
-source ./gangliaLib.sh;
-
-GMETAD_BIN=/usr/sbin/gmetad;
-GMETAD_CONF_FILE=${GANGLIA_CONF_DIR}/gmetad.conf;
-GMETAD_PID_FILE=${GANGLIA_RUNTIME_DIR}/gmetad.pid;
-
-function getGmetadLoggedPid()
-{
- if [ -e "${GMETAD_PID_FILE}" ]
- then
- echo `cat ${GMETAD_PID_FILE}`;
- fi
-}
-
-function getGmetadRunningPid()
-{
- gmetadLoggedPid=`getGmetadLoggedPid`;
-
- if [ -n "${gmetadLoggedPid}" ]
- then
- echo `ps -o pid=MYPID -p ${gmetadLoggedPid} | tail -1 | awk '{print $1}' | grep -v MYPID`;
- fi
-}
-
-function generateGmetadConf()
-{
- now=`date`;
-
- cat <<END_OF_GMETAD_CONF_1
-#################### Generated by ${0} on ${now} ####################
-#
-#-------------------------------------------------------------------------------
-# Setting the debug_level to 1 will keep daemon in the forground and
-# show only error messages. Setting this value higher than 1 will make
-# gmetad output debugging information and stay in the foreground.
-# default: 0
-# debug_level 10
-#
-#-------------------------------------------------------------------------------
-# What to monitor. The most important section of this file.
-#
-# The data_source tag specifies either a cluster or a grid to
-# monitor. If we detect the source is a cluster, we will maintain a complete
-# set of RRD databases for it, which can be used to create historical
-# graphs of the metrics. If the source is a grid (it comes from another gmetad),
-# we will only maintain summary RRDs for it.
-#
-# Format:
-# data_source "my cluster" [polling interval] address1:port addreses2:port ...
-#
-# The keyword 'data_source' must immediately be followed by a unique
-# string which identifies the source, then an optional polling interval in
-# seconds. The source will be polled at this interval on average.
-# If the polling interval is omitted, 15sec is asssumed.
-#
-# If you choose to set the polling interval to something other than the default,
-# note that the web frontend determines a host as down if its TN value is less
-# than 4 * TMAX (20sec by default). Therefore, if you set the polling interval
-# to something around or greater than 80sec, this will cause the frontend to
-# incorrectly display hosts as down even though they are not.
-#
-# A list of machines which service the data source follows, in the
-# format ip:port, or name:port. If a port is not specified then 8649
-# (the default gmond port) is assumed.
-# default: There is no default value
-#
-# data_source "my cluster" 10 localhost my.machine.edu:8649 1.2.3.5:8655
-# data_source "my grid" 50 1.3.4.7:8655 grid.org:8651 grid-backup.org:8651
-# data_source "another source" 1.3.4.7:8655 1.3.4.8
-END_OF_GMETAD_CONF_1
-
- # Get info about all the configured Ganglia clusters.
- getGangliaClusterInfo | while read gangliaClusterInfoLine
- do
- # From each, parse out ${gmondClusterName}, ${gmondMasterIP} and ${gmondPort}...
- read gmondClusterName gmondMasterIP gmondPort <<<`echo ${gangliaClusterInfoLine}`;
- # ...and generate a corresponding data_source line for gmetad.conf.
- echo "data_source \"${gmondClusterName}\" ${gmondMasterIP}:${gmondPort}";
- done
-
- cat <<END_OF_GMETAD_CONF_2
-#
-# Round-Robin Archives
-# You can specify custom Round-Robin archives here (defaults are listed below)
-#
-# Old Default RRA: Keep 1 hour of metrics at 15 second resolution. 1 day at 6 minute
-# RRAs "RRA:AVERAGE:0.5:1:244" "RRA:AVERAGE:0.5:24:244" "RRA:AVERAGE:0.5:168:244" "RRA:AVERAGE:0.5:672:244" \
-# "RRA:AVERAGE:0.5:5760:374"
-# New Default RRA
-# Keep 5856 data points at 15 second resolution assuming 15 second (default) polling. That's 1 day
-# Two weeks of data points at 1 minute resolution (average)
-#RRAs "RRA:AVERAGE:0.5:1:5856" "RRA:AVERAGE:0.5:4:20160" "RRA:AVERAGE:0.5:40:52704"
-# Retaining existing resolution
-RRAs "RRA:AVERAGE:0.5:1:244" "RRA:AVERAGE:0.5:24:244" "RRA:AVERAGE:0.5:168:244" "RRA:AVERAGE:0.5:672:244" \
- "RRA:AVERAGE:0.5:5760:374"
-#
-#-------------------------------------------------------------------------------
-# Scalability mode. If on, we summarize over downstream grids, and respect
-# authority tags. If off, we take on 2.5.0-era behavior: we do not wrap our output
-# in <GRID></GRID> tags, we ignore all <GRID> tags we see, and always assume
-# we are the "authority" on data source feeds. This approach does not scale to
-# large groups of clusters, but is provided for backwards compatibility.
-# default: on
-# scalable off
-#
-#-------------------------------------------------------------------------------
-# The name of this Grid. All the data sources above will be wrapped in a GRID
-# tag with this name.
-# default: unspecified
-gridname "HDP_GRID"
-#
-#-------------------------------------------------------------------------------
-# The authority URL for this grid. Used by other gmetads to locate graphs
-# for our data sources. Generally points to a ganglia/
-# website on this machine.
-# default: "http://hostname/ganglia/",
-# where hostname is the name of this machine, as defined by gethostname().
-# authority "http://mycluster.org/newprefix/"
-#
-#-------------------------------------------------------------------------------
-# List of machines this gmetad will share XML with. Localhost
-# is always trusted.
-# default: There is no default value
-# trusted_hosts 127.0.0.1 169.229.50.165 my.gmetad.org
-#
-#-------------------------------------------------------------------------------
-# If you want any host which connects to the gmetad XML to receive
-# data, then set this value to "on"
-# default: off
-# all_trusted on
-#
-#-------------------------------------------------------------------------------
-# If you don't want gmetad to setuid then set this to off
-# default: on
-# setuid off
-#
-#-------------------------------------------------------------------------------
-# User gmetad will setuid to (defaults to "nobody")
-# default: "nobody"
-setuid_username "${GMETAD_USER}"
-#
-#-------------------------------------------------------------------------------
-# Umask to apply to created rrd files and grid directory structure
-# default: 0 (files are public)
-# umask 022
-#
-#-------------------------------------------------------------------------------
-# The port gmetad will answer requests for XML
-# default: 8651
-# xml_port 8651
-#
-#-------------------------------------------------------------------------------
-# The port gmetad will answer queries for XML. This facility allows
-# simple subtree and summation views of the XML tree.
-# default: 8652
-# interactive_port 8652
-#
-#-------------------------------------------------------------------------------
-# The number of threads answering XML requests
-# default: 4
-# server_threads 10
-#
-#-------------------------------------------------------------------------------
-# Where gmetad stores its round-robin databases
-# default: "/var/lib/ganglia/rrds"
-rrd_rootdir "${RRD_ROOTDIR}"
-#
-#-------------------------------------------------------------------------------
-# In earlier versions of gmetad, hostnames were handled in a case
-# sensitive manner
-# If your hostname directories have been renamed to lower case,
-# set this option to 0 to disable backward compatibility.
-# From version 3.2, backwards compatibility will be disabled by default.
-# default: 1 (for gmetad < 3.2)
-# default: 0 (for gmetad >= 3.2)
-case_sensitive_hostnames 1
-END_OF_GMETAD_CONF_2
-}