You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by at...@apache.org on 2014/01/16 14:56:59 UTC

[01/11] git commit: AMBARI-4312 Storm: Install wizard. Add master components Storm UI Server, DRPC Server, LogViewer Server. (Denys Buzhor via atkach)

Updated Branches:
  refs/heads/trunk 5f48900bd -> 1c9b15735


AMBARI-4312 Storm: Install wizard. Add master components Storm UI Server, DRPC Server, LogViewer Server. (Denys Buzhor via atkach)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/1c9b1573
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/1c9b1573
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/1c9b1573

Branch: refs/heads/trunk
Commit: 1c9b15735cb85a5b9044de45575e8d7c3cf23836
Parents: 8c0e377
Author: atkach <at...@hortonworks.com>
Authored: Thu Jan 16 15:56:32 2014 +0200
Committer: atkach <at...@hortonworks.com>
Committed: Thu Jan 16 15:56:53 2014 +0200

----------------------------------------------------------------------
 .../configurations/cluster_level_configs.json   |  5 ++-
 .../app/controllers/wizard/step5_controller.js  |  3 ++
 ambari-web/app/data/HDP2/global_properties.js   | 39 ++++++++++++++++++++
 ambari-web/app/data/service_components.js       | 26 ++++++++++++-
 ambari-web/app/data/service_configs.js          |  3 ++
 ambari-web/app/mappers/server_data_mapper.js    |  5 ++-
 ambari-web/app/models/host_component.js         |  3 ++
 ambari-web/app/models/quick_links.js            | 10 ++---
 ambari-web/app/models/service_config.js         | 12 ++++++
 9 files changed, 97 insertions(+), 9 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/assets/data/configurations/cluster_level_configs.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/configurations/cluster_level_configs.json b/ambari-web/app/assets/data/configurations/cluster_level_configs.json
index 8e958df..9fc68d6 100644
--- a/ambari-web/app/assets/data/configurations/cluster_level_configs.json
+++ b/ambari-web/app/assets/data/configurations/cluster_level_configs.json
@@ -423,6 +423,7 @@
         "dfs_name_dir" : "/hadoop/hdfs/namenode",
         "hfile_blockcache_size" : "0.25",
         "kadmin_pw" : "",
+        "logviewerserver_host" : "c6401.ambari.apache.org",
         "gmond_user" : "nobody",
         "hregion_majorcompaction" : "86400000",
         "user_group" : "hadoop",
@@ -546,7 +547,9 @@
         "oozie_database_name": "XE",
         "oozie_metastore_user_name" : "oozieuser",
         "oozie_metastore_user_passwd" : "ooziepassword",
-        "oozie_jdbc_connection_url": "jdbc:oracle:thin:@//oozie-orcl-host:1521/XE"
+        "oozie_jdbc_connection_url": "jdbc:oracle:thin:@//oozie-orcl-host:1521/XE",
+        "stormuiserver_host" : "c6401.ambari.apache.org",
+        "drpcserver_host" : "c6401.ambari.apache.org"
       }
     },
     {

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/controllers/wizard/step5_controller.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/wizard/step5_controller.js b/ambari-web/app/controllers/wizard/step5_controller.js
index 2e3164c..a93d325 100644
--- a/ambari-web/app/controllers/wizard/step5_controller.js
+++ b/ambari-web/app/controllers/wizard/step5_controller.js
@@ -450,6 +450,9 @@ App.WizardStep5Controller = Em.Controller.extend({
           "31" : 1,
           "else" : 2
         }).host_name;
+      case 'STORM_UI_SERVER':
+      case 'LOGVIEWER_SERVER':
+      case 'DRPC_SERVER':
       case 'NIMBUS':
         return this.getNimbusServer(noOfHosts);
       default:

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/data/HDP2/global_properties.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/global_properties.js b/ambari-web/app/data/HDP2/global_properties.js
index 8276a1a..04632ac 100644
--- a/ambari-web/app/data/HDP2/global_properties.js
+++ b/ambari-web/app/data/HDP2/global_properties.js
@@ -1323,6 +1323,45 @@ module.exports =
       "serviceName": "STORM",
       "category": "General"
     },
+    {
+      "id": "puppet var",
+      "name": "stormuiserver_host",
+      "displayName": "Storm UI Server host",
+      "description": "",
+      "defaultValue": "",
+      "displayType": "masterHost",
+      "isReconfigurable": false,
+      "isVisible": true,
+      "isRequiredByAgent": true,
+      "serviceName": "STORM",
+      "category": "StormUIServer"
+    },
+    {
+      "id": "puppet var",
+      "name": "logviewerserver_host",
+      "displayName": "Logviewer Server host",
+      "description": "",
+      "defaultValue": "",
+      "displayType": "masterHost",
+      "isReconfigurable": false,
+      "isVisible": true,
+      "isRequiredByAgent": true,
+      "serviceName": "STORM",
+      "category": "LogviewerServer"
+    },
+    {
+      "id": "puppet var",
+      "name": "drpcserver_host",
+      "displayName": "DRPC Server host",
+      "description": "",
+      "defaultValue": "",
+      "displayType": "masterHost",
+      "isReconfigurable": false,
+      "isVisible": true,
+      "isRequiredByAgent": true,
+      "serviceName": "STORM",
+      "category": "DRPCServer"
+    },
   /**********************************************MISC***************************************/
     {
       "id": "puppet var",

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/data/service_components.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/service_components.js b/ambari-web/app/data/service_components.js
index d897805..b4c7e0e 100644
--- a/ambari-web/app/data/service_components.js
+++ b/ambari-web/app/data/service_components.js
@@ -338,7 +338,6 @@ module.exports = new Ember.Set([
     isClient: false,
     description: 'Falcon Server for mirroring'
   },
-  // @todo change with real info about service component
   {
     service_name: 'STORM',
     component_name: 'NIMBUS',
@@ -347,7 +346,6 @@ module.exports = new Ember.Set([
     isClient: false,
     description: 'Master component for STORM'
   },
-  // @todo change with real info about service component
   {
     service_name: 'STORM',
     component_name: 'SUPERVISOR',
@@ -355,5 +353,29 @@ module.exports = new Ember.Set([
     isMaster: false,
     isClient: true,
     description: 'Client component for STORM'
+  },
+  {
+    service_name: 'STORM',
+    component_name: 'STORM_UI_SERVER',
+    display_name: 'Storm UI Server',
+    isMaster: true,
+    isClient: false,
+    description: 'Master component for STORM'
+  },
+  {
+    service_name: 'STORM',
+    component_name: 'LOGVIEWER_SERVER',
+    display_name: 'Logviewer Server',
+    isMaster: true,
+    isClient: false,
+    description: 'Master component for STORM'
+  },
+  {
+    service_name: 'STORM',
+    component_name: 'DRPC_SERVER',
+    display_name: 'DRPC Server',
+    isMaster: true,
+    isClient: false,
+    description: 'Master component for STORM'
   }
 ]);

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/data/service_configs.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/service_configs.js b/ambari-web/app/data/service_configs.js
index d20a4ec..af6c4f4 100644
--- a/ambari-web/app/data/service_configs.js
+++ b/ambari-web/app/data/service_configs.js
@@ -233,6 +233,9 @@ module.exports = [
     configCategories: [
       App.ServiceConfigCategory.create({ name: 'Nimbus', displayName : 'Nimbus'}),
       App.ServiceConfigCategory.create({ name: 'Supervisor', displayName : 'SuperVisor'}),
+      App.ServiceConfigCategory.create({ name: 'StormUIServer', displayName : 'Storm UI Server'}),
+      App.ServiceConfigCategory.create({ name: 'DRPCServer', displayName : 'DRPC Server'}),
+      App.ServiceConfigCategory.create({ name: 'LogviewerServer', displayName : 'Logviewer Server'}),
       App.ServiceConfigCategory.create({ name: 'General', displayName : 'General'}),
       App.ServiceConfigCategory.create({ name: 'Advanced', displayName : 'Advanced'})
     ],

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/mappers/server_data_mapper.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/mappers/server_data_mapper.js b/ambari-web/app/mappers/server_data_mapper.js
index 5f16a86..0ce6cf4 100644
--- a/ambari-web/app/mappers/server_data_mapper.js
+++ b/ambari-web/app/mappers/server_data_mapper.js
@@ -216,6 +216,9 @@ App.QuickDataMapper.componentServiceMap = function () {
     'GLUSTERFS_CLIENT': 'GLUSTERFS',
     'FALCON_SERVER': 'FALCON',
     'NIMBUS': 'STORM',
-    'SUPERVISOR': 'STORM'
+    'SUPERVISOR': 'STORM',
+    'STORM_UI_SERVER': 'STORM',
+    'DRPC_SERVER': 'STORM',
+    'LOGVIEWER_SERVER': 'STORM'
   }
 };

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/models/host_component.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/models/host_component.js b/ambari-web/app/models/host_component.js
index a353b0a..17887ac 100644
--- a/ambari-web/app/models/host_component.js
+++ b/ambari-web/app/models/host_component.js
@@ -79,6 +79,9 @@ App.HostComponent = DS.Model.extend({
       case 'FLUME_SERVER':
       case 'FALCON_SERVER':
       case 'NIMBUS':
+      case 'STORM_UI_SERVER':
+      case 'LOGVIEWER_SERVER':
+      case 'DRPC_SERVER':
       case 'RESOURCEMANAGER':
         return true;
       default:

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/models/quick_links.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/models/quick_links.js b/ambari-web/app/models/quick_links.js
index 5b98c04..97e728a 100644
--- a/ambari-web/app/models/quick_links.js
+++ b/ambari-web/app/models/quick_links.js
@@ -393,14 +393,14 @@ App.QuickLinks.FIXTURES = [
   {
     id:31,
     label:'Nimbus UI',
-    url:'%@://%@:%@/stacks',
+    url:'%@://%@:%@/',
     service_id: 'STORM',
-    template:'%@://%@:%@/nimbus',
-    http_config: 'mapreduce.jobhistory.webapp.address',
-    https_config: 'mapreduce.jobhistory.webapp.https.address',
+    template:'%@://%@:%@/',
+    http_config: 'nimbus.host',
+    https_config: 'nimbus.host',
     site: 'storm-site',
     regex: portRegex,
-    default_http_port: 6627
+    default_http_port: 8744
   },
   {
     id:32,

http://git-wip-us.apache.org/repos/asf/ambari/blob/1c9b1573/ambari-web/app/models/service_config.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/models/service_config.js b/ambari-web/app/models/service_config.js
index a875265..2c6928a 100644
--- a/ambari-web/app/models/service_config.js
+++ b/ambari-web/app/models/service_config.js
@@ -399,6 +399,18 @@ App.ServiceConfigProperty = Ember.Object.extend({
         if (!App.supports.storm) return; // @todo remove test mode check after Storm service integration
         this.set('value', masterComponentHostsInDB.findProperty('component', 'NIMBUS').hostName);
         break;
+      case 'logviewerserver_host':
+        if (!App.supports.storm) return; // @todo remove test mode check after Storm service integration
+        this.set('value', masterComponentHostsInDB.findProperty('component', 'LOGVIEWER_SERVER').hostName);
+        break;
+      case 'drpcserver_host':
+        if (!App.supports.storm) return; // @todo remove test mode check after Storm service integration
+        this.set('value', masterComponentHostsInDB.findProperty('component', 'DRPC_SERVER').hostName);
+        break;
+      case 'stormuiserver_host':
+        if (!App.supports.storm) return; // @todo remove test mode check after Storm service integration
+        this.set('value', masterComponentHostsInDB.findProperty('component', 'STORM_UI_SERVER').hostName);
+        break;
       case 'hbase.zookeeper.quorum':
         var zkHosts = masterComponentHostsInDB.filterProperty('component', 'ZOOKEEPER_SERVER').mapProperty('hostName');
         this.setDefaultValue("(\\w*)", zkHosts);


[05/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HDFS.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HDFS.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HDFS.json
index 002de94..bd1cb3c 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HDFS.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HDFS.json
@@ -1,1181 +1,1181 @@
 {
-  "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations?fields=*&_=1389714316446",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/ambari.dfs.datanode.http.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/ambari.dfs.datanode.http.port",
       "StackConfigurations" : {
         "property_description" : "\n      The datanode http port. This property is effective only if referenced from dfs.datanode.http.address property.\n    ",
         "property_name" : "ambari.dfs.datanode.http.port",
         "property_value" : "50075",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/ambari.dfs.datanode.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/ambari.dfs.datanode.port",
       "StackConfigurations" : {
         "property_description" : "\n      The datanode port for data transfer. This property is effective only if referenced from dfs.datanode.address property.\n    ",
         "property_name" : "ambari.dfs.datanode.port",
         "property_value" : "50010",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/datanode_du_reserved",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/datanode_du_reserved",
       "StackConfigurations" : {
         "property_description" : "Reserved space for HDFS",
         "property_name" : "datanode_du_reserved",
         "property_value" : "1073741824",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.block.access.token.enable",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.block.access.token.enable",
       "StackConfigurations" : {
-        "property_description" : "\n      If \"true\", access tokens are used as capabilities for accessing datanodes.\n      If \"false\", no access tokens are checked on accessing datanodes.\n    ",
+        "property_description" : "\nIf \"true\", access tokens are used as capabilities for accessing datanodes.\nIf \"false\", no access tokens are checked on accessing datanodes.\n",
         "property_name" : "dfs.block.access.token.enable",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.blockreport.initialDelay",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.blockreport.initialDelay",
       "StackConfigurations" : {
         "property_description" : "Delay for first block report in seconds.",
         "property_name" : "dfs.blockreport.initialDelay",
         "property_value" : "120",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.blocksize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.blocksize",
       "StackConfigurations" : {
         "property_description" : "The default block size for new files.",
         "property_name" : "dfs.blocksize",
         "property_value" : "134217728",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.client.read.shortcircuit",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.client.read.shortcircuit",
       "StackConfigurations" : {
         "property_description" : "\n      This configuration parameter turns on short-circuit local reads.\n    ",
         "property_name" : "dfs.client.read.shortcircuit",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.client.read.shortcircuit.streams.cache.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.client.read.shortcircuit.streams.cache.size",
       "StackConfigurations" : {
         "property_description" : "\n      The DFSClient maintains a cache of recently opened file descriptors. This\n      parameter controls the size of that cache. Setting this higher will use\n      more file descriptors, but potentially provide better performance on\n      workloads involving lots of seeks.\n    ",
         "property_name" : "dfs.client.read.shortcircuit.streams.cache.size",
         "property_value" : "4096",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.cluster.administrators",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.cluster.administrators",
       "StackConfigurations" : {
         "property_description" : "ACL for who all can view the default servlets in the HDFS",
         "property_name" : "dfs.cluster.administrators",
         "property_value" : " hdfs",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.address",
       "StackConfigurations" : {
         "property_description" : "\n      The datanode server address and port for data transfer.\n    ",
         "property_name" : "dfs.datanode.address",
         "property_value" : "0.0.0.0:${ambari.dfs.datanode.port}",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.balance.bandwidthPerSec",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.balance.bandwidthPerSec",
       "StackConfigurations" : {
-        "property_description" : "\n      Specifies the maximum amount of bandwidth that each datanode\n      can utilize for the balancing purpose in term of\n      the number of bytes per second.\n    ",
+        "property_description" : "\n        Specifies the maximum amount of bandwidth that each datanode\n        can utilize for the balancing purpose in term of\n        the number of bytes per second.\n  ",
         "property_name" : "dfs.datanode.balance.bandwidthPerSec",
         "property_value" : "6250000",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.data.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.data.dir",
       "StackConfigurations" : {
-        "property_description" : "Determines where on the local filesystem an DFS data node\n      should store its blocks.  If this is a comma-delimited\n      list of directories, then data will be stored in all named\n      directories, typically on different devices.\n      Directories that do not exist are ignored.\n    ",
+        "property_description" : "Determines where on the local filesystem an DFS data node\n  should store its blocks.  If this is a comma-delimited\n  list of directories, then data will be stored in all named\n  directories, typically on different devices.\n  Directories that do not exist are ignored.\n  ",
         "property_name" : "dfs.datanode.data.dir",
         "property_value" : "/hadoop/hdfs/data",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.data.dir.perm",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.data.dir.perm",
       "StackConfigurations" : {
-        "property_description" : "The permissions that should be there on dfs.datanode.data.dir\n      directories. The datanode will not come up if the permissions are\n      different on existing dfs.datanode.data.dir directories. If the directories\n      don't exist, they will be created with this permission.",
+        "property_description" : "The permissions that should be there on dfs.datanode.data.dir\ndirectories. The datanode will not come up if the permissions are\ndifferent on existing dfs.datanode.data.dir directories. If the directories\ndon't exist, they will be created with this permission.",
         "property_name" : "dfs.datanode.data.dir.perm",
         "property_value" : "750",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.du.reserved",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.du.reserved",
       "StackConfigurations" : {
-        "property_description" : "Reserved space in bytes per volume. Always leave this much space free for non dfs use.\n    ",
+        "property_description" : "Reserved space in bytes per volume. Always leave this much space free for non dfs use.\n",
         "property_name" : "dfs.datanode.du.reserved",
         "property_value" : "1073741824",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.failed.volumes.tolerated",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.failed.volumes.tolerated",
       "StackConfigurations" : {
         "property_description" : " Number of failed disks a DataNode would tolerate before it stops offering service",
         "property_name" : "dfs.datanode.failed.volumes.tolerated",
         "property_value" : "0",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.http.address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.http.address",
       "StackConfigurations" : {
         "property_description" : "\n      The datanode http server address and port.\n    ",
         "property_name" : "dfs.datanode.http.address",
         "property_value" : "0.0.0.0:${ambari.dfs.datanode.http.port}",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.ipc.address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.ipc.address",
       "StackConfigurations" : {
-        "property_description" : "\n      The datanode ipc server address and port.\n      If the port is 0 then the server will start on a free port.\n    ",
+        "property_description" : "\nThe datanode ipc server address and port.\nIf the port is 0 then the server will start on a free port.\n",
         "property_name" : "dfs.datanode.ipc.address",
         "property_value" : "0.0.0.0:8010",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.datanode.max.transfer.threads",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.datanode.max.transfer.threads",
       "StackConfigurations" : {
         "property_description" : "PRIVATE CONFIG VARIABLE",
         "property_name" : "dfs.datanode.max.transfer.threads",
         "property_value" : "1024",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.domain.socket.path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.domain.socket.path",
       "StackConfigurations" : {
         "property_description" : null,
         "property_name" : "dfs.domain.socket.path",
         "property_value" : "/var/lib/hadoop-hdfs/dn_socket",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.heartbeat.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.heartbeat.interval",
       "StackConfigurations" : {
         "property_description" : "Determines datanode heartbeat interval in seconds.",
         "property_name" : "dfs.heartbeat.interval",
         "property_value" : "3",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.hosts.exclude",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.hosts.exclude",
       "StackConfigurations" : {
-        "property_description" : "Names a file that contains a list of hosts that are\n      not permitted to connect to the namenode.  The full pathname of the\n      file must be specified.  If the value is empty, no hosts are\n      excluded.",
+        "property_description" : "Names a file that contains a list of hosts that are\n    not permitted to connect to the namenode.  The full pathname of the\n    file must be specified.  If the value is empty, no hosts are\n    excluded.",
         "property_name" : "dfs.hosts.exclude",
         "property_value" : "/etc/hadoop/conf/dfs.exclude",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.https.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.https.port",
       "StackConfigurations" : {
         "property_description" : "\n      This property is used by HftpFileSystem.\n    ",
         "property_name" : "dfs.https.port",
         "property_value" : "50470",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.journalnode.edits.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.journalnode.edits.dir",
       "StackConfigurations" : {
         "property_description" : "The path where the JournalNode daemon will store its local state. ",
         "property_name" : "dfs.journalnode.edits.dir",
         "property_value" : "/grid/0/hdfs/journal",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.journalnode.http-address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.journalnode.http-address",
       "StackConfigurations" : {
-        "property_description" : "The address and port the JournalNode web UI listens on.\n      If the port is 0 then the server will start on a free port. ",
+        "property_description" : "The address and port the JournalNode web UI listens on.\n     If the port is 0 then the server will start on a free port. ",
         "property_name" : "dfs.journalnode.http-address",
         "property_value" : "0.0.0.0:8480",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.accesstime.precision",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.accesstime.precision",
       "StackConfigurations" : {
-        "property_description" : "The access time for HDFS file is precise upto this value.\n      The default value is 1 hour. Setting a value of 0 disables\n      access times for HDFS.\n    ",
+        "property_description" : "The access time for HDFS file is precise upto this value.\n                 The default value is 1 hour. Setting a value of 0 disables\n                 access times for HDFS.\n    ",
         "property_name" : "dfs.namenode.accesstime.precision",
         "property_value" : "0",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.avoid.read.stale.datanode",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.avoid.read.stale.datanode",
       "StackConfigurations" : {
         "property_description" : "\n      Indicate whether or not to avoid reading from stale datanodes whose\n      heartbeat messages have not been received by the namenode for more than a\n      specified time interval.\n    ",
         "property_name" : "dfs.namenode.avoid.read.stale.datanode",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.avoid.write.stale.datanode",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.avoid.write.stale.datanode",
       "StackConfigurations" : {
         "property_description" : "\n      Indicate whether or not to avoid writing to stale datanodes whose\n      heartbeat messages have not been received by the namenode for more than a\n      specified time interval.\n    ",
         "property_name" : "dfs.namenode.avoid.write.stale.datanode",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.checkpoint.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.checkpoint.dir",
       "StackConfigurations" : {
         "property_description" : "Determines where on the local filesystem the DFS secondary\n      name node should store the temporary images to merge.\n      If this is a comma-delimited list of directories then the image is\n      replicated in all of the directories for redundancy.\n    ",
         "property_name" : "dfs.namenode.checkpoint.dir",
         "property_value" : "/hadoop/hdfs/namesecondary",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.checkpoint.edits.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.checkpoint.edits.dir",
       "StackConfigurations" : {
         "property_description" : "Determines where on the local filesystem the DFS secondary\n      name node should store the temporary edits to merge.\n      If this is a comma-delimited list of directoires then teh edits is\n      replicated in all of the directoires for redundancy.\n      Default value is same as dfs.namenode.checkpoint.dir\n    ",
         "property_name" : "dfs.namenode.checkpoint.edits.dir",
         "property_value" : "${dfs.namenode.checkpoint.dir}",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.checkpoint.period",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.checkpoint.period",
       "StackConfigurations" : {
         "property_description" : "The number of seconds between two periodic checkpoints.\n    ",
         "property_name" : "dfs.namenode.checkpoint.period",
         "property_value" : "21600",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.handler.count",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.handler.count",
       "StackConfigurations" : {
         "property_description" : "Added to grow Queue size so that more client connections are allowed",
         "property_name" : "dfs.namenode.handler.count",
         "property_value" : "100",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.handler.count",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.handler.count",
       "StackConfigurations" : {
         "property_description" : "The number of server threads for the namenode.",
         "property_name" : "dfs.namenode.handler.count",
         "property_value" : "40",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.http-address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.http-address",
       "StackConfigurations" : {
-        "property_description" : "The name of the default file system.  Either the\n      literal string \"local\" or a host:port for NDFS.",
+        "property_description" : "The name of the default file system.  Either the\nliteral string \"local\" or a host:port for NDFS.",
         "property_name" : "dfs.namenode.http-address",
         "property_value" : "localhost:50070",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.https-address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.https-address",
       "StackConfigurations" : {
         "property_description" : "The https address where namenode binds",
         "property_name" : "dfs.namenode.https-address",
         "property_value" : "localhost:50470",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.name.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.name.dir",
       "StackConfigurations" : {
         "property_description" : "Determines where on the local filesystem the DFS name node\n      should store the name table.  If this is a comma-delimited list\n      of directories then the name table is replicated in all of the\n      directories, for redundancy. ",
         "property_name" : "dfs.namenode.name.dir",
         "property_value" : "/hadoop/hdfs/namenode",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.name.dir.restore",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.name.dir.restore",
       "StackConfigurations" : {
         "property_description" : "Set to true to enable NameNode to attempt recovering a previously failed dfs.namenode.name.dir.\n      When enabled, a recovery of any failed directory is attempted during checkpoint.",
         "property_name" : "dfs.namenode.name.dir.restore",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.safemode.threshold-pct",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.safemode.threshold-pct",
       "StackConfigurations" : {
-        "property_description" : "\n      Specifies the percentage of blocks that should satisfy\n      the minimal replication requirement defined by dfs.namenode.replication.min.\n      Values less than or equal to 0 mean not to start in safe mode.\n      Values greater than 1 will make safe mode permanent.\n    ",
+        "property_description" : "\n        Specifies the percentage of blocks that should satisfy\n        the minimal replication requirement defined by dfs.namenode.replication.min.\n        Values less than or equal to 0 mean not to start in safe mode.\n        Values greater than 1 will make safe mode permanent.\n        ",
         "property_name" : "dfs.namenode.safemode.threshold-pct",
         "property_value" : "1.0f",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.secondary.http-address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.secondary.http-address",
       "StackConfigurations" : {
         "property_description" : "Address of secondary namenode web server",
         "property_name" : "dfs.namenode.secondary.http-address",
         "property_value" : "localhost:50090",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.stale.datanode.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.stale.datanode.interval",
       "StackConfigurations" : {
         "property_description" : "Datanode is stale after not getting a heartbeat in this interval in ms",
         "property_name" : "dfs.namenode.stale.datanode.interval",
         "property_value" : "30000",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.namenode.write.stale.datanode.ratio",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.namenode.write.stale.datanode.ratio",
       "StackConfigurations" : {
         "property_description" : "When the ratio of number stale datanodes to total datanodes marked is greater\n      than this ratio, stop avoiding writing to stale nodes so as to prevent causing hotspots.\n    ",
         "property_name" : "dfs.namenode.write.stale.datanode.ratio",
         "property_value" : "1.0f",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.permissions.enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.permissions.enabled",
       "StackConfigurations" : {
-        "property_description" : "\n      If \"true\", enable permission checking in HDFS.\n      If \"false\", permission checking is turned off,\n      but all other behavior is unchanged.\n      Switching from one parameter value to the other does not change the mode,\n      owner or group of files or directories.\n    ",
+        "property_description" : "\nIf \"true\", enable permission checking in HDFS.\nIf \"false\", permission checking is turned off,\nbut all other behavior is unchanged.\nSwitching from one parameter value to the other does not change the mode,\nowner or group of files or directories.\n",
         "property_name" : "dfs.permissions.enabled",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.permissions.superusergroup",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.permissions.superusergroup",
       "StackConfigurations" : {
         "property_description" : "The name of the group of super-users.",
         "property_name" : "dfs.permissions.superusergroup",
         "property_value" : "hdfs",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.replication",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.replication",
       "StackConfigurations" : {
-        "property_description" : "Default block replication.\n    ",
+        "property_description" : "Default block replication.\n  ",
         "property_name" : "dfs.replication",
         "property_value" : "3",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.replication.max",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.replication.max",
       "StackConfigurations" : {
-        "property_description" : "Maximal block replication.\n    ",
+        "property_description" : "Maximal block replication.\n  ",
         "property_name" : "dfs.replication.max",
         "property_value" : "50",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.support.append",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.support.append",
       "StackConfigurations" : {
         "property_description" : "to enable dfs append",
         "property_name" : "dfs.support.append",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs.webhdfs.enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs.webhdfs.enabled",
       "StackConfigurations" : {
         "property_description" : "Whether to enable WebHDFS feature",
         "property_name" : "dfs.webhdfs.enabled",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_block_local_path_access_user",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_block_local_path_access_user",
       "StackConfigurations" : {
         "property_description" : "Default Block Replication.",
         "property_name" : "dfs_block_local_path_access_user",
         "property_value" : "hbase",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_datanode_address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_datanode_address",
       "StackConfigurations" : {
         "property_description" : "Port for datanode address.",
         "property_name" : "dfs_datanode_address",
         "property_value" : "50010",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_datanode_data_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_datanode_data_dir",
       "StackConfigurations" : {
         "property_description" : "Data directories for Data Nodes.",
         "property_name" : "dfs_datanode_data_dir",
         "property_value" : "/hadoop/hdfs/data",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_datanode_data_dir_perm",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_datanode_data_dir_perm",
       "StackConfigurations" : {
         "property_description" : "Datanode dir perms.",
         "property_name" : "dfs_datanode_data_dir_perm",
         "property_value" : "750",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_datanode_failed_volume_tolerated",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_datanode_failed_volume_tolerated",
       "StackConfigurations" : {
         "property_description" : "DataNode volumes failure toleration",
         "property_name" : "dfs_datanode_failed_volume_tolerated",
         "property_value" : "0",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_datanode_http_address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_datanode_http_address",
       "StackConfigurations" : {
         "property_description" : "Port for datanode address.",
         "property_name" : "dfs_datanode_http_address",
         "property_value" : "50075",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_namenode_checkpoint_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_namenode_checkpoint_dir",
       "StackConfigurations" : {
         "property_description" : "Secondary NameNode checkpoint dir.",
         "property_name" : "dfs_namenode_checkpoint_dir",
         "property_value" : "/hadoop/hdfs/namesecondary",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_namenode_checkpoint_period",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_namenode_checkpoint_period",
       "StackConfigurations" : {
         "property_description" : "HDFS Maximum Checkpoint Delay",
         "property_name" : "dfs_namenode_checkpoint_period",
         "property_value" : "21600",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_namenode_name_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_namenode_name_dir",
       "StackConfigurations" : {
         "property_description" : "NameNode Directories.",
         "property_name" : "dfs_namenode_name_dir",
         "property_value" : "/hadoop/hdfs/namenode",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_replication",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_replication",
       "StackConfigurations" : {
         "property_description" : "Default Block Replication.",
         "property_name" : "dfs_replication",
         "property_value" : "3",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dfs_webhdfs_enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dfs_webhdfs_enabled",
       "StackConfigurations" : {
         "property_description" : "WebHDFS enabled",
         "property_name" : "dfs_webhdfs_enabled",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/dtnode_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/dtnode_heapsize",
       "StackConfigurations" : {
         "property_description" : "DataNode maximum Java heap size",
         "property_name" : "dtnode_heapsize",
         "property_value" : "1024",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/fs.checkpoint.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/fs.checkpoint.size",
       "StackConfigurations" : {
         "property_description" : "The size of the current edit log (in bytes) that triggers\n      a periodic checkpoint even if the maximum checkpoint delay is not reached\n    ",
         "property_name" : "fs.checkpoint.size",
         "property_value" : "67108864",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/fs.defaultFS",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/fs.defaultFS",
       "StackConfigurations" : {
         "property_description" : "The name of the default file system.  Either the\n  literal string \"local\" or a host:port for NDFS.",
         "property_name" : "fs.defaultFS",
         "property_value" : "hdfs://localhost:8020",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/fs.permissions.umask-mode",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/fs.permissions.umask-mode",
       "StackConfigurations" : {
-        "property_description" : "\n      The octal umask used when creating files and directories.\n    ",
+        "property_description" : "\nThe octal umask used when creating files and directories.\n",
         "property_name" : "fs.permissions.umask-mode",
         "property_value" : "022",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/fs.trash.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/fs.trash.interval",
       "StackConfigurations" : {
         "property_description" : "Number of minutes between trash checkpoints.\n  If zero, the trash feature is disabled.\n  ",
         "property_name" : "fs.trash.interval",
         "property_value" : "360",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/fs_checkpoint_size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/fs_checkpoint_size",
       "StackConfigurations" : {
         "property_description" : "FS Checkpoint Size.",
         "property_name" : "fs_checkpoint_size",
         "property_value" : "0.5",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/hadoop.security.auth_to_local",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/hadoop.security.auth_to_local",
       "StackConfigurations" : {
         "property_description" : "The mapping from kerberos principal names to local OS mapreduce.job.user.names.\n  So the default rule is just \"DEFAULT\" which takes all principals in your default domain to their first component.\n  \"omalley@APACHE.ORG\" and \"omalley/admin@APACHE.ORG\" to \"omalley\", if your default domain is APACHE.ORG.\nThe translations rules have 3 sections:\n      base     filter    substitution\nThe base consists of a number that represents the number of components in the principal name excluding the realm and the pattern for building the name from the sections of the principal name. The base uses $0 to mean the realm, $1 to mean the first component and $2 to mean the second component.\n\n[1:$1@$0] translates \"omalley@APACHE.ORG\" to \"omalley@APACHE.ORG\"\n[2:$1] translates \"omalley/admin@APACHE.ORG\" to \"omalley\"\n[2:$1%$2] translates \"omalley/admin@APACHE.ORG\" to \"omalley%admin\"\n\nThe filter is a regex in parens that must the generated string 
 for the rule to apply.\n\n\"(.*%admin)\" will take any string that ends in \"%admin\"\n\"(.*@ACME.COM)\" will take any string that ends in \"@ACME.COM\"\n\nFinally, the substitution is a sed rule to translate a regex into a fixed string.\n\n\"s/@ACME\\.COM//\" removes the first instance of \"@ACME.COM\".\n\"s/@[A-Z]*\\.COM//\" removes the first instance of \"@\" followed by a name followed by \".COM\".\n\"s/X/Y/g\" replaces all of the \"X\" in the name with \"Y\"\n\nSo, if your default realm was APACHE.ORG, but you also wanted to take all principals from ACME.COM that had a single component \"joe@ACME.COM\", you'd do:\n\nRULE:[1:$1@$0](.@ACME.ORG)s/@.//\nDEFAULT\n\nTo also translate the names with a second component, you'd make the rules:\n\nRULE:[1:$1@$0](.@ACME.ORG)s/@.//\nRULE:[2:$1@$0](.@ACME.ORG)s/@.//\nDEFAULT\n\nIf you want to treat all principals from APACHE.ORG with /admin as \"admin\", your rules would look like:\n\nRULE[2:$1%$2@$0](.%admin@APACHE.ORG)s/./admin/\nDEFAULT\n
     ",
         "property_name" : "hadoop.security.auth_to_local",
         "property_value" : "\n        RULE:[2:$1@$0]([rn]m@.*)s/.*/yarn/\n        RULE:[2:$1@$0](jhs@.*)s/.*/mapred/\n        RULE:[2:$1@$0]([nd]n@.*)s/.*/hdfs/\n        RULE:[2:$1@$0](hm@.*)s/.*/hbase/\n        RULE:[2:$1@$0](rs@.*)s/.*/hbase/\n        DEFAULT\n    ",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/hadoop.security.authentication",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/hadoop.security.authentication",
       "StackConfigurations" : {
         "property_description" : "\n   Set the authentication for the cluster. Valid values are: simple or\n   kerberos.\n   ",
         "property_name" : "hadoop.security.authentication",
         "property_value" : "simple",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/hadoop.security.authorization",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/hadoop.security.authorization",
       "StackConfigurations" : {
         "property_description" : "\n     Enable authorization for different protocols.\n  ",
         "property_name" : "hadoop.security.authorization",
         "property_value" : "false",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/hadoop_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/hadoop_heapsize",
       "StackConfigurations" : {
         "property_description" : "Hadoop maximum Java heap size",
         "property_name" : "hadoop_heapsize",
         "property_value" : "1024",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/hadoop_pid_dir_prefix",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/hadoop_pid_dir_prefix",
       "StackConfigurations" : {
         "property_description" : "Hadoop PID Dir Prefix",
         "property_name" : "hadoop_pid_dir_prefix",
         "property_value" : "/var/run/hadoop",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/hdfs_log_dir_prefix",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/hdfs_log_dir_prefix",
       "StackConfigurations" : {
         "property_description" : "Hadoop Log Dir Prefix",
         "property_name" : "hdfs_log_dir_prefix",
         "property_value" : "/var/log/hadoop",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/hdfs_user",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/hdfs_user",
       "StackConfigurations" : {
         "property_description" : "User and Groups.",
         "property_name" : "hdfs_user",
         "property_value" : "hdfs",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/io.compression.codecs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/io.compression.codecs",
       "StackConfigurations" : {
         "property_description" : "A list of the compression codec classes that can be used\n                 for compression/decompression.",
         "property_name" : "io.compression.codecs",
         "property_value" : "org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/io.file.buffer.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/io.file.buffer.size",
       "StackConfigurations" : {
         "property_description" : "The size of buffer for use in sequence files.\n  The size of this buffer should probably be a multiple of hardware\n  page size (4096 on Intel x86), and it determines how much data is\n  buffered during read and write operations.",
         "property_name" : "io.file.buffer.size",
         "property_value" : "131072",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/io.serializations",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/io.serializations",
       "StackConfigurations" : {
         "property_description" : null,
         "property_name" : "io.serializations",
         "property_value" : "org.apache.hadoop.io.serializer.WritableSerialization",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/ipc.client.connect.max.retries",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/ipc.client.connect.max.retries",
       "StackConfigurations" : {
         "property_description" : "Defines the maximum number of retries for IPC connections.",
         "property_name" : "ipc.client.connect.max.retries",
         "property_value" : "50",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/ipc.client.connection.maxidletime",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/ipc.client.connection.maxidletime",
       "StackConfigurations" : {
         "property_description" : "The maximum time after which a client will bring down the\n               connection to the server.\n  ",
         "property_name" : "ipc.client.connection.maxidletime",
         "property_value" : "30000",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/ipc.client.idlethreshold",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/ipc.client.idlethreshold",
       "StackConfigurations" : {
         "property_description" : "Defines the threshold number of connections after which\n               connections will be inspected for idleness.\n  ",
         "property_name" : "ipc.client.idlethreshold",
         "property_value" : "8000",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/kerberos_domain",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/kerberos_domain",
       "StackConfigurations" : {
         "property_description" : "Kerberos realm.",
         "property_name" : "kerberos_domain",
         "property_value" : "EXAMPLE.COM",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/keytab_path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/keytab_path",
       "StackConfigurations" : {
         "property_description" : "Kerberos keytab path.",
         "property_name" : "keytab_path",
         "property_value" : "/etc/security/keytabs",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/keytab_path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/keytab_path",
       "StackConfigurations" : {
         "property_description" : "KeyTab Directory.",
         "property_name" : "keytab_path",
         "property_value" : "/etc/security/keytabs",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/lzo_enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/lzo_enabled",
       "StackConfigurations" : {
         "property_description" : "LZO compression enabled",
         "property_name" : "lzo_enabled",
         "property_value" : "true",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/mapreduce.jobtracker.webinterface.trusted",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/mapreduce.jobtracker.webinterface.trusted",
       "StackConfigurations" : {
         "property_description" : " If set to true, the web interfaces of JT and NN may contain\n                actions, such as kill job, delete file, etc., that should\n                not be exposed to public. Enable this option if the interfaces\n                are only reachable by those who have the right authorization.\n  ",
         "property_name" : "mapreduce.jobtracker.webinterface.trusted",
         "property_value" : "false",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/namenode_formatted_mark_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/namenode_formatted_mark_dir",
       "StackConfigurations" : {
         "property_description" : "Formatteed Mark Directory.",
         "property_name" : "namenode_formatted_mark_dir",
         "property_value" : "/var/run/hadoop/hdfs/namenode/formatted/",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/namenode_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/namenode_heapsize",
       "StackConfigurations" : {
         "property_description" : "NameNode Java heap size",
         "property_name" : "namenode_heapsize",
         "property_value" : "1024",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/namenode_opt_maxnewsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/namenode_opt_maxnewsize",
       "StackConfigurations" : {
         "property_description" : "NameNode maximum new generation size",
         "property_name" : "namenode_opt_maxnewsize",
         "property_value" : "200",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/namenode_opt_newsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/namenode_opt_newsize",
       "StackConfigurations" : {
         "property_description" : "NameNode new generation size",
         "property_name" : "namenode_opt_newsize",
         "property_value" : "200",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/proxyuser_group",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/proxyuser_group",
       "StackConfigurations" : {
         "property_description" : "Proxy user group.",
         "property_name" : "proxyuser_group",
         "property_value" : "users",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.admin.operations.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.admin.operations.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for AdminOperationsProtocol. Used for admin commands.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.admin.operations.protocol.acl",
         "property_value" : "hadoop",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.client.datanode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.client.datanode.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for ClientDatanodeProtocol, the client-to-datanode protocol\n    for block recovery.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.client.datanode.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.client.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.client.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for ClientProtocol, which is used by user code\n    via the DistributedFileSystem.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.client.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.datanode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.datanode.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for DatanodeProtocol, which is used by datanodes to\n    communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.datanode.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.inter.datanode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.inter.datanode.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for InterDatanodeProtocol, the inter-datanode protocol\n    for updating generation timestamp.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.inter.datanode.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.inter.tracker.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.inter.tracker.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for InterTrackerProtocol, used by the tasktrackers to\n    communicate with the jobtracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.inter.tracker.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.job.client.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.job.client.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for JobSubmissionProtocol, used by job clients to\n    communciate with the jobtracker for job submission, querying job status etc.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.job.client.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.job.task.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.job.task.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for TaskUmbilicalProtocol, used by the map and reduce\n    tasks to communicate with the parent tasktracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.job.task.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.namenode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.namenode.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for NamenodeProtocol, the protocol used by the secondary\n    namenode to communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.namenode.protocol.acl",
         "property_value" : "*",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.refresh.policy.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.refresh.policy.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for RefreshAuthorizationPolicyProtocol, used by the\n    dfsadmin and mradmin commands to refresh the security policy in-effect.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.refresh.policy.protocol.acl",
         "property_value" : "hadoop",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security.refresh.usertogroups.mappings.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security.refresh.usertogroups.mappings.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for RefreshUserMappingsProtocol. Used to refresh\n    users mappings. The ACL is a comma-separated list of user and\n    group names. The user and group list is separated by a blank. For\n    e.g. \"alice,bob users,wheel\".  A special value of \"*\" means all\n    users are allowed.",
         "property_name" : "security.refresh.usertogroups.mappings.protocol.acl",
         "property_value" : "hadoop",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://c6401.ambari.apache.org:8080/api/v1/stacks2/HDP/versions/2.0.8/stackServices/HDFS/configurations/security_enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HDFS/configurations/security_enabled",
       "StackConfigurations" : {
         "property_description" : "Hadoop Security",
         "property_name" : "security_enabled",
         "property_value" : "false",
         "service_name" : "HDFS",
         "stack_name" : "HDP",
-        "stack_version" : "2.0.8",
+        "stack_version" : "2.0.1",
         "type" : "global.xml"
       }
     }
   ]
-}
+}
\ No newline at end of file


[07/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/NAGIOS.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/NAGIOS.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/NAGIOS.json
index 58dbfd3..d7223c0 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/NAGIOS.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/NAGIOS.json
@@ -1,41 +1,41 @@
 {
-  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations?fields=*&_=1368459065260",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations?fields=*",
   "items" : [
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations/nagios_group",
       "StackConfigurations" : {
         "property_description" : "Nagios Group.",
-        "property_value" : "nagios",
-        "stack_version" : "1.3.0",
         "property_name" : "nagios_group",
+        "property_value" : "nagios",
         "service_name" : "NAGIOS",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
-      }
-    },
-    {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations/nagios_web_login",
-      "StackConfigurations" : {
-        "property_description" : "Nagios web user.",
-        "property_value" : "nagiosadmin",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "nagios_web_login",
-        "service_name" : "NAGIOS",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "type" : "global.xml"
       }
     },
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations/nagios_user",
       "StackConfigurations" : {
         "property_description" : "Nagios Username.",
+        "property_name" : "nagios_user",
         "property_value" : "nagios",
+        "service_name" : "NAGIOS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "nagios_user",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations/nagios_web_login",
+      "StackConfigurations" : {
+        "property_description" : "Nagios web user.",
+        "property_name" : "nagios_web_login",
+        "property_value" : "nagiosadmin",
         "service_name" : "NAGIOS",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     }
   ]
-}
+}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/OOZIE.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/OOZIE.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/OOZIE.json
index 25d3049..ba29153 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/OOZIE.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/OOZIE.json
@@ -1,316 +1,436 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.systemmode",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.kerberos.name.rules",
       "StackConfigurations" : {
-        "property_description" : "\n     System mode for  Oozie at startup.\n     ",
-        "property_value" : "NORMAL",
+        "property_description" : "The mapping from kerberos principal names to local OS user names.",
+        "property_name" : "oozie.authentication.kerberos.name.rules",
+        "property_value" : "DEFAULT",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.systemmode",
-        "filename" : "oozie-site.xml",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.type",
+      "StackConfigurations" : {
+        "property_description" : "\n      ",
+        "property_name" : "oozie.authentication.type",
+        "property_value" : "simple",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.kerberos.name.rules",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.base.url",
       "StackConfigurations" : {
-        "property_description" : "The mapping from kerberos principal names to local OS user names.",
-        "property_value" : "RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/\nRULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/\nRULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\nRULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\nDEFAULT",
+        "property_description" : "Base Oozie URL.",
+        "property_name" : "oozie.base.url",
+        "property_value" : "http://localhost:11000/oozie",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.authentication.kerberos.name.rules",
-        "filename" : "oozie-site.xml",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.db.schema.name",
+      "StackConfigurations" : {
+        "property_description" : "\n      Oozie DataBase Name\n     ",
+        "property_name" : "oozie.db.schema.name",
+        "property_value" : "oozie",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.ActionService.executor.ext.classes",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "oozie.service.ActionService.executor.ext.classes",
+        "property_value" : "org.apache.oozie.action.email.EmailActionExecutor,\norg.apache.oozie.action.hadoop.HiveActionExecutor,\norg.apache.oozie.action.hadoop.ShellActionExecutor,\norg.apache.oozie.action.hadoop.SqoopActionExecutor,\norg.apache.oozie.action.hadoop.DistcpActionExecutor",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.AuthorizationService.security.enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.AuthorizationService.authorization.enabled",
       "StackConfigurations" : {
         "property_description" : "\n     Specifies whether security (user name/admin role) is enabled or not.\n     If disabled any user can manage Oozie system and manage any job.\n     ",
+        "property_name" : "oozie.service.AuthorizationService.authorization.enabled",
         "property_value" : "true",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.AuthorizationService.security.enabled",
-        "filename" : "oozie-site.xml",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.callable.concurrency",
+      "StackConfigurations" : {
+        "property_description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
+        "property_name" : "oozie.service.CallableQueueService.callable.concurrency",
+        "property_value" : "3",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.queue.size",
+      "StackConfigurations" : {
+        "property_description" : "Max callable queue size",
+        "property_name" : "oozie.service.CallableQueueService.queue.size",
+        "property_value" : "1000",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.threads",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.threads",
       "StackConfigurations" : {
         "property_description" : "Number of threads used for executing callables",
+        "property_name" : "oozie.service.CallableQueueService.threads",
         "property_value" : "10",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.CallableQueueService.threads",
-        "filename" : "oozie-site.xml",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.hadoop.configurations",
+      "StackConfigurations" : {
+        "property_description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
+        "property_name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
+        "property_value" : "*=/etc/hadoop/conf",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.jobTracker.whitelist",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.jobTracker.whitelist",
       "StackConfigurations" : {
         "property_description" : "\n      Whitelisted job tracker for Oozie service.\n      ",
-        "property_value" : " ",
-        "stack_version" : "1.3.0",
         "property_name" : "oozie.service.HadoopAccessorService.jobTracker.whitelist",
-        "filename" : "oozie-site.xml",
+        "property_value" : " ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.pool.max.active.conn",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.nameNode.whitelist",
       "StackConfigurations" : {
-        "property_description" : "\n             Max number of connections.\n        ",
-        "property_value" : "10",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.pool.max.active.conn",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n      ",
+        "property_name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
+        "property_value" : " ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.base.url",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.create.db.schema",
       "StackConfigurations" : {
-        "property_description" : "Base Oozie URL.",
-        "property_value" : "http://localhost:11000/oozie",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.base.url",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
+        "property_name" : "oozie.service.JPAService.create.db.schema",
+        "property_value" : "false",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.SchemaService.wf.ext.schemas",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.driver",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.SchemaService.wf.ext.schemas",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n            JDBC driver class.\n        ",
+        "property_name" : "oozie.service.JPAService.jdbc.driver",
+        "property_value" : "org.apache.derby.jdbc.EmbeddedDriver",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.older.than",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.password",
       "StackConfigurations" : {
-        "property_description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
-        "property_value" : "30",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.PurgeService.older.than",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n            DB user password.\n\n            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,\n                       if empty Configuration assumes it is NULL.\n        ",
+        "property_name" : "oozie.service.JPAService.jdbc.password",
+        "property_value" : " ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.nameNode.whitelist",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.url",
       "StackConfigurations" : {
-        "property_description" : "\n      ",
-        "property_value" : " ",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n            JDBC URL.\n        ",
+        "property_name" : "oozie.service.JPAService.jdbc.url",
+        "property_value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.coord.normal.default.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.username",
       "StackConfigurations" : {
-        "property_description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
-        "property_value" : "120",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.coord.normal.default.timeout",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n            DB user name.\n        ",
+        "property_name" : "oozie.service.JPAService.jdbc.username",
+        "property_value" : "oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/use.system.libpath.for.mapreduce.and.pig.jobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.pool.max.active.conn",
       "StackConfigurations" : {
-        "property_description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
-        "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n             Max number of connections.\n        ",
+        "property_name" : "oozie.service.JPAService.pool.max.active.conn",
+        "property_value" : "10",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.queue.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.older.than",
       "StackConfigurations" : {
-        "property_description" : "Max callable queue size",
-        "property_value" : "1000",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.CallableQueueService.queue.size",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
+        "property_name" : "oozie.service.PurgeService.older.than",
+        "property_value" : "30",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.ActionService.executor.ext.classes",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.purge.interval",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "org.apache.oozie.action.email.EmailActionExecutor,\norg.apache.oozie.action.hadoop.HiveActionExecutor,\norg.apache.oozie.action.hadoop.ShellActionExecutor,\norg.apache.oozie.action.hadoop.SqoopActionExecutor,\norg.apache.oozie.action.hadoop.DistcpActionExecutor",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.ActionService.executor.ext.classes",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n     Interval at which the purge service will run, in seconds.\n     ",
+        "property_name" : "oozie.service.PurgeService.purge.interval",
+        "property_value" : "3600",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.password",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.SchemaService.wf.ext.schemas",
       "StackConfigurations" : {
-        "property_description" : "\n            DB user password.\n\n            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,\n                       if empty Configuration assumes it is NULL.\n        ",
-        "property_value" : " ",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.jdbc.password",
-        "filename" : "oozie-site.xml",
+        "property_description" : null,
+        "property_name" : "oozie.service.SchemaService.wf.ext.schemas",
+        "property_value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.WorkflowAppService.system.libpath",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.WorkflowAppService.system.libpath",
       "StackConfigurations" : {
         "property_description" : "\n      System library path to use for workflow applications.\n      This path is added to workflow application if their job properties sets\n      the property 'oozie.use.system.libpath' to true.\n      ",
-        "property_value" : "/user/${user.name}/share/lib",
-        "stack_version" : "1.3.0",
         "property_name" : "oozie.service.WorkflowAppService.system.libpath",
-        "filename" : "oozie-site.xml",
+        "property_value" : "/user/${user.name}/share/lib",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.driver",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.coord.normal.default.timeout",
       "StackConfigurations" : {
-        "property_description" : "\n            JDBC driver class.\n        ",
-        "property_value" : "org.apache.derby.jdbc.EmbeddedDriver",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.jdbc.driver",
-        "filename" : "oozie-site.xml",
+        "property_description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
+        "property_name" : "oozie.service.coord.normal.default.timeout",
+        "property_value" : "120",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.hadoop.configurations",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.system.id",
       "StackConfigurations" : {
-        "property_description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
-        "property_value" : "*=/etc/hadoop/conf",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n    The Oozie system ID.\n    ",
+        "property_name" : "oozie.system.id",
+        "property_value" : "oozie-${user.name}",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.purge.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.systemmode",
       "StackConfigurations" : {
-        "property_description" : "\n     Interval at which the purge service will run, in seconds.\n     ",
-        "property_value" : "3600",
+        "property_description" : "\n     System mode for  Oozie at startup.\n     ",
+        "property_name" : "oozie.systemmode",
+        "property_value" : "NORMAL",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.PurgeService.purge.interval",
-        "filename" : "oozie-site.xml",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_ambari_database",
+      "StackConfigurations" : {
+        "property_description" : "Database default.",
+        "property_name" : "oozie_ambari_database",
+        "property_value" : "MySQL",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.create.db.schema",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_data_dir",
       "StackConfigurations" : {
-        "property_description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
-        "property_value" : "false",
+        "property_description" : "Data directory in which the Oozie DB exists",
+        "property_name" : "oozie_data_dir",
+        "property_value" : "/hadoop/oozie/data",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.create.db.schema",
-        "filename" : "oozie-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_database_name",
+      "StackConfigurations" : {
+        "property_description" : "Database name used for the Oozie.",
+        "property_name" : "oozie_database_name",
+        "property_value" : "oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.system.id",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_derby_database",
       "StackConfigurations" : {
-        "property_description" : "\n    The Oozie system ID.\n    ",
-        "property_value" : "oozie-${user.name}",
+        "property_description" : "Oozie Derby Database.",
+        "property_name" : "oozie_derby_database",
+        "property_value" : "Derby",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.system.id",
-        "filename" : "oozie-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_existing_mysql_database",
+      "StackConfigurations" : {
+        "property_description" : "Oozie MySQL Database.",
+        "property_name" : "oozie_existing_mysql_database",
+        "property_value" : "MySQL",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.username",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_existing_oracle_database",
       "StackConfigurations" : {
-        "property_description" : "\n            DB user name.\n        ",
-        "property_value" : "sa",
+        "property_description" : "Oracle Database",
+        "property_name" : "oozie_existing_oracle_database",
+        "property_value" : "Oracle",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.jdbc.username",
-        "filename" : "oozie-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_log_dir",
+      "StackConfigurations" : {
+        "property_description" : "Directory for oozie logs",
+        "property_name" : "oozie_log_dir",
+        "property_value" : "/var/log/oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.db.schema.name",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_metastore_user_name",
       "StackConfigurations" : {
-        "property_description" : "\n      Oozie DataBase Name\n     ",
+        "property_description" : "Database user name to use to connect to the database",
+        "property_name" : "oozie_metastore_user_name",
         "property_value" : "oozie",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.db.schema.name",
-        "filename" : "oozie-site.xml",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.type",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_pid_dir",
       "StackConfigurations" : {
-        "property_description" : "\n      ",
-        "property_value" : "simple",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.authentication.type",
-        "filename" : "oozie-site.xml",
+        "property_description" : "Directory in which the pid files for oozie reside.",
+        "property_name" : "oozie_pid_dir",
+        "property_value" : "/var/run/oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.url",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie_user",
       "StackConfigurations" : {
-        "property_description" : "\n            JDBC URL.\n        ",
-        "property_value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.jdbc.url",
-        "filename" : "oozie-site.xml",
+        "property_description" : "Oozie User.",
+        "property_name" : "oozie_user",
+        "property_value" : "oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.callable.concurrency",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/use.system.libpath.for.mapreduce.and.pig.jobs",
       "StackConfigurations" : {
-        "property_description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
-        "property_value" : "3",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.CallableQueueService.callable.concurrency",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
+        "property_name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
+        "property_value" : "false",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "oozie-site.xml"
       }
     }
   ]

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/WEBHCAT.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/WEBHCAT.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/WEBHCAT.json
index 2bf81aa..89f6e30 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/WEBHCAT.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/WEBHCAT.json
@@ -1,172 +1,184 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.streaming.jar",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.exec.timeout",
       "StackConfigurations" : {
-        "property_description" : "The hdfs path to the Hadoop streaming jar file.",
-        "property_value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.streaming.jar",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "Time out for templeton api",
+        "property_name" : "templeton.exec.timeout",
+        "property_value" : "60000",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
-      }
-    },
-    {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.override.enabled",
-      "StackConfigurations" : {
-        "property_description" : "\n     Enable the override path in templeton.override.jars\n   ",
-        "property_value" : "false",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "templeton.override.enabled",
-        "filename" : "webhcat-site.xml",
-        "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.archive",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop",
       "StackConfigurations" : {
-        "property_description" : "The path to the Pig archive.",
-        "property_value" : "hdfs:///apps/webhcat/pig.tar.gz",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.pig.archive",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Hadoop executable.",
+        "property_name" : "templeton.hadoop",
+        "property_value" : "/usr/bin/hadoop",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop.conf.dir",
       "StackConfigurations" : {
-        "property_description" : "The path to the Hadoop executable.",
-        "property_value" : "/usr/bin/hadoop",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.hadoop",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Hadoop configuration.",
+        "property_name" : "templeton.hadoop.conf.dir",
+        "property_value" : "/etc/hadoop/conf",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hcat",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hcat",
       "StackConfigurations" : {
         "property_description" : "The path to the hcatalog executable.",
-        "property_value" : "/usr/bin/hcat",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.hcat",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "/usr/bin/hcat",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.exec.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.archive",
       "StackConfigurations" : {
-        "property_description" : "Time out for templeton api",
-        "property_value" : "60000",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.exec.timeout",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Hive archive.",
+        "property_name" : "templeton.hive.archive",
+        "property_value" : "hdfs:///apps/webhcat/hive.tar.gz",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop.conf.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.path",
       "StackConfigurations" : {
-        "property_description" : "The path to the Hadoop configuration.",
-        "property_value" : "/etc/hadoop/conf",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.hadoop.conf.dir",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Hive executable.",
+        "property_name" : "templeton.hive.path",
+        "property_value" : "hive.tar.gz/hive/bin/hive",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.jar",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.jar",
       "StackConfigurations" : {
         "property_description" : "The path to the Templeton jar file.",
-        "property_value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.jar",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.libjars",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.libjars",
       "StackConfigurations" : {
         "property_description" : "Jars to add the the classpath.",
-        "property_value" : "/usr/lib/zookeeper/zookeeper.jar",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.libjars",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "/usr/lib/zookeeper/zookeeper.jar",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.override.enabled",
       "StackConfigurations" : {
-        "property_description" : "The HTTP port for the main server.",
-        "property_value" : "50111",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.port",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "\n     Enable the override path in templeton.override.jars\n   ",
+        "property_name" : "templeton.override.enabled",
+        "property_value" : "false",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.storage.class",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.archive",
       "StackConfigurations" : {
-        "property_description" : "The class to use as storage",
-        "property_value" : "org.apache.hcatalog.templeton.tool.ZooKeeperStorage",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.storage.class",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Pig archive.",
+        "property_name" : "templeton.pig.archive",
+        "property_value" : "hdfs:///apps/webhcat/pig.tar.gz",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.archive",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.path",
       "StackConfigurations" : {
-        "property_description" : "The path to the Hive archive.",
-        "property_value" : "hdfs:///apps/webhcat/hive.tar.gz",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.hive.archive",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Pig executable.",
+        "property_name" : "templeton.pig.path",
+        "property_value" : "pig.tar.gz/pig/bin/pig",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.port",
       "StackConfigurations" : {
-        "property_description" : "The path to the Hive executable.",
-        "property_value" : "hive.tar.gz/hive/bin/hive",
+        "property_description" : "The HTTP port for the main server.",
+        "property_name" : "templeton.port",
+        "property_value" : "50111",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "templeton.hive.path",
-        "filename" : "webhcat-site.xml",
+        "type" : "webhcat-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.storage.class",
+      "StackConfigurations" : {
+        "property_description" : "The class to use as storage",
+        "property_name" : "templeton.storage.class",
+        "property_value" : "org.apache.hcatalog.templeton.tool.ZooKeeperStorage",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.streaming.jar",
       "StackConfigurations" : {
-        "property_description" : "The path to the Pig executable.",
-        "property_value" : "pig.tar.gz/pig/bin/pig",
+        "property_description" : "The hdfs path to the Hadoop streaming jar file.",
+        "property_name" : "templeton.streaming.jar",
+        "property_value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "templeton.pig.path",
-        "filename" : "webhcat-site.xml",
+        "type" : "webhcat-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.zookeeper.hosts",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper servers, as comma separated host:port pairs",
+        "property_name" : "templeton.zookeeper.hosts",
+        "property_value" : "localhost:2181",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "webhcat-site.xml"
       }
     }
   ]

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/ZOOKEEPER.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/ZOOKEEPER.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/ZOOKEEPER.json
index aeeac38..a488527 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/ZOOKEEPER.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/ZOOKEEPER.json
@@ -1,4 +1,113 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations?fields=*",
-  "items" : [ ]
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations?fields=*",
+  "items" : [
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/clientPort",
+      "StackConfigurations" : {
+        "property_description" : "Port for running ZK Server.",
+        "property_name" : "clientPort",
+        "property_value" : "2181",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/initLimit",
+      "StackConfigurations" : {
+        "property_description" : "Ticks to allow for sync at Init.",
+        "property_name" : "initLimit",
+        "property_value" : "10",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/syncLimit",
+      "StackConfigurations" : {
+        "property_description" : "Ticks to allow for sync at Runtime.",
+        "property_name" : "syncLimit",
+        "property_value" : "5",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/tickTime",
+      "StackConfigurations" : {
+        "property_description" : "The length of a single tick in milliseconds, which is the basic time unit used by ZooKeeper",
+        "property_name" : "tickTime",
+        "property_value" : "2000",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/zk_data_dir",
+      "StackConfigurations" : {
+        "property_description" : "Data directory for ZooKeeper.",
+        "property_name" : "zk_data_dir",
+        "property_value" : "/hadoop/zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/zk_log_dir",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Log Dir",
+        "property_name" : "zk_log_dir",
+        "property_value" : "/var/log/zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/zk_pid_dir",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Pid Dir",
+        "property_name" : "zk_pid_dir",
+        "property_value" : "/var/run/zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/zk_pid_file",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Pid File",
+        "property_name" : "zk_pid_file",
+        "property_value" : "/var/run/zookeeper/zookeeper_server.pid",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations/zk_user",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper User.",
+        "property_name" : "zk_user",
+        "property_value" : "zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    }
+  ]
 }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/GANGLIA.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/GANGLIA.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/GANGLIA.json
index 97b2c77..a7941c1 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/GANGLIA.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/GANGLIA.json
@@ -1,65 +1,65 @@
 {
-  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations?fields=*&_=1368459065278",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/GANGLIA/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/ganglia_runtime_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/GANGLIA/configurations/ganglia_conf_dir",
       "StackConfigurations" : {
-        "property_description" : "Run directories for Ganglia",
-        "property_value" : "/var/run/ganglia/hdp",
-        "stack_version" : "1.3.0",
-        "property_name" : "ganglia_runtime_dir",
+        "property_description" : "",
+        "property_name" : "ganglia_conf_dir",
+        "property_value" : "/etc/ganglia/hdp",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/gmetad_user",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/GANGLIA/configurations/ganglia_runtime_dir",
       "StackConfigurations" : {
-        "property_description" : "User ",
-        "property_value" : "nobody",
-        "stack_version" : "1.3.0",
-        "property_name" : "gmetad_user",
+        "property_description" : "",
+        "property_name" : "ganglia_runtime_dir",
+        "property_value" : "/var/run/ganglia/hdp",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/rrdcached_base_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/GANGLIA/configurations/gmetad_user",
       "StackConfigurations" : {
-        "property_description" : "Default directory for saving the rrd files on ganglia server",
-        "property_value" : "/var/lib/ganglia/rrds",
-        "stack_version" : "1.3.0",
-        "property_name" : "rrdcached_base_dir",
+        "property_description" : "",
+        "property_name" : "gmetad_user",
+        "property_value" : "nobody",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/gmond_user",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/GANGLIA/configurations/gmond_user",
       "StackConfigurations" : {
-        "property_description" : "User ",
-        "property_value" : "nobody",
-        "stack_version" : "1.3.0",
+        "property_description" : "",
         "property_name" : "gmond_user",
+        "property_value" : "nobody",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/ganglia_conf_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/GANGLIA/configurations/rrdcached_base_dir",
       "StackConfigurations" : {
-        "property_description" : "Config directory for Ganglia",
-        "property_value" : "/etc/ganglia/hdp",
-        "stack_version" : "1.3.0",
-        "property_name" : "ganglia_conf_dir",
+        "property_description" : "Location of rrd files.",
+        "property_name" : "rrdcached_base_dir",
+        "property_value" : "/var/lib/ganglia/rrds",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     }
   ]
-}
+}
\ No newline at end of file


[04/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HIVE.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HIVE.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HIVE.json
index 5596881..c943ac9 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HIVE.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HIVE.json
@@ -1,208 +1,424 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/javax.jdo.option.ConnectionDriverName",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/ambari.hive.db.schema.name",
       "StackConfigurations" : {
-        "property_description" : "Driver class name for a JDBC metastore",
-        "property_value" : "com.mysql.jdbc.Driver",
-        "stack_version" : "1.3.0",
-        "property_name" : "javax.jdo.option.ConnectionDriverName",
-        "filename" : "hive-site.xml",
+        "property_description" : "Database name used as the Hive Metastore",
+        "property_name" : "ambari.hive.db.schema.name",
+        "property_value" : "hive",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_pid_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/fs.file.impl.disable.cache",
       "StackConfigurations" : {
-        "property_description" : "Hive PID Dir.",
-        "property_value" : "/var/run/hive",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive_pid_dir",
-        "filename" : "global.xml",
+        "property_description" : null,
+        "property_name" : "fs.file.impl.disable.cache",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/fs.hdfs.impl.disable.cache",
       "StackConfigurations" : {
-        "property_description" : "enable or disable the hive client authorization",
+        "property_description" : null,
+        "property_name" : "fs.hdfs.impl.disable.cache",
         "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.security.authorization.enabled",
-        "filename" : "hive-site.xml",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_conf_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.auto.convert.join",
       "StackConfigurations" : {
-        "property_description" : "Hive Conf Dir.",
-        "property_value" : "/etc/hive/conf",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive_conf_dir",
-        "filename" : "global.xml",
+        "property_description" : "Whether Hive enable the optimization about converting common\n      join into mapjoin based on the input file size.",
+        "property_name" : "hive.auto.convert.join",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hadoop.clientside.fs.operations",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.auto.convert.join.noconditionaltask",
       "StackConfigurations" : {
-        "property_description" : "FS operations are owned by client",
+        "property_description" : "Whether Hive enable the optimization about converting common join into mapjoin based on the input file\n      size. If this paramater is on, and the sum of size for n-1 of the tables/partitions for a n-way join is smaller than the\n      specified size, the join is directly converted to a mapjoin (there is no conditional task).\n    ",
+        "property_name" : "hive.auto.convert.join.noconditionaltask",
         "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "hadoop.clientside.fs.operations",
-        "filename" : "hive-site.xml",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.manager",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.auto.convert.join.noconditionaltask.size",
       "StackConfigurations" : {
-        "property_description" : "the hive client authorization manager class name.\n    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
-        "property_value" : "org.apache.hcatalog.security.HdfsAuthorizationProvider",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.security.authorization.manager",
-        "filename" : "hive-site.xml",
+        "property_description" : "If hive.auto.convert.join.noconditionaltask is off, this parameter does not take affect. However, if it\n      is on, and the sum of size for n-1 of the tables/partitions for a n-way join is smaller than this size, the join is directly\n      converted to a mapjoin(there is no conditional task). The default is 10MB.\n    ",
+        "property_name" : "hive.auto.convert.join.noconditionaltask.size",
+        "property_value" : "1000000000",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/fs.hdfs.impl.disable.cache",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.auto.convert.sortmerge.join",
+      "StackConfigurations" : {
+        "property_description" : "Will the join be automatically converted to a sort-merge join, if the joined tables pass\n      the criteria for sort-merge join.\n    ",
+        "property_name" : "hive.auto.convert.sortmerge.join",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.auto.convert.sortmerge.join.noconditionaltask",
       "StackConfigurations" : {
         "property_description" : null,
+        "property_name" : "hive.auto.convert.sortmerge.join.noconditionaltask",
         "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs.hdfs.impl.disable.cache",
-        "filename" : "hive-site.xml",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.warehouse.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.enforce.bucketing",
       "StackConfigurations" : {
-        "property_description" : "location of default database for the warehouse",
-        "property_value" : "/apps/hive/warehouse",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.warehouse.dir",
-        "filename" : "hive-site.xml",
+        "property_description" : "Whether bucketing is enforced. If true, while inserting into the table, bucketing is enforced.",
+        "property_name" : "hive.enforce.bucketing",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.semantic.analyzer.factory.impl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.enforce.sorting",
       "StackConfigurations" : {
-        "property_description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
-        "property_value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.semantic.analyzer.factory.impl",
-        "filename" : "hive-site.xml",
+        "property_description" : "Whether sorting is enforced. If true, while inserting into the table, sorting is enforced.",
+        "property_name" : "hive.enforce.sorting",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.map.aggr",
+      "StackConfigurations" : {
+        "property_description" : "Whether to use map-side aggregation in Hive Group By queries.",
+        "property_name" : "hive.map.aggr",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.mapjoin.bucket.cache.size",
+      "StackConfigurations" : {
+        "property_description" : "\n      Size per reducer.The default is 1G, i.e if the input size is 10G, it\n      will use 10 reducers.\n    ",
+        "property_name" : "hive.mapjoin.bucket.cache.size",
+        "property_value" : "10000",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_aux_jars_path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.mapred.reduce.tasks.speculative.execution",
       "StackConfigurations" : {
-        "property_description" : "Hive auxiliary jar path.",
-        "property_value" : "/usr/lib/hcatalog/share/hcatalog/hcatalog-core.jar",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive_aux_jars_path",
-        "filename" : "global.xml",
+        "property_description" : "Whether speculative execution for reducers should be turned on.",
+        "property_name" : "hive.mapred.reduce.tasks.speculative.execution",
+        "property_value" : "false",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.metastore.cache.pinobjtypes",
+      "StackConfigurations" : {
+        "property_description" : "List of comma separated metastore object types that should be pinned in the cache",
+        "property_name" : "hive.metastore.cache.pinobjtypes",
+        "property_value" : "Table,Database,Type,FieldSchema,Order",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.client.socket.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.metastore.client.socket.timeout",
       "StackConfigurations" : {
         "property_description" : "MetaStore Client socket timeout in seconds",
-        "property_value" : "60",
-        "stack_version" : "1.3.0",
         "property_name" : "hive.metastore.client.socket.timeout",
-        "filename" : "hive-site.xml",
+        "property_value" : "60",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.metastore.execute.setugi",
+      "StackConfigurations" : {
+        "property_description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
+        "property_name" : "hive.metastore.execute.setugi",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.server2.enable.doAs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.metastore.uris",
+      "StackConfigurations" : {
+        "property_description" : "URI for client to contact metastore server",
+        "property_name" : "hive.metastore.uris",
+        "property_value" : "thrift://localhost:9083",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.metastore.warehouse.dir",
+      "StackConfigurations" : {
+        "property_description" : "location of default database for the warehouse",
+        "property_name" : "hive.metastore.warehouse.dir",
+        "property_value" : "/apps/hive/warehouse",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.optimize.bucketmapjoin",
       "StackConfigurations" : {
         "property_description" : null,
+        "property_name" : "hive.optimize.bucketmapjoin",
         "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.server2.enable.doAs",
-        "filename" : "hive-site.xml",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.cache.pinobjtypes",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.optimize.bucketmapjoin.sortedmerge",
       "StackConfigurations" : {
-        "property_description" : "List of comma separated metastore object types that should be pinned in the cache",
-        "property_value" : "Table,Database,Type,FieldSchema,Order",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.cache.pinobjtypes",
-        "filename" : "hive-site.xml",
+        "property_description" : null,
+        "property_name" : "hive.optimize.bucketmapjoin.sortedmerge",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.execute.setugi",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.optimize.index.filter",
       "StackConfigurations" : {
-        "property_description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
+        "property_description" : "\n    Whether to enable automatic use of indexes\n    ",
+        "property_name" : "hive.optimize.index.filter",
         "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.execute.setugi",
-        "filename" : "hive-site.xml",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.optimize.mapjoin.mapreduce",
+      "StackConfigurations" : {
+        "property_description" : "If hive.auto.convert.join is off, this parameter does not take\n      affect. If it is on, and if there are map-join jobs followed by a map-reduce\n      job (for e.g a group by), each map-only job is merged with the following\n      map-reduce job.\n    ",
+        "property_name" : "hive.optimize.mapjoin.mapreduce",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.optimize.reducededuplication",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hive.optimize.reducededuplication",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.optimize.reducededuplication.min.reducer",
+      "StackConfigurations" : {
+        "property_description" : "Reduce deduplication merges two RSs by moving key/parts/reducer-num of the child RS to parent RS.\n      That means if reducer-num of the child RS is fixed (order by or forced bucketing) and small, it can make very slow, single MR.\n      The optimization will be disabled if number of reducers is less than specified value.\n    ",
+        "property_name" : "hive.optimize.reducededuplication.min.reducer",
+        "property_value" : "1",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/mysql_connector_url",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.security.authenticator.manager",
       "StackConfigurations" : {
-        "property_description" : "Hive PID Dir.",
-        "property_value" : "${download_url}/mysql-connector-java-5.1.18.zip",
-        "stack_version" : "1.3.0",
-        "property_name" : "mysql_connector_url",
-        "filename" : "global.xml",
+        "property_description" : "Hive client authenticator manager class name. The user-defined authenticator class should implement interface org.apache.hadoop.hive.ql.security.HiveAuthenticationProvider.  ",
+        "property_name" : "hive.security.authenticator.manager",
+        "property_value" : "org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.local",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.security.authorization.enabled",
       "StackConfigurations" : {
-        "property_description" : "controls whether to connect to remove metastore server or\n    open a new metastore server in Hive Client JVM",
+        "property_description" : "enable or disable the hive client authorization",
+        "property_name" : "hive.security.authorization.enabled",
         "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.local",
-        "filename" : "hive-site.xml",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.security.authorization.manager",
+      "StackConfigurations" : {
+        "property_description" : "the hive client authorization manager class name.\n    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
+        "property_name" : "hive.security.authorization.manager",
+        "property_value" : "org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.security.metastore.authorization.manager",
+      "StackConfigurations" : {
+        "property_description" : "The authorization manager class name to be used in the metastore for authorization. The user-defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveMetastoreAuthorizationProvider.  ",
+        "property_name" : "hive.security.metastore.authorization.manager",
+        "property_value" : "org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.semantic.analyzer.factory.impl",
+      "StackConfigurations" : {
+        "property_description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
+        "property_name" : "hive.semantic.analyzer.factory.impl",
+        "property_value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.server2.enable.doAs",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hive.server2.enable.doAs",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/hive.vectorized.execution.enabled",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hive.vectorized.execution.enabled",
+        "property_value" : "false",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/javax.jdo.option.ConnectionDriverName",
+      "StackConfigurations" : {
+        "property_description" : "Driver class name for a JDBC metastore",
+        "property_name" : "javax.jdo.option.ConnectionDriverName",
+        "property_value" : "com.mysql.jdbc.Driver",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/javax.jdo.option.ConnectionPassword",
+      "StackConfigurations" : {
+        "property_description" : "password to use against metastore database",
+        "property_name" : "javax.jdo.option.ConnectionPassword",
+        "property_value" : " ",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/javax.jdo.option.ConnectionURL",
+      "StackConfigurations" : {
+        "property_description" : "JDBC connect string for a JDBC metastore",
+        "property_name" : "javax.jdo.option.ConnectionURL",
+        "property_value" : "jdbc",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_log_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HIVE/configurations/javax.jdo.option.ConnectionUserName",
       "StackConfigurations" : {
-        "property_description" : "Directory for Hive Log files.",
-        "property_value" : "/var/log/hive",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive_log_dir",
-        "filename" : "global.xml",
+        "property_description" : "username to use against metastore database",
+        "property_name" : "javax.jdo.option.ConnectionUserName",
+        "property_value" : "hive",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hive-site.xml"
       }
     }
   ]

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/MAPREDUCE2.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/MAPREDUCE2.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/MAPREDUCE2.json
index e036803..8115132 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/MAPREDUCE2.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/MAPREDUCE2.json
@@ -2,544 +2,508 @@
   "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.system.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.queue.default.acl-administer-jobs",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "/mapred/system",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.system.dir",
+        "property_description" : null,
+        "property_name" : "mapred.queue.default.acl-administer-jobs",
+        "property_value" : "*",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "mapred-queue-acls.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.reduce.parallel.copies",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.queue.default.acl-submit-job",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "30",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.reduce.parallel.copies",
+        "property_description" : null,
+        "property_name" : "mapred.queue.default.acl-submit-job",
+        "property_value" : "*",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "mapred-queue-acls.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.tracker.persist.jobstatus.active",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred_log_dir_prefix",
       "StackConfigurations" : {
-        "property_description" : "Indicates if persistency of job status information is\n  active or not.\n  ",
-        "property_value" : "false",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.tracker.persist.jobstatus.active",
+        "property_description" : "Mapreduce Log Dir Prefix",
+        "property_name" : "mapred_log_dir_prefix",
+        "property_value" : "/var/log/hadoop-mapreduce",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.jobtracker.blacklist.fault-timeout-window",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred_pid_dir_prefix",
       "StackConfigurations" : {
-        "property_description" : "\n    3-hour sliding window (value is in minutes)\n  ",
-        "property_value" : "180",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.jobtracker.blacklist.fault-timeout-window",
+        "property_description" : "Mapreduce PID Dir Prefix",
+        "property_name" : "mapred_pid_dir_prefix",
+        "property_value" : "/var/run/hadoop-mapreduce",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.cluster.administrators",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred_user",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : " hadoop",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.cluster.administrators",
+        "property_description" : "Mapreduce User",
+        "property_name" : "mapred_user",
+        "property_value" : "mapred",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.fileoutputcommitter.marksuccessfuljobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.admin.map.child.java.opts",
       "StackConfigurations" : {
         "property_description" : null,
-        "property_value" : "false",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.fileoutputcommitter.marksuccessfuljobs",
+        "property_name" : "mapreduce.admin.map.child.java.opts",
+        "property_value" : "-Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.shuffle.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.admin.reduce.child.java.opts",
       "StackConfigurations" : {
-        "property_description" : "Default port that the ShuffleHandler will run on. ShuffleHandler is a service run at the NodeManager to facilitate transfers of intermediate Map outputs to requesting Reducers.",
-        "property_value" : "8081",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.shuffle.port",
+        "property_description" : null,
+        "property_name" : "mapreduce.admin.reduce.child.java.opts",
+        "property_value" : "-Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.tasktracker.tasks.sleeptime-before-sigkill",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.admin.user.env",
       "StackConfigurations" : {
-        "property_description" : "Normally, this is the amount of time before killing\n  processes, and the recommended-default is 5.000 seconds - a value of\n  5000 here.  In this case, we are using it solely to blast tasks before\n  killing them, and killing them very quickly (1/4 second) to guarantee\n  that we do not leave VMs around for later jobs.\n  ",
-        "property_value" : "250",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.tasktracker.tasks.sleeptime-before-sigkill",
+        "property_description" : "\n      Additional execution environment entries for map and reduce task processes.\n      This is not an additive property. You must preserve the original value if\n      you want your map and reduce tasks to have access to native libraries (compression, etc)\n    ",
+        "property_name" : "mapreduce.admin.user.env",
+        "property_value" : "LD_LIBRARY_PATH=/usr/lib/hadoop/lib/native:/usr/lib/hadoop/lib/native/`$JAVA_HOME/bin/java -d32 -version &> /dev/null;if [ $? -eq 0 ]; then echo Linux-i386-32; else echo Linux-amd64-64;fi`",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.input.limit",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.am.max-attempts",
       "StackConfigurations" : {
-        "property_description" : "The limit on the input size of the reduce. (This value\n  is 10 Gb.)  If the estimated input size of the reduce is greater than\n  this value, job is failed. A value of -1 means that there is no limit\n  set. ",
-        "property_value" : "10737418240",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.reduce.input.limit",
+        "property_description" : "\n      The maximum number of application attempts. It is a\n      application-specific setting. It should not be larger than the global number\n      set by resourcemanager. Otherwise, it will be override. The default number is\n      set to 2, to allow at least one retry for AM.\n    ",
+        "property_name" : "mapreduce.am.max-attempts",
+        "property_value" : "2",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.reduce.tasks.speculative.execution",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.application.classpath",
       "StackConfigurations" : {
-        "property_description" : "If true, then multiple instances of some reduce tasks\n               may be executed in parallel.",
-        "property_value" : "false",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.reduce.tasks.speculative.execution",
+        "property_description" : "\n      CLASSPATH for MR applications. A comma-separated list of CLASSPATH\n      entries.\n    ",
+        "property_name" : "mapreduce.application.classpath",
+        "property_value" : "$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*,$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.jobtracker.blacklist.fault-bucket-width",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.cluster.administrators",
       "StackConfigurations" : {
-        "property_description" : "\n    15-minute bucket size (value is in minutes)\n  ",
-        "property_value" : "15",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.jobtracker.blacklist.fault-bucket-width",
+        "property_description" : "\n      Administrators for MapReduce applications.\n    ",
+        "property_name" : "mapreduce.cluster.administrators",
+        "property_value" : " hadoop",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/io.sort.record.percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.framework.name",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : ".2",
-        "stack_version" : "2.0.1",
-        "property_name" : "io.sort.record.percent",
+        "property_description" : "\n      The runtime framework for executing MapReduce jobs. Can be one of local,\n      classic or yarn.\n    ",
+        "property_name" : "mapreduce.framework.name",
+        "property_value" : "yarn",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.history.server.embedded",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.job.reduce.slowstart.completedmaps",
       "StackConfigurations" : {
-        "property_description" : "Should job history server be embedded within Job tracker\nprocess",
-        "property_value" : "false",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.history.server.embedded",
+        "property_description" : "\n      Fraction of the number of maps in the job which should be complete before\n      reduces are scheduled for the job.\n    ",
+        "property_name" : "mapreduce.job.reduce.slowstart.completedmaps",
+        "property_value" : "0.05",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/hadoop.job.history.user.location",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobhistory.address",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "none",
-        "stack_version" : "2.0.1",
-        "property_name" : "hadoop.job.history.user.location",
+        "property_description" : "Enter your JobHistoryServer hostname.",
+        "property_name" : "mapreduce.jobhistory.address",
+        "property_value" : "localhost:10020",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.shuffle.input.buffer.percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobhistory.done-dir",
       "StackConfigurations" : {
-        "property_description" : "The percentage of memory to be allocated from the maximum heap\n  size to storing map outputs during the shuffle.\n  ",
-        "property_value" : "0.7",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.shuffle.input.buffer.percent",
+        "property_description" : "\n      Directory where history files are managed by the MR JobHistory Server.\n    ",
+        "property_name" : "mapreduce.jobhistory.done-dir",
+        "property_value" : "/mr-history/done",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.tracker.handler.count",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobhistory.intermediate-done-dir",
       "StackConfigurations" : {
-        "property_description" : "\n    The number of server threads for the JobTracker. This should be roughly\n    4% of the number of tasktracker nodes.\n    ",
-        "property_value" : "50",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.tracker.handler.count",
+        "property_description" : "\n      Directory where history files are written by MapReduce jobs.\n    ",
+        "property_name" : "mapreduce.jobhistory.intermediate-done-dir",
+        "property_value" : "/mr-history/tmp",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobtracker.split.metainfo.maxsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobhistory.webapp.address",
       "StackConfigurations" : {
-        "property_description" : "If the size of the split metainfo file is larger than this, the JobTracker will fail the job during\n    initialize.\n   ",
-        "property_value" : "50000000",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.jobtracker.split.metainfo.maxsize",
+        "property_description" : "Enter your JobHistoryServer hostname.",
+        "property_name" : "mapreduce.jobhistory.webapp.address",
+        "property_value" : "localhost:19888",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.healthChecker.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.map.java.opts",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "135000",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.healthChecker.interval",
+        "property_description" : "\n      Larger heap-size for child jvms of maps.\n    ",
+        "property_name" : "mapreduce.map.java.opts",
+        "property_value" : "-Xmx756m",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.output.compression.type",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.map.log.level",
       "StackConfigurations" : {
-        "property_description" : "If the job outputs are to compressed as SequenceFiles, how should\n               they be compressed? Should be one of NONE, RECORD or BLOCK.\n  ",
-        "property_value" : "BLOCK",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.output.compression.type",
+        "property_description" : "\n      The logging level for the map task. The allowed levels are:\n      OFF, FATAL, ERROR, WARN, INFO, DEBUG, TRACE and ALL.\n    ",
+        "property_name" : "mapreduce.map.log.level",
+        "property_value" : "INFO",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobtracker.staging.root.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.map.memory.mb",
       "StackConfigurations" : {
-        "property_description" : "The Path prefix for where the staging directories should be placed. The next level is always the user's\n   name. It is a path in the default file system.",
-        "property_value" : "/user",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.jobtracker.staging.root.dir",
+        "property_description" : "Virtual memory for single Map task",
+        "property_name" : "mapreduce.map.memory.mb",
+        "property_value" : "1024",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.queue.default.acl-administer-jobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.map.output.compress",
       "StackConfigurations" : {
         "property_description" : null,
-        "property_value" : "*",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.queue.default.acl-administer-jobs",
+        "property_name" : "mapreduce.map.output.compress",
+        "property_value" : "false",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-queue-acls.xml"
-      }
-    },
-    {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.child.root.logger",
-      "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "INFO,TLA",
         "stack_version" : "2.0.1",
-        "property_name" : "mapred.child.root.logger",
-        "service_name" : "MAPREDUCE2",
-        "stack_name" : "HDP",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobhistory.done-dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.map.sort.spill.percent",
       "StackConfigurations" : {
-        "property_description" : "Directory where history files are managed by the MR JobHistory Server.",
-        "property_value" : "/mr-history/done",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.jobhistory.done-dir",
+        "property_description" : "\n      The soft limit in the serialization buffer. Once reached, a thread will\n      begin to spill the contents to disk in the background. Note that\n      collection will not block if this threshold is exceeded while a spill\n      is already in progress, so spills may be larger than this threshold when\n      it is set to less than .5\n    ",
+        "property_name" : "mapreduce.map.sort.spill.percent",
+        "property_value" : "0.7",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.map.tasks.speculative.execution",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.map.speculative",
       "StackConfigurations" : {
-        "property_description" : "If true, then multiple instances of some map tasks\n               may be executed in parallel.",
+        "property_description" : "\n      If true, then multiple instances of some map tasks\n      may be executed in parallel.\n    ",
+        "property_name" : "mapreduce.map.speculative",
         "property_value" : "false",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.map.tasks.speculative.execution",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-site.xml"
-      }
-    },
-    {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.task.timeout",
-      "StackConfigurations" : {
-        "property_description" : "The number of milliseconds before a task will be\n  terminated if it neither reads an input, writes an output, nor\n  updates its status string.\n  ",
-        "property_value" : "600000",
         "stack_version" : "2.0.1",
-        "property_name" : "mapred.task.timeout",
-        "service_name" : "MAPREDUCE2",
-        "stack_name" : "HDP",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.queue.names",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.output.fileoutputformat.compress",
       "StackConfigurations" : {
-        "property_description" : " Comma separated list of queues configured for this jobtracker.",
-        "property_value" : "default",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.queue.names",
+        "property_description" : "\n      Should the job outputs be compressed?\n    ",
+        "property_name" : "mapreduce.output.fileoutputformat.compress",
+        "property_value" : "false",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-site.xml"
-      }
-    },
-    {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.tasktracker.group",
-      "StackConfigurations" : {
-        "property_description" : "The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.",
-        "property_value" : "hadoop",
         "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.tasktracker.group",
-        "service_name" : "MAPREDUCE2",
-        "stack_name" : "HDP",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.shuffle.merge.percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.output.fileoutputformat.compress.type",
       "StackConfigurations" : {
-        "property_description" : "The usage threshold at which an in-memory merge will be\n  initiated, expressed as a percentage of the total memory allocated to\n  storing in-memory map outputs, as defined by\n  mapred.job.shuffle.input.buffer.percent.\n  ",
-        "property_value" : "0.66",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.shuffle.merge.percent",
+        "property_description" : "\n      If the job outputs are to compressed as SequenceFiles, how should\n      they be compressed? Should be one of NONE, RECORD or BLOCK.\n    ",
+        "property_name" : "mapreduce.output.fileoutputformat.compress.type",
+        "property_value" : "BLOCK",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.jobtracker.retirejob.check",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.input.buffer.percent",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "10000",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.jobtracker.retirejob.check",
+        "property_description" : "\n      The percentage of memory- relative to the maximum heap size- to\n      retain map outputs during the reduce. When the shuffle is concluded, any\n      remaining map outputs in memory must consume less than this threshold before\n      the reduce can begin.\n    ",
+        "property_name" : "mapreduce.reduce.input.buffer.percent",
+        "property_value" : "0.0",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.jobtracker.retirejob.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.java.opts",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "21600000",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.jobtracker.retirejob.interval",
+        "property_description" : "\n      Larger heap-size for child jvms of reduces.\n    ",
+        "property_name" : "mapreduce.reduce.java.opts",
+        "property_value" : "-Xmx756m",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.jobtracker.completeuserjobs.maximum",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.log.level",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "5",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.jobtracker.completeuserjobs.maximum",
+        "property_description" : "\n      The logging level for the reduce task. The allowed levels are:\n      OFF, FATAL, ERROR, WARN, INFO, DEBUG, TRACE and ALL.\n    ",
+        "property_name" : "mapreduce.reduce.log.level",
+        "property_value" : "INFO",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.reduce.input.buffer.percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.memory.mb",
       "StackConfigurations" : {
-        "property_description" : "The percentage of memory- relative to the maximum heap size- to\n  retain map outputs during the reduce. When the shuffle is concluded, any\n  remaining map outputs in memory must consume less than this threshold before\n  the reduce can begin.\n  ",
-        "property_value" : "0.0",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.reduce.input.buffer.percent",
+        "property_description" : "Virtual memory for single Reduce task",
+        "property_name" : "mapreduce.reduce.memory.mb",
+        "property_value" : "1024",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.jobhistory.intermediate-done-dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.shuffle.input.buffer.percent",
       "StackConfigurations" : {
-        "property_description" : "Directory where history files are written by MapReduce jobs.",
-        "property_value" : "/mr-history/tmp",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapreduce.jobhistory.intermediate-done-dir",
+        "property_description" : "\n      The percentage of memory to be allocated from the maximum heap\n      size to storing map outputs during the shuffle.\n    ",
+        "property_name" : "mapreduce.reduce.shuffle.input.buffer.percent",
+        "property_value" : "0.7",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.inmem.merge.threshold",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.shuffle.merge.percent",
       "StackConfigurations" : {
-        "property_description" : "The threshold, in terms of the number of files\n  for the in-memory merge process. When we accumulate threshold number of files\n  we initiate the in-memory merge and spill to disk. A value of 0 or less than\n  0 indicates we want to DON'T have any threshold and instead depend only on\n  the ramfs's memory consumption to trigger the merge.\n  ",
-        "property_value" : "1000",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.inmem.merge.threshold",
+        "property_description" : "\n      The usage threshold at which an in-memory merge will be\n      initiated, expressed as a percentage of the total memory allocated to\n      storing in-memory map outputs, as defined by\n      mapreduce.reduce.shuffle.input.buffer.percent.\n    ",
+        "property_name" : "mapreduce.reduce.shuffle.merge.percent",
+        "property_value" : "0.66",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.reuse.jvm.num.tasks",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.shuffle.parallelcopies",
       "StackConfigurations" : {
-        "property_description" : "\n    How many tasks to run per jvm. If set to -1, there is no limit\n  ",
-        "property_value" : "1",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.reuse.jvm.num.tasks",
+        "property_description" : "\n      The default number of parallel transfers run by reduce during\n      the copy(shuffle) phase.\n    ",
+        "property_name" : "mapreduce.reduce.shuffle.parallelcopies",
+        "property_value" : "30",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.tracker.history.completed.location",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.reduce.speculative",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "/mapred/history/done",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.tracker.history.completed.location",
+        "property_description" : "\n      If true, then multiple instances of some reduce tasks may be\n      executed in parallel.\n    ",
+        "property_name" : "mapreduce.reduce.speculative",
+        "property_value" : "false",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/jetty.connector",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.shuffle.port",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "org.mortbay.jetty.nio.SelectChannelConnector",
-        "stack_version" : "2.0.1",
-        "property_name" : "jetty.connector",
+        "property_description" : "\n      Default port that the ShuffleHandler will run on.\n      ShuffleHandler is a service run at the NodeManager to facilitate\n      transfers of intermediate Map outputs to requesting Reducers.\n    ",
+        "property_name" : "mapreduce.shuffle.port",
+        "property_value" : "13562",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.reduce.slowstart.completed.maps",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.task.io.sort.factor",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "0.05",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.reduce.slowstart.completed.maps",
+        "property_description" : "\n      The number of streams to merge at once while sorting files.\n      This determines the number of open file handles.\n    ",
+        "property_name" : "mapreduce.task.io.sort.factor",
+        "property_value" : "100",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/tasktracker.http.threads",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.task.io.sort.mb",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "50",
-        "stack_version" : "2.0.1",
-        "property_name" : "tasktracker.http.threads",
+        "property_description" : "\n      The total amount of buffer memory to use while sorting files, in megabytes.\n      By default, gives each merge stream 1MB, which should minimize seeks.\n    ",
+        "property_name" : "mapreduce.task.io.sort.mb",
+        "property_value" : "200",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/io.sort.factor",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapreduce.task.timeout",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "100",
-        "stack_version" : "2.0.1",
-        "property_name" : "io.sort.factor",
+        "property_description" : "\n      The number of milliseconds before a task will be\n      terminated if it neither reads an input, writes an output, nor\n      updates its status string.\n    ",
+        "property_name" : "mapreduce.task.timeout",
+        "property_value" : "300000",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.job.tracker.persist.jobstatus.hours",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/yarn.app.mapreduce.am.admin-command-opts",
       "StackConfigurations" : {
-        "property_description" : "The number of hours job status information is persisted in DFS.\n    The job status information will be available after it drops of the memory\n    queue and between jobtracker restarts. With a zero value the job status\n    information is not persisted at all in DFS.\n  ",
-        "property_value" : "1",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.job.tracker.persist.jobstatus.hours",
+        "property_description" : "\n      Java opts for the MR App Master processes for admin purposes.\n      It will appears before the opts set by yarn.app.mapreduce.am.command-opts and\n      thus its options can be overridden user.\n\n      Usage of -Djava.library.path can cause programs to no longer function if\n      hadoop native libraries are used. These values should instead be set as part\n      of LD_LIBRARY_PATH in the map / reduce JVM env using the mapreduce.map.env and\n      mapreduce.reduce.env config settings.\n    ",
+        "property_name" : "yarn.app.mapreduce.am.admin-command-opts",
+        "property_value" : "-Djava.net.preferIPv4Stack=true -Dhadoop.metrics.log.level=WARN",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.jobtracker.restart.recover",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/yarn.app.mapreduce.am.command-opts",
       "StackConfigurations" : {
-        "property_description" : "\"true\" to enable (job) recovery upon restart,\n               \"false\" to start afresh\n    ",
-        "property_value" : "false",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.jobtracker.restart.recover",
+        "property_description" : "\n      Java opts for the MR App Master processes.\n      The following symbol, if present, will be interpolated: @taskid@ is replaced\n      by current TaskID. Any other occurrences of '@' will go unchanged.\n      For example, to enable verbose gc logging to a file named for the taskid in\n      /tmp and to set the heap maximum to be a gigabyte, pass a 'value' of:\n      -Xmx1024m -verbose:gc -Xloggc:/tmp/@taskid@.gc\n\n      Usage of -Djava.library.path can cause programs to no longer function if\n      hadoop native libraries are used. These values should instead be set as part\n      of LD_LIBRARY_PATH in the map / reduce JVM env using the mapreduce.map.env and\n      mapreduce.reduce.env config settings.\n    ",
+        "property_name" : "yarn.app.mapreduce.am.command-opts",
+        "property_value" : "-Xmx312m",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.max.tracker.blacklists",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/yarn.app.mapreduce.am.log.level",
       "StackConfigurations" : {
-        "property_description" : "\n    if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted\n  ",
-        "property_value" : "16",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.max.tracker.blacklists",
+        "property_description" : "MR App Master process log level.",
+        "property_name" : "yarn.app.mapreduce.am.log.level",
+        "property_value" : "INFO",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.healthChecker.script.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/yarn.app.mapreduce.am.resource.mb",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "60000",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.healthChecker.script.timeout",
+        "property_description" : "The amount of memory the MR AppMaster needs.",
+        "property_name" : "yarn.app.mapreduce.am.resource.mb",
+        "property_value" : "512",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/mapred.queue.default.acl-submit-job",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/MAPREDUCE2/configurations/yarn.app.mapreduce.am.staging-dir",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "*",
-        "stack_version" : "2.0.1",
-        "property_name" : "mapred.queue.default.acl-submit-job",
+        "property_description" : "\n      The staging dir used while submitting jobs.\n    ",
+        "property_name" : "yarn.app.mapreduce.am.staging-dir",
+        "property_value" : "/user",
         "service_name" : "MAPREDUCE2",
         "stack_name" : "HDP",
-        "type" : "mapred-queue-acls.xml"
+        "stack_version" : "2.0.1",
+        "type" : "mapred-site.xml"
       }
     }
   ]
-}
+}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/NAGIOS.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/NAGIOS.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/NAGIOS.json
index 58dbfd3..190e677 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/NAGIOS.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/NAGIOS.json
@@ -1,41 +1,41 @@
 {
-  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations?fields=*&_=1368459065260",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/NAGIOS/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations/nagios_group",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/NAGIOS/configurations/nagios_group",
       "StackConfigurations" : {
-        "property_description" : "Nagios Group.",
-        "property_value" : "nagios",
-        "stack_version" : "1.3.0",
+        "property_description" : "",
         "property_name" : "nagios_group",
+        "property_value" : "nagios",
         "service_name" : "NAGIOS",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations/nagios_web_login",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/NAGIOS/configurations/nagios_user",
       "StackConfigurations" : {
-        "property_description" : "Nagios web user.",
-        "property_value" : "nagiosadmin",
-        "stack_version" : "1.3.0",
-        "property_name" : "nagios_web_login",
+        "property_description" : "Nagios process user.",
+        "property_name" : "nagios_user",
+        "property_value" : "nagios",
         "service_name" : "NAGIOS",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS/configurations/nagios_user",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/NAGIOS/configurations/nagios_web_login",
       "StackConfigurations" : {
-        "property_description" : "Nagios Username.",
-        "property_value" : "nagios",
-        "stack_version" : "1.3.0",
-        "property_name" : "nagios_user",
+        "property_description" : "Web user name.",
+        "property_name" : "nagios_web_login",
+        "property_value" : "nagiosadmin",
         "service_name" : "NAGIOS",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     }
   ]
-}
+}
\ No newline at end of file


[03/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/OOZIE.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/OOZIE.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/OOZIE.json
index d32a7b9..e1c5b18 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/OOZIE.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/OOZIE.json
@@ -1,316 +1,376 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.systemmode",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.authentication.kerberos.name.rules",
       "StackConfigurations" : {
-        "property_description" : "\n     System mode for  Oozie at startup.\n     ",
-        "property_value" : "NORMAL",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.systemmode",
-        "filename" : "oozie-site.xml",
+        "property_description" : "The mapping from kerberos principal names to local OS user names.",
+        "property_name" : "oozie.authentication.kerberos.name.rules",
+        "property_value" : "\n        RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/\n        RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/\n        RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        DEFAULT\n        ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.kerberos.name.rules",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.authentication.type",
       "StackConfigurations" : {
-        "property_description" : "The mapping from kerberos principal names to local OS user names.",
-        "property_value" : "\n        RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/\n        RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/\n        RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        DEFAULT\n        ",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.authentication.kerberos.name.rules",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n      ",
+        "property_name" : "oozie.authentication.type",
+        "property_value" : "simple",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.AuthorizationService.security.enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.base.url",
       "StackConfigurations" : {
-        "property_description" : "\n     Specifies whether security (user name/admin role) is enabled or not.\n     If disabled any user can manage Oozie system and manage any job.\n     ",
-        "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.AuthorizationService.security.enabled",
-        "filename" : "oozie-site.xml",
+        "property_description" : "Base Oozie URL.",
+        "property_name" : "oozie.base.url",
+        "property_value" : "http://localhost:11000/oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.threads",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.credentials.credentialclasses",
       "StackConfigurations" : {
-        "property_description" : "Number of threads used for executing callables",
-        "property_value" : "10",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.CallableQueueService.threads",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n        Credential Class to be used for HCat.\n      ",
+        "property_name" : "oozie.credentials.credentialclasses",
+        "property_value" : "hcat=org.apache.oozie.action.hadoop.HCatCredentials",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.jobTracker.whitelist",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.db.schema.name",
       "StackConfigurations" : {
-        "property_description" : "\n      Whitelisted job tracker for Oozie service.\n      ",
-        "property_value" : " ",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.HadoopAccessorService.jobTracker.whitelist",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n      Oozie DataBase Name\n     ",
+        "property_name" : "oozie.db.schema.name",
+        "property_value" : "oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.pool.max.active.conn",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.ActionService.executor.ext.classes",
       "StackConfigurations" : {
-        "property_description" : "\n             Max number of connections.\n        ",
-        "property_value" : "10",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.pool.max.active.conn",
-        "filename" : "oozie-site.xml",
+        "property_description" : null,
+        "property_name" : "oozie.service.ActionService.executor.ext.classes",
+        "property_value" : "\n            org.apache.oozie.action.email.EmailActionExecutor,\n            org.apache.oozie.action.hadoop.HiveActionExecutor,\n            org.apache.oozie.action.hadoop.ShellActionExecutor,\n            org.apache.oozie.action.hadoop.SqoopActionExecutor,\n            org.apache.oozie.action.hadoop.DistcpActionExecutor\n        ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.base.url",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.AuthorizationService.security.enabled",
       "StackConfigurations" : {
-        "property_description" : "Base Oozie URL.",
-        "property_value" : "http://localhost:11000/oozie",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.base.url",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n     Specifies whether security (user name/admin role) is enabled or not.\n     If disabled any user can manage Oozie system and manage any job.\n     ",
+        "property_name" : "oozie.service.AuthorizationService.security.enabled",
+        "property_value" : "true",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.SchemaService.wf.ext.schemas",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.callable.concurrency",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.SchemaService.wf.ext.schemas",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
+        "property_name" : "oozie.service.CallableQueueService.callable.concurrency",
+        "property_value" : "3",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.older.than",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.queue.size",
       "StackConfigurations" : {
-        "property_description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
-        "property_value" : "30",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.PurgeService.older.than",
-        "filename" : "oozie-site.xml",
+        "property_description" : "Max callable queue size",
+        "property_name" : "oozie.service.CallableQueueService.queue.size",
+        "property_value" : "1000",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.nameNode.whitelist",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.threads",
       "StackConfigurations" : {
-        "property_description" : "\n      ",
-        "property_value" : " ",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
-        "filename" : "oozie-site.xml",
+        "property_description" : "Number of threads used for executing callables",
+        "property_name" : "oozie.service.CallableQueueService.threads",
+        "property_value" : "10",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.coord.normal.default.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.hadoop.configurations",
       "StackConfigurations" : {
-        "property_description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
-        "property_value" : "120",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.coord.normal.default.timeout",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
+        "property_name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
+        "property_value" : "*=/etc/hadoop/conf",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/use.system.libpath.for.mapreduce.and.pig.jobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.jobTracker.whitelist",
       "StackConfigurations" : {
-        "property_description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
-        "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n      Whitelisted job tracker for Oozie service.\n      ",
+        "property_name" : "oozie.service.HadoopAccessorService.jobTracker.whitelist",
+        "property_value" : " ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.queue.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.nameNode.whitelist",
       "StackConfigurations" : {
-        "property_description" : "Max callable queue size",
-        "property_value" : "1000",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.CallableQueueService.queue.size",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n      ",
+        "property_name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
+        "property_value" : " ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.ActionService.executor.ext.classes",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.JPAService.create.db.schema",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "\n            org.apache.oozie.action.email.EmailActionExecutor,\n            org.apache.oozie.action.hadoop.HiveActionExecutor,\n            org.apache.oozie.action.hadoop.ShellActionExecutor,\n            org.apache.oozie.action.hadoop.SqoopActionExecutor,\n            org.apache.oozie.action.hadoop.DistcpActionExecutor\n        ",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.ActionService.executor.ext.classes",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
+        "property_name" : "oozie.service.JPAService.create.db.schema",
+        "property_value" : "false",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.driver",
+      "StackConfigurations" : {
+        "property_description" : "\n            JDBC driver class.\n        ",
+        "property_name" : "oozie.service.JPAService.jdbc.driver",
+        "property_value" : "org.apache.derby.jdbc.EmbeddedDriver",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.password",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.password",
       "StackConfigurations" : {
         "property_description" : "\n            DB user password.\n\n            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,\n                       if empty Configuration assumes it is NULL.\n        ",
-        "property_value" : " ",
-        "stack_version" : "1.3.0",
         "property_name" : "oozie.service.JPAService.jdbc.password",
-        "filename" : "oozie-site.xml",
+        "property_value" : " ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.WorkflowAppService.system.libpath",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.url",
       "StackConfigurations" : {
-        "property_description" : "\n      System library path to use for workflow applications.\n      This path is added to workflow application if their job properties sets\n      the property 'oozie.use.system.libpath' to true.\n      ",
-        "property_value" : "/user/${user.name}/share/lib",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.WorkflowAppService.system.libpath",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n            JDBC URL.\n        ",
+        "property_name" : "oozie.service.JPAService.jdbc.url",
+        "property_value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.driver",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.username",
       "StackConfigurations" : {
-        "property_description" : "\n            JDBC driver class.\n        ",
-        "property_value" : "org.apache.derby.jdbc.EmbeddedDriver",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.jdbc.driver",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n          Database user name to use to connect to the database\n        ",
+        "property_name" : "oozie.service.JPAService.jdbc.username",
+        "property_value" : "oozie",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.hadoop.configurations",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.JPAService.pool.max.active.conn",
       "StackConfigurations" : {
-        "property_description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
-        "property_value" : "*=/etc/hadoop/conf",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n             Max number of connections.\n        ",
+        "property_name" : "oozie.service.JPAService.pool.max.active.conn",
+        "property_value" : "10",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.purge.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.PurgeService.older.than",
+      "StackConfigurations" : {
+        "property_description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
+        "property_name" : "oozie.service.PurgeService.older.than",
+        "property_value" : "30",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.PurgeService.purge.interval",
       "StackConfigurations" : {
         "property_description" : "\n     Interval at which the purge service will run, in seconds.\n     ",
-        "property_value" : "3600",
-        "stack_version" : "1.3.0",
         "property_name" : "oozie.service.PurgeService.purge.interval",
-        "filename" : "oozie-site.xml",
+        "property_value" : "3600",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.create.db.schema",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.SchemaService.wf.ext.schemas",
       "StackConfigurations" : {
-        "property_description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
-        "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.create.db.schema",
-        "filename" : "oozie-site.xml",
+        "property_description" : null,
+        "property_name" : "oozie.service.SchemaService.wf.ext.schemas",
+        "property_value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd,shell-action-0.2.xsd,oozie-sla-0.1.xsd,oozie-sla-0.2.xsd",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.system.id",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.URIHandlerService.uri.handlers",
       "StackConfigurations" : {
-        "property_description" : "\n    The Oozie system ID.\n    ",
-        "property_value" : "oozie-${user.name}",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.system.id",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n        Enlist the different uri handlers supported for data availability checks.\n      ",
+        "property_name" : "oozie.service.URIHandlerService.uri.handlers",
+        "property_value" : "org.apache.oozie.dependency.FSURIHandler,org.apache.oozie.dependency.HCatURIHandler",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.username",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.WorkflowAppService.system.libpath",
       "StackConfigurations" : {
-        "property_description" : "\n            DB user name.\n        ",
-        "property_value" : "sa",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.jdbc.username",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n      System library path to use for workflow applications.\n      This path is added to workflow application if their job properties sets\n      the property 'oozie.use.system.libpath' to true.\n      ",
+        "property_name" : "oozie.service.WorkflowAppService.system.libpath",
+        "property_value" : "/user/${user.name}/share/lib",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.db.schema.name",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.coord.normal.default.timeout",
       "StackConfigurations" : {
-        "property_description" : "\n      Oozie DataBase Name\n     ",
-        "property_value" : "oozie",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.db.schema.name",
-        "filename" : "oozie-site.xml",
+        "property_description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
+        "property_name" : "oozie.service.coord.normal.default.timeout",
+        "property_value" : "120",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.type",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.service.coord.push.check.requeue.interval",
       "StackConfigurations" : {
-        "property_description" : "\n      ",
-        "property_value" : "simple",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.authentication.type",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n        Command re-queue interval for push dependencies (in millisecond).\n    ",
+        "property_name" : "oozie.service.coord.push.check.requeue.interval",
+        "property_value" : "30000",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.url",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.services",
       "StackConfigurations" : {
-        "property_description" : "\n            JDBC URL.\n        ",
-        "property_value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.JPAService.jdbc.url",
-        "filename" : "oozie-site.xml",
+        "property_description" : "List of Oozie services",
+        "property_name" : "oozie.services",
+        "property_value" : "\n        org.apache.oozie.service.SchedulerService,\n        org.apache.oozie.service.InstrumentationService,\n        org.apache.oozie.service.CallableQueueService,\n        org.apache.oozie.service.UUIDService,\n        org.apache.oozie.service.ELService,\n        org.apache.oozie.service.AuthorizationService,\n        org.apache.oozie.service.UserGroupInformationService,\n        org.apache.oozie.service.HadoopAccessorService,\n        org.apache.oozie.service.URIHandlerService,\n        org.apache.oozie.service.MemoryLocksService,\n        org.apache.oozie.service.DagXLogInfoService,\n        org.apache.oozie.service.SchemaService,\n        org.apache.oozie.service.LiteWorkflowAppService,\n        org.apache.oozie.service.JPAService,\n        org.apache.oozie.service.StoreService,\n        org.apache.oozie.service.CoordinatorStoreService,\n        org.apache.oozie.service.SLAStoreService,\n        org.apache.oozie.service.DBLiteWorkflowStoreService,\
 n        org.apache.oozie.service.CallbackService,\n        org.apache.oozie.service.ActionService,\n        org.apache.oozie.service.ActionCheckerService,\n        org.apache.oozie.service.RecoveryService,\n        org.apache.oozie.service.PurgeService,\n        org.apache.oozie.service.CoordinatorEngineService,\n        org.apache.oozie.service.BundleEngineService,\n        org.apache.oozie.service.DagEngineService,\n        org.apache.oozie.service.CoordMaterializeTriggerService,\n        org.apache.oozie.service.StatusTransitService,\n        org.apache.oozie.service.PauseTransitService,\n        org.apache.oozie.service.GroupsService,\n        org.apache.oozie.service.ProxyUserService\n      ",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.callable.concurrency",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.services.ext",
       "StackConfigurations" : {
-        "property_description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
-        "property_value" : "3",
-        "stack_version" : "1.3.0",
-        "property_name" : "oozie.service.CallableQueueService.callable.concurrency",
-        "filename" : "oozie-site.xml",
+        "property_description" : "\n       To add/replace services defined in 'oozie.services' with custom implementations.\n       Class names must be separated by commas.\n    ",
+        "property_name" : "oozie.services.ext",
+        "property_value" : "org.apache.oozie.service.PartitionDependencyManagerService,org.apache.oozie.service.HCatAccessorService",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.system.id",
+      "StackConfigurations" : {
+        "property_description" : "\n    The Oozie system ID.\n    ",
+        "property_name" : "oozie.system.id",
+        "property_value" : "oozie-${user.name}",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/oozie.systemmode",
+      "StackConfigurations" : {
+        "property_description" : "\n     System mode for  Oozie at startup.\n     ",
+        "property_name" : "oozie.systemmode",
+        "property_value" : "NORMAL",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/OOZIE/configurations/use.system.libpath.for.mapreduce.and.pig.jobs",
+      "StackConfigurations" : {
+        "property_description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
+        "property_name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
+        "property_value" : "false",
         "service_name" : "OOZIE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "oozie-site.xml"
       }
     }
   ]

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/WEBHCAT.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/WEBHCAT.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/WEBHCAT.json
index 2bf81aa..f73e1ef 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/WEBHCAT.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/WEBHCAT.json
@@ -1,172 +1,232 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.streaming.jar",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/hcat_log_dir",
       "StackConfigurations" : {
-        "property_description" : "The hdfs path to the Hadoop streaming jar file.",
-        "property_value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.streaming.jar",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "",
+        "property_name" : "hcat_log_dir",
+        "property_value" : "/var/log/webhcat",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.override.enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/hcat_pid_dir",
       "StackConfigurations" : {
-        "property_description" : "\n     Enable the override path in templeton.override.jars\n   ",
-        "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.override.enabled",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "",
+        "property_name" : "hcat_pid_dir",
+        "property_value" : "/var/run/webhcat",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.archive",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/hcat_user",
       "StackConfigurations" : {
-        "property_description" : "The path to the Pig archive.",
-        "property_value" : "hdfs:///apps/webhcat/pig.tar.gz",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.pig.archive",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "",
+        "property_name" : "hcat_user",
+        "property_value" : "hcat",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.exec.timeout",
+      "StackConfigurations" : {
+        "property_description" : "Time out for templeton api",
+        "property_name" : "templeton.exec.timeout",
+        "property_value" : "60000",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.hadoop",
       "StackConfigurations" : {
         "property_description" : "The path to the Hadoop executable.",
-        "property_value" : "/usr/bin/hadoop",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.hadoop",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "/usr/bin/hadoop",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.hadoop.conf.dir",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Hadoop configuration.",
+        "property_name" : "templeton.hadoop.conf.dir",
+        "property_value" : "/etc/hadoop/conf",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hcat",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.hcat",
       "StackConfigurations" : {
         "property_description" : "The path to the hcatalog executable.",
-        "property_value" : "/usr/bin/hcat",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.hcat",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "/usr/bin/hcat",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.exec.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.hive.archive",
       "StackConfigurations" : {
-        "property_description" : "Time out for templeton api",
-        "property_value" : "60000",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.exec.timeout",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Hive archive.",
+        "property_name" : "templeton.hive.archive",
+        "property_value" : "hdfs:///apps/webhcat/hive.tar.gz",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop.conf.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.hive.path",
       "StackConfigurations" : {
-        "property_description" : "The path to the Hadoop configuration.",
-        "property_value" : "/etc/hadoop/conf",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.hadoop.conf.dir",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The path to the Hive executable.",
+        "property_name" : "templeton.hive.path",
+        "property_value" : "hive.tar.gz/hive/bin/hive",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.jar",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.jar",
       "StackConfigurations" : {
         "property_description" : "The path to the Templeton jar file.",
-        "property_value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.jar",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.libjars",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.libjars",
       "StackConfigurations" : {
         "property_description" : "Jars to add the the classpath.",
-        "property_value" : "/usr/lib/zookeeper/zookeeper.jar",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.libjars",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "/usr/lib/zookeeper/zookeeper.jar",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.override.enabled",
+      "StackConfigurations" : {
+        "property_description" : "\n     Enable the override path in templeton.override.jars\n   ",
+        "property_name" : "templeton.override.enabled",
+        "property_value" : "false",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.pig.archive",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Pig archive.",
+        "property_name" : "templeton.pig.archive",
+        "property_value" : "hdfs:///apps/webhcat/pig.tar.gz",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.pig.path",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Pig executable.",
+        "property_name" : "templeton.pig.path",
+        "property_value" : "pig.tar.gz/pig/bin/pig",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.port",
       "StackConfigurations" : {
         "property_description" : "The HTTP port for the main server.",
-        "property_value" : "50111",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.port",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "50111",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.storage.class",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.storage.class",
       "StackConfigurations" : {
         "property_description" : "The class to use as storage",
-        "property_value" : "org.apache.hcatalog.templeton.tool.ZooKeeperStorage",
-        "stack_version" : "1.3.0",
         "property_name" : "templeton.storage.class",
-        "filename" : "webhcat-site.xml",
+        "property_value" : "org.apache.hive.hcatalog.templeton.tool.ZooKeeperStorage",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.archive",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.streaming.jar",
       "StackConfigurations" : {
-        "property_description" : "The path to the Hive archive.",
-        "property_value" : "hdfs:///apps/webhcat/hive.tar.gz",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.hive.archive",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "The hdfs path to the Hadoop streaming jar file.",
+        "property_name" : "templeton.streaming.jar",
+        "property_value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/templeton.zookeeper.hosts",
       "StackConfigurations" : {
-        "property_description" : "The path to the Hive executable.",
-        "property_value" : "hive.tar.gz/hive/bin/hive",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.hive.path",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "ZooKeeper servers, as comma separated host:port pairs",
+        "property_name" : "templeton.zookeeper.hosts",
+        "property_value" : "localhost:2181",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "webhcat-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/WEBHCAT/configurations/webhcat_user",
       "StackConfigurations" : {
-        "property_description" : "The path to the Pig executable.",
-        "property_value" : "pig.tar.gz/pig/bin/pig",
-        "stack_version" : "1.3.0",
-        "property_name" : "templeton.pig.path",
-        "filename" : "webhcat-site.xml",
+        "property_description" : "",
+        "property_name" : "webhcat_user",
+        "property_value" : "hcat",
         "service_name" : "WEBHCAT",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     }
   ]


[11/11] git commit: AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
AMBARI-4310 Update mock data of services configurations. (atkach)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/8c0e3771
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/8c0e3771
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/8c0e3771

Branch: refs/heads/trunk
Commit: 8c0e377195f2c58727011d299a0da2b093508a72
Parents: 5f48900
Author: atkach <at...@hortonworks.com>
Authored: Thu Jan 16 14:03:18 2014 +0200
Committer: atkach <at...@hortonworks.com>
Committed: Thu Jan 16 15:56:53 2014 +0200

----------------------------------------------------------------------
 .../wizard/stack/hdp/version1.3.0/GANGLIA.json  |   64 +-
 .../wizard/stack/hdp/version1.3.0/HBASE.json    |  626 ++++++--
 .../wizard/stack/hdp/version1.3.0/HDFS.json     | 1210 ++++++++++-----
 .../wizard/stack/hdp/version1.3.0/HIVE.json     |  526 +++++--
 .../stack/hdp/version1.3.0/MAPREDUCE.json       | 1398 ++++++++++++------
 .../wizard/stack/hdp/version1.3.0/NAGIOS.json   |   40 +-
 .../wizard/stack/hdp/version1.3.0/OOZIE.json    |  452 +++---
 .../wizard/stack/hdp/version1.3.0/WEBHCAT.json  |  202 +--
 .../stack/hdp/version1.3.0/ZOOKEEPER.json       |  113 +-
 .../wizard/stack/hdp/version2.0.1/GANGLIA.json  |   72 +-
 .../wizard/stack/hdp/version2.0.1/HBASE.json    |  656 ++++++--
 .../wizard/stack/hdp/version2.0.1/HDFS.json     |  426 +++---
 .../wizard/stack/hdp/version2.0.1/HIVE.json     |  436 ++++--
 .../stack/hdp/version2.0.1/MAPREDUCE2.json      |  458 +++---
 .../wizard/stack/hdp/version2.0.1/NAGIOS.json   |   44 +-
 .../wizard/stack/hdp/version2.0.1/OOZIE.json    |  418 +++---
 .../wizard/stack/hdp/version2.0.1/WEBHCAT.json  |  234 +--
 .../wizard/stack/hdp/version2.0.1/YARN.json     |  564 ++++---
 .../stack/hdp/version2.0.1/ZOOKEEPER.json       |  113 +-
 19 files changed, 5455 insertions(+), 2597 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/GANGLIA.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/GANGLIA.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/GANGLIA.json
index 97b2c77..fb98dd2 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/GANGLIA.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/GANGLIA.json
@@ -1,65 +1,65 @@
 {
-  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations?fields=*&_=1368459065278",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations?fields=*",
   "items" : [
     {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/ganglia_conf_dir",
+      "StackConfigurations" : {
+        "property_description" : "Config directory for Ganglia",
+        "property_name" : "ganglia_conf_dir",
+        "property_value" : "/etc/ganglia/hdp",
+        "service_name" : "GANGLIA",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/ganglia_runtime_dir",
       "StackConfigurations" : {
         "property_description" : "Run directories for Ganglia",
-        "property_value" : "/var/run/ganglia/hdp",
-        "stack_version" : "1.3.0",
         "property_name" : "ganglia_runtime_dir",
+        "property_value" : "/var/run/ganglia/hdp",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/gmetad_user",
       "StackConfigurations" : {
         "property_description" : "User ",
-        "property_value" : "nobody",
-        "stack_version" : "1.3.0",
         "property_name" : "gmetad_user",
+        "property_value" : "nobody",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
-      }
-    },
-    {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/rrdcached_base_dir",
-      "StackConfigurations" : {
-        "property_description" : "Default directory for saving the rrd files on ganglia server",
-        "property_value" : "/var/lib/ganglia/rrds",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "rrdcached_base_dir",
-        "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "type" : "global.xml"
       }
     },
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/gmond_user",
       "StackConfigurations" : {
         "property_description" : "User ",
-        "property_value" : "nobody",
-        "stack_version" : "1.3.0",
         "property_name" : "gmond_user",
+        "property_value" : "nobody",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/ganglia_conf_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA/configurations/rrdcached_base_dir",
       "StackConfigurations" : {
-        "property_description" : "Config directory for Ganglia",
-        "property_value" : "/etc/ganglia/hdp",
-        "stack_version" : "1.3.0",
-        "property_name" : "ganglia_conf_dir",
+        "property_description" : "Default directory for saving the rrd files on ganglia server",
+        "property_name" : "rrdcached_base_dir",
+        "property_value" : "/var/lib/ganglia/rrds",
         "service_name" : "GANGLIA",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     }
   ]
-}
+}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HBASE.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HBASE.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HBASE.json
index ff0b600..c58ce95 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HBASE.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HBASE.json
@@ -1,292 +1,640 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/zookeeper_sessiontimeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/client_scannercaching",
       "StackConfigurations" : {
-        "property_description" : "ZooKeeper Session Timeout",
-        "property_value" : "60000",
-        "stack_version" : "1.3.0",
-        "property_name" : "zookeeper_sessiontimeout",
+        "property_description" : "Base Client Scanner Caching",
+        "property_name" : "client_scannercaching",
+        "property_value" : "100",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_master_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/dfs.client.read.shortcircuit",
       "StackConfigurations" : {
-        "property_description" : "HBase Master Heap Size",
-        "property_value" : "1024",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase_master_heapsize",
+        "property_description" : "Enable/Disable short circuit read for your client.\n    Hadoop servers should be configured to allow short circuit read\n    for the hbase user for this to take effect\n    ",
+        "property_name" : "dfs.client.read.shortcircuit",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hstore_compactionthreshold",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/dfs.support.append",
       "StackConfigurations" : {
-        "property_description" : "HBase HStore compaction threshold.",
-        "property_value" : "3",
+        "property_description" : "Does HDFS allow appends to files?\n    This is an hdfs config. set in here so the hdfs client will do append support.\n    You must ensure that this config. is true serverside too when running hbase\n    (You will have to restart your cluster after setting it).\n    ",
+        "property_name" : "dfs.support.append",
+        "property_value" : "true",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hstore_compactionthreshold",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.client.keyvalue.maxsize",
+      "StackConfigurations" : {
+        "property_description" : "Specifies the combined maximum allowed size of a KeyValue\n    instance. This is to set an upper boundary for a single entry saved in a\n    storage file. Since they cannot be split it helps avoiding that a region\n    cannot be split any further because the data is too large. It seems wise\n    to set this to a fraction of the maximum region size. Setting it to zero\n    or less disables the check.\n    ",
+        "property_name" : "hbase.client.keyvalue.maxsize",
+        "property_value" : "10485760",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hfile_blockcache_size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.client.scanner.caching",
       "StackConfigurations" : {
-        "property_description" : "HFile block cache size.",
-        "property_value" : "0.25",
+        "property_description" : "Number of rows that will be fetched when calling next\n    on a scanner if it is not served from (local, client) memory. Higher\n    caching values will enable faster scanners but will eat up more memory\n    and some calls of next may take longer and longer times when the cache is empty.\n    Do not set this value such that the time between invocations is greater\n    than the scanner timeout; i.e. hbase.regionserver.lease.period\n    ",
+        "property_name" : "hbase.client.scanner.caching",
+        "property_value" : "100",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hfile_blockcache_size",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.cluster.distributed",
+      "StackConfigurations" : {
+        "property_description" : "The mode the cluster will be in. Possible values are\n      false for standalone mode and true for distributed mode.  If\n      false, startup will run all HBase and ZooKeeper daemons together\n      in the one JVM.\n    ",
+        "property_name" : "hbase.cluster.distributed",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.client.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.hregion.majorcompaction",
       "StackConfigurations" : {
-        "property_description" : "ACL for HRegionInterface protocol implementations (ie. \n    clients talking to HRegionServers)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
+        "property_description" : "The time (in milliseconds) between 'major' compactions of all\n    HStoreFiles in a region.  Default: 1 day.\n    Set to 0 to disable automated major compactions.\n    ",
+        "property_name" : "hbase.hregion.majorcompaction",
+        "property_value" : "86400000",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "security.client.protocol.acl",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.hregion.max.filesize",
+      "StackConfigurations" : {
+        "property_description" : "\n    Maximum HStoreFile size. If any one of a column families' HStoreFiles has\n    grown to exceed this value, the hosting HRegion is split in two.\n    Default: 1G.\n    ",
+        "property_name" : "hbase.hregion.max.filesize",
+        "property_value" : "10737418240",
         "service_name" : "HBASE",
-        "type" : "hbase-policy.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hregion_blockmultiplier",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.hregion.memstore.block.multiplier",
       "StackConfigurations" : {
-        "property_description" : "HBase Region Block Multiplier",
+        "property_description" : "Block updates if memstore has hbase.hregion.memstore.block.multiplier\n    time hbase.hregion.flush.size bytes.  Useful preventing\n    runaway memstore during spikes in update traffic.  Without an\n    upper-bound, memstore fills such that when it flushes the\n    resultant flush files take a long time to compact or split, or\n    worse, we OOME\n    ",
+        "property_name" : "hbase.hregion.memstore.block.multiplier",
         "property_value" : "2",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hregion_blockmultiplier",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.hregion.memstore.flush.size",
+      "StackConfigurations" : {
+        "property_description" : "\n    Memstore will be flushed to disk if size of the memstore\n    exceeds this number of bytes.  Value is checked by a thread that runs\n    every hbase.server.thread.wakefrequency.\n    ",
+        "property_name" : "hbase.hregion.memstore.flush.size",
+        "property_value" : "134217728",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.zookeeper.useMulti",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.hregion.memstore.mslab.enabled",
       "StackConfigurations" : {
-        "property_description" : "Instructs HBase to make use of ZooKeeper's multi-update functionality.\n    This allows certain ZooKeeper operations to complete more quickly and prevents some issues\n    with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).В·\n    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+\n    and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will\n    not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).\n    ",
+        "property_description" : "\n      Enables the MemStore-Local Allocation Buffer,\n      a feature which works to prevent heap fragmentation under\n      heavy write loads. This can reduce the frequency of stop-the-world\n      GC pauses on large heaps.\n    ",
+        "property_name" : "hbase.hregion.memstore.mslab.enabled",
         "property_value" : "true",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase.zookeeper.useMulti",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.hstore.blockingStoreFiles",
+      "StackConfigurations" : {
+        "property_description" : "\n    If more than this number of StoreFiles in any one Store\n    (one StoreFile is written per flush of MemStore) then updates are\n    blocked for this HRegion until a compaction is completed, or\n    until hbase.hstore.blockingWaitTime has been exceeded.\n    ",
+        "property_name" : "hbase.hstore.blockingStoreFiles",
+        "property_value" : "10",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.superuser",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.hstore.compactionThreshold",
       "StackConfigurations" : {
-        "property_description" : "List of users or groups (comma-separated), who are allowed full privileges, regardless of stored ACLs, across the cluster.Only used when HBase security is enabled.\n",
-        "property_value" : "hbase",
+        "property_description" : "\n    If more than this number of HStoreFiles in any one HStore\n    (one HStoreFile is written per flush of memstore) then a compaction\n    is run to rewrite all HStoreFiles files as one.  Larger numbers\n    put off compaction but when it runs, it takes longer to complete.\n    ",
+        "property_name" : "hbase.hstore.compactionThreshold",
+        "property_value" : "3",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase.superuser",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.regionserver.global.memstore.lowerLimit",
+      "StackConfigurations" : {
+        "property_description" : "When memstores are being forced to flush to make room in\n      memory, keep flushing until we hit this mark. Defaults to 35% of heap.\n      This value equal to hbase.regionserver.global.memstore.upperLimit causes\n      the minimum possible flushing to occur when updates are blocked due to\n      memstore limiting.\n    ",
+        "property_name" : "hbase.regionserver.global.memstore.lowerLimit",
+        "property_value" : "0.38",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_conf_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.regionserver.global.memstore.upperLimit",
       "StackConfigurations" : {
-        "property_description" : "Config Directory for HBase.",
-        "property_value" : "/etc/hbase",
+        "property_description" : "Maximum size of all memstores in a region server before new\n      updates are blocked and flushes are forced. Defaults to 40% of heap\n    ",
+        "property_name" : "hbase.regionserver.global.memstore.upperLimit",
+        "property_value" : "0.4",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase_conf_dir",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.regionserver.handler.count",
+      "StackConfigurations" : {
+        "property_description" : "Count of RPC Listener instances spun up on RegionServers.\n    Same property is used by the Master for count of master handlers.\n    Default is 10.\n    ",
+        "property_name" : "hbase.regionserver.handler.count",
+        "property_value" : "60",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.rootdir",
+      "StackConfigurations" : {
+        "property_description" : "The directory shared by region servers and into\n    which HBase persists.  The URL should be 'fully-qualified'\n    to include the filesystem scheme.  For example, to specify the\n    HDFS directory '/hbase' where the HDFS instance's namenode is\n    running at namenode.example.org on port 9000, set this value to:\n    hdfs://namenode.example.org:9000/hbase.  By default HBase writes\n    into /tmp.  Change this configuration else all data will be lost\n    on machine restart.\n    ",
+        "property_name" : "hbase.rootdir",
+        "property_value" : "hdfs://localhost:8020/apps/hbase/data",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.rpc.engine",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hbase.rpc.engine",
+        "property_value" : "org.apache.hadoop.hbase.ipc.WritableRpcEngine",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.security.authentication",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hbase.security.authentication",
+        "property_value" : "simple",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.security.authorization",
+      "StackConfigurations" : {
+        "property_description" : "Enables HBase authorization. Set the value of this property to false to disable HBase authorization.\n    ",
+        "property_name" : "hbase.security.authorization",
+        "property_value" : "false",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.superuser",
+      "StackConfigurations" : {
+        "property_description" : "List of users or groups (comma-separated), who are allowed\n    full privileges, regardless of stored ACLs, across the cluster.\n    Only used when HBase security is enabled.\n    ",
+        "property_name" : "hbase.superuser",
+        "property_value" : "hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.tmp.dir",
+      "StackConfigurations" : {
+        "property_description" : "Temporary directory on the local filesystem.\n    Change this setting to point to a location more permanent\n    than '/tmp' (The '/tmp' directory is often cleared on\n    machine restart).\n    ",
+        "property_name" : "hbase.tmp.dir",
+        "property_value" : "/hadoop/hbase",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.zookeeper.property.clientPort",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.zookeeper.property.clientPort",
       "StackConfigurations" : {
         "property_description" : "Property from ZooKeeper's config zoo.cfg.\n    The port at which the clients will connect.\n    ",
+        "property_name" : "hbase.zookeeper.property.clientPort",
         "property_value" : "2181",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase.zookeeper.property.clientPort",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.zookeeper.quorum",
+      "StackConfigurations" : {
+        "property_description" : "Comma separated list of servers in the ZooKeeper Quorum.\n    For example, \"host1.mydomain.com,host2.mydomain.com,host3.mydomain.com\".\n    By default this is set to localhost for local and pseudo-distributed modes\n    of operation. For a fully-distributed setup, this should be set to a full\n    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh\n    this is the list of servers which we will start/stop ZooKeeper on.\n    ",
+        "property_name" : "hbase.zookeeper.quorum",
+        "property_value" : "localhost",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_log_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.zookeeper.useMulti",
       "StackConfigurations" : {
-        "property_description" : "Log Directories for HBase.",
-        "property_value" : "/var/log/hbase",
+        "property_description" : "Instructs HBase to make use of ZooKeeper's multi-update functionality.\n    This allows certain ZooKeeper operations to complete more quickly and prevents some issues\n    with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).·\n    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+\n    and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will\n    not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).\n    ",
+        "property_name" : "hbase.zookeeper.useMulti",
+        "property_value" : "true",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase_log_dir",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_conf_dir",
+      "StackConfigurations" : {
+        "property_description" : "Config Directory for HBase.",
+        "property_name" : "hbase_conf_dir",
+        "property_value" : "/etc/hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_tmp_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_conf_dir",
       "StackConfigurations" : {
-        "property_description" : "Hbase temp directory",
-        "property_value" : "/var/log/hbase",
+        "property_description" : "HBase conf dir.",
+        "property_name" : "hbase_conf_dir",
+        "property_value" : "/etc/hbase",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_hdfs_root_dir",
+      "StackConfigurations" : {
+        "property_description" : "HBase Relative Path to HDFS.",
+        "property_name" : "hbase_hdfs_root_dir",
+        "property_value" : "/apps/hbase/data",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase_tmp_dir",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_log_dir",
+      "StackConfigurations" : {
+        "property_description" : "Log Directories for HBase.",
+        "property_name" : "hbase_log_dir",
+        "property_value" : "/var/log/hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hfile_max_keyvalue_size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_master_heapsize",
       "StackConfigurations" : {
-        "property_description" : "HBase Client Maximum key-value Size",
-        "property_value" : "10485760",
+        "property_description" : "HBase Master Heap Size",
+        "property_name" : "hbase_master_heapsize",
+        "property_value" : "1024",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hfile_max_keyvalue_size",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_pid_dir",
+      "StackConfigurations" : {
+        "property_description" : "Log Directories for HBase.",
+        "property_name" : "hbase_pid_dir",
+        "property_value" : "/var/run/hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_regionserver_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_regionserver_heapsize",
       "StackConfigurations" : {
         "property_description" : "Log Directories for HBase.",
+        "property_name" : "hbase_regionserver_heapsize",
         "property_value" : "1024",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase_regionserver_heapsize",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_user",
+      "StackConfigurations" : {
+        "property_description" : "HBase User Name.",
+        "property_name" : "hbase_user",
+        "property_value" : "hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hregion_majorcompaction",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hdfs_enable_shortcircuit_read",
       "StackConfigurations" : {
-        "property_description" : "HBase Major Compaction.",
-        "property_value" : "86400000",
+        "property_description" : "HDFS Short Circuit Read",
+        "property_name" : "hdfs_enable_shortcircuit_read",
+        "property_value" : "true",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hregion_majorcompaction",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hdfs_support_append",
+      "StackConfigurations" : {
+        "property_description" : "HDFS append support",
+        "property_name" : "hdfs_support_append",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/client_scannercaching",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hfile.block.cache.size",
       "StackConfigurations" : {
-        "property_description" : "Base Client Scanner Caching",
-        "property_value" : "100",
+        "property_description" : "\n        Percentage of maximum heap (-Xmx setting) to allocate to block cache\n        used by HFile/StoreFile. Default of 0.25 means allocate 25%.\n        Set to 0 to disable but it's not recommended.\n    ",
+        "property_name" : "hfile.block.cache.size",
+        "property_value" : "0.40",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "client_scannercaching",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hfile_blockcache_size",
+      "StackConfigurations" : {
+        "property_description" : "HFile block cache size.",
+        "property_name" : "hfile_blockcache_size",
+        "property_value" : "0.25",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.masterregion.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hfile_max_keyvalue_size",
       "StackConfigurations" : {
-        "property_description" : "ACL for HMasterRegionInterface protocol implementations\n    (for HRegionServers communicating with HMaster)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
+        "property_description" : "HBase Client Maximum key-value Size",
+        "property_name" : "hfile_max_keyvalue_size",
+        "property_value" : "10485760",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "security.masterregion.protocol.acl",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hregion_blockmultiplier",
+      "StackConfigurations" : {
+        "property_description" : "HBase Region Block Multiplier",
+        "property_name" : "hregion_blockmultiplier",
+        "property_value" : "2",
         "service_name" : "HBASE",
-        "type" : "hbase-policy.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.cluster.distributed",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hregion_majorcompaction",
       "StackConfigurations" : {
-        "property_description" : "The mode the cluster will be in. Possible values are\n      false for standalone mode and true for distributed mode.  If\n      false, startup will run all HBase and ZooKeeper daemons together\n      in the one JVM.\n    ",
-        "property_value" : "true",
+        "property_description" : "HBase Major Compaction.",
+        "property_name" : "hregion_majorcompaction",
+        "property_value" : "86400000",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase.cluster.distributed",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hstore_blockingstorefiles",
+      "StackConfigurations" : {
+        "property_description" : "HStore blocking storefiles.",
+        "property_name" : "hstore_blockingstorefiles",
+        "property_value" : "7",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.master.lease.thread.wakefrequency",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hstore_compactionthreshold",
       "StackConfigurations" : {
-        "property_description" : "The interval between checks for expired region server leases.\n    This value has been reduced due to the other reduced values above so that\n    the master will notice a dead region server sooner. The default is 15 seconds.\n    ",
-        "property_value" : "3000",
+        "property_description" : "HBase HStore compaction threshold.",
+        "property_name" : "hstore_compactionthreshold",
+        "property_value" : "3",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase.master.lease.thread.wakefrequency",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hstorefile_maxsize",
+      "StackConfigurations" : {
+        "property_description" : "Maximum HStoreFile Size",
+        "property_name" : "hstorefile_maxsize",
+        "property_value" : "10737418240",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/regionserver_handlers",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/regionserver_handlers",
       "StackConfigurations" : {
         "property_description" : "HBase RegionServer Handler",
+        "property_name" : "regionserver_handlers",
         "property_value" : "30",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "regionserver_handlers",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/regionserver_memstore_lab",
+      "StackConfigurations" : {
+        "property_description" : "Region Server memstore.",
+        "property_name" : "regionserver_memstore_lab",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.regionserver.optionalcacheflushinterval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/regionserver_memstore_lowerlimit",
       "StackConfigurations" : {
-        "property_description" : "\n      Amount of time to wait since the last time a region was flushed before\n      invoking an optional cache flush. Default 60,000.\n    ",
-        "property_value" : "10000",
+        "property_description" : "Region Server memstore lower limit.",
+        "property_name" : "regionserver_memstore_lowerlimit",
+        "property_value" : "0.35",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase.regionserver.optionalcacheflushinterval",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/regionserver_memstore_upperlimit",
+      "StackConfigurations" : {
+        "property_description" : "Region Server memstore upper limit.",
+        "property_name" : "regionserver_memstore_upperlimit",
+        "property_value" : "0.4",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_pid_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.admin.protocol.acl",
       "StackConfigurations" : {
-        "property_description" : "Log Directories for HBase.",
-        "property_value" : "/var/run/hbase",
+        "property_description" : "ACL for HMasterInterface protocol implementation (ie. \n    clients talking to HMaster for admin operations).\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.admin.protocol.acl",
+        "property_value" : "*",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hbase_pid_dir",
+        "type" : "hbase-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.client.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for HRegionInterface protocol implementations (ie. \n    clients talking to HRegionServers)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.client.protocol.acl",
+        "property_value" : "*",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-policy.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hstorefile_maxsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.masterregion.protocol.acl",
       "StackConfigurations" : {
-        "property_description" : "Maximum HStoreFile Size",
-        "property_value" : "1073741824",
+        "property_description" : "ACL for HMasterRegionInterface protocol implementations\n    (for HRegionServers communicating with HMaster)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.masterregion.protocol.acl",
+        "property_value" : "*",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hstorefile_maxsize",
+        "type" : "hbase-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/zookeeper.session.timeout",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper session timeout.\n      HBase passes this to the zk quorum as suggested maximum time for a\n      session (This setting becomes zookeeper's 'maxSessionTimeout').  See\n      http://hadoop.apache.org/zookeeper/docs/current/zookeeperProgrammers.html#ch_zkSessions\n      \"The client sends a requested timeout, the server responds with the\n      timeout that it can give the client. \" In milliseconds.\n    ",
+        "property_name" : "zookeeper.session.timeout",
+        "property_value" : "60000",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.admin.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/zookeeper.znode.parent",
       "StackConfigurations" : {
-        "property_description" : "ACL for HMasterInterface protocol implementation (ie. \n    clients talking to HMaster for admin operations).\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
+        "property_description" : "Root ZNode for HBase in ZooKeeper. All of HBase's ZooKeeper\n      files that are configured with a relative path will go under this node.\n      By default, all of HBase's ZooKeeper file path are configured with a\n      relative path, so they will all go under this directory unless changed.\n    ",
+        "property_name" : "zookeeper.znode.parent",
+        "property_value" : "/hbase-unsecure",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "security.admin.protocol.acl",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/zookeeper_sessiontimeout",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Session Timeout",
+        "property_name" : "zookeeper_sessiontimeout",
+        "property_value" : "60000",
         "service_name" : "HBASE",
-        "type" : "hbase-policy.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     }
   ]


[10/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HDFS.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HDFS.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HDFS.json
index dbfc16a..90293a0 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HDFS.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HDFS.json
@@ -1,736 +1,1156 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.blockreport.initialDelay",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ambari.dfs.datanode.http.port",
       "StackConfigurations" : {
-        "property_description" : "Delay for first block report in seconds.",
-        "property_value" : "120",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.blockreport.initialDelay",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "\n      The datanode http port. This property is effective only if referenced from dfs.datanode.http.address property.\n    ",
+        "property_name" : "ambari.dfs.datanode.http.port",
+        "property_value" : "50075",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.safemode.threshold.pct",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ambari.dfs.datanode.port",
       "StackConfigurations" : {
-        "property_description" : "\n        Specifies the percentage of blocks that should satisfy\n        the minimal replication requirement defined by dfs.replication.min.\n        Values less than or equal to 0 mean not to start in safe mode.\n        Values greater than 1 will make safe mode permanent.\n        ",
-        "property_value" : "1.0f",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.safemode.threshold.pct",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "\n      The datanode port for data transfer. This property is effective only if referenced from dfs.datanode.address property.\n    ",
+        "property_name" : "ambari.dfs.datanode.port",
+        "property_value" : "50010",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/datanode_du_reserved",
       "StackConfigurations" : {
-        "property_description" : "Secondary NameNode checkpoint dir.",
-        "property_value" : "/hadoop/hdfs/namesecondary",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs_checkpoint_dir",
-        "filename" : "global.xml",
+        "property_description" : "Reserved space for HDFS",
+        "property_name" : "datanode_du_reserved",
+        "property_value" : "1",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.umaskmode",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.access.time.precision",
       "StackConfigurations" : {
-        "property_description" : "\nThe octal umask used when creating files and directories.\n",
-        "property_value" : "077",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.umaskmode",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "The access time for HDFS file is precise upto this value.\n      The default value is 1 hour. Setting a value of 0 disables\n      access times for HDFS.\n    ",
+        "property_name" : "dfs.access.time.precision",
+        "property_value" : "0",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.compression.codec.lzo.class",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.balance.bandwidthPerSec",
       "StackConfigurations" : {
-        "property_description" : "The implementation for lzo codec.",
-        "property_value" : "com.hadoop.compression.lzo.LzoCodec",
-        "stack_version" : "1.3.0",
-        "property_name" : "io.compression.codec.lzo.class",
-        "filename" : "core-site.xml",
+        "property_description" : "\n      Specifies the maximum amount of bandwidth that each datanode\n      can utilize for the balancing purpose in term of\n      the number of bytes per second.\n    ",
+        "property_name" : "dfs.balance.bandwidthPerSec",
+        "property_value" : "6250000",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.heartbeat.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.block.access.token.enable",
       "StackConfigurations" : {
-        "property_description" : "Determines datanode heartbeat interval in seconds.",
-        "property_value" : "3",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.heartbeat.interval",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "\n      If \"true\", access tokens are used as capabilities for accessing datanodes.\n      If \"false\", no access tokens are checked on accessing datanodes.\n    ",
+        "property_name" : "dfs.block.access.token.enable",
+        "property_value" : "true",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_datanode_failed_volume_tolerated",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.block.local-path-access.user",
       "StackConfigurations" : {
-        "property_description" : "DataNode volumes failure toleration",
-        "property_value" : "0",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs_datanode_failed_volume_tolerated",
-        "filename" : "global.xml",
+        "property_description" : "the user who is allowed to perform short\n      circuit reads.\n    ",
+        "property_name" : "dfs.block.local-path-access.user",
+        "property_value" : "hbase",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_name_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.block.size",
       "StackConfigurations" : {
-        "property_description" : "NameNode Directories.",
-        "property_value" : "/hadoop/hdfs/namenode",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs_name_dir",
-        "filename" : "global.xml",
+        "property_description" : "The default block size for new files.",
+        "property_name" : "dfs.block.size",
+        "property_value" : "134217728",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.blockreport.initialDelay",
       "StackConfigurations" : {
-        "property_description" : "FS Checkpoint Size.",
-        "property_value" : "0.5",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs_checkpoint_size",
-        "filename" : "global.xml",
+        "property_description" : "Delay for first block report in seconds.",
+        "property_name" : "dfs.blockreport.initialDelay",
+        "property_value" : "120",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.balance.bandwidthPerSec",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.cluster.administrators",
       "StackConfigurations" : {
-        "property_description" : "\n        Specifies the maximum amount of bandwidth that each datanode\n        can utilize for the balancing purpose in term of\n        the number of bytes per second.\n  ",
-        "property_value" : "6250000",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.balance.bandwidthPerSec",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "ACL for who all can view the default servlets in the HDFS",
+        "property_name" : "dfs.cluster.administrators",
+        "property_value" : " hdfs",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_opt_maxnewsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.data.dir",
       "StackConfigurations" : {
-        "property_description" : "NameNode maximum new generation size",
-        "property_value" : "640",
-        "stack_version" : "1.3.0",
-        "property_name" : "namenode_opt_maxnewsize",
-        "filename" : "global.xml",
+        "property_description" : "Determines where on the local filesystem an DFS data node\n      should store its blocks.  If this is a comma-delimited\n      list of directories, then data will be stored in all named\n      directories, typically on different devices.\n      Directories that do not exist are ignored.\n    ",
+        "property_name" : "dfs.data.dir",
+        "property_value" : "/hadoop/hdfs/data",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.edits.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.address",
       "StackConfigurations" : {
-        "property_description" : "Determines where on the local filesystem the DFS secondary\n        name node should store the temporary edits to merge.\n        If this is a comma-delimited list of directoires then teh edits is\n        replicated in all of the directoires for redundancy.\n        Default value is same as fs.checkpoint.dir\n    ",
-        "property_value" : "${fs.checkpoint.dir}",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs.checkpoint.edits.dir",
-        "filename" : "core-site.xml",
+        "property_description" : "\n      The datanode server address and port for data transfer.\n    ",
+        "property_name" : "dfs.datanode.address",
+        "property_value" : "0.0.0.0:${ambari.dfs.datanode.port}",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/keytab_path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.data.dir.perm",
       "StackConfigurations" : {
-        "property_description" : "KeyTab Directory.",
-        "property_value" : "/etc/security/keytabs",
-        "stack_version" : "1.3.0",
-        "property_name" : "keytab_path",
-        "filename" : "global.xml",
+        "property_description" : "The permissions that should be there on dfs.data.dir\n      directories. The datanode will not come up if the permissions are\n      different on existing dfs.data.dir directories. If the directories\n      don't exist, they will be created with this permission.",
+        "property_name" : "dfs.datanode.data.dir.perm",
+        "property_value" : "750",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.block.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.du.pct",
       "StackConfigurations" : {
-        "property_description" : "The default block size for new files.",
-        "property_value" : "134217728",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.block.size",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "When calculating remaining space, only use this percentage of the real available space\n    ",
+        "property_name" : "dfs.datanode.du.pct",
+        "property_value" : "0.85f",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security_enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.du.reserved",
       "StackConfigurations" : {
-        "property_description" : "Hadoop Security",
-        "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "security_enabled",
-        "filename" : "global.xml",
+        "property_description" : "Reserved space in bytes per volume. Always leave this much space free for non dfs use.\n    ",
+        "property_name" : "dfs.datanode.du.reserved",
+        "property_value" : "1073741824",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.serializations",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.failed.volumes.tolerated",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "org.apache.hadoop.io.serializer.WritableSerialization",
-        "stack_version" : "1.3.0",
-        "property_name" : "io.serializations",
-        "filename" : "core-site.xml",
+        "property_description" : "#of failed disks dn would tolerate",
+        "property_name" : "dfs.datanode.failed.volumes.tolerated",
+        "property_value" : "0",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.task.umbilical.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.failed.volumes.tolerated",
       "StackConfigurations" : {
-        "property_description" : "ACL for TaskUmbilicalProtocol, used by the map and reduce\n    tasks to communicate with the parent tasktracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
-        "property_name" : "security.task.umbilical.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "property_description" : "Number of failed disks datanode would tolerate",
+        "property_name" : "dfs.datanode.failed.volumes.tolerated",
+        "property_value" : "0",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/kerberos_domain",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.http.address",
       "StackConfigurations" : {
-        "property_description" : "Kerberos realm.",
-        "property_value" : "EXAMPLE.COM",
-        "stack_version" : "1.3.0",
-        "property_name" : "kerberos_domain",
-        "filename" : "global.xml",
+        "property_description" : "\n      The datanode http server address and port.\n    ",
+        "property_name" : "dfs.datanode.http.address",
+        "property_value" : "0.0.0.0:${ambari.dfs.datanode.http.port}",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_data_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.ipc.address",
       "StackConfigurations" : {
-        "property_description" : "Data directories for Data Nodes.",
-        "property_value" : "/hadoop/hdfs/data",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs_data_dir",
-        "filename" : "global.xml",
+        "property_description" : "\n      The datanode ipc server address and port.\n      If the port is 0 then the server will start on a free port.\n    ",
+        "property_name" : "dfs.datanode.ipc.address",
+        "property_value" : "0.0.0.0:8010",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hdfs_log_dir_prefix",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.max.xcievers",
       "StackConfigurations" : {
-        "property_description" : "Hadoop Log Dir Prefix",
-        "property_value" : "/var/log/hadoop",
-        "stack_version" : "1.3.0",
-        "property_name" : "hdfs_log_dir_prefix",
-        "filename" : "global.xml",
+        "property_description" : "PRIVATE CONFIG VARIABLE",
+        "property_name" : "dfs.datanode.max.xcievers",
+        "property_value" : "4096",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.datanode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.socket.write.timeout",
       "StackConfigurations" : {
-        "property_description" : "ACL for DatanodeProtocol, which is used by datanodes to\n    communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
-        "property_name" : "security.datanode.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "property_description" : "DFS Client write socket timeout",
+        "property_name" : "dfs.datanode.socket.write.timeout",
+        "property_value" : "0",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.connect.max.retries",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.heartbeat.interval",
       "StackConfigurations" : {
-        "property_description" : "Defines the maximum number of retries for IPC connections.",
-        "property_value" : "50",
-        "stack_version" : "1.3.0",
-        "property_name" : "ipc.client.connect.max.retries",
-        "filename" : "core-site.xml",
+        "property_description" : "Determines datanode heartbeat interval in seconds.",
+        "property_name" : "dfs.heartbeat.interval",
+        "property_value" : "3",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_period",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.hosts",
       "StackConfigurations" : {
-        "property_description" : "HDFS Maximum Checkpoint Delay",
-        "property_value" : "21600",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs_checkpoint_period",
-        "filename" : "global.xml",
+        "property_description" : "Names a file that contains a list of hosts that are\n      permitted to connect to the namenode. The full pathname of the file\n      must be specified.  If the value is empty, all hosts are\n      permitted.",
+        "property_name" : "dfs.hosts",
+        "property_value" : "/etc/hadoop/conf/dfs.include",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.handler.count",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.hosts.exclude",
       "StackConfigurations" : {
-        "property_description" : "The number of server threads for the namenode.",
-        "property_value" : "40",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.namenode.handler.count",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "Names a file that contains a list of hosts that are\n      not permitted to connect to the namenode.  The full pathname of the\n      file must be specified.  If the value is empty, no hosts are\n      excluded.",
+        "property_name" : "dfs.hosts.exclude",
+        "property_value" : "/etc/hadoop/conf/dfs.exclude",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.ipc.address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.http.address",
       "StackConfigurations" : {
-        "property_description" : "\nThe datanode ipc server address and port.\nIf the port is 0 then the server will start on a free port.\n",
-        "property_value" : "0.0.0.0:8010",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.datanode.ipc.address",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "The name of the default file system.  Either the\n      literal string \"local\" or a host:port for NDFS.",
+        "property_name" : "dfs.http.address",
+        "property_value" : "localhost:50070",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_webhdfs_enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.https.address",
       "StackConfigurations" : {
-        "property_description" : "WebHDFS enabled",
-        "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs_webhdfs_enabled",
-        "filename" : "global.xml",
+        "property_description" : "The https address where namenode binds",
+        "property_name" : "dfs.https.address",
+        "property_value" : "localhost:50470",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.trash.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.https.port",
       "StackConfigurations" : {
-        "property_description" : "Number of minutes between trash checkpoints.\n  If zero, the trash feature is disabled.\n  ",
-        "property_value" : "360",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs.trash.interval",
-        "filename" : "core-site.xml",
+        "property_description" : "\n      This property is used by HftpFileSystem.\n    ",
+        "property_name" : "dfs.https.port",
+        "property_value" : "50070",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.secondary.https.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.https.port",
       "StackConfigurations" : {
-        "property_description" : "The https port where secondary-namenode binds",
-        "property_value" : "50490",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.secondary.https.port",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "The https port where namenode binds",
+        "property_name" : "dfs.https.port",
+        "property_value" : "50470",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/datanode_du_reserved",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.name.dir",
       "StackConfigurations" : {
-        "property_description" : "Reserved space for HDFS",
-        "property_value" : "1",
-        "stack_version" : "1.3.0",
-        "property_name" : "datanode_du_reserved",
-        "filename" : "global.xml",
+        "property_description" : "Determines where on the local filesystem the DFS name node\n      should store the name table.  If this is a comma-delimited list\n      of directories then the name table is replicated in all of the\n      directories, for redundancy. ",
+        "property_name" : "dfs.name.dir",
+        "property_value" : "/hadoop/hdfs/namenode",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.file.buffer.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.avoid.read.stale.datanode",
       "StackConfigurations" : {
-        "property_description" : "The size of buffer for use in sequence files.\n  The size of this buffer should probably be a multiple of hardware\n  page size (4096 on Intel x86), and it determines how much data is\n  buffered during read and write operations.",
-        "property_value" : "131072",
-        "stack_version" : "1.3.0",
-        "property_name" : "io.file.buffer.size",
-        "filename" : "core-site.xml",
+        "property_description" : "\n      Indicate whether or not to avoid reading from stale datanodes whose\n      heartbeat messages have not been received by the namenode for more than a\n      specified time interval.\n    ",
+        "property_name" : "dfs.namenode.avoid.read.stale.datanode",
+        "property_value" : "true",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.replication.max",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.avoid.write.stale.datanode",
       "StackConfigurations" : {
-        "property_description" : "Maximal block replication.\n  ",
-        "property_value" : "50",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.replication.max",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "\n      Indicate whether or not to avoid writing to stale datanodes whose\n      heartbeat messages have not been received by the namenode for more than a\n      specified time interval.\n    ",
+        "property_name" : "dfs.namenode.avoid.write.stale.datanode",
+        "property_value" : "true",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hadoop_pid_dir_prefix",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.handler.count",
       "StackConfigurations" : {
-        "property_description" : "Hadoop PID Dir Prefix",
-        "property_value" : "/var/run/hadoop",
-        "stack_version" : "1.3.0",
-        "property_name" : "hadoop_pid_dir_prefix",
-        "filename" : "global.xml",
+        "property_description" : "Added to grow Queue size so that more client connections are allowed",
+        "property_name" : "dfs.namenode.handler.count",
+        "property_value" : "100",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.inter.datanode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.handler.count",
       "StackConfigurations" : {
-        "property_description" : "ACL for InterDatanodeProtocol, the inter-datanode protocol\n    for updating generation timestamp.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
-        "property_name" : "security.inter.datanode.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "property_description" : "The number of server threads for the namenode.",
+        "property_name" : "dfs.namenode.handler.count",
+        "property_value" : "40",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.socket.write.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.stale.datanode.interval",
       "StackConfigurations" : {
-        "property_description" : "DFS Client write socket timeout",
-        "property_value" : "0",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.datanode.socket.write.timeout",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "Datanode is stale after not getting a heartbeat in this interval in ms",
+        "property_name" : "dfs.namenode.stale.datanode.interval",
+        "property_value" : "30000",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.max.xcievers",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.write.stale.datanode.ratio",
       "StackConfigurations" : {
-        "property_description" : "PRIVATE CONFIG VARIABLE",
-        "property_value" : "4096",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.datanode.max.xcievers",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "When the ratio of number stale datanodes to total datanodes marked is greater\n      than this ratio, stop avoiding writing to stale nodes so as to prevent causing hotspots.\n    ",
+        "property_name" : "dfs.namenode.write.stale.datanode.ratio",
+        "property_value" : "1.0f",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.server.max.response.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.permissions",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "5242880",
-        "stack_version" : "1.3.0",
-        "property_name" : "ipc.server.max.response.size",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "\n      If \"true\", enable permission checking in HDFS.\n      If \"false\", permission checking is turned off,\n      but all other behavior is unchanged.\n      Switching from one parameter value to the other does not change the mode,\n      owner or group of files or directories.\n    ",
+        "property_name" : "dfs.permissions",
+        "property_value" : "true",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.permissions.supergroup",
       "StackConfigurations" : {
-        "property_description" : "The size of the current edit log (in bytes) that triggers\n       a periodic checkpoint even if the fs.checkpoint.period hasn't expired.\n  ",
-        "property_value" : "536870912",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs.checkpoint.size",
-        "filename" : "core-site.xml",
+        "property_description" : "The name of the group of super-users.",
+        "property_name" : "dfs.permissions.supergroup",
+        "property_value" : "hdfs",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.namenode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.replication",
       "StackConfigurations" : {
-        "property_description" : "ACL for NamenodeProtocol, the protocol used by the secondary\n    namenode to communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
-        "property_name" : "security.namenode.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "property_description" : "Default block replication.\n    ",
+        "property_name" : "dfs.replication",
+        "property_value" : "3",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.permissions",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.replication.max",
       "StackConfigurations" : {
-        "property_description" : "\nIf \"true\", enable permission checking in HDFS.\nIf \"false\", permission checking is turned off,\nbut all other behavior is unchanged.\nSwitching from one parameter value to the other does not change the mode,\nowner or group of files or directories.\n",
-        "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.permissions",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "Maximal block replication.\n    ",
+        "property_name" : "dfs.replication.max",
+        "property_value" : "50",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.https.port",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.safemode.threshold.pct",
       "StackConfigurations" : {
-        "property_description" : "The https port where namenode binds",
-        "property_value" : "50470",
-        "stack_version" : "1.3.0",
-        "property_name" : "dfs.https.port",
-        "filename" : "hdfs-site.xml",
+        "property_description" : "\n      Specifies the percentage of blocks that should satisfy\n      the minimal replication requirement defined by dfs.replication.min.\n      Values less than or equal to 0 mean not to start in safe mode.\n      Values greater than 1 will make safe mode permanent.\n    ",
+        "property_name" : "dfs.safemode.threshold.pct",
+        "property_value" : "1.0f",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.secondary.http.address",
       "StackConfigurations" : {
-        "property_description" : "NameNode Java heap size",
-        "property_value" : "1024",
+        "property_description" : "Address of secondary namenode web server",
+        "property_name" : "dfs.secondary.http.address",
+        "property_value" : "localhost:50090",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "namenode_heapsize",
-        "filename" : "global.xml",
+        "type" : "hdfs-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.secondary.https.port",
+      "StackConfigurations" : {
+        "property_description" : "The https port where secondary-namenode binds",
+        "property_name" : "dfs.secondary.https.port",
+        "property_value" : "50490",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.handler.count",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.support.append",
       "StackConfigurations" : {
-        "property_description" : "Added to grow Queue size so that more client connections are allowed",
-        "property_value" : "100",
+        "property_description" : "to enable dfs append",
+        "property_name" : "dfs.support.append",
+        "property_value" : "true",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.namenode.handler.count",
-        "filename" : "hdfs-site.xml",
+        "type" : "hdfs-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.umaskmode",
+      "StackConfigurations" : {
+        "property_description" : "\n      The octal umask used when creating files and directories.\n    ",
+        "property_name" : "dfs.umaskmode",
+        "property_value" : "077",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.web.ugi",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.web.ugi",
       "StackConfigurations" : {
-        "property_description" : "The user account used by the web interface.\nSyntax: USERNAME,GROUP1,GROUP2, ...\n",
+        "property_description" : "The user account used by the web interface.\n      Syntax: USERNAME,GROUP1,GROUP2, ...\n    ",
+        "property_name" : "dfs.web.ugi",
         "property_value" : "gopher,gopher",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.web.ugi",
-        "filename" : "hdfs-site.xml",
+        "type" : "hdfs-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.webhdfs.enabled",
+      "StackConfigurations" : {
+        "property_description" : "to enable webhdfs",
+        "property_name" : "dfs.webhdfs.enabled",
+        "property_value" : "true",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.du.pct",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_block_local_path_access_user",
       "StackConfigurations" : {
-        "property_description" : "When calculating remaining space, only use this percentage of the real available space\n",
-        "property_value" : "0.85f",
+        "property_description" : "Default Block Replication.",
+        "property_name" : "dfs_block_local_path_access_user",
+        "property_value" : "hbase",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.datanode.du.pct",
-        "filename" : "hdfs-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_data_dir",
+      "StackConfigurations" : {
+        "property_description" : "Data directories for Data Nodes.",
+        "property_name" : "dfs_data_dir",
+        "property_value" : "/hadoop/hdfs/data",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.period",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_datanode_address",
       "StackConfigurations" : {
-        "property_description" : "The number of seconds between two periodic checkpoints.\n  ",
-        "property_value" : "21600",
+        "property_description" : "Port for datanode address.",
+        "property_name" : "dfs_datanode_address",
+        "property_value" : "50010",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "fs.checkpoint.period",
-        "filename" : "core-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_datanode_data_dir_perm",
+      "StackConfigurations" : {
+        "property_description" : "Datanode dir perms.",
+        "property_name" : "dfs_datanode_data_dir_perm",
+        "property_value" : "750",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.block.access.token.enable",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_datanode_failed_volume_tolerated",
       "StackConfigurations" : {
-        "property_description" : "\nIf \"true\", access tokens are used as capabilities for accessing datanodes.\nIf \"false\", no access tokens are checked on accessing datanodes.\n",
-        "property_value" : "true",
+        "property_description" : "DataNode volumes failure toleration",
+        "property_name" : "dfs_datanode_failed_volume_tolerated",
+        "property_value" : "0",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.block.access.token.enable",
-        "filename" : "hdfs-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_datanode_http_address",
+      "StackConfigurations" : {
+        "property_description" : "Port for datanode address.",
+        "property_name" : "dfs_datanode_http_address",
+        "property_value" : "50075",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.cluster.administrators",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_name_dir",
       "StackConfigurations" : {
-        "property_description" : "ACL for who all can view the default servlets in the HDFS",
-        "property_value" : " hdfs",
+        "property_description" : "NameNode Directories.",
+        "property_name" : "dfs_name_dir",
+        "property_value" : "/hadoop/hdfs/namenode",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.cluster.administrators",
-        "filename" : "hdfs-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_replication",
+      "StackConfigurations" : {
+        "property_description" : "Default Block Replication.",
+        "property_name" : "dfs_replication",
+        "property_value" : "3",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dtnode_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_webhdfs_enabled",
+      "StackConfigurations" : {
+        "property_description" : "WebHDFS enabled",
+        "property_name" : "dfs_webhdfs_enabled",
+        "property_value" : "true",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dtnode_heapsize",
       "StackConfigurations" : {
         "property_description" : "DataNode maximum Java heap size",
+        "property_name" : "dtnode_heapsize",
         "property_value" : "1024",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dtnode_heapsize",
-        "filename" : "global.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.dir",
+      "StackConfigurations" : {
+        "property_description" : "Determines where on the local filesystem the DFS secondary\n        name node should store the temporary images to merge.\n        If this is a comma-delimited list of directories then the image is\n        replicated in all of the directories for redundancy.\n    ",
+        "property_name" : "fs.checkpoint.dir",
+        "property_value" : "/hadoop/hdfs/namesecondary",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.job.submission.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.edits.dir",
       "StackConfigurations" : {
-        "property_description" : "ACL for JobSubmissionProtocol, used by job clients to\n    communciate with the jobtracker for job submission, querying job status etc.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
+        "property_description" : "Determines where on the local filesystem the DFS secondary\n        name node should store the temporary edits to merge.\n        If this is a comma-delimited list of directoires then teh edits is\n        replicated in all of the directoires for redundancy.\n        Default value is same as fs.checkpoint.dir\n    ",
+        "property_name" : "fs.checkpoint.edits.dir",
+        "property_value" : "${fs.checkpoint.dir}",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "security.job.submission.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.period",
+      "StackConfigurations" : {
+        "property_description" : "The number of seconds between two periodic checkpoints.\n  ",
+        "property_name" : "fs.checkpoint.period",
+        "property_value" : "21600",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/webinterface.private.actions",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.size",
       "StackConfigurations" : {
-        "property_description" : " If set to true, the web interfaces of JT and NN may contain\n                actions, such as kill job, delete file, etc., that should\n                not be exposed to public. Enable this option if the interfaces\n                are only reachable by those who have the right authorization.\n  ",
-        "property_value" : "false",
+        "property_description" : "The size of the current edit log (in bytes) that triggers\n       a periodic checkpoint even if the fs.checkpoint.period hasn't expired.\n  ",
+        "property_name" : "fs.checkpoint.size",
+        "property_value" : "67108864",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "webinterface.private.actions",
-        "filename" : "core-site.xml",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.default.name",
+      "StackConfigurations" : {
+        "property_description" : "The name of the default file system.  Either the\n  literal string \"local\" or a host:port for NDFS.",
+        "property_name" : "fs.default.name",
+        "property_value" : "hdfs://localhost:8020",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.connection.maxidletime",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.trash.interval",
       "StackConfigurations" : {
-        "property_description" : "The maximum time after which a client will bring down the\n               connection to the server.\n  ",
-        "property_value" : "30000",
+        "property_description" : "Number of minutes between trash checkpoints.\n  If zero, the trash feature is disabled.\n  ",
+        "property_name" : "fs.trash.interval",
+        "property_value" : "360",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "ipc.client.connection.maxidletime",
-        "filename" : "core-site.xml",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_dir",
+      "StackConfigurations" : {
+        "property_description" : "Secondary NameNode checkpoint dir.",
+        "property_name" : "fs_checkpoint_dir",
+        "property_value" : "/hadoop/hdfs/namesecondary",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.permissions.supergroup",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_period",
       "StackConfigurations" : {
-        "property_description" : "The name of the group of super-users.",
-        "property_value" : "hdfs",
+        "property_description" : "HDFS Maximum Checkpoint Delay",
+        "property_name" : "fs_checkpoint_period",
+        "property_value" : "21600",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.permissions.supergroup",
-        "filename" : "hdfs-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_size",
+      "StackConfigurations" : {
+        "property_description" : "FS Checkpoint Size.",
+        "property_name" : "fs_checkpoint_size",
+        "property_value" : "0.5",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hadoop_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hadoop.security.authentication",
+      "StackConfigurations" : {
+        "property_description" : "\n   Set the authentication for the cluster. Valid values are: simple or\n   kerberos.\n   ",
+        "property_name" : "hadoop.security.authentication",
+        "property_value" : "simple",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hadoop_heapsize",
       "StackConfigurations" : {
         "property_description" : "Hadoop maximum Java heap size",
+        "property_name" : "hadoop_heapsize",
         "property_value" : "1024",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hadoop_heapsize",
-        "filename" : "global.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hadoop_pid_dir_prefix",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop PID Dir Prefix",
+        "property_name" : "hadoop_pid_dir_prefix",
+        "property_value" : "/var/run/hadoop",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hdfs_log_dir_prefix",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop Log Dir Prefix",
+        "property_name" : "hdfs_log_dir_prefix",
+        "property_value" : "/var/log/hadoop",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.idlethreshold",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hdfs_user",
+      "StackConfigurations" : {
+        "property_description" : "User and Groups.",
+        "property_name" : "hdfs_user",
+        "property_value" : "hdfs",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.compression.codec.lzo.class",
+      "StackConfigurations" : {
+        "property_description" : "The implementation for lzo codec.",
+        "property_name" : "io.compression.codec.lzo.class",
+        "property_value" : "com.hadoop.compression.lzo.LzoCodec",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.compression.codecs",
+      "StackConfigurations" : {
+        "property_description" : "A list of the compression codec classes that can be used\n                 for compression/decompression.",
+        "property_name" : "io.compression.codecs",
+        "property_value" : "org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,com.hadoop.compression.lzo.LzoCodec,com.hadoop.compression.lzo.LzopCodec,org.apache.hadoop.io.compress.SnappyCodec",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.file.buffer.size",
+      "StackConfigurations" : {
+        "property_description" : "The size of buffer for use in sequence files.\n  The size of this buffer should probably be a multiple of hardware\n  page size (4096 on Intel x86), and it determines how much data is\n  buffered during read and write operations.",
+        "property_name" : "io.file.buffer.size",
+        "property_value" : "131072",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.serializations",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "io.serializations",
+        "property_value" : "org.apache.hadoop.io.serializer.WritableSerialization",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.connect.max.retries",
+      "StackConfigurations" : {
+        "property_description" : "Defines the maximum number of retries for IPC connections.",
+        "property_name" : "ipc.client.connect.max.retries",
+        "property_value" : "50",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.connection.maxidletime",
+      "StackConfigurations" : {
+        "property_description" : "The maximum time after which a client will bring down the\n               connection to the server.\n  ",
+        "property_name" : "ipc.client.connection.maxidletime",
+        "property_value" : "30000",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.idlethreshold",
       "StackConfigurations" : {
         "property_description" : "Defines the threshold number of connections after which\n               connections will be inspected for idleness.\n  ",
+        "property_name" : "ipc.client.idlethreshold",
         "property_value" : "8000",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "ipc.client.idlethreshold",
-        "filename" : "core-site.xml",
+        "type" : "core-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.server.max.response.size",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "ipc.server.max.response.size",
+        "property_value" : "5242880",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hdfs-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.inter.tracker.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.server.read.threadpool.size",
       "StackConfigurations" : {
-        "property_description" : "ACL for InterTrackerProtocol, used by the tasktrackers to\n    communicate with the jobtracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
+        "property_description" : "",
+        "property_name" : "ipc.server.read.threadpool.size",
+        "property_value" : "5",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "security.inter.tracker.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "type" : "hdfs-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/kerberos_domain",
+      "StackConfigurations" : {
+        "property_description" : "Kerberos realm.",
+        "property_name" : "kerberos_domain",
+        "property_value" : "EXAMPLE.COM",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.failed.volumes.tolerated",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/keytab_path",
       "StackConfigurations" : {
-        "property_description" : "Number of failed disks datanode would tolerate",
-        "property_value" : "0",
+        "property_description" : "Kerberos keytab path.",
+        "property_name" : "keytab_path",
+        "property_value" : "/etc/security/keytabs",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.datanode.failed.volumes.tolerated",
-        "filename" : "hdfs-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_formatted_mark_dir",
+      "StackConfigurations" : {
+        "property_description" : "Formatteed Mark Directory.",
+        "property_name" : "namenode_formatted_mark_dir",
+        "property_value" : "/var/run/hadoop/hdfs/namenode/formatted/",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_opt_newsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_heapsize",
       "StackConfigurations" : {
-        "property_description" : "NameNode new generation size",
+        "property_description" : "NameNode Java heap size",
+        "property_name" : "namenode_heapsize",
+        "property_value" : "1024",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_opt_maxnewsize",
+      "StackConfigurations" : {
+        "property_description" : "NameNode maximum new generation size",
+        "property_name" : "namenode_opt_maxnewsize",
         "property_value" : "200",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_opt_newsize",
+      "StackConfigurations" : {
+        "property_description" : "NameNode new generation size",
         "property_name" : "namenode_opt_newsize",
-        "filename" : "global.xml",
+        "property_value" : "200",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.client.datanode.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/proxyuser_group",
       "StackConfigurations" : {
-        "property_description" : "ACL for ClientDatanodeProtocol, the client-to-datanode protocol\n    for block recovery.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
+        "property_description" : "Proxy user group.",
+        "property_name" : "proxyuser_group",
+        "property_value" : "users",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.client.datanode.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for ClientDatanodeProtocol, the client-to-datanode protocol\n    for block recovery.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
         "property_name" : "security.client.datanode.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "property_value" : "*",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.client.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.client.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for ClientProtocol, which is used by user code\n    via the DistributedFileSystem.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.client.protocol.acl",
         "property_value" : "*",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "security.client.protocol.acl",
-        "filename" : "hadoop-policy.xml",
+        "type" : "hadoop-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.datanode.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for DatanodeProtocol, which is used by datanodes to\n    communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.datanode.protocol.acl",
+        "property_value" : "*",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.server.read.threadpool.size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.inter.datanode.protocol.acl",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "5",
+        "property_description" : "ACL for InterDatanodeProtocol, the inter-datanode protocol\n    for updating generation timestamp.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.inter.datanode.protocol.acl",
+        "property_value" : "*",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "ipc.server.read.threadpool.size",
-        "filename" : "hdfs-site.xml",
+        "type" : "hadoop-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.inter.tracker.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for InterTrackerProtocol, used by the tasktrackers to\n    communicate with the jobtracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.inter.tracker.protocol.acl",
+        "property_value" : "*",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hadoop-policy.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.access.time.precision",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.job.submission.protocol.acl",
       "StackConfigurations" : {
-        "property_description" : "The access time for HDFS file is precise upto this value.\n               The default value is 1 hour. Setting a value of 0 disables\n               access times for HDFS.\n  ",
-        "property_value" : "0",
+        "property_description" : "ACL for JobSubmissionProtocol, used by job clients to\n    communciate with the jobtracker for job submission, querying job status etc.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.job.submission.protocol.acl",
+        "property_value" : "*",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "dfs.access.time.precision",
-        "filename" : "hdfs-site.xml",
+        "type" : "hadoop-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.namenode.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for NamenodeProtocol, the protocol used by the secondary\n    namenode to communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.namenode.protocol.acl",
+        "property_value" : "*",
         "service_name" : "HDFS",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hadoop-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.task.umbilical.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for TaskUmbilicalProtocol, used by the map and reduce\n    tasks to communicate with the parent tasktracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.task.umbilical.protocol.acl",
+        "property_value" : "*",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hadoop-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security_enabled",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop Security",
+        "property_name" : "security_enabled",
+        "property_value" : "false",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/webinterface.private.actions",
+      "StackConfigurations" : {
+        "property_description" : " If set to true, the web interfaces of JT and NN may contain\n                actions, such as kill job, delete file, etc., that should\n                not be exposed to public. Enable this option if the interfaces\n                are only reachable by those who have the right authorization.\n  ",
+        "property_name" : "webinterface.private.actions",
+        "property_value" : "false",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "core-site.xml"
       }
     }
   ]


[02/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/YARN.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/YARN.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/YARN.json
index 6f2798d..0c6b9f8 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/YARN.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/YARN.json
@@ -2,459 +2,663 @@
   "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.container-monitor.interval-ms",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/nodemanager_heapsize",
       "StackConfigurations" : {
-        "property_description" : "The interval, in milliseconds, for which the node manager\n    waits  between two cycles of monitoring its containers' memory usage. \n    ",
-        "property_value" : "3000",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.container-monitor.interval-ms",
+        "property_description" : "Max heapsize for NodeManager using a numerical value in the scale of MB",
+        "property_name" : "nodemanager_heapsize",
+        "property_value" : "1024",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "yarn-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.scheduler.class",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/resourcemanager_heapsize",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.resourcemanager.scheduler.class",
+        "property_description" : "Max heapsize for ResourceManager using a numerical value in the scale of MB",
+        "property_name" : "resourcemanager_heapsize",
+        "property_value" : "1024",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "yarn-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.admin.address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.acl.enable",
       "StackConfigurations" : {
         "property_description" : null,
-        "property_value" : "TODO-RMNODE-HOSTNAME:8141",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.resourcemanager.admin.address",
+        "property_name" : "yarn.acl.enable",
+        "property_value" : "true",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.queues",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.admin.acl",
       "StackConfigurations" : {
-        "property_description" : "\n      The queues at the this level (root is the root queue).\n    ",
-        "property_value" : "default",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.queues",
+        "property_description" : null,
+        "property_name" : "yarn.admin.acl",
+        "property_value" : "*",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.capacity",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.application.classpath",
       "StackConfigurations" : {
-        "property_description" : "\n      The total capacity as a percentage out of 100 for this queue.\n      If it has child queues then this includes their capacity as well.\n      The child queues capacity should add up to their parent queue's capacity\n      or less.\n    ",
-        "property_value" : "100",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.capacity",
+        "property_description" : "Classpath for typical applications.",
+        "property_name" : "yarn.application.classpath",
+        "property_value" : "/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.acl_submit_jobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.log-aggregation-enable",
       "StackConfigurations" : {
-        "property_description" : "\n      The ACL of who can submit jobs to the default queue.\n    ",
-        "property_value" : "*",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.default.acl_submit_jobs",
+        "property_description" : "Whether to enable log aggregation",
+        "property_name" : "yarn.log-aggregation-enable",
+        "property_value" : "true",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.log-aggregation.compression-type",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.log-aggregation.retain-seconds",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "gz",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.log-aggregation.compression-type",
+        "property_description" : "\n      How long to keep aggregation logs before deleting them. -1 disables.\n      Be careful set this too small and you will spam the name node.\n    ",
+        "property_name" : "yarn.log-aggregation.retain-seconds",
+        "property_value" : "2592000",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.delete.debug-delay-sec",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.log.server.url",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "36000",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.delete.debug-delay-sec",
+        "property_description" : "\n      URI for the HistoryServer's log resource\n    ",
+        "property_name" : "yarn.log.server.url",
+        "property_value" : "http://localhost:19888/jobhistory/logs",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.address",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "0.0.0.0:45454",
-        "stack_version" : "2.0.1",
+        "property_description" : "The address of the container manager in the NM.",
         "property_name" : "yarn.nodemanager.address",
+        "property_value" : "0.0.0.0:45454",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.resource.memory-mb",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.admin-env",
       "StackConfigurations" : {
-        "property_description" : "Amount of physical memory, in MB, that can be allocated\n      for containers.",
-        "property_value" : "8192",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.resource.memory-mb",
+        "property_description" : "\n      Environment variables that should be forwarded from the NodeManager's\n      environment to the container's.\n    ",
+        "property_name" : "yarn.nodemanager.admin-env",
+        "property_value" : "MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.maximum-capacity",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.aux-services",
       "StackConfigurations" : {
-        "property_description" : "\n      The maximum capacity of the default queue. \n    ",
-        "property_value" : "100",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.default.maximum-capacity",
+        "property_description" : "Auxilliary services of NodeManager. A valid service name should only contain a-zA-Z0-9_ and can\n      not start with numbers",
+        "property_name" : "yarn.nodemanager.aux-services",
+        "property_value" : "mapreduce_shuffle",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.resource-tracker.address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.aux-services.mapreduce_shuffle.class",
       "StackConfigurations" : {
         "property_description" : null,
-        "property_value" : "TODO-RMNODE-HOSTNAME:8025",
+        "property_name" : "yarn.nodemanager.aux-services.mapreduce_shuffle.class",
+        "property_value" : "org.apache.hadoop.mapred.ShuffleHandler",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.resourcemanager.resource-tracker.address",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.container-executor.class",
+      "StackConfigurations" : {
+        "property_description" : "ContainerExecutor for launching containers",
+        "property_name" : "yarn.nodemanager.container-executor.class",
+        "property_value" : "org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.vmem-pmem-ratio",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.container-monitor.interval-ms",
       "StackConfigurations" : {
-        "property_description" : "Ratio between virtual memory to physical memory when\n    setting memory limits for containers. Container allocations are\n    expressed in terms of physical memory, and virtual memory usage\n    is allowed to exceed this allocation by this ratio.\n    ",
-        "property_value" : "2.1",
+        "property_description" : "\n      The interval, in milliseconds, for which the node manager\n      waits  between two cycles of monitoring its containers' memory usage.\n    ",
+        "property_name" : "yarn.nodemanager.container-monitor.interval-ms",
+        "property_value" : "3000",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.vmem-pmem-ratio",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.delete.debug-delay-sec",
+      "StackConfigurations" : {
+        "property_description" : "\n      Number of seconds after an application finishes before the nodemanager's\n      DeletionService will delete the application's localized file directory\n      and log directory.\n\n      To diagnose Yarn application problems, set this property's value large\n      enough (for example, to 600 = 10 minutes) to permit examination of these\n      directories. After changing the property's value, you must restart the\n      nodemanager in order for it to have an effect.\n\n      The roots of Yarn applications' work directories is configurable with\n      the yarn.nodemanager.local-dirs property (see below), and the roots\n      of the Yarn applications' log directories is configurable with the\n      yarn.nodemanager.log-dirs property (see also below).\n    ",
+        "property_name" : "yarn.nodemanager.delete.debug-delay-sec",
+        "property_value" : "0",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.remote-app-log-dir-suffix",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.disk-health-checker.min-healthy-disks",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "logs",
+        "property_description" : "\n      The minimum fraction of number of disks to be healthy for the nodemanager\n      to launch new containers. This correspond to both\n      yarn-nodemanager.local-dirs and yarn.nodemanager.log-dirs. i.e.\n      If there are less number of healthy local-dirs (or log-dirs) available,\n      then new containers will not be launched on this node.\n    ",
+        "property_name" : "yarn.nodemanager.disk-health-checker.min-healthy-disks",
+        "property_value" : "0.25",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.remote-app-log-dir-suffix",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.health-checker.interval-ms",
+      "StackConfigurations" : {
+        "property_description" : "Frequency of running node health script.",
+        "property_name" : "yarn.nodemanager.health-checker.interval-ms",
+        "property_value" : "135000",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.health-checker.script.timeout-ms",
       "StackConfigurations" : {
-        "property_description" : null,
+        "property_description" : "Script time out period.",
+        "property_name" : "yarn.nodemanager.health-checker.script.timeout-ms",
         "property_value" : "60000",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.health-checker.script.timeout-ms",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.linux-container-executor.group",
+      "StackConfigurations" : {
+        "property_description" : "Unix group of the NodeManager",
+        "property_name" : "yarn.nodemanager.linux-container-executor.group",
+        "property_value" : "hadoop",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.local-dirs",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "TODO-YARN-LOCAL-DIR",
-        "stack_version" : "2.0.1",
+        "property_description" : "\n      List of directories to store localized files in. An\n      application's localized file directory will be found in:\n      ${yarn.nodemanager.local-dirs}/usercache/${user}/appcache/application_${appid}.\n      Individual containers' work directories, called container_${contid}, will\n      be subdirectories of this.\n    ",
         "property_name" : "yarn.nodemanager.local-dirs",
+        "property_value" : "/hadoop/yarn/local",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.user-limit-factor",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.log-aggregation.compression-type",
       "StackConfigurations" : {
-        "property_description" : "\n      Default queue user limit a percentage from 0.0 to 1.0.\n    ",
-        "property_value" : "1",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.default.user-limit-factor",
+        "property_description" : "\n      T-file compression types used to compress aggregated logs.\n    ",
+        "property_name" : "yarn.nodemanager.log-aggregation.compression-type",
+        "property_value" : "gz",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.maximum-am-resource-percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.log-dirs",
       "StackConfigurations" : {
-        "property_description" : "\n      Maximum percent of resources in the cluster which can be used to run \n      application masters i.e. controls number of concurrent running\n      applications.\n    ",
-        "property_value" : "0.1",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.maximum-am-resource-percent",
+        "property_description" : "\n      Where to store container logs. An application's localized log directory\n      will be found in ${yarn.nodemanager.log-dirs}/application_${appid}.\n      Individual containers' log directories will be below this, in directories\n      named container_{$contid}. Each container directory will contain the files\n      stderr, stdin, and syslog generated by that container.\n    ",
+        "property_name" : "yarn.nodemanager.log-dirs",
+        "property_value" : "/hadoop/yarn/log",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.state",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.log.retain-second",
       "StackConfigurations" : {
-        "property_description" : "\n      The state of the default queue. State can be one of RUNNING or STOPPED.\n    ",
-        "property_value" : "RUNNING",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.default.state",
+        "property_description" : "\n      Time in seconds to retain user logs. Only applicable if\n      log aggregation is disabled.\n    ",
+        "property_name" : "yarn.nodemanager.log.retain-second",
+        "property_value" : "604800",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.acl_administer_queues",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.remote-app-log-dir",
       "StackConfigurations" : {
-        "property_description" : "\n      The ACL for who can administer this queue i.e. change sub-queue \n      allocations.\n    ",
-        "property_value" : "*",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.acl_administer_queues",
+        "property_description" : null,
+        "property_name" : "yarn.nodemanager.remote-app-log-dir",
+        "property_value" : "/app-logs",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.log.retain-second",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.remote-app-log-dir-suffix",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "604800",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.log.retain-second",
+        "property_description" : "\n      The remote log dir will be created at\n      {yarn.nodemanager.remote-app-log-dir}/${user}/{thisParam}.\n    ",
+        "property_name" : "yarn.nodemanager.remote-app-log-dir-suffix",
+        "property_value" : "logs",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.acl_administer_jobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.resource.memory-mb",
       "StackConfigurations" : {
-        "property_description" : "\n      The ACL of who can administer jobs on the default queue.\n    ",
-        "property_value" : "*",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.default.acl_administer_jobs",
+        "property_description" : "Amount of physical memory, in MB, that can be allocated\n      for containers.",
+        "property_name" : "yarn.nodemanager.resource.memory-mb",
+        "property_value" : "5120",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.application.classpath",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.vmem-check-enabled",
       "StackConfigurations" : {
-        "property_description" : "Classpath for typical applications.",
-        "property_value" : "/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*",
+        "property_description" : "\n      Whether virtual memory limits will be enforced for containers.\n    ",
+        "property_name" : "yarn.nodemanager.vmem-check-enabled",
+        "property_value" : "false",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.application.classpath",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.vmem-pmem-ratio",
+      "StackConfigurations" : {
+        "property_description" : "Ratio between virtual memory to physical memory when\n      setting memory limits for containers. Container allocations are\n      expressed in terms of physical memory, and virtual memory usage\n      is allowed to exceed this allocation by this ratio.\n    ",
+        "property_name" : "yarn.nodemanager.vmem-pmem-ratio",
+        "property_value" : "2.1",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
       "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.address",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "TODO-RMNODE-HOSTNAME:8050",
-        "stack_version" : "2.0.1",
+        "property_description" : "\n      The address of the applications manager interface in the\n      RM.\n    ",
         "property_name" : "yarn.resourcemanager.address",
+        "property_value" : "localhost:8050",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.log-dirs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.admin.address",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "TODO-YARN-LOG-DIR",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.log-dirs",
+        "property_description" : "The address of the RM admin interface.",
+        "property_name" : "yarn.resourcemanager.admin.address",
+        "property_value" : "localhost:8141",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.maximum-allocation-mb",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.am.max-attempts",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "8192",
+        "property_description" : "\n      The maximum number of application attempts. It's a global\n      setting for all application masters. Each application master can specify\n      its individual maximum number of application attempts via the API, but the\n      individual number cannot be more than the global upper bound. If it is,\n      the resourcemanager will override it. The default number is set to 2, to\n      allow at least one retry for AM.\n    ",
+        "property_name" : "yarn.resourcemanager.am.max-attempts",
+        "property_value" : "2",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.maximum-allocation-mb",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.hostname",
+      "StackConfigurations" : {
+        "property_description" : "The hostname of the RM.",
+        "property_name" : "yarn.resourcemanager.hostname",
+        "property_value" : "localhost",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.container-executor.class",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.resource-tracker.address",
       "StackConfigurations" : {
-        "property_description" : "ContainerExecutor for launching containers",
-        "property_value" : "org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor",
+        "property_description" : null,
+        "property_name" : "yarn.resourcemanager.resource-tracker.address",
+        "property_value" : "localhost:8025",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.container-executor.class",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.scheduler.address",
+      "StackConfigurations" : {
+        "property_description" : "The address of the scheduler interface.",
+        "property_name" : "yarn.resourcemanager.scheduler.address",
+        "property_value" : "localhost:8030",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.remote-app-log-dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.scheduler.class",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "/app-logs",
+        "property_description" : "The class to use as the resource scheduler.",
+        "property_name" : "yarn.resourcemanager.scheduler.class",
+        "property_value" : "org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.remote-app-log-dir",
+        "type" : "yarn-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.webapp.address",
+      "StackConfigurations" : {
+        "property_description" : "\n      The address of the RM web application.\n    ",
+        "property_name" : "yarn.resourcemanager.webapp.address",
+        "property_value" : "localhost:8088",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.unfunded.capacity",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.maximum-am-resource-percent",
       "StackConfigurations" : {
-        "property_description" : "\n      No description\n    ",
-        "property_value" : "50",
+        "property_description" : "\n      Maximum percent of resources in the cluster which can be used to run \n      application masters i.e. controls number of concurrent running\n      applications.\n    ",
+        "property_name" : "yarn.scheduler.capacity.maximum-am-resource-percent",
+        "property_value" : "0.2",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.unfunded.capacity",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.maximum-applications",
+      "StackConfigurations" : {
+        "property_description" : "\n      Maximum number of applications that can be pending and running.\n    ",
+        "property_name" : "yarn.scheduler.capacity.maximum-applications",
+        "property_value" : "10000",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.health-checker.interval-ms",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.node-locality-delay",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "135000",
+        "property_description" : "\n      No description\n    ",
+        "property_name" : "yarn.scheduler.capacity.node-locality-delay",
+        "property_value" : "40",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.health-checker.interval-ms",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.acl_administer_queues",
+      "StackConfigurations" : {
+        "property_description" : "\n      The ACL for who can administer this queue i.e. change sub-queue \n      allocations.\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.acl_administer_queues",
+        "property_value" : "*",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "yarn-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.minimum-allocation-mb",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.capacity",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "1024",
+        "property_description" : "\n      The total capacity as a percentage out of 100 for this queue.\n      If it has child queues then this includes their capacity as well.\n      The child queues capacity should add up to their parent queue's capacity\n      or less.\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.capacity",
+        "property_value" : "100",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.minimum-allocation-mb",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.acl_administer_jobs",
+      "StackConfigurations" : {
+        "property_description" : "\n      The ACL of who can administer jobs on the default queue.\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.default.acl_administer_jobs",
+        "property_value" : "*",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "yarn-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.maximum-applications",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.acl_submit_jobs",
       "StackConfigurations" : {
-        "property_description" : "\n      Maximum number of applications that can be pending and running.\n    ",
-        "property_value" : "10000",
+        "property_description" : "\n      The ACL of who can submit jobs to the default queue.\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.default.acl_submit_jobs",
+        "property_value" : "*",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.maximum-applications",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.capacity",
+      "StackConfigurations" : {
+        "property_description" : "Default queue target capacity.",
+        "property_name" : "yarn.scheduler.capacity.root.default.capacity",
+        "property_value" : "100",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.aux-services",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.maximum-capacity",
       "StackConfigurations" : {
-        "property_description" : "Auxilliary services of NodeManager",
-        "property_value" : "mapreduce_shuffle",
+        "property_description" : "\n      The maximum capacity of the default queue. \n    ",
+        "property_name" : "yarn.scheduler.capacity.root.default.maximum-capacity",
+        "property_value" : "100",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.aux-services",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.state",
+      "StackConfigurations" : {
+        "property_description" : "\n      The state of the default queue. State can be one of RUNNING or STOPPED.\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.default.state",
+        "property_value" : "RUNNING",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "yarn-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.health-checker.script.path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.user-limit-factor",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "/etc/hadoop/conf/health_check",
+        "property_description" : "\n      Default queue user limit a percentage from 0.0 to 1.0.\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.default.user-limit-factor",
+        "property_value" : "1",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.health-checker.script.path",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.queues",
+      "StackConfigurations" : {
+        "property_description" : "\n      The queues at the this level (root is the root queue).\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.queues",
+        "property_value" : "default",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "yarn-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.log-aggregation-enable",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.unfunded.capacity",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "true",
+        "property_description" : "\n      No description\n    ",
+        "property_name" : "yarn.scheduler.capacity.root.unfunded.capacity",
+        "property_value" : "50",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.log-aggregation-enable",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.maximum-allocation-mb",
+      "StackConfigurations" : {
+        "property_description" : "\n      The maximum allocation for every container request at the RM,\n      in MBs. Memory requests higher than this won't take effect,\n      and will get capped to this value.\n    ",
+        "property_name" : "yarn.scheduler.maximum-allocation-mb",
+        "property_value" : "2048",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.nodemanager.aux-services.class",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.minimum-allocation-mb",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "org.apache.hadoop.mapred.ShuffleHandler",
-        "stack_version" : "2.0.1",
-        "property_name" : "yarn.nodemanager.aux-services.class",
+        "property_description" : "\n      TThe minimum allocation for every container request at the RM,\n      in MBs. Memory requests lower than this won't take effect,\n      and the specified value will get allocated at minimum.\n    ",
+        "property_name" : "yarn.scheduler.minimum-allocation-mb",
+        "property_value" : "512",
         "service_name" : "YARN",
         "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
         "type" : "yarn-site.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.resourcemanager.scheduler.address",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn_heapsize",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "TODO-RMNODE-HOSTNAME:8030",
+        "property_description" : "Max heapsize for all YARN components using a numerical value in the scale of MB",
+        "property_name" : "yarn_heapsize",
+        "property_value" : "1024",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.resourcemanager.scheduler.address",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn_log_dir_prefix",
+      "StackConfigurations" : {
+        "property_description" : "YARN Log Dir Prefix",
+        "property_name" : "yarn_log_dir_prefix",
+        "property_value" : "/var/log/hadoop-yarn",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "yarn-site.xml"
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn.scheduler.capacity.root.default.capacity",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn_pid_dir_prefix",
       "StackConfigurations" : {
-        "property_description" : "Default queue target capacity.",
-        "property_value" : "100",
+        "property_description" : "YARN PID Dir Prefix",
+        "property_name" : "yarn_pid_dir_prefix",
+        "property_value" : "/var/run/hadoop-yarn",
+        "service_name" : "YARN",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
-        "property_name" : "yarn.scheduler.capacity.root.default.capacity",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/YARN/configurations/yarn_user",
+      "StackConfigurations" : {
+        "property_description" : "YARN User",
+        "property_name" : "yarn_user",
+        "property_value" : "yarn",
         "service_name" : "YARN",
         "stack_name" : "HDP",
-        "type" : "capacity-scheduler.xml"
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     }
   ]

http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/ZOOKEEPER.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/ZOOKEEPER.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/ZOOKEEPER.json
index aeeac38..4763412 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/ZOOKEEPER.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/ZOOKEEPER.json
@@ -1,4 +1,113 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations?fields=*",
-  "items" : [ ]
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations?fields=*",
+  "items" : [
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/clientPort",
+      "StackConfigurations" : {
+        "property_description" : "Port for running ZK Server.",
+        "property_name" : "clientPort",
+        "property_value" : "2181",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/initLimit",
+      "StackConfigurations" : {
+        "property_description" : "Ticks to allow for sync at Init.",
+        "property_name" : "initLimit",
+        "property_value" : "10",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/syncLimit",
+      "StackConfigurations" : {
+        "property_description" : "Ticks to allow for sync at Runtime.",
+        "property_name" : "syncLimit",
+        "property_value" : "5",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/tickTime",
+      "StackConfigurations" : {
+        "property_description" : "The length of a single tick in milliseconds, which is the basic time unit used by ZooKeeper",
+        "property_name" : "tickTime",
+        "property_value" : "2000",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/zk_data_dir",
+      "StackConfigurations" : {
+        "property_description" : "Data directory for ZooKeeper.",
+        "property_name" : "zk_data_dir",
+        "property_value" : "/hadoop/zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/zk_log_dir",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Log Dir",
+        "property_name" : "zk_log_dir",
+        "property_value" : "/var/log/zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/zk_pid_dir",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Pid Dir",
+        "property_name" : "zk_pid_dir",
+        "property_value" : "/var/run/zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/zk_pid_file",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Pid File",
+        "property_name" : "zk_pid_file",
+        "property_value" : "/var/run/zookeeper/zookeeper_server.pid",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/ZOOKEEPER/configurations/zk_user",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper User.",
+        "property_name" : "zk_user",
+        "property_value" : "zookeeper",
+        "service_name" : "ZOOKEEPER",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    }
+  ]
 }
\ No newline at end of file


[06/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HBASE.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HBASE.json b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HBASE.json
index eb43f8c..a4bd060 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HBASE.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version2.0.1/HBASE.json
@@ -1,292 +1,640 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/zookeeper_sessiontimeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/client_scannercaching",
       "StackConfigurations" : {
-        "property_description" : "ZooKeeper Session Timeout",
-        "property_value" : "60000",
-        "stack_version" : "1.3.0",
-        "property_name" : "zookeeper_sessiontimeout",
+        "property_description" : "Base Client Scanner Caching",
+        "property_name" : "client_scannercaching",
+        "property_value" : "100",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_master_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/dfs.domain.socket.path",
       "StackConfigurations" : {
-        "property_description" : "HBase Master Heap Size",
-        "property_value" : "1024",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase_master_heapsize",
+        "property_description" : "Path to domain socket.",
+        "property_name" : "dfs.domain.socket.path",
+        "property_value" : "/var/lib/hadoop-hdfs/dn_socket",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hstore_compactionthreshold",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.client.keyvalue.maxsize",
       "StackConfigurations" : {
-        "property_description" : "HBase HStore compaction threshold.",
-        "property_value" : "3",
-        "stack_version" : "1.3.0",
-        "property_name" : "hstore_compactionthreshold",
+        "property_description" : "Specifies the combined maximum allowed size of a KeyValue\n    instance. This is to set an upper boundary for a single entry saved in a\n    storage file. Since they cannot be split it helps avoiding that a region\n    cannot be split any further because the data is too large. It seems wise\n    to set this to a fraction of the maximum region size. Setting it to zero\n    or less disables the check.\n    ",
+        "property_name" : "hbase.client.keyvalue.maxsize",
+        "property_value" : "10485760",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hfile_blockcache_size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.client.scanner.caching",
       "StackConfigurations" : {
-        "property_description" : "HFile block cache size.",
-        "property_value" : "0.25",
-        "stack_version" : "1.3.0",
-        "property_name" : "hfile_blockcache_size",
+        "property_description" : "Number of rows that will be fetched when calling next\n    on a scanner if it is not served from (local, client) memory. Higher\n    caching values will enable faster scanners but will eat up more memory\n    and some calls of next may take longer and longer times when the cache is empty.\n    Do not set this value such that the time between invocations is greater\n    than the scanner timeout; i.e. hbase.regionserver.lease.period\n    ",
+        "property_name" : "hbase.client.scanner.caching",
+        "property_value" : "100",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.client.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.cluster.distributed",
       "StackConfigurations" : {
-        "property_description" : "ACL for HRegionInterface protocol implementations (ie. \n    clients talking to HRegionServers)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
-        "property_name" : "security.client.protocol.acl",
+        "property_description" : "The mode the cluster will be in. Possible values are\n      false for standalone mode and true for distributed mode.  If\n      false, startup will run all HBase and ZooKeeper daemons together\n      in the one JVM.\n    ",
+        "property_name" : "hbase.cluster.distributed",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "hbase-policy.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hregion_blockmultiplier",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.defaults.for.version.skip",
       "StackConfigurations" : {
-        "property_description" : "HBase Region Block Multiplier",
+        "property_description" : "Disables version verification.",
+        "property_name" : "hbase.defaults.for.version.skip",
+        "property_value" : "true",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hregion.majorcompaction",
+      "StackConfigurations" : {
+        "property_description" : "The time (in milliseconds) between 'major' compactions of all\n    HStoreFiles in a region.  Default: 1 day.\n    Set to 0 to disable automated major compactions.\n    ",
+        "property_name" : "hbase.hregion.majorcompaction",
+        "property_value" : "86400000",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hregion.max.filesize",
+      "StackConfigurations" : {
+        "property_description" : "\n    Maximum HStoreFile size. If any one of a column families' HStoreFiles has\n    grown to exceed this value, the hosting HRegion is split in two.\n    Default: 1G.\n    ",
+        "property_name" : "hbase.hregion.max.filesize",
+        "property_value" : "10737418240",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hregion.memstore.block.multiplier",
+      "StackConfigurations" : {
+        "property_description" : "Block updates if memstore has hbase.hregion.memstore.block.multiplier\n    time hbase.hregion.flush.size bytes.  Useful preventing\n    runaway memstore during spikes in update traffic.  Without an\n    upper-bound, memstore fills such that when it flushes the\n    resultant flush files take a long time to compact or split, or\n    worse, we OOME\n    ",
+        "property_name" : "hbase.hregion.memstore.block.multiplier",
         "property_value" : "2",
-        "stack_version" : "1.3.0",
-        "property_name" : "hregion_blockmultiplier",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.zookeeper.useMulti",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hregion.memstore.flush.size",
       "StackConfigurations" : {
-        "property_description" : "Instructs HBase to make use of ZooKeeper's multi-update functionality.\n    This allows certain ZooKeeper operations to complete more quickly and prevents some issues\n    with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).В·\n    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+\n    and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will\n    not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).\n    ",
+        "property_description" : "\n    Memstore will be flushed to disk if size of the memstore\n    exceeds this number of bytes.  Value is checked by a thread that runs\n    every hbase.server.thread.wakefrequency.\n    ",
+        "property_name" : "hbase.hregion.memstore.flush.size",
+        "property_value" : "134217728",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hregion.memstore.mslab.enabled",
+      "StackConfigurations" : {
+        "property_description" : "\n      Enables the MemStore-Local Allocation Buffer,\n      a feature which works to prevent heap fragmentation under\n      heavy write loads. This can reduce the frequency of stop-the-world\n      GC pauses on large heaps.\n    ",
+        "property_name" : "hbase.hregion.memstore.mslab.enabled",
         "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase.zookeeper.useMulti",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.superuser",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hstore.blockingStoreFiles",
       "StackConfigurations" : {
-        "property_description" : "List of users or groups (comma-separated), who are allowed full privileges, regardless of stored ACLs, across the cluster.Only used when HBase security is enabled.\n",
-        "property_value" : "hbase",
+        "property_description" : "\n    If more than this number of StoreFiles in any one Store\n    (one StoreFile is written per flush of MemStore) then updates are\n    blocked for this HRegion until a compaction is completed, or\n    until hbase.hstore.blockingWaitTime has been exceeded.\n    ",
+        "property_name" : "hbase.hstore.blockingStoreFiles",
+        "property_value" : "10",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hstore.compactionThreshold",
+      "StackConfigurations" : {
+        "property_description" : "\n    If more than this number of HStoreFiles in any one HStore\n    (one HStoreFile is written per flush of memstore) then a compaction\n    is run to rewrite all HStoreFiles files as one.  Larger numbers\n    put off compaction but when it runs, it takes longer to complete.\n    ",
+        "property_name" : "hbase.hstore.compactionThreshold",
+        "property_value" : "3",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
         "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.hstore.flush.retries.number",
+      "StackConfigurations" : {
+        "property_description" : "\n    The number of times the region flush operation will be retried.\n    ",
+        "property_name" : "hbase.hstore.flush.retries.number",
+        "property_value" : "120",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.regionserver.global.memstore.lowerLimit",
+      "StackConfigurations" : {
+        "property_description" : "When memstores are being forced to flush to make room in\n      memory, keep flushing until we hit this mark. Defaults to 35% of heap.\n      This value equal to hbase.regionserver.global.memstore.upperLimit causes\n      the minimum possible flushing to occur when updates are blocked due to\n      memstore limiting.\n    ",
+        "property_name" : "hbase.regionserver.global.memstore.lowerLimit",
+        "property_value" : "0.38",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.regionserver.global.memstore.upperLimit",
+      "StackConfigurations" : {
+        "property_description" : "Maximum size of all memstores in a region server before new\n      updates are blocked and flushes are forced. Defaults to 40% of heap\n    ",
+        "property_name" : "hbase.regionserver.global.memstore.upperLimit",
+        "property_value" : "0.4",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.regionserver.handler.count",
+      "StackConfigurations" : {
+        "property_description" : "Count of RPC Listener instances spun up on RegionServers.\n    Same property is used by the Master for count of master handlers.\n    Default is 10.\n    ",
+        "property_name" : "hbase.regionserver.handler.count",
+        "property_value" : "60",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.rootdir",
+      "StackConfigurations" : {
+        "property_description" : "The directory shared by region servers and into\n    which HBase persists.  The URL should be 'fully-qualified'\n    to include the filesystem scheme.  For example, to specify the\n    HDFS directory '/hbase' where the HDFS instance's namenode is\n    running at namenode.example.org on port 9000, set this value to:\n    hdfs://namenode.example.org:9000/hbase.  By default HBase writes\n    into /tmp.  Change this configuration else all data will be lost\n    on machine restart.\n    ",
+        "property_name" : "hbase.rootdir",
+        "property_value" : "hdfs://localhost:8020/apps/hbase/data",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.security.authentication",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hbase.security.authentication",
+        "property_value" : "simple",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.security.authorization",
+      "StackConfigurations" : {
+        "property_description" : "Enables HBase authorization. Set the value of this property to false to disable HBase authorization.\n    ",
+        "property_name" : "hbase.security.authorization",
+        "property_value" : "false",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.superuser",
+      "StackConfigurations" : {
+        "property_description" : "List of users or groups (comma-separated), who are allowed\n    full privileges, regardless of stored ACLs, across the cluster.\n    Only used when HBase security is enabled.\n    ",
         "property_name" : "hbase.superuser",
+        "property_value" : "hbase",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.tmp.dir",
+      "StackConfigurations" : {
+        "property_description" : "Temporary directory on the local filesystem.\n    Change this setting to point to a location more permanent\n    than '/tmp' (The '/tmp' directory is often cleared on\n    machine restart).\n    ",
+        "property_name" : "hbase.tmp.dir",
+        "property_value" : "/hadoop/hbase",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.zookeeper.property.clientPort",
+      "StackConfigurations" : {
+        "property_description" : "Property from ZooKeeper's config zoo.cfg.\n    The port at which the clients will connect.\n    ",
+        "property_name" : "hbase.zookeeper.property.clientPort",
+        "property_value" : "2181",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.zookeeper.quorum",
+      "StackConfigurations" : {
+        "property_description" : "Comma separated list of servers in the ZooKeeper Quorum.\n    For example, \"host1.mydomain.com,host2.mydomain.com,host3.mydomain.com\".\n    By default this is set to localhost for local and pseudo-distributed modes\n    of operation. For a fully-distributed setup, this should be set to a full\n    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh\n    this is the list of servers which we will start/stop ZooKeeper on.\n    ",
+        "property_name" : "hbase.zookeeper.quorum",
+        "property_value" : "localhost",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase.zookeeper.useMulti",
+      "StackConfigurations" : {
+        "property_description" : "Instructs HBase to make use of ZooKeeper's multi-update functionality.\n    This allows certain ZooKeeper operations to complete more quickly and prevents some issues\n    with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).·\n    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+\n    and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will\n    not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).\n    ",
+        "property_name" : "hbase.zookeeper.useMulti",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_conf_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_conf_dir",
       "StackConfigurations" : {
         "property_description" : "Config Directory for HBase.",
+        "property_name" : "hbase_conf_dir",
         "property_value" : "/etc/hbase",
-        "stack_version" : "1.3.0",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_conf_dir",
+      "StackConfigurations" : {
+        "property_description" : "HBase conf dir.",
         "property_name" : "hbase_conf_dir",
+        "property_value" : "/etc/hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.zookeeper.property.clientPort",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_hdfs_root_dir",
       "StackConfigurations" : {
-        "property_description" : "Property from ZooKeeper's config zoo.cfg.\n    The port at which the clients will connect.\n    ",
-        "property_value" : "2181",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase.zookeeper.property.clientPort",
+        "property_description" : "HBase Relative Path to HDFS.",
+        "property_name" : "hbase_hdfs_root_dir",
+        "property_value" : "/apps/hbase/data",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_log_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_log_dir",
       "StackConfigurations" : {
         "property_description" : "Log Directories for HBase.",
-        "property_value" : "/var/log/hbase",
-        "stack_version" : "1.3.0",
         "property_name" : "hbase_log_dir",
+        "property_value" : "/var/log/hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_tmp_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_master_heapsize",
       "StackConfigurations" : {
-        "property_description" : "Hbase temp directory",
-        "property_value" : "/var/log/hbase",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase_tmp_dir",
+        "property_description" : "HBase Master Heap Size",
+        "property_name" : "hbase_master_heapsize",
+        "property_value" : "1024",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hfile_max_keyvalue_size",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_pid_dir",
       "StackConfigurations" : {
-        "property_description" : "HBase Client Maximum key-value Size",
-        "property_value" : "10485760",
-        "stack_version" : "1.3.0",
-        "property_name" : "hfile_max_keyvalue_size",
+        "property_description" : "Log Directories for HBase.",
+        "property_name" : "hbase_pid_dir",
+        "property_value" : "/var/run/hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_regionserver_heapsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_regionserver_heapsize",
       "StackConfigurations" : {
         "property_description" : "Log Directories for HBase.",
-        "property_value" : "1024",
-        "stack_version" : "1.3.0",
         "property_name" : "hbase_regionserver_heapsize",
+        "property_value" : "1024",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hregion_majorcompaction",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hbase_user",
       "StackConfigurations" : {
-        "property_description" : "HBase Major Compaction.",
-        "property_value" : "86400000",
-        "stack_version" : "1.3.0",
-        "property_name" : "hregion_majorcompaction",
+        "property_description" : "HBase User Name.",
+        "property_name" : "hbase_user",
+        "property_value" : "hbase",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/client_scannercaching",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hdfs_enable_shortcircuit_read",
       "StackConfigurations" : {
-        "property_description" : "Base Client Scanner Caching",
-        "property_value" : "100",
-        "stack_version" : "1.3.0",
-        "property_name" : "client_scannercaching",
+        "property_description" : "HDFS Short Circuit Read",
+        "property_name" : "hdfs_enable_shortcircuit_read",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.masterregion.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hdfs_support_append",
       "StackConfigurations" : {
-        "property_description" : "ACL for HMasterRegionInterface protocol implementations\n    (for HRegionServers communicating with HMaster)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
-        "property_name" : "security.masterregion.protocol.acl",
+        "property_description" : "HDFS append support",
+        "property_name" : "hdfs_support_append",
+        "property_value" : "true",
         "service_name" : "HBASE",
-        "type" : "hbase-policy.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.cluster.distributed",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hfile.block.cache.size",
       "StackConfigurations" : {
-        "property_description" : "The mode the cluster will be in. Possible values are\n      false for standalone mode and true for distributed mode.  If\n      false, startup will run all HBase and ZooKeeper daemons together\n      in the one JVM.\n    ",
-        "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase.cluster.distributed",
+        "property_description" : "\n        Percentage of maximum heap (-Xmx setting) to allocate to block cache\n        used by HFile/StoreFile. Default of 0.25 means allocate 25%.\n        Set to 0 to disable but it's not recommended.\n    ",
+        "property_name" : "hfile.block.cache.size",
+        "property_value" : "0.40",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.master.lease.thread.wakefrequency",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hfile_blockcache_size",
       "StackConfigurations" : {
-        "property_description" : "The interval between checks for expired region server leases.\n    This value has been reduced due to the other reduced values above so that\n    the master will notice a dead region server sooner. The default is 15 seconds.\n    ",
-        "property_value" : "3000",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase.master.lease.thread.wakefrequency",
+        "property_description" : "HFile block cache size.",
+        "property_name" : "hfile_blockcache_size",
+        "property_value" : "0.40",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/regionserver_handlers",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hfile_max_keyvalue_size",
       "StackConfigurations" : {
-        "property_description" : "HBase RegionServer Handler",
-        "property_value" : "30",
-        "stack_version" : "1.3.0",
-        "property_name" : "regionserver_handlers",
+        "property_description" : "HBase Client Maximum key-value Size",
+        "property_name" : "hfile_max_keyvalue_size",
+        "property_value" : "10485760",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase.regionserver.optionalcacheflushinterval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hregion_blockmultiplier",
       "StackConfigurations" : {
-        "property_description" : "\n      Amount of time to wait since the last time a region was flushed before\n      invoking an optional cache flush. Default 60,000.\n    ",
-        "property_value" : "10000",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase.regionserver.optionalcacheflushinterval",
+        "property_description" : "HBase Region Block Multiplier",
+        "property_name" : "hregion_blockmultiplier",
+        "property_value" : "2",
         "service_name" : "HBASE",
-        "type" : "hbase-site.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hbase_pid_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hregion_majorcompaction",
       "StackConfigurations" : {
-        "property_description" : "Log Directories for HBase.",
-        "property_value" : "/var/run/hbase",
-        "stack_version" : "1.3.0",
-        "property_name" : "hbase_pid_dir",
+        "property_description" : "The time between major compactions of all HStoreFiles in a region. Set to 0 to disable automated major compactions.",
+        "property_name" : "hregion_majorcompaction",
+        "property_value" : "604800000",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hstore_blockingstorefiles",
+      "StackConfigurations" : {
+        "property_description" : "HStore blocking storefiles.",
+        "property_name" : "hstore_blockingstorefiles",
+        "property_value" : "10",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/hstorefile_maxsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hstore_compactionthreshold",
+      "StackConfigurations" : {
+        "property_description" : "HBase HStore compaction threshold.",
+        "property_name" : "hstore_compactionthreshold",
+        "property_value" : "3",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/hstorefile_maxsize",
       "StackConfigurations" : {
         "property_description" : "Maximum HStoreFile Size",
-        "property_value" : "1073741824",
-        "stack_version" : "1.3.0",
         "property_name" : "hstorefile_maxsize",
+        "property_value" : "10737418240",
         "service_name" : "HBASE",
-        "type" : "global.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE/configurations/security.admin.protocol.acl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/regionserver_handlers",
+      "StackConfigurations" : {
+        "property_description" : "HBase RegionServer Handler",
+        "property_name" : "regionserver_handlers",
+        "property_value" : "60",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/regionserver_memstore_lab",
+      "StackConfigurations" : {
+        "property_description" : "Region Server memstore.",
+        "property_name" : "regionserver_memstore_lab",
+        "property_value" : "true",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/regionserver_memstore_lowerlimit",
+      "StackConfigurations" : {
+        "property_description" : "Region Server memstore lower limit.",
+        "property_name" : "regionserver_memstore_lowerlimit",
+        "property_value" : "0.38",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/regionserver_memstore_upperlimit",
+      "StackConfigurations" : {
+        "property_description" : "Region Server memstore upper limit.",
+        "property_name" : "regionserver_memstore_upperlimit",
+        "property_value" : "0.4",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/security.admin.protocol.acl",
       "StackConfigurations" : {
         "property_description" : "ACL for HMasterInterface protocol implementation (ie. \n    clients talking to HMaster for admin operations).\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
         "property_name" : "security.admin.protocol.acl",
+        "property_value" : "*",
         "service_name" : "HBASE",
-        "type" : "hbase-policy.xml",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/security.client.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for HRegionInterface protocol implementations (ie. \n    clients talking to HRegionServers)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.client.protocol.acl",
+        "property_value" : "*",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/security.masterregion.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for HMasterRegionInterface protocol implementations\n    (for HRegionServers communicating with HMaster)\n    The ACL is a comma-separated list of user and group names. The user and \n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\". \n    A special value of \"*\" means all users are allowed.",
+        "property_name" : "security.masterregion.protocol.acl",
+        "property_value" : "*",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-policy.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/zookeeper.session.timeout",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper session timeout.\n      HBase passes this to the zk quorum as suggested maximum time for a\n      session (This setting becomes zookeeper's 'maxSessionTimeout').  See\n      http://hadoop.apache.org/zookeeper/docs/current/zookeeperProgrammers.html#ch_zkSessions\n      \"The client sends a requested timeout, the server responds with the\n      timeout that it can give the client. \" In milliseconds.\n    ",
+        "property_name" : "zookeeper.session.timeout",
+        "property_value" : "30000",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/zookeeper.znode.parent",
+      "StackConfigurations" : {
+        "property_description" : "Root ZNode for HBase in ZooKeeper. All of HBase's ZooKeeper\n      files that are configured with a relative path will go under this node.\n      By default, all of HBase's ZooKeeper file path are configured with a\n      relative path, so they will all go under this directory unless changed.\n    ",
+        "property_name" : "zookeeper.znode.parent",
+        "property_value" : "/hbase-unsecure",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "hbase-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/2.0.1/stackServices/HBASE/configurations/zookeeper_sessiontimeout",
+      "StackConfigurations" : {
+        "property_description" : "ZooKeeper Session Timeout",
+        "property_name" : "zookeeper_sessiontimeout",
+        "property_value" : "30000",
+        "service_name" : "HBASE",
+        "stack_name" : "HDP",
+        "stack_version" : "2.0.1",
+        "type" : "global.xml"
       }
     }
   ]


[08/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/MAPREDUCE.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/MAPREDUCE.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/MAPREDUCE.json
index 4ab1584..2b15b0e 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/MAPREDUCE.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/MAPREDUCE.json
@@ -1,772 +1,1324 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.tasks.speculative.execution",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/ambari.mapred.child.java.opts.memory",
       "StackConfigurations" : {
-        "property_description" : "If true, then multiple instances of some reduce tasks\n               may be executed in parallel.",
-        "property_value" : "false",
+        "property_description" : "Java options Memory for the TaskTracker child processes",
+        "property_name" : "ambari.mapred.child.java.opts.memory",
+        "property_value" : "768",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.reduce.tasks.speculative.execution",
-        "filename" : "mapred-site.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/hadoop.job.history.user.location",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hadoop.job.history.user.location",
+        "property_value" : "none",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/hadoop_heapsize",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop maximum Java heap size",
+        "property_name" : "hadoop_heapsize",
+        "property_value" : "1024",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.factor",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_name" : "io.sort.factor",
+        "property_value" : "100",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.mb",
+      "StackConfigurations" : {
+        "property_description" : "\n      The total amount of Map-side buffer memory to use while sorting files\n    ",
+        "property_name" : "io.sort.mb",
+        "property_value" : "200",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.record.percent",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_name" : "io.sort.record.percent",
+        "property_value" : ".2",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.spill.percent",
+      "StackConfigurations" : {
+        "property_description" : "Percentage of sort buffer used for record collection",
+        "property_name" : "io.sort.spill.percent",
+        "property_value" : "0.9",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io_sort_mb",
+      "StackConfigurations" : {
+        "property_description" : "The total amount of Map-side buffer memory to use while sorting files (Expert-only configuration).",
+        "property_name" : "io_sort_mb",
+        "property_value" : "200",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io_sort_spill_percent",
+      "StackConfigurations" : {
+        "property_description" : "Percentage of sort buffer used for record collection (Expert-only configuration.",
+        "property_name" : "io_sort_spill_percent",
+        "property_value" : "0.9",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jetty.connector",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_name" : "jetty.connector",
+        "property_value" : "org.mortbay.jetty.nio.SelectChannelConnector",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jtnode_heapsize",
+      "StackConfigurations" : {
+        "property_description" : "Maximum Java heap size for JobTracker in MB (Java option -Xmx)",
+        "property_name" : "jtnode_heapsize",
+        "property_value" : "1024",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jtnode_opt_maxnewsize",
+      "StackConfigurations" : {
+        "property_description" : "Max New size.",
+        "property_name" : "jtnode_opt_maxnewsize",
+        "property_value" : "200",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jtnode_opt_newsize",
+      "StackConfigurations" : {
+        "property_description" : "Mem New Size.",
+        "property_name" : "jtnode_opt_newsize",
+        "property_value" : "200",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/lzo_enabled",
+      "StackConfigurations" : {
+        "property_description" : "LZO compression enabled",
+        "property_name" : "lzo_enabled",
+        "property_value" : "true",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-init-accept-jobs-factor",
+      "StackConfigurations" : {
+        "property_description" : "The default multipe of (maximum-system-jobs * queue-capacity) \n    used to determine the number of jobs which are accepted by the scheduler.  \n    ",
+        "property_name" : "mapred.capacity-scheduler.default-init-accept-jobs-factor",
+        "property_value" : "10",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-maximum-active-tasks-per-queue",
+      "StackConfigurations" : {
+        "property_description" : "The default maximum number of tasks, across all jobs in the \n    queue, which can be initialized concurrently. Once the queue's jobs exceed \n    this limit they will be queued on disk.  \n    ",
+        "property_name" : "mapred.capacity-scheduler.default-maximum-active-tasks-per-queue",
+        "property_value" : "200000",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-maximum-active-tasks-per-user",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-maximum-active-tasks-per-user",
       "StackConfigurations" : {
         "property_description" : "The default maximum number of tasks per-user, across all the of \n    the user's jobs in the queue, which can be initialized concurrently. Once \n    the user's jobs exceed this limit they will be queued on disk.  \n    ",
+        "property_name" : "mapred.capacity-scheduler.default-maximum-active-tasks-per-user",
         "property_value" : "100000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.default-maximum-active-tasks-per-user",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-minimum-user-limit-percent",
+      "StackConfigurations" : {
+        "property_description" : "The percentage of the resources limited to a particular user\n      for the job queue at any given point of time by default.\n    ",
+        "property_name" : "mapred.capacity-scheduler.default-minimum-user-limit-percent",
+        "property_value" : "100",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.init-accept-jobs-factor",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-supports-priority",
       "StackConfigurations" : {
-        "property_description" : "The multipe of (maximum-system-jobs * queue-capacity) used to \n    determine the number of jobs which are accepted by the scheduler.  \n    ",
-        "property_value" : "10",
+        "property_description" : "If true, priorities of jobs will be taken into \n      account in scheduling decisions by default in a job queue.\n    ",
+        "property_name" : "mapred.capacity-scheduler.default-supports-priority",
+        "property_value" : "false",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.queue.default.init-accept-jobs-factor",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-user-limit-factor",
+      "StackConfigurations" : {
+        "property_description" : "The default multiple of queue-capacity which is used to \n    determine the amount of slots a single user can consume concurrently.\n    ",
+        "property_name" : "mapred.capacity-scheduler.default-user-limit-factor",
+        "property_value" : "1",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.init-poll-interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.init-poll-interval",
       "StackConfigurations" : {
         "property_description" : "The amount of time in miliseconds which is used to poll \n    the job queues for jobs to initialize.\n    ",
+        "property_name" : "mapred.capacity-scheduler.init-poll-interval",
         "property_value" : "5000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.init-poll-interval",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.init-worker-threads",
+      "StackConfigurations" : {
+        "property_description" : "Number of worker threads which would be used by\n    Initialization poller to initialize jobs in a set of queue.\n    If number mentioned in property is equal to number of job queues\n    then a single thread would initialize jobs in a queue. If lesser\n    then a thread would get a set of queues assigned. If the number\n    is greater then number of threads would be equal to number of \n    job queues.\n    ",
+        "property_name" : "mapred.capacity-scheduler.init-worker-threads",
+        "property_value" : "5",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.history.server.embedded",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.maximum-system-jobs",
       "StackConfigurations" : {
-        "property_description" : "Should job history server be embedded within Job tracker\nprocess",
-        "property_value" : "false",
+        "property_description" : "Maximum number of jobs in the system which can be initialized,\n     concurrently, by the CapacityScheduler.\n    ",
+        "property_name" : "mapred.capacity-scheduler.maximum-system-jobs",
+        "property_value" : "3000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapreduce.history.server.embedded",
-        "filename" : "mapred-site.xml",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.capacity",
+      "StackConfigurations" : {
+        "property_description" : "Percentage of the number of slots in the cluster that are\n      to be available for jobs in this queue.\n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.capacity",
+        "property_value" : "100",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.blacklist.fault-bucket-width",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.init-accept-jobs-factor",
       "StackConfigurations" : {
-        "property_description" : "\n    15-minute bucket size (value is in minutes)\n  ",
-        "property_value" : "15",
+        "property_description" : "The multipe of (maximum-system-jobs * queue-capacity) used to \n    determine the number of jobs which are accepted by the scheduler.  \n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.init-accept-jobs-factor",
+        "property_value" : "10",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.jobtracker.blacklist.fault-bucket-width",
-        "filename" : "mapred-site.xml",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-capacity",
+      "StackConfigurations" : {
+        "property_description" : "\n\tmaximum-capacity defines a limit beyond which a queue cannot use the capacity of the cluster.\n\tThis provides a means to limit how much excess capacity a queue can use. By default, there is no limit.\n\tThe maximum-capacity of a queue can only be greater than or equal to its minimum capacity.\n        Default value of -1 implies a queue can use complete capacity of the cluster.\n\n        This property could be to curtail certain jobs which are long running in nature from occupying more than a \n        certain percentage of the cluster, which in the absence of pre-emption, could lead to capacity guarantees of \n        other queues being affected.\n        \n        One important thing to note is that maximum-capacity is a percentage , so based on the cluster's capacity\n        the max capacity would change. So if large no of nodes or racks get added to the cluster , max Capacity in \n        absolute terms would increase accordingly.\n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-capacity",
+        "property_value" : "-1",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.map.tasks.speculative.execution",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks",
       "StackConfigurations" : {
-        "property_description" : "If true, then multiple instances of some map tasks\n               may be executed in parallel.",
-        "property_value" : "false",
+        "property_description" : "The maximum number of tasks, across all jobs in the queue, \n    which can be initialized concurrently. Once the queue's jobs exceed this \n    limit they will be queued on disk.  \n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks",
+        "property_value" : "200000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.map.tasks.speculative.execution",
-        "filename" : "mapred-site.xml",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks-per-user",
+      "StackConfigurations" : {
+        "property_description" : "The maximum number of tasks per-user, across all the of the \n    user's jobs in the queue, which can be initialized concurrently. Once the \n    user's jobs exceed this limit they will be queued on disk.  \n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks-per-user",
+        "property_value" : "100000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.minimum-user-limit-percent",
+      "StackConfigurations" : {
+        "property_description" : " Each queue enforces a limit on the percentage of resources \n    allocated to a user at any given time, if there is competition for them. \n    This user limit can vary between a minimum and maximum value. The former\n    depends on the number of users who have submitted jobs, and the latter is\n    set to this property value. For example, suppose the value of this \n    property is 25. If two users have submitted jobs to a queue, no single \n    user can use more than 50% of the queue resources. If a third user submits\n    a job, no single user can use more than 33% of the queue resources. With 4 \n    or more users, no user can use more than 25% of the queue's resources. A \n    value of 100 implies no user limits are imposed. \n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.minimum-user-limit-percent",
+        "property_value" : "100",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.supports-priority",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.supports-priority",
       "StackConfigurations" : {
         "property_description" : "If true, priorities of jobs will be taken into \n      account in scheduling decisions.\n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.supports-priority",
         "property_value" : "false",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.queue.default.supports-priority",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "capacity-scheduler.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.user-limit-factor",
+      "StackConfigurations" : {
+        "property_description" : "The multiple of the queue capacity which can be configured to \n    allow a single user to acquire more slots. \n    ",
+        "property_name" : "mapred.capacity-scheduler.queue.default.user-limit-factor",
+        "property_value" : "1",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "capacity-scheduler.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.blacklist.fault-timeout-window",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.child.java.opts",
       "StackConfigurations" : {
-        "property_description" : "\n    3-hour sliding window (value is in minutes)\n  ",
-        "property_value" : "180",
+        "property_description" : "Java options for the TaskTracker child processes",
+        "property_name" : "mapred.child.java.opts",
+        "property_value" : "-server -Xmx${ambari.mapred.child.java.opts.memory}m -Djava.net.preferIPv4Stack=true",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.jobtracker.blacklist.fault-timeout-window",
-        "filename" : "mapred-site.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.child.root.logger",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "mapred.child.root.logger",
+        "property_value" : "INFO,TLA",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.persist.jobstatus.active",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.cluster.map.memory.mb",
       "StackConfigurations" : {
-        "property_description" : "Indicates if persistency of job status information is\n  active or not.\n  ",
-        "property_value" : "false",
+        "property_description" : "\n      The virtual memory size of a single Map slot in the MapReduce framework\n    ",
+        "property_name" : "mapred.cluster.map.memory.mb",
+        "property_value" : "1536",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.tracker.persist.jobstatus.active",
-        "filename" : "mapred-site.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.cluster.max.map.memory.mb",
+      "StackConfigurations" : {
+        "property_description" : "\n      Upper limit on virtual memory size for a single Map task of any MapReduce job\n    ",
+        "property_name" : "mapred.cluster.max.map.memory.mb",
+        "property_value" : "6144",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-supports-priority",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.cluster.max.reduce.memory.mb",
       "StackConfigurations" : {
-        "property_description" : "If true, priorities of jobs will be taken into \n      account in scheduling decisions by default in a job queue.\n    ",
-        "property_value" : "false",
+        "property_description" : "\n      Upper limit on virtual memory size for a single Reduce task of any MapReduce job\n    ",
+        "property_name" : "mapred.cluster.max.reduce.memory.mb",
+        "property_value" : "4096",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.default-supports-priority",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.cluster.reduce.memory.mb",
+      "StackConfigurations" : {
+        "property_description" : "\n      The virtual memory size of a single Reduce slot in the MapReduce framework\n    ",
+        "property_name" : "mapred.cluster.reduce.memory.mb",
+        "property_value" : "2048",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-user-limit-factor",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.healthChecker.interval",
       "StackConfigurations" : {
-        "property_description" : "The default multiple of queue-capacity which is used to \n    determine the amount of slots a single user can consume concurrently.\n    ",
+        "property_description" : null,
+        "property_name" : "mapred.healthChecker.interval",
+        "property_value" : "135000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.healthChecker.script.path",
+      "StackConfigurations" : {
+        "property_description" : "\n      Directory path to view job status\n    ",
+        "property_name" : "mapred.healthChecker.script.path",
+        "property_value" : "/etc/hadoop/conf/health_check",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.healthChecker.script.timeout",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "mapred.healthChecker.script.timeout",
+        "property_value" : "60000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.hosts",
+      "StackConfigurations" : {
+        "property_description" : "\n      Names a file that contains the list of nodes that may\n      connect to the jobtracker.  If the value is empty, all hosts are\n      permitted.\n    ",
+        "property_name" : "mapred.hosts",
+        "property_value" : "/etc/hadoop/conf/mapred.include",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.hosts.exclude",
+      "StackConfigurations" : {
+        "property_description" : "\n      Names a file that contains the list of hosts that\n      should be excluded by the jobtracker.  If the value is empty, no\n      hosts are excluded.\n    ",
+        "property_name" : "mapred.hosts.exclude",
+        "property_value" : "/etc/hadoop/conf/mapred.exclude",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.inmem.merge.threshold",
+      "StackConfigurations" : {
+        "property_description" : "The threshold, in terms of the number of files\n      for the in-memory merge process. When we accumulate threshold number of files\n      we initiate the in-memory merge and spill to disk. A value of 0 or less than\n      0 indicates we want to DON'T have any threshold and instead depend only on\n      the ramfs's memory consumption to trigger the merge.\n    ",
+        "property_name" : "mapred.inmem.merge.threshold",
+        "property_value" : "1000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.map.memory.mb",
+      "StackConfigurations" : {
+        "property_description" : "\n      Virtual memory for single Map task\n    ",
+        "property_name" : "mapred.job.map.memory.mb",
+        "property_value" : "1536",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.reduce.input.buffer.percent",
+      "StackConfigurations" : {
+        "property_description" : "The percentage of memory- relative to the maximum heap size- to\n      retain map outputs during the reduce. When the shuffle is concluded, any\n      remaining map outputs in memory must consume less than this threshold before\n      the reduce can begin.\n    ",
+        "property_name" : "mapred.job.reduce.input.buffer.percent",
+        "property_value" : "0.0",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.reduce.memory.mb",
+      "StackConfigurations" : {
+        "property_description" : "\n      Virtual memory for single Reduce task\n    ",
+        "property_name" : "mapred.job.reduce.memory.mb",
+        "property_value" : "2048",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.reuse.jvm.num.tasks",
+      "StackConfigurations" : {
+        "property_description" : "\n      How many tasks to run per jvm. If set to -1, there is no limit\n    ",
+        "property_name" : "mapred.job.reuse.jvm.num.tasks",
         "property_value" : "1",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.default-user-limit-factor",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.shuffle.input.buffer.percent",
+      "StackConfigurations" : {
+        "property_description" : "The percentage of memory to be allocated from the maximum heap\n      size to storing map outputs during the shuffle.\n    ",
+        "property_name" : "mapred.job.shuffle.input.buffer.percent",
+        "property_value" : "0.7",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.shuffle.merge.percent",
+      "StackConfigurations" : {
+        "property_description" : "The usage threshold at which an in-memory merge will be\n      initiated, expressed as a percentage of the total memory allocated to\n      storing in-memory map outputs, as defined by\n      mapred.job.shuffle.input.buffer.percent.\n    ",
+        "property_name" : "mapred.job.shuffle.merge.percent",
+        "property_value" : "0.66",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker",
+      "StackConfigurations" : {
+        "property_description" : "JobTracker address",
+        "property_name" : "mapred.job.tracker",
+        "property_value" : "localhost:50300",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.record.percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.handler.count",
+      "StackConfigurations" : {
+        "property_description" : "\n      The number of server threads for the JobTracker. This should be roughly\n      4% of the number of tasktracker nodes.\n    ",
+        "property_name" : "mapred.job.tracker.handler.count",
+        "property_value" : "50",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.history.completed.location",
       "StackConfigurations" : {
         "property_description" : "No description",
-        "property_value" : ".2",
+        "property_name" : "mapred.job.tracker.history.completed.location",
+        "property_value" : "/mapred/history/done",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "io.sort.record.percent",
-        "filename" : "mapred-site.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.http.address",
+      "StackConfigurations" : {
+        "property_description" : "JobTracker host and http port address",
+        "property_name" : "mapred.job.tracker.http.address",
+        "property_value" : "localhost:50030",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.retirejob.check",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.persist.jobstatus.active",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "10000",
+        "property_description" : "Indicates if persistency of job status information is\n      active or not.\n    ",
+        "property_name" : "mapred.job.tracker.persist.jobstatus.active",
+        "property_value" : "false",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.jobtracker.retirejob.check",
-        "filename" : "mapred-site.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.persist.jobstatus.dir",
+      "StackConfigurations" : {
+        "property_description" : "The directory where the job status information is persisted\n      in a file system to be available after it drops of the memory queue and\n      between jobtracker restarts.\n    ",
+        "property_name" : "mapred.job.tracker.persist.jobstatus.dir",
+        "property_value" : "/mapred/jobstatus",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.system.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.persist.jobstatus.hours",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "/mapred/system",
+        "property_description" : "The number of hours job status information is persisted in DFS.\n      The job status information will be available after it drops of the memory\n      queue and between jobtracker restarts. With a zero value the job status\n      information is not persisted at all in DFS.\n    ",
+        "property_name" : "mapred.job.tracker.persist.jobstatus.hours",
+        "property_value" : "1",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.system.dir",
-        "filename" : "mapred-site.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.blacklist.fault-bucket-width",
+      "StackConfigurations" : {
+        "property_description" : "\n      15-minute bucket size (value is in minutes)\n    ",
+        "property_name" : "mapred.jobtracker.blacklist.fault-bucket-width",
+        "property_value" : "15",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-capacity",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.blacklist.fault-timeout-window",
       "StackConfigurations" : {
-        "property_description" : "\n\tmaximum-capacity defines a limit beyond which a queue cannot use the capacity of the cluster.\n\tThis provides a means to limit how much excess capacity a queue can use. By default, there is no limit.\n\tThe maximum-capacity of a queue can only be greater than or equal to its minimum capacity.\n        Default value of -1 implies a queue can use complete capacity of the cluster.\n\n        This property could be to curtail certain jobs which are long running in nature from occupying more than a \n        certain percentage of the cluster, which in the absence of pre-emption, could lead to capacity guarantees of \n        other queues being affected.\n        \n        One important thing to note is that maximum-capacity is a percentage , so based on the cluster's capacity\n        the max capacity would change. So if large no of nodes or racks get added to the cluster , max Capacity in \n        absolute terms would increase accordingly.\n    ",
-        "property_value" : "-1",
+        "property_description" : "\n      3-hour sliding window (value is in minutes)\n    ",
+        "property_name" : "mapred.jobtracker.blacklist.fault-timeout-window",
+        "property_value" : "180",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-capacity",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.completeuserjobs.maximum",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "mapred.jobtracker.completeuserjobs.maximum",
+        "property_value" : "0",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jetty.connector",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.maxtasks.per.job",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "org.mortbay.jetty.nio.SelectChannelConnector",
+        "property_description" : "The maximum number of tasks for a single job.\n      A value of -1 indicates that there is no maximum.  ",
+        "property_name" : "mapred.jobtracker.maxtasks.per.job",
+        "property_value" : "-1",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "jetty.connector",
-        "filename" : "mapred-site.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.restart.recover",
+      "StackConfigurations" : {
+        "property_description" : "\"true\" to enable (job) recovery upon restart,\n      \"false\" to start afresh\n    ",
+        "property_name" : "mapred.jobtracker.restart.recover",
+        "property_value" : "false",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-maximum-active-tasks-per-queue",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.retirejob.check",
       "StackConfigurations" : {
-        "property_description" : "The default maximum number of tasks, across all jobs in the \n    queue, which can be initialized concurrently. Once the queue's jobs exceed \n    this limit they will be queued on disk.  \n    ",
-        "property_value" : "200000",
+        "property_description" : null,
+        "property_name" : "mapred.jobtracker.retirejob.check",
+        "property_value" : "10000",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.default-maximum-active-tasks-per-queue",
-        "filename" : "capacity-scheduler.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.retirejob.interval",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "mapred.jobtracker.retirejob.interval",
+        "property_value" : "21600000",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/scheduler_name",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.taskScheduler",
       "StackConfigurations" : {
-        "property_description" : "MapRed Capacity Scheduler.",
+        "property_description" : null,
+        "property_name" : "mapred.jobtracker.taskScheduler",
         "property_value" : "org.apache.hadoop.mapred.CapacityTaskScheduler",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "scheduler_name",
-        "filename" : "global.xml",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.local.dir",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_name" : "mapred.local.dir",
+        "property_value" : "/hadoop/mapred",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.map.output.compression.codec",
+      "StackConfigurations" : {
+        "property_description" : "If the map outputs are compressed, how should they be\n      compressed\n    ",
+        "property_name" : "mapred.map.output.compression.codec",
+        "property_value" : "org.apache.hadoop.io.compress.SnappyCodec",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.map.tasks.speculative.execution",
+      "StackConfigurations" : {
+        "property_description" : "If true, then multiple instances of some map tasks\n      may be executed in parallel.",
+        "property_name" : "mapred.map.tasks.speculative.execution",
+        "property_value" : "false",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.max.tracker.blacklists",
+      "StackConfigurations" : {
+        "property_description" : "\n      if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted\n    ",
+        "property_name" : "mapred.max.tracker.blacklists",
+        "property_value" : "16",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.reuse.jvm.num.tasks",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.output.compression.type",
       "StackConfigurations" : {
-        "property_description" : "\n    How many tasks to run per jvm. If set to -1, there is no limit\n  ",
-        "property_value" : "1",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.reuse.jvm.num.tasks",
-        "filename" : "mapred-site.xml",
+        "property_description" : "If the job outputs are to compressed as SequenceFiles, how should\n      they be compressed? Should be one of NONE, RECORD or BLOCK.\n    ",
+        "property_name" : "mapred.output.compression.type",
+        "property_value" : "BLOCK",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.max.tracker.blacklists",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.default.acl-administer-jobs",
       "StackConfigurations" : {
-        "property_description" : "\n    if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted\n  ",
-        "property_value" : "16",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.max.tracker.blacklists",
-        "filename" : "mapred-site.xml",
+        "property_description" : null,
+        "property_name" : "mapred.queue.default.acl-administer-jobs",
+        "property_value" : "*",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-queue-acls.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.default.acl-submit-job",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.default.acl-submit-job",
       "StackConfigurations" : {
         "property_description" : null,
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
         "property_name" : "mapred.queue.default.acl-submit-job",
-        "filename" : "mapred-queue-acls.xml",
+        "property_value" : "*",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-queue-acls.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.names",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.names",
       "StackConfigurations" : {
         "property_description" : " Comma separated list of queues configured for this jobtracker.",
-        "property_value" : "default",
-        "stack_version" : "1.3.0",
         "property_name" : "mapred.queue.names",
-        "filename" : "mapred-site.xml",
+        "property_value" : "default",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
-      }
-    },
-    {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.reduce.input.buffer.percent",
-      "StackConfigurations" : {
-        "property_description" : "The percentage of memory- relative to the maximum heap size- to\n  retain map outputs during the reduce. When the shuffle is concluded, any\n  remaining map outputs in memory must consume less than this threshold before\n  the reduce can begin.\n  ",
-        "property_value" : "0.0",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.reduce.input.buffer.percent",
-        "filename" : "mapred-site.xml",
-        "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobtracker.split.metainfo.maxsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.parallel.copies",
       "StackConfigurations" : {
-        "property_description" : "If the size of the split metainfo file is larger than this, the JobTracker will fail the job during\n    initialize.\n   ",
-        "property_value" : "50000000",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapreduce.jobtracker.split.metainfo.maxsize",
-        "filename" : "mapred-site.xml",
+        "property_description" : "No description",
+        "property_name" : "mapred.reduce.parallel.copies",
+        "property_value" : "30",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/hadoop.job.history.user.location",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.slowstart.completed.maps",
       "StackConfigurations" : {
         "property_description" : null,
-        "property_value" : "none",
-        "stack_version" : "1.3.0",
-        "property_name" : "hadoop.job.history.user.location",
-        "filename" : "mapred-site.xml",
+        "property_name" : "mapred.reduce.slowstart.completed.maps",
+        "property_value" : "0.05",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.completeuserjobs.maximum",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.tasks.speculative.execution",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "100",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.jobtracker.completeuserjobs.maximum",
-        "filename" : "mapred-site.xml",
+        "property_description" : "If true, then multiple instances of some reduce tasks\n      may be executed in parallel.",
+        "property_name" : "mapred.reduce.tasks.speculative.execution",
+        "property_value" : "false",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.output.compression.type",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.system.dir",
       "StackConfigurations" : {
-        "property_description" : "If the job outputs are to compressed as SequenceFiles, how should\n               they be compressed? Should be one of NONE, RECORD or BLOCK.\n  ",
-        "property_value" : "BLOCK",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.output.compression.type",
-        "filename" : "mapred-site.xml",
+        "property_description" : "Path on the HDFS where where the MapReduce framework stores system files",
+        "property_name" : "mapred.system.dir",
+        "property_value" : "/mapred/system",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.retirejob.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.task.timeout",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "0",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.jobtracker.retirejob.interval",
-        "filename" : "mapred-site.xml",
+        "property_description" : "The number of milliseconds before a task will be\n      terminated if it neither reads an input, writes an output, nor\n      updates its status string.\n    ",
+        "property_name" : "mapred.task.timeout",
+        "property_value" : "600000",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.healthChecker.interval",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.task.tracker.task-controller",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "135000",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.healthChecker.interval",
-        "filename" : "mapred-site.xml",
+        "property_description" : "\n      TaskController which is used to launch and manage task execution.\n    ",
+        "property_name" : "mapred.task.tracker.task-controller",
+        "property_value" : "org.apache.hadoop.mapred.DefaultTaskController",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jtnode_opt_newsize",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.tasktracker.map.tasks.maximum",
       "StackConfigurations" : {
-        "property_description" : "MapRed Capacity Scheduler.",
-        "property_value" : "200",
-        "stack_version" : "1.3.0",
-        "property_name" : "jtnode_opt_newsize",
-        "filename" : "global.xml",
+        "property_description" : "No description",
+        "property_name" : "mapred.tasktracker.map.tasks.maximum",
+        "property_value" : "4",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.task.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.tasktracker.reduce.tasks.maximum",
       "StackConfigurations" : {
-        "property_description" : "The number of milliseconds before a task will be\n  terminated if it neither reads an input, writes an output, nor\n  updates its status string.\n  ",
-        "property_value" : "600000",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.task.timeout",
-        "filename" : "mapred-site.xml",
+        "property_description" : "No description",
+        "property_name" : "mapred.tasktracker.reduce.tasks.maximum",
+        "property_value" : "2",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.persist.jobstatus.hours",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.tasktracker.tasks.sleeptime-before-sigkill",
       "StackConfigurations" : {
-        "property_description" : "The number of hours job status information is persisted in DFS.\n    The job status information will be available after it drops of the memory\n    queue and between jobtracker restarts. With a zero value the job status\n    information is not persisted at all in DFS.\n  ",
-        "property_value" : "1",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.tracker.persist.jobstatus.hours",
-        "filename" : "mapred-site.xml",
+        "property_description" : "Normally, this is the amount of time before killing\n      processes, and the recommended-default is 5.000 seconds - a value of\n      5000 here.  In this case, we are using it solely to blast tasks before\n      killing them, and killing them very quickly (1/4 second) to guarantee\n      that we do not leave VMs around for later jobs.\n    ",
+        "property_name" : "mapred.tasktracker.tasks.sleeptime-before-sigkill",
+        "property_value" : "250",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.shuffle.merge.percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.userlog.retain.hours",
       "StackConfigurations" : {
-        "property_description" : "The usage threshold at which an in-memory merge will be\n  initiated, expressed as a percentage of the total memory allocated to\n  storing in-memory map outputs, as defined by\n  mapred.job.shuffle.input.buffer.percent.\n  ",
-        "property_value" : "0.66",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.shuffle.merge.percent",
-        "filename" : "mapred-site.xml",
+        "property_description" : "\n      The maximum time, in hours, for which the user-logs are to be retained after the job completion.\n    ",
+        "property_name" : "mapred.userlog.retain.hours",
+        "property_value" : "24",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.restart.recover",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_child_java_opts_sz",
       "StackConfigurations" : {
-        "property_description" : "\"true\" to enable (job) recovery upon restart,\n               \"false\" to start afresh\n    ",
-        "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.jobtracker.restart.recover",
-        "filename" : "mapred-site.xml",
+        "property_description" : "Java options for the TaskTracker child processes.",
+        "property_name" : "mapred_child_java_opts_sz",
+        "property_value" : "768",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.handler.count",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_cluster_map_mem_mb",
       "StackConfigurations" : {
-        "property_description" : "\n    The number of server threads for the JobTracker. This should be roughly\n    4% of the number of tasktracker nodes.\n    ",
-        "property_value" : "50",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.tracker.handler.count",
-        "filename" : "mapred-site.xml",
+        "property_description" : "The virtual memory size of a single Map slot in the MapReduce framework",
+        "property_name" : "mapred_cluster_map_mem_mb",
+        "property_value" : "-1",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks-per-user",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_cluster_red_mem_mb",
       "StackConfigurations" : {
-        "property_description" : "The maximum number of tasks per-user, across all the of the \n    user's jobs in the queue, which can be initialized concurrently. Once the \n    user's jobs exceed this limit they will be queued on disk.  \n    ",
-        "property_value" : "100000",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks-per-user",
-        "filename" : "capacity-scheduler.xml",
+        "property_description" : "The virtual memory size of a single Reduce slot in the MapReduce framework",
+        "property_name" : "mapred_cluster_red_mem_mb",
+        "property_value" : "-1",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.tasktracker.group",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_job_map_mem_mb",
       "StackConfigurations" : {
-        "property_description" : "The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.",
-        "property_value" : "hadoop",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapreduce.tasktracker.group",
-        "filename" : "mapred-site.xml",
+        "property_description" : "Virtual memory for single Map task",
+        "property_name" : "mapred_job_map_mem_mb",
+        "property_value" : "-1",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_jobstatus_dir",
       "StackConfigurations" : {
-        "property_description" : "The maximum number of tasks, across all jobs in the queue, \n    which can be initialized concurrently. Once the queue's jobs exceed this \n    limit they will be queued on disk.  \n    ",
-        "property_value" : "200000",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks",
-        "filename" : "capacity-scheduler.xml",
+        "property_description" : "Job Status directory",
+        "property_name" : "mapred_jobstatus_dir",
+        "property_value" : "/mapred/jobstatus",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_local_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_local_dir",
       "StackConfigurations" : {
         "property_description" : "MapRed Local Directories.",
-        "property_value" : "/hadoop/mapred",
-        "stack_version" : "1.3.0",
         "property_name" : "mapred_local_dir",
-        "filename" : "global.xml",
+        "property_value" : "/hadoop/mapred",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
-      }
-    },
-    {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.inmem.merge.threshold",
-      "StackConfigurations" : {
-        "property_description" : "The threshold, in terms of the number of files\n  for the in-memory merge process. When we accumulate threshold number of files\n  we initiate the in-memory merge and spill to disk. A value of 0 or less than\n  0 indicates we want to DON'T have any threshold and instead depend only on\n  the ramfs's memory consumption to trigger the merge.\n  ",
-        "property_value" : "1000",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapred.inmem.merge.threshold",
-        "filename" : "mapred-site.xml",
-        "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.default.acl-administer-jobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_map_tasks_max",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "*",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.queue.default.acl-administer-jobs",
-        "filename" : "mapred-queue-acls.xml",
+        "property_description" : "Number of slots that Map tasks that run simultaneously can occupy on a TaskTracker",
+        "property_name" : "mapred_map_tasks_max",
+        "property_value" : "4",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
-      }
-    },
-    {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.reduce.input.limit",
-      "StackConfigurations" : {
-        "property_description" : "The limit on the input size of the reduce. (This value\n  is 10 Gb.)  If the estimated input size of the reduce is greater than\n  this value, job is failed. A value of -1 means that there is no limit\n  set. ",
-        "property_value" : "10737418240",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mapreduce.reduce.input.limit",
-        "filename" : "mapred-site.xml",
-        "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.cluster.administrators",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_red_tasks_max",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : " hadoop",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapreduce.cluster.administrators",
-        "filename" : "mapred-site.xml",
+        "property_description" : "Number of slots that Reduce tasks that run simultaneously can occupy on a TaskTracker",
+        "property_name" : "mapred_red_tasks_max",
+        "property_value" : "2",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.healthChecker.script.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_system_dir",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "60000",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.healthChecker.script.timeout",
-        "filename" : "mapred-site.xml",
+        "property_description" : "MapRed System Directories.",
+        "property_name" : "mapred_system_dir",
+        "property_value" : "/mapred/system",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.parallel.copies",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_user",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "30",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.reduce.parallel.copies",
-        "filename" : "mapred-site.xml",
+        "property_description" : "MapReduce User.",
+        "property_name" : "mapred_user",
+        "property_value" : "mapred",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.history.completed.location",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.cluster.administrators",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "/mapred/history/done",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.tracker.history.completed.location",
-        "filename" : "mapred-site.xml",
+        "property_description" : null,
+        "property_name" : "mapreduce.cluster.administrators",
+        "property_value" : " hadoop",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.slowstart.completed.maps",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.fileoutputcommitter.marksuccessfuljobs",
       "StackConfigurations" : {
         "property_description" : null,
-        "property_value" : "0.05",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.reduce.slowstart.completed.maps",
-        "filename" : "mapred-site.xml",
+        "property_name" : "mapreduce.fileoutputcommitter.marksuccessfuljobs",
+        "property_value" : "false",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.tasktracker.tasks.sleeptime-before-sigkill",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.history.server.embedded",
       "StackConfigurations" : {
-        "property_description" : "Normally, this is the amount of time before killing\n  processes, and the recommended-default is 5.000 seconds - a value of\n  5000 here.  In this case, we are using it solely to blast tasks before\n  killing them, and killing them very quickly (1/4 second) to guarantee\n  that we do not leave VMs around for later jobs.\n  ",
-        "property_value" : "250",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.tasktracker.tasks.sleeptime-before-sigkill",
-        "filename" : "mapred-site.xml",
+        "property_description" : "Should job history server be embedded within Job tracker\n      process",
+        "property_name" : "mapreduce.history.server.embedded",
+        "property_value" : "false",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.capacity",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.history.server.http.address",
       "StackConfigurations" : {
-        "property_description" : "Percentage of the number of slots in the cluster that are\n      to be available for jobs in this queue.\n    ",
-        "property_value" : "100",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.queue.default.capacity",
-        "filename" : "capacity-scheduler.xml",
+        "property_description" : "Http address of the history server",
+        "property_name" : "mapreduce.history.server.http.address",
+        "property_value" : "localhost:51111",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.shuffle.input.buffer.percent",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobhistory.done-dir",
       "StackConfigurations" : {
-        "property_description" : "The percentage of memory to be allocated from the maximum heap\n  size to storing map outputs during the shuffle.\n  ",
-        "property_value" : "0.7",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.job.shuffle.input.buffer.percent",
-        "filename" : "mapred-site.xml",
+        "property_description" : "\n      Directory where history files are managed by the MR JobHistory Server.\n    ",
+        "property_name" : "mapreduce.jobhistory.done-dir",
+        "property_value" : "/mr-history/done",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.child.root.logger",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobhistory.intermediate-done-dir",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "INFO,TLA",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.child.root.logger",
-        "filename" : "mapred-site.xml",
+        "property_description" : "\n      Directory where history files are written by MapReduce jobs.\n    ",
+        "property_name" : "mapreduce.jobhistory.intermediate-done-dir",
+        "property_value" : "/mr-history/tmp",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.fileoutputcommitter.marksuccessfuljobs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobhistory.webapp.address",
       "StackConfigurations" : {
-        "property_description" : null,
-        "property_value" : "false",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapreduce.fileoutputcommitter.marksuccessfuljobs",
-        "filename" : "mapred-site.xml",
+        "property_description" : "Enter your JobHistoryServer hostname.",
+        "property_name" : "mapreduce.jobhistory.webapp.address",
+        "property_value" : "localhost:19888",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-init-accept-jobs-factor",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobtracker.split.metainfo.maxsize",
       "StackConfigurations" : {
-        "property_description" : "The default multipe of (maximum-system-jobs * queue-capacity) \n    used to determine the number of jobs which are accepted by the scheduler.  \n    ",
-        "property_value" : "10",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred.capacity-scheduler.default-init-accept-jobs-factor",
-        "filename" : "capacity-scheduler.xml",
+        "property_description" : "If the size of the split metainfo file is larger than this, the JobTracker will fail the job during\n      initialize.\n    ",
+        "property_name" : "mapreduce.jobtracker.split.metainfo.maxsize",
+        "property_value" : "50000000",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_system_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobtracker.staging.root.dir",
       "StackConfigurations" : {
-        "property_description" : "MapRed System Directories.",
-        "property_value" : "/mapred/system",
-        "stack_version" : "1.3.0",
-        "property_name" : "mapred_system_dir",
-        "filename" : "global.xml",
+        "property_description" : "The Path prefix for where the staging directories should be placed. The next level is always the user's\n      name. It is a path in the default file system.",
+        "property_name" : "mapreduce.jobtracker.staging.root.dir",
+        "property_value" : "/user",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.factor",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.reduce.input.limit",
       "StackConfigurations" : {
-        "property_description" : "No description",
-        "property_value" : "100",
-        "stack_version" : "1.3.0",
-        "property_name" : "io.sort.factor",
-        "filename" : "mapred-site.xml",
+        "property_description" : "The limit on the input size of the reduce. (This value\n      is 10 Gb.)  If the estimated input size of the reduce is greater than\n      this value, job is failed. A value of -1 means that there is no limit\n      set. ",
+        "property_name" : "mapreduce.reduce.input.limit",
+        "property_value" : "10737418240",
         "service_name" : "MAPREDUCE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "mapred-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/ver

<TRUNCATED>

[09/11] AMBARI-4310 Update mock data of services configurations. (atkach)

Posted by at...@apache.org.
http://git-wip-us.apache.org/repos/asf/ambari/blob/8c0e3771/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HIVE.json
----------------------------------------------------------------------
diff --git a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HIVE.json b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HIVE.json
index 5596881..12c9b86 100644
--- a/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HIVE.json
+++ b/ambari-web/app/assets/data/wizard/stack/hdp/version1.3.0/HIVE.json
@@ -1,208 +1,544 @@
 {
-  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations?fields=*",
+  "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations?fields=*",
   "items" : [
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/javax.jdo.option.ConnectionDriverName",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/ambari.hive.db.schema.name",
       "StackConfigurations" : {
-        "property_description" : "Driver class name for a JDBC metastore",
-        "property_value" : "com.mysql.jdbc.Driver",
-        "stack_version" : "1.3.0",
-        "property_name" : "javax.jdo.option.ConnectionDriverName",
-        "filename" : "hive-site.xml",
+        "property_description" : "Database name used as the Hive Metastore",
+        "property_name" : "ambari.hive.db.schema.name",
+        "property_value" : "hive",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_pid_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/fs.file.impl.disable.cache",
       "StackConfigurations" : {
-        "property_description" : "Hive PID Dir.",
-        "property_value" : "/var/run/hive",
-        "stack_version" : "1.3.0",
-        "property_name" : "hive_pid_dir",
-        "filename" : "global.xml",
+        "property_description" : null,
+        "property_name" : "fs.file.impl.disable.cache",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.enabled",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/fs.hdfs.impl.disable.cache",
       "StackConfigurations" : {
-        "property_description" : "enable or disable the hive client authorization",
+        "property_description" : null,
+        "property_name" : "fs.hdfs.impl.disable.cache",
         "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.security.authorization.enabled",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hcat_log_dir",
+      "StackConfigurations" : {
+        "property_description" : "WebHCat Log Dir.",
+        "property_name" : "hcat_log_dir",
+        "property_value" : "/var/log/webhcat",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_conf_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hcat_pid_dir",
       "StackConfigurations" : {
-        "property_description" : "Hive Conf Dir.",
-        "property_value" : "/etc/hive/conf",
+        "property_description" : "WebHCat Pid Dir.",
+        "property_name" : "hcat_pid_dir",
+        "property_value" : "/etc/run/webhcat",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive_conf_dir",
-        "filename" : "global.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hcat_user",
+      "StackConfigurations" : {
+        "property_description" : "HCat User.",
+        "property_name" : "hcat_user",
+        "property_value" : "hcat",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hadoop.clientside.fs.operations",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.auto.convert.join",
       "StackConfigurations" : {
-        "property_description" : "FS operations are owned by client",
+        "property_description" : "Whether Hive enable the optimization about converting common\n      join into mapjoin based on the input file size.",
+        "property_name" : "hive.auto.convert.join",
         "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hadoop.clientside.fs.operations",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.auto.convert.join.noconditionaltask",
+      "StackConfigurations" : {
+        "property_description" : "Whether Hive enable the optimization about converting common join into mapjoin based on the input file\n      size. If this paramater is on, and the sum of size for n-1 of the tables/partitions for a n-way join is smaller than the\n      specified size, the join is directly converted to a mapjoin (there is no conditional task).\n    ",
+        "property_name" : "hive.auto.convert.join.noconditionaltask",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.manager",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.auto.convert.join.noconditionaltask.size",
       "StackConfigurations" : {
-        "property_description" : "the hive client authorization manager class name.\n    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
-        "property_value" : "org.apache.hcatalog.security.HdfsAuthorizationProvider",
+        "property_description" : "If hive.auto.convert.join.noconditionaltask is off, this parameter does not take affect. However, if it\n      is on, and the sum of size for n-1 of the tables/partitions for a n-way join is smaller than this size, the join is directly\n      converted to a mapjoin(there is no conditional task). The default is 10MB.\n    ",
+        "property_name" : "hive.auto.convert.join.noconditionaltask.size",
+        "property_value" : "1000000000",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.security.authorization.manager",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.auto.convert.sortmerge.join",
+      "StackConfigurations" : {
+        "property_description" : "Will the join be automatically converted to a sort-merge join, if the joined tables pass\n      the criteria for sort-merge join.\n    ",
+        "property_name" : "hive.auto.convert.sortmerge.join",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/fs.hdfs.impl.disable.cache",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.auto.convert.sortmerge.join.noconditionaltask",
       "StackConfigurations" : {
         "property_description" : null,
+        "property_name" : "hive.auto.convert.sortmerge.join.noconditionaltask",
         "property_value" : "true",
-        "stack_version" : "1.3.0",
-        "property_name" : "fs.hdfs.impl.disable.cache",
-        "filename" : "hive-site.xml",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.warehouse.dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.enforce.bucketing",
       "StackConfigurations" : {
-        "property_description" : "location of default database for the warehouse",
-        "property_value" : "/apps/hive/warehouse",
+        "property_description" : "Whether bucketing is enforced. If true, while inserting into the table, bucketing is enforced.",
+        "property_name" : "hive.enforce.bucketing",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.warehouse.dir",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.enforce.sorting",
+      "StackConfigurations" : {
+        "property_description" : "Whether sorting is enforced. If true, while inserting into the table, sorting is enforced.",
+        "property_name" : "hive.enforce.sorting",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.semantic.analyzer.factory.impl",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.map.aggr",
       "StackConfigurations" : {
-        "property_description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
-        "property_value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
+        "property_description" : "Whether to use map-side aggregation in Hive Group By queries.",
+        "property_name" : "hive.map.aggr",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.semantic.analyzer.factory.impl",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.mapjoin.bucket.cache.size",
+      "StackConfigurations" : {
+        "property_description" : "\n      Size per reducer.The default is 1G, i.e if the input size is 10G, it\n      will use 10 reducers.\n    ",
+        "property_name" : "hive.mapjoin.bucket.cache.size",
+        "property_value" : "10000",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_aux_jars_path",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.mapred.reduce.tasks.speculative.execution",
       "StackConfigurations" : {
-        "property_description" : "Hive auxiliary jar path.",
-        "property_value" : "/usr/lib/hcatalog/share/hcatalog/hcatalog-core.jar",
+        "property_description" : "Whether speculative execution for reducers should be turned on.",
+        "property_name" : "hive.mapred.reduce.tasks.speculative.execution",
+        "property_value" : "false",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive_aux_jars_path",
-        "filename" : "global.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.cache.pinobjtypes",
+      "StackConfigurations" : {
+        "property_description" : "List of comma separated metastore object types that should be pinned in the cache",
+        "property_name" : "hive.metastore.cache.pinobjtypes",
+        "property_value" : "Table,Database,Type,FieldSchema,Order",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.client.socket.timeout",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.client.socket.timeout",
       "StackConfigurations" : {
         "property_description" : "MetaStore Client socket timeout in seconds",
+        "property_name" : "hive.metastore.client.socket.timeout",
         "property_value" : "60",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.client.socket.timeout",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.execute.setugi",
+      "StackConfigurations" : {
+        "property_description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
+        "property_name" : "hive.metastore.execute.setugi",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.uris",
+      "StackConfigurations" : {
+        "property_description" : "URI for client to contact metastore server",
+        "property_name" : "hive.metastore.uris",
+        "property_value" : "thrift://localhost:9083",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.warehouse.dir",
+      "StackConfigurations" : {
+        "property_description" : "location of default database for the warehouse",
+        "property_name" : "hive.metastore.warehouse.dir",
+        "property_value" : "/apps/hive/warehouse",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.server2.enable.doAs",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.optimize.bucketmapjoin",
       "StackConfigurations" : {
         "property_description" : null,
+        "property_name" : "hive.optimize.bucketmapjoin",
         "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.server2.enable.doAs",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.optimize.bucketmapjoin.sortedmerge",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hive.optimize.bucketmapjoin.sortedmerge",
+        "property_value" : "true",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.cache.pinobjtypes",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.optimize.mapjoin.mapreduce",
       "StackConfigurations" : {
-        "property_description" : "List of comma separated metastore object types that should be pinned in the cache",
-        "property_value" : "Table,Database,Type,FieldSchema,Order",
+        "property_description" : "If hive.auto.convert.join is off, this parameter does not take\n      affect. If it is on, and if there are map-join jobs followed by a map-reduce\n      job (for e.g a group by), each map-only job is merged with the following\n      map-reduce job.\n    ",
+        "property_name" : "hive.optimize.mapjoin.mapreduce",
+        "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.cache.pinobjtypes",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.optimize.reducededuplication.min.reducer",
+      "StackConfigurations" : {
+        "property_description" : "Reduce deduplication merges two RSs by moving key/parts/reducer-num of the child RS to parent RS.\n      That means if reducer-num of the child RS is fixed (order by or forced bucketing) and small, it can make very slow, single MR.\n      The optimization will be disabled if number of reducers is less than specified value.\n    ",
+        "property_name" : "hive.optimize.reducededuplication.min.reducer",
+        "property_value" : "1",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.enabled",
+      "StackConfigurations" : {
+        "property_description" : "enable or disable the hive client authorization",
+        "property_name" : "hive.security.authorization.enabled",
+        "property_value" : "false",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.execute.setugi",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.manager",
       "StackConfigurations" : {
-        "property_description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
+        "property_description" : "the hive client authorization manager class name.\n      The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
+        "property_name" : "hive.security.authorization.manager",
+        "property_value" : "org.apache.hcatalog.security.HdfsAuthorizationProvider",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.semantic.analyzer.factory.impl",
+      "StackConfigurations" : {
+        "property_description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
+        "property_name" : "hive.semantic.analyzer.factory.impl",
+        "property_value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.server2.enable.doAs",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_name" : "hive.server2.enable.doAs",
         "property_value" : "true",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.execute.setugi",
-        "filename" : "hive-site.xml",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_ambari_database",
+      "StackConfigurations" : {
+        "property_description" : "Database type.",
+        "property_name" : "hive_ambari_database",
+        "property_value" : "MySQL",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/mysql_connector_url",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_aux_jars_path",
       "StackConfigurations" : {
-        "property_description" : "Hive PID Dir.",
-        "property_value" : "${download_url}/mysql-connector-java-5.1.18.zip",
+        "property_description" : "Hive auxiliary jar path.",
+        "property_name" : "hive_aux_jars_path",
+        "property_value" : "/usr/lib/hcatalog/share/hcatalog/hcatalog-core.jar",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "mysql_connector_url",
-        "filename" : "global.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_conf_dir",
+      "StackConfigurations" : {
+        "property_description" : "Hive Conf Dir.",
+        "property_name" : "hive_conf_dir",
+        "property_value" : "/etc/hive/conf",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.local",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_dbroot",
       "StackConfigurations" : {
-        "property_description" : "controls whether to connect to remove metastore server or\n    open a new metastore server in Hive Client JVM",
-        "property_value" : "false",
+        "property_description" : "Hive DB Directory.",
+        "property_name" : "hive_dbroot",
+        "property_value" : "/usr/lib/hive/lib/",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive.metastore.local",
-        "filename" : "hive-site.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_lib",
+      "StackConfigurations" : {
+        "property_description" : "Hive Library.",
+        "property_name" : "hive_lib",
+        "property_value" : "/usr/lib/hive/lib/",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     },
     {
-      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_log_dir",
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_log_dir",
       "StackConfigurations" : {
         "property_description" : "Directory for Hive Log files.",
+        "property_name" : "hive_log_dir",
         "property_value" : "/var/log/hive",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
         "stack_version" : "1.3.0",
-        "property_name" : "hive_log_dir",
-        "filename" : "global.xml",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_metastore_port",
+      "StackConfigurations" : {
+        "property_description" : "Hive Metastore port.",
+        "property_name" : "hive_metastore_port",
+        "property_value" : "9083",
         "service_name" : "HIVE",
-        "stack_name" : "HDP"
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_metastore_user_name",
+      "StackConfigurations" : {
+        "property_description" : "Database username to use to connect to the database.",
+        "property_name" : "hive_metastore_user_name",
+        "property_value" : "hive",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_pid_dir",
+      "StackConfigurations" : {
+        "property_description" : "Hive PID Dir.",
+        "property_name" : "hive_pid_dir",
+        "property_value" : "/var/run/hive",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_user",
+      "StackConfigurations" : {
+        "property_description" : "Hive User.",
+        "property_name" : "hive_user",
+        "property_value" : "hive",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/javax.jdo.option.ConnectionDriverName",
+      "StackConfigurations" : {
+        "property_description" : "Driver class name for a JDBC metastore",
+        "property_name" : "javax.jdo.option.ConnectionDriverName",
+        "property_value" : "com.mysql.jdbc.Driver",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/javax.jdo.option.ConnectionPassword",
+      "StackConfigurations" : {
+        "property_description" : "password to use against metastore database",
+        "property_name" : "javax.jdo.option.ConnectionPassword",
+        "property_value" : " ",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/javax.jdo.option.ConnectionURL",
+      "StackConfigurations" : {
+        "property_description" : "JDBC connect string for a JDBC metastore",
+        "property_name" : "javax.jdo.option.ConnectionURL",
+        "property_value" : "jdbc",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/javax.jdo.option.ConnectionUserName",
+      "StackConfigurations" : {
+        "property_description" : "username to use against metastore database",
+        "property_name" : "javax.jdo.option.ConnectionUserName",
+        "property_value" : "hive",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "hive-site.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/mysql_connector_url",
+      "StackConfigurations" : {
+        "property_description" : "Hive PID Dir.",
+        "property_name" : "mysql_connector_url",
+        "property_value" : "${download_url}/mysql-connector-java-5.1.18.zip",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
+      }
+    },
+    {
+      "href" : "http://192.168.56.101:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/webhcat_user",
+      "StackConfigurations" : {
+        "property_description" : "WebHCat User.",
+        "property_name" : "webhcat_user",
+        "property_value" : "hcat",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP",
+        "stack_version" : "1.3.0",
+        "type" : "global.xml"
       }
     }
   ]