You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by ja...@apache.org on 2013/10/23 00:56:56 UTC

[1/4] AMBARI-3568: Clean up unnecessary globals from Ambari Web. (jaimin)

Updated Branches:
  refs/heads/trunk 63d563854 -> 88f513259


http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/data/site_properties.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/site_properties.js b/ambari-web/app/data/site_properties.js
index bc68d25..502bca1 100644
--- a/ambari-web/app/data/site_properties.js
+++ b/ambari-web/app/data/site_properties.js
@@ -18,5 +18,629 @@
 module.exports =
 {
   "configProperties": [
+  /**********************************************HDFS***************************************/
+    {
+      "id": "site property",
+      "name": "fs.checkpoint.dir",
+      "displayName": "SecondaryNameNode Checkpoint directory",
+      "description": "Directory on the local filesystem where the Secondary NameNode should store the temporary images to merge",
+      "defaultValue": "",
+      "defaultDirectory": "/hadoop/hdfs/namesecondary",
+      "displayType": "directory",
+      "isOverridable": false,
+      "isVisible": true,
+      "serviceName": "HDFS",
+      "category": "SNameNode",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "fs.checkpoint.period",
+      "displayName": "HDFS Maximum Checkpoint Delay",
+      "description": "Maximum delay between two consecutive checkpoints for HDFS",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "seconds",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HDFS",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "fs.checkpoint.size",
+      "displayName": "HDFS Maximum Edit Log Size for Checkpointing",
+      "description": "Maximum size of the edits log file that forces an urgent checkpoint even if the maximum checkpoint delay is not reached",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "isVisible": true,
+      "serviceName": "HDFS",
+      "category": "General",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "dfs.name.dir",
+      "displayName": "NameNode directories",
+      "description": "NameNode directories for HDFS to store the file system image",
+      "defaultValue": "",
+      "defaultDirectory": "/hadoop/hdfs/namenode",
+      "displayType": "directories",
+      "isOverridable": false,
+      "isVisible": true,
+      "serviceName": "HDFS",
+      "category": "NameNode",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "dfs.webhdfs.enabled",
+      "displayName": "WebHDFS enabled",
+      "description": "Whether to enable WebHDFS feature",
+      "defaultValue": "",
+      "displayType": "checkbox",
+      "isOverridable": false,
+      "category": "General",
+      "serviceName": "HDFS",
+      "index": 0
+    },
+
+    {
+      "id": "site property",
+      "name": "dfs.datanode.failed.volumes.tolerated",
+      "displayName": "DataNode volumes failure toleration",
+      "description": "The number of volumes that are allowed to fail before a DataNode stops offering service",
+      "defaultValue": "",
+      "displayType": "int",
+      "isVisible": true,
+      "category": "DataNode",
+      "serviceName": "HDFS",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "dfs.data.dir",
+      "displayName": "DataNode directories",
+      "description": "DataNode directories for HDFS to store the data blocks",
+      "defaultValue": "",
+      "defaultDirectory": "/hadoop/hdfs/data",
+      "displayType": "directories",
+      "isVisible": true,
+      "category": "DataNode",
+      "serviceName": "HDFS",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "dfs.datanode.data.dir.perm",
+      "displayName": "DataNode directories permission",
+      "description": "",
+      "defaultValue": "",
+      "displayType": "int",
+      "isVisible": false,
+      "category": "DataNode",
+      "serviceName": "HDFS"
+    },
+    {
+      "id": "site property",
+      "name": "dfs.replication",
+      "displayName": "Block replication",
+      "description": "Default block replication.",
+      "displayType": "int",
+      "defaultValue": "",
+      "isVisible": true,
+      "category": "Advanced",
+      "serviceName": "HDFS"
+    },
+    {
+      "id": "site property",
+      "name": "dfs.datanode.du.reserved",
+      "displayName": "Reserved space for HDFS",
+      "description": "Reserved space in bytes per volume. Always leave this much space free for non dfs use.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HDFS",
+      "index": 2
+    },
+
+  /******************************************MAPREDUCE***************************************/
+    {
+      "id": "site property",
+      "name": "mapred.local.dir",
+      "displayName": "MapReduce local directories",
+      "description": "Directories for MapReduce to store intermediate data files",
+      "defaultValue": "",
+      "defaultDirectory": "/hadoop/mapred",
+      "displayType": "directories",
+      "serviceName": "MAPREDUCE",
+      "category": "TaskTracker",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "mapred.jobtracker.taskScheduler",
+      "displayName": "MapReduce Capacity Scheduler",
+      "description": "The scheduler to use for scheduling of MapReduce jobs",
+      "defaultValue": "",
+      "displayType": "advanced",
+      "isOverridable": false,
+      "serviceName": "MAPREDUCE",
+      "index": 0
+    },
+    {
+      "id": "site property",
+      "name": "mapred.tasktracker.map.tasks.maximum",
+      "displayName": "Number of Map slots per node",
+      "description": "Number of slots that Map tasks that run simultaneously can occupy on a TaskTracker",
+      "defaultValue": "",
+      "displayType": "int",
+      "serviceName": "MAPREDUCE",
+      "category": "TaskTracker",
+      "index": 2
+    },
+    {
+      "id": "site property",
+      "name": "mapred.tasktracker.reduce.tasks.maximum",
+      "displayName": "Number of Reduce slots per node",
+      "description": "Number of slots that Reduce tasks that run simultaneously can occupy on a TaskTracker.",
+      "defaultValue": "",
+      "displayType": "int",
+      "serviceName": "MAPREDUCE",
+      "category": "TaskTracker",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "mapred.cluster.reduce.memory.mb",
+      "displayName": "Cluster's Reduce slot size (virtual memory)",
+      "description": "The virtual memory size of a single Reduce slot in the MapReduce framework",
+      "defaultValue": "2048",
+      "displayType": "int",
+      "unit": "MB",
+      "serviceName": "MAPREDUCE",
+      "index": 2
+    },
+    {
+      "id": "site property",
+      "name": "mapred.job.map.memory.mb",
+      "displayName": "Default virtual memory for a job's map-task",
+      "description": "Virtual memory for single Map task",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "MB",
+      "serviceName": "MAPREDUCE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "mapred.cluster.max.map.memory.mb",
+      "displayName": "Upper limit on virtual memory for single Map task",
+      "description": "Upper limit on virtual memory size for a single Map task of any MapReduce job",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "MB",
+      "serviceName": "MAPREDUCE",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "mapred.cluster.max.reduce.memory.mb",
+      "displayName": "Upper limit on virtual memory for single Reduce task",
+      "description": "Upper limit on virtual memory size for a single Reduce task of any MapReduce job",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "MB",
+      "serviceName": "MAPREDUCE",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "mapred.job.reduce.memory.mb",
+      "displayName": "Default virtual memory for a job's reduce-task",
+      "description": "Virtual memory for single Reduce task",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "MB",
+      "serviceName": "MAPREDUCE",
+      "index": 6
+    },
+    {
+      "id": "site property",
+      "name": "mapred.cluster.map.memory.mb",
+      "displayName": "Cluster's Map slot size (virtual memory)",
+      "description": "The virtual memory size of a single Map slot in the MapReduce framework",
+      "defaultValue": "1536",
+      "displayType": "int",
+      "unit": "MB",
+      "serviceName": "MAPREDUCE",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "mapred.healthChecker.script.path",
+      "displayName": "Job Status directory",
+      "description": "Directory path to view job status",
+      "defaultValue": "",
+      "displayType": "advanced",
+      "serviceName": "MAPREDUCE",
+      "category": "Advanced"
+    },
+    {
+      "id": "site property",
+      "name": "io.sort.mb",
+      "displayName": "Map-side sort buffer memory",
+      "description": "The total amount of Map-side buffer memory to use while sorting files (Expert-only configuration)",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "MB",
+      "serviceName": "MAPREDUCE",
+      "index": 7
+    },
+    {
+      "id": "site property",
+      "name": "io.sort.spill.percent",
+      "displayName": "Limit on buffer",
+      "description": "Percentage of sort buffer used for record collection",
+      "defaultValue": "",
+      "displayType": "float",
+      "serviceName": "MAPREDUCE",
+      "index": 8
+    },
+    {
+      "id": "site property",
+      "name": "mapred.system.dir",
+      "displayName": "MapReduce system directories",
+      "description": "Path on the HDFS where where the MapReduce framework stores system files",
+      "defaultValue": "/mapred/system",
+      "displayType": "directories",
+      "serviceName": "MAPREDUCE",
+      "category": "Advanced"
+    },
+    {
+      "id": "site property",
+      "name": "mapred.userlog.retain.hours",
+      "displayName": "Job log retention (hours)",
+      "description": "The maximum time, in hours, for which the user-logs are to be retained after the job completion.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "hours",
+      "serviceName": "MAPREDUCE",
+      "index": 9
+    },
+    {
+      "id": "site property",
+      "name": "mapred.jobtracker.maxtasks.per.job",
+      "displayName": "Maximum number tasks for a Job",
+      "description": "Maximum number of tasks for a single Job",
+      "defaultValue": "",
+      "displayType": "int",
+      "serviceName": "MAPREDUCE",
+      "index": 10
+    },
+
+  /**********************************************oozie-site***************************************/
+    {
+      "id": "site property",
+      "name": "oozie.db.schema.name",
+      "displayName": "Database Name",
+      "description": "Database name used for the Oozie",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "host",
+      "isVisible": true,
+      "isObserved": true,
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.username",
+      "displayName": "Database Username",
+      "description": "Database user name to use to connect to the database",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "host",
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.password",
+      "displayName": "Database Password",
+      "description": "Database password to use to connect to the database",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "password",
+      "isVisible": true,
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.driver",
+      "displayName": "JDBC driver class",
+      "defaultValue": "",
+      "value": "",     // the value is overwritten in code
+      "isVisible": false,
+      "description": "Database name used for the Oozie",
+      "category": "Advanced",
+      "serviceName": "OOZIE"
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.url",
+      "displayName": "Database URL",
+      "description": "The JDBC connection URL to the database",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "advanced",
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 6
+    },
+
+  /**********************************************hive-site***************************************/
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionDriverName",
+      "displayName": "JDBC driver class",
+      "defaultValue": "",
+      "value": "",     // the value is overwritten in code
+      "isVisible": false,
+      "description": "Driver class name for a JDBC metastore",
+      "category": "Advanced",
+      "serviceName": "HIVE"
+    },
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionUserName",
+      "displayName": "Database Username",
+      "description": "Database user name to use to connect to the database",
+      "defaultValue": "hive",
+      "displayType": "host",
+      "isOverridable": false,
+      "category": "Hive Metastore",
+      "serviceName": "HIVE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionPassword",
+      "displayName": "Database Password",
+      "description": "Database password to use to connect to the metastore database",
+      "defaultValue": "",
+      "displayType": "password",
+      "isOverridable": false,
+      "isVisible": true,
+      "category": "Hive Metastore",
+      "serviceName": "HIVE",
+      "index": 6
+    },
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionURL",
+      "displayName": "Database URL",
+      "value": "",
+      "defaultValue": "", // set to a 'jdbc' to not include this in initial error count
+      "description": "The JDBC connection URL to the database",
+      "displayType": "advanced",
+      "isOverridable": false,
+      "isVisible": true,
+      "category": "Hive Metastore",
+      "serviceName": "HIVE",
+      "index": 7
+    },
+
+  /**********************************************hbase-site***************************************/
+    {
+      "id": "site property",
+      "name": "hbase.tmp.dir",
+      "displayName": "HBase local directory",
+      "description": "Temporary directory on the local filesystem",
+      "defaultDirectory": "/hadoop/hbase",
+      "defaultValue": "",
+      "displayType": "directory",
+      "isVisible": true,
+      "category": "Advanced",
+      "serviceName": "HBASE"
+
+    },
+    {
+      "id": "site property",
+      "name": "hbase.regionserver.global.memstore.upperLimit",
+      "displayName": "hbase.regionserver.global.memstore.upperLimit",
+      "description": "",
+      "defaultValue": "",
+      "displayType": "float",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.regionserver.global.memstore.lowerLimit",
+      "displayName": "hbase.regionserver.global.memstore.lowerLimit",
+      "defaultValue": "",
+      "description": "When memstores are being forced to flush to make room in\
+      memory, keep flushing until we hit this mark. Defaults to 35% of heap.\
+      This value equal to hbase.regionserver.global.memstore.upperLimit causes\
+      the minimum possible flushing to occur when updates are blocked due to\
+      memstore limiting.",
+      "displayType": "float",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hstore.blockingStoreFiles",
+      "displayName": "hstore blocking storefiles",
+      "description": "If more than this number of StoreFiles in any one Store (one StoreFile is written per flush of " +
+        "MemStore) then updates are blocked for this HRegion until a compaction is completed, or until " +
+        "hbase.hstore.blockingWaitTime has been exceeded.",
+      "defaultValue": "",
+      "isRequired": true,
+      "displayType": "int",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hstore.compactionThreshold",
+      "displayName": "HBase HStore compaction threshold",
+      "description": "If more than this number of HStoreFiles in any one HStore then a compaction is run to rewrite all HStoreFiles files as one.",
+      "defaultValue": "3",
+      "displayType": "int",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 0
+    },
+    {
+      "id": "site property",
+      "name": "hfile.block.cache.size",
+      "displayName": "HFile block cache size ",
+      "description": "Percentage of maximum heap (-Xmx setting) to allocate to block cache used by HFile/StoreFile. Set to 0 to disable but it's not recommended.",
+      "defaultValue": "0.40",
+      "displayType": "float",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.max.filesize",
+      "displayName": "Maximum HStoreFile Size",
+      "description": "If any one of a column families' HStoreFiles has grown to exceed this value, the hosting HRegion is split in two.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 2
+    },
+    {
+      "id": "site property",
+      "name": "hbase.regionserver.handler.count",
+      "displayName": "HBase RegionServer Handler",
+      "description": "Count of RPC Listener instances spun up on RegionServers",
+      "defaultValue": "60",
+      "displayType": "int",
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 2
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.majorcompaction",
+      "displayName": "HBase Region Major Compaction",
+      "description": "The time between major compactions of all HStoreFiles in a region. Set to 0 to disable automated major compactions.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "ms",
+      "isVisible": true,
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.memstore.block.multiplier",
+      "displayName": "HBase Region Block Multiplier",
+      "description": "Block updates if memstore has \"Multiplier * HBase Region Memstore Flush Size\" bytes. Useful preventing runaway memstore during spikes in update traffic",
+      "defaultValue": "",
+      "displayType": "int",
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.memstore.mslab.enabled",
+      "displayName": "hbase.hregion.memstore.mslab.enabled",
+      "description": "Enables the MemStore-Local Allocation Buffer,\
+      a feature which works to prevent heap fragmentation under\
+      heavy write loads. This can reduce the frequency of stop-the-world\
+      GC pauses on large heaps.",
+      "defaultValue": "",
+      "displayType": "checkbox",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.memstore.flush.size",
+      "displayName": "HBase Region Memstore Flush Size",
+      "description": "Memstore will be flushed to disk if size of the memstore exceeds this number of bytes.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "hbase.client.scanner.caching",
+      "displayName": "HBase Client Scanner Caching",
+      "description": "Number of rows that will be fetched when calling next on a scanner if it is not served from \
+      (local, client) memory. Do not set this value such that the time between invocations is greater than the scanner timeout",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "rows",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "zookeeper.session.timeout",
+      "displayName": "Zookeeper timeout for HBase Session",
+      "description": "HBase passes this to the zk quorum as suggested maximum time for a session",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "ms",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "hbase.client.keyvalue.maxsize",
+      "displayName": "HBase Client Maximum key-value Size",
+      "description": "Specifies the combined maximum allowed size of a KeyValue instance. It should be set to a fraction of the maximum region size.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "dfs.client.read.shortcircuit",
+      "displayName": "HDFS Short-circuit read",
+      "description": "This configuration parameter turns on short-circuit local reads.",
+      "defaultValue": "",
+      "displayType": "checkbox",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "dfs.support.append",
+      "displayName": "HDFS append support",
+      "description": "HDFS append support",
+      "defaultValue": "",
+      "displayType": "checkbox",
+      "serviceName": "HBASE",
+      "category": "Advanced"
+    }
   ]
 };

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/models/service_config.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/models/service_config.js b/ambari-web/app/models/service_config.js
index e606bd9..ae7df60 100644
--- a/ambari-web/app/models/service_config.js
+++ b/ambari-web/app/models/service_config.js
@@ -287,21 +287,20 @@ App.ServiceConfigProperty = Ember.Object.extend({
       case 'zookeeperserver_hosts':
         this.set('value', masterComponentHostsInDB.filterProperty('component', 'ZOOKEEPER_SERVER').mapProperty('hostName'));
         break;
-      case 'dfs_name_dir':
-      case 'dfs_namenode_name_dir':
-      case 'dfs_data_dir':
-      case 'dfs_datanode_data_dir':
+      case 'dfs.name.dir':
+      case 'dfs.namenode.name.dir':
+      case 'dfs.data.dir':
+      case 'dfs.datanode.data.dir':
       case 'yarn.nodemanager.local-dirs':
       case 'yarn.nodemanager.log-dirs':
-      case 'mapred_local_dir':
-      case 'mapreduce_cluster_local_dir':
+      case 'mapred.local.dir':
         this.unionAllMountPoints(!isOnlyFirstOneNeeded, localDB);
         break;
-      case 'fs_checkpoint_dir':
-      case 'dfs_namenode_checkpoint_dir':
+      case 'fs.checkpoint.dir':
+      case 'dfs.namenode.checkpoint.dir':
       case 'zk_data_dir':
       case 'oozie_data_dir':
-      case 'hbase_tmp_dir':
+      case 'hbase.tmp.dir':
         this.unionAllMountPoints(isOnlyFirstOneNeeded, localDB);
         break;
     }
@@ -329,29 +328,28 @@ App.ServiceConfigProperty = Ember.Object.extend({
     var temp = '';
     var setOfHostNames = [];
     switch (this.get('name')) {
-      case 'dfs_namenode_name_dir':
-      case 'dfs_name_dir':
+      case 'dfs.namenode.name.dir':
+      case 'dfs.name.dir':
         var components = masterComponentHostsInDB.filterProperty('component', 'NAMENODE');
         components.forEach(function (component) {
           setOfHostNames.push(component.hostName);
         }, this);
         break;
-      case 'fs_checkpoint_dir':
-      case 'dfs_namenode_checkpoint_dir':
+      case 'fs.checkpoint.dir':
+      case 'dfs.namenode.checkpoint.dir':
         var components = masterComponentHostsInDB.filterProperty('component', 'SECONDARY_NAMENODE');
         components.forEach(function (component) {
           setOfHostNames.push(component.hostName);
         }, this);
         break;
-      case 'dfs_data_dir':
-      case 'dfs_datanode_data_dir':
+      case 'dfs.data.dir':
+      case 'dfs.datanode.data.dir':
         temp = slaveComponentHostsInDB.findProperty('componentName', 'DATANODE');
         temp.hosts.forEach(function (host) {
           setOfHostNames.push(host.hostName);
         }, this);
         break;
-      case 'mapred_local_dir':
-      case 'mapreduce_cluster_local_dir':
+      case 'mapred.local.dir':
         temp = slaveComponentHostsInDB.findProperty('componentName', 'TASKTRACKER') || slaveComponentHostsInDB.findProperty('componentName', 'NODEMANAGER');
         temp.hosts.forEach(function (host) {
           setOfHostNames.push(host.hostName);
@@ -376,7 +374,7 @@ App.ServiceConfigProperty = Ember.Object.extend({
           setOfHostNames.push(component.hostName);
         }, this);
         break;
-      case 'hbase_tmp_dir':
+      case 'hbase.tmp.dir':
         var temp = slaveComponentHostsInDB.findProperty('componentName', 'HBASE_REGIONSERVER');
         temp.hosts.forEach(function (host) {
           setOfHostNames.push(host.hostName);
@@ -571,7 +569,7 @@ App.ServiceConfigProperty = Ember.Object.extend({
           }
           break;
         case 'advanced':
-          if(this.get('name')=='hive_jdbc_connection_url' || this.get('name')=='oozie_jdbc_connection_url') {
+          if(this.get('name')=='javax.jdo.option.ConnectionURL' || this.get('name')=='oozie.service.JPAService.jdbc.url') {
             if (validator.isNotTrimmed(value)) {
               this.set('errorMessage', Em.I18n.t('host.trimspacesValidation'));
               isError = true;

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/utils/config.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/utils/config.js b/ambari-web/app/utils/config.js
index 731a988..47c2150 100644
--- a/ambari-web/app/utils/config.js
+++ b/ambari-web/app/utils/config.js
@@ -985,7 +985,7 @@ App.config = Em.Object.create({
       case 'password':
         break;
       case 'advanced':
-        if(name == 'hive_jdbc_connection_url' || name == 'oozie_jdbc_connection_url') {
+        if(name == 'javax.jdo.option.ConnectionURL' || name == 'oozie.service.JPAService.jdbc.url') {
           rez = value.trim();
         }
       default:

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/views/common/configs/services_config.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/views/common/configs/services_config.js b/ambari-web/app/views/common/configs/services_config.js
index c49e38c..cbda51c 100644
--- a/ambari-web/app/views/common/configs/services_config.js
+++ b/ambari-web/app/views/common/configs/services_config.js
@@ -129,14 +129,14 @@ App.ServiceConfigsByCategoryView = Ember.View.extend({
         }
       } else if (changedProperty.get("name") == "hbase_user" && !App.get('isHadoop2Stack')) {
         curConfigs = stepConfigs.findProperty("serviceName", "HDFS").get("configs");
-        if (newValue != curConfigs.findProperty("name", "dfs_block_local_path_access_user").get("value")) {
+        if (newValue != curConfigs.findProperty("name", "dfs.block.local-path-access.user").get("value")) {
           this.affectedProperties.push(
             {
               serviceName: "HDFS",
-              propertyName: "dfs_block_local_path_access_user",
+              propertyName: "dfs.block.local-path-access.user",
               propertyDisplayName: "dfs.block.local-path-access.user",
               newValue: newValue,
-              curValue: curConfigs.findProperty("name", "dfs_block_local_path_access_user").get("value"),
+              curValue: curConfigs.findProperty("name", "dfs.block.local-path-access.user").get("value"),
               changedPropertyName: "hbase_user"
             }
           );

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/views/wizard/controls_view.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/views/wizard/controls_view.js b/ambari-web/app/views/wizard/controls_view.js
index c9f542d..de0ec4f 100644
--- a/ambari-web/app/views/wizard/controls_view.js
+++ b/ambari-web/app/views/wizard/controls_view.js
@@ -250,7 +250,7 @@ App.ServiceConfigRadioButtons = Ember.View.extend({
       case 'HIVE':
         return this.get('categoryConfigsAll').findProperty('name', 'hive_database_name').get('value');
       case 'OOZIE':
-        return this.get('categoryConfigsAll').findProperty('name', 'oozie_database_name').get('value');
+        return this.get('categoryConfigsAll').findProperty('name', 'oozie.db.schema.name').get('value');
       default:
         return null;
     }
@@ -283,9 +283,9 @@ App.ServiceConfigRadioButtons = Ember.View.extend({
 
   connectionUrl: function () {
     if (this.get('serviceConfig.serviceName') === 'HIVE') {
-      return this.get('categoryConfigsAll').findProperty('name', 'hive_jdbc_connection_url');
+      return this.get('categoryConfigsAll').findProperty('name', 'javax.jdo.option.ConnectionURL');
     } else {
-      return this.get('categoryConfigsAll').findProperty('name', 'oozie_jdbc_connection_url');
+      return this.get('categoryConfigsAll').findProperty('name', 'oozie.service.JPAService.jdbc.url');
     }
   }.property('serviceConfig.serviceName'),
 


[2/4] AMBARI-3568: Clean up unnecessary globals from Ambari Web. (jaimin)

Posted by ja...@apache.org.
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/data/HDP2/global_properties.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/global_properties.js b/ambari-web/app/data/HDP2/global_properties.js
index fc2a670..e96ccfc 100644
--- a/ambari-web/app/data/HDP2/global_properties.js
+++ b/ambari-web/app/data/HDP2/global_properties.js
@@ -93,21 +93,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_namenode_name_dir",
-      "displayName": "NameNode directories",
-      "description": "NameNode directories for HDFS to store the file system image",
-      "defaultValue": "",
-      "defaultDirectory": "/hadoop/hdfs/namenode",
-      "displayType": "directories",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "NameNode",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
       "name": "namenode_heapsize",
       "displayName": "NameNode Java heap size",
       "description": "Initial and maximum Java heap size for NameNode (Java options -Xms and -Xmx).  This also applies to the Secondary NameNode.",
@@ -153,21 +138,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_namenode_checkpoint_dir",
-      "displayName": "SecondaryNameNode Checkpoint directory",
-      "description": "Directory on the local filesystem where the Secondary NameNode should store the temporary images to merge",
-      "defaultValue": "",
-      "defaultDirectory": "/hadoop/hdfs/namesecondary",
-      "displayType": "directory",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "SNameNode",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
       "name": "datanode_hosts", //not in the schema. For UI purpose
       "displayName": "DataNode hosts",
       "value": "",
@@ -184,20 +154,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_datanode_data_dir",
-      "displayName": "DataNode directories",
-      "description": "DataNode directories for HDFS to store the data blocks",
-      "defaultValue": "",
-      "defaultDirectory": "/hadoop/hdfs/data",
-      "displayType": "directories",
-      "isVisible": true,
-      "domain": "datanode-global",
-      "serviceName": "HDFS",
-      "category": "DataNode",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
       "name": "dtnode_heapsize",
       "displayName": "DataNode maximum Java heap size",
       "description": "Maximum Java heap size for DataNode (Java option -Xmx)",
@@ -212,32 +168,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_datanode_failed_volume_tolerated",
-      "displayName": "DataNode volumes failure toleration",
-      "description": "The number of volumes that are allowed to fail before a DataNode stops offering service",
-      "defaultValue": "0",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "datanode-global",
-      "serviceName": "HDFS",
-      "category": "DataNode",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "dfs_webhdfs_enabled",
-      "displayName": "WebHDFS enabled",
-      "description": "Whether to enable WebHDFS feature",
-      "defaultValue": true,
-      "displayType": "checkbox",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "index": 0
-    },
-    {
-      "id": "puppet var",
       "name": "hadoop_heapsize",
       "displayName": "Hadoop maximum Java heap size",
       "description": "Maximum Java heap size for daemons such as Balancer (Java option -Xmx)",
@@ -251,47 +181,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "datanode_du_reserved",
-      "displayName": "Reserved space for HDFS",
-      "description": "Reserved space in GB per volume",
-      "defaultValue": "1073741824",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "dfs_namenode_checkpoint_period",
-      "displayName": "HDFS Maximum Checkpoint Delay",
-      "description": "Maximum delay between two consecutive checkpoints for HDFS",
-      "defaultValue": "21600",
-      "displayType": "int",
-      "unit": "seconds",
-      "isVisible": true,
-      "domain": "global",
-      "filename": "hdfs-site.xml",
-      "serviceName": "HDFS",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "fs_checkpoint_size",
-      "displayName": "HDFS Maximum Edit Log Size for Checkpointing",
-      "description": "Maximum size of the edits log file that forces an urgent checkpoint even if the maximum checkpoint delay is not reached",
-      "defaultValue": "0.5",
-      "displayType": "float",
-      "unit": "GB",
-      "isVisible": true,
-      "domain": "global",
-      "filename": "core-site.xml",
-      "serviceName": "HDFS",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
       "name": "hdfs_log_dir_prefix",
       "displayName": "Hadoop Log Dir Prefix",
       "description": "The parent directory for Hadoop log files.  The HDFS log directory will be ${hadoop_log_dir_prefix} / ${hdfs_user} and the MapReduce log directory will be ${hadoop_log_dir_prefix} / ${mapred_user}.",
@@ -347,33 +236,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_replication",
-      "displayName": "Block replication",
-      "description": "Default block replication.",
-      "displayType": "int",
-      "defaultValue": "3",
-      "isVisible": true,
-      "filename": "hdfs-site.xml",
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "dfs_datanode_data_dir_perm",
-      "displayName": "dfs_datanode_data_dir_perm",
-      "description": "",
-      "defaultValue": "750",
-      "isReconfigurable": true,
-      "displayType": "int",
-      "isVisible": false,
-      "filename": "hdfs-site.xml",
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
       "name": "security_enabled",
       "displayName": "Hadoop Security",
       "description": "Enable hadoop security",
@@ -645,136 +507,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "regionserver_handlers",
-      "displayName": "HBase RegionServer Handler",
-      "description": "Count of RPC Listener instances spun up on RegionServers",
-      "defaultValue": "60",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "hregion_majorcompaction",
-      "displayName": "HBase Region Major Compaction",
-      "description": "The time between major compactions of all HStoreFiles in a region. Set to 0 to disable automated major compactions.",
-      "defaultValue": "604800000",
-      "displayType": "int",
-      "unit": "ms",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "hregion_blockmultiplier",
-      "displayName": "HBase Region Block Multiplier",
-      "description": "Block updates if memstore has \"Multiplier * HBase Region Memstore Flush Size\" bytes. Useful preventing runaway memstore during spikes in update traffic",
-      "defaultValue": "2",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
-      "name": "hregion_memstoreflushsize",
-      "displayName": "HBase Region Memstore Flush Size",
-      "description": "Memstore will be flushed to disk if size of the memstore exceeds this number of bytes.",
-      "defaultValue": "134217728",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
-      "name": "hstore_compactionthreshold",
-      "displayName": "HBase HStore compaction threshold",
-      "description": "If more than this number of HStoreFiles in any one HStore then a compaction is run to rewrite all HStoreFiles files as one.",
-      "defaultValue": "3",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 0
-    },
-    {
-      "id": "puppet var",
-      "name": "hfile_blockcache_size",
-      "displayName": "HFile block cache size ",
-      "description": "Percentage of maximum heap (-Xmx setting) to allocate to block cache used by HFile/StoreFile. Set to 0 to disable but it's not recommended.",
-      "defaultValue": "0.40",
-      "displayType": "float",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
-      "name": "hstorefile_maxsize",
-      "displayName": "Maximum HStoreFile Size",
-      "description": "If any one of a column families' HStoreFiles has grown to exceed this value, the hosting HRegion is split in two.",
-      "defaultValue": "10737418240",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "client_scannercaching",
-      "displayName": "HBase Client Scanner Caching",
-      "description": "Number of rows that will be fetched when calling next on a scanner if it is not served from (local, client) memory. Do not set this value such that the time between invocations is greater than the scanner timeout",
-      "defaultValue": "100",
-      "displayType": "int",
-      "unit": "rows",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "zookeeper_sessiontimeout",
-      "displayName": "Zookeeper timeout for HBase Session",
-      "description": "HBase passes this to the zk quorum as suggested maximum time for a session",
-      "defaultValue": "30000",
-      "displayType": "int",
-      "unit": "ms",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
-      "name": "hfile_max_keyvalue_size",
-      "displayName": "HBase Client Maximum key-value Size",
-      "description": "Specifies the combined maximum allowed size of a KeyValue instance. It should be set to a fraction of the maximum region size.",
-      "defaultValue": "10485760",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
       "name": "hbase_log_dir",
       "displayName": "HBase Log Dir",
       "description": "Directory for HBase logs",
@@ -814,72 +546,6 @@ module.exports =
       "serviceName": "HBASE",
       "category": "Advanced"
     },
-    {
-      "id": "puppet var",
-      "name": "hbase_tmp_dir",
-      "displayName": "HBase local directory",
-      "description": "Temporary directory on the local filesystem",
-      "defaultDirectory": "/hadoop/hbase",
-      "defaultValue": "",
-      "displayType": "directory",
-      "isVisible": true,
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "hstore_blockingstorefiles",
-      "displayName": "hstore blocking storefiles",
-      "description": "If more than this number of StoreFiles in any one Store (one StoreFile is written per flush of " +
-        "MemStore) then updates are blocked for this HRegion until a compaction is completed, or until " +
-        "hbase.hstore.blockingWaitTime has been exceeded.",
-      "defaultValue": 10,
-      "isRequired": true,
-      "displayType": "init",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "regionserver_memstore_lab",
-      "displayName": "regionserver_memstore_lab",
-      "description": "",
-      "defaultValue": true,
-      "isRequired": false,
-      "displayType": "checkbox",
-      "isVisible": false,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "regionserver_memstore_lowerlimit",
-      "displayName": "regionserver_memstore_lowerlimit",
-      "description": "",
-      "defaultValue": "0.38",
-      "isRequired": false,
-      "displayType": "float",
-      "isVisible": false,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "regionserver_memstore_upperlimit",
-      "displayName": "regionserver_memstore_upperlimit",
-      "description": "",
-      "defaultValue": "0.4",
-      "isRequired": true,
-      "displayType": "float",
-      "isVisible": false,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
   /**********************************************HIVE***************************************/
     {
       "id": "puppet var",
@@ -1060,50 +726,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "hive_metastore_user_name",
-      "displayName": "Database Username",
-      "description": "Database user name to use to connect to the database",
-      "defaultValue": "hive",
-      "isReconfigurable": true,
-      "displayType": "host",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HIVE",
-      "category": "Hive Metastore",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
-      "name": "hive_metastore_user_passwd",
-      "displayName": "Database Password",
-      "description": "Database password to use to connect to the PostgreSQL database",
-      "defaultValue": "",
-      "isReconfigurable": true,
-      "displayType": "password",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HIVE",
-      "category": "Hive Metastore",
-      "index": 6
-    },
-    {
-      "id": "puppet var",
-      "name": "hive_jdbc_connection_url",
-      "displayName": "Database URL",
-      "value": "",
-      "defaultValue": "jdbc", // set to a 'jdbc' to not include this in initial error count
-      "description": "The JDBC connection URL to the database",
-      "displayType": "advanced",
-      "isOverridable": false,
-      "isVisible": true,
-      "serviceName": "HIVE",
-      "category": "Hive Metastore",
-      "index": 7
-    },
-    {
-      "id": "puppet var",
       "name": "hive_metastore_port",
       "displayName": "Hive metastore port",
       "description": "",
@@ -1376,66 +998,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "oozie_database_name",
-      "displayName": "Database Name",
-      "description": "Database name used for the Oozie",
-      "defaultValue": "oozie",
-      "isReconfigurable": true,
-      "isOverridable": false,
-      "displayType": "host",
-      "isVisible": true,
-      "isObserved": true,
-      //"domain": "global",
-      "serviceName": "OOZIE",
-      "category": "Oozie Server",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "oozie_metastore_user_name",
-      "displayName": "Database Username",
-      "description": "Database user name to use to connect to the database",
-      "defaultValue": "oozie",
-      "isReconfigurable": true,
-      "isOverridable": false,
-      "displayType": "host",
-      "isVisible": true,
-      //"domain": "global",
-      "serviceName": "OOZIE",
-      "category": "Oozie Server",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
-      "name": "oozie_metastore_user_passwd",
-      "displayName": "Database Password",
-      "description": "Database password to use to connect to the database",
-      "defaultValue": "",
-      "isReconfigurable": true,
-      "isOverridable": false,
-      "displayType": "password",
-      "isVisible": true,
-      //"domain": "global",
-      "serviceName": "OOZIE",
-      "category": "Oozie Server",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
-      "name": "oozie_jdbc_connection_url",
-      "displayName": "Database URL",
-      "description": "The JDBC connection URL to the database",
-      "defaultValue": "jdbc",  // set to a 'jdbc' to not include this in initial error count
-      "isReconfigurable": true,
-      "isOverridable": false,
-      "displayType": "advanced",
-      "isVisible": true,
-      "serviceName": "OOZIE",
-      "category": "Oozie Server",
-      "index": 6
-    },
-    {
-      "id": "puppet var",
       "name": "oozie_data_dir",
       "displayName": "Oozie Data Dir",
       "description": "Data directory in which the Oozie DB exists",

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/data/HDP2/site_properties.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/site_properties.js b/ambari-web/app/data/HDP2/site_properties.js
index c30660d..7326dc6 100644
--- a/ambari-web/app/data/HDP2/site_properties.js
+++ b/ambari-web/app/data/HDP2/site_properties.js
@@ -18,6 +18,147 @@
 module.exports =
 {
   "configProperties": [
+  /**********************************************HDFS***************************************/
+    {
+      "id": "site property",
+      "name": "dfs.namenode.checkpoint.dir",
+      "displayName": "SecondaryNameNode Checkpoint directory",
+      "description": "Directory on the local filesystem where the Secondary NameNode should store the temporary images to merge",
+      "defaultValue": "",
+      "defaultDirectory": "/hadoop/hdfs/namesecondary",
+      "displayType": "directory",
+      "isOverridable": false,
+      "isVisible": true,
+      "serviceName": "HDFS",
+      "category": "SNameNode",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "dfs.namenode.checkpoint.period",
+      "displayName": "HDFS Maximum Checkpoint Delay",
+      "description": "Maximum delay between two consecutive checkpoints for HDFS",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "seconds",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HDFS",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "fs.checkpoint.size",
+      "displayName": "HDFS Maximum Edit Log Size for Checkpointing",
+      "description": "Maximum size of the edits log file that forces an urgent checkpoint even if the maximum checkpoint delay is not reached",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "isVisible": true,
+      "serviceName": "HDFS",
+      "category": "General",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "dfs.namenode.name.dir",
+      "displayName": "NameNode directories",
+      "description": "NameNode directories for HDFS to store the file system image",
+      "defaultValue": "",
+      "defaultDirectory": "/hadoop/hdfs/namenode",
+      "displayType": "directories",
+      "isOverridable": false,
+      "isVisible": true,
+      "serviceName": "HDFS",
+      "category": "NameNode",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "dfs.webhdfs.enabled",
+      "displayName": "WebHDFS enabled",
+      "description": "Whether to enable WebHDFS feature",
+      "defaultValue": true,
+      "displayType": "checkbox",
+      "isOverridable": false,
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HDFS",
+      "index": 0
+    },
+    {
+      "id": "site property",
+      "name": "dfs.datanode.failed.volumes.tolerated",
+      "displayName": "DataNode volumes failure toleration",
+      "description": "The number of volumes that are allowed to fail before a DataNode stops offering service",
+      "defaultValue": "0",
+      "displayType": "int",
+      "isVisible": true,
+      "category": "DataNode",
+      "serviceName": "HDFS",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "dfs.datanode.data.dir",
+      "displayName": "DataNode directories",
+      "description": "DataNode directories for HDFS to store the data blocks",
+      "defaultValue": "",
+      "defaultDirectory": "/hadoop/hdfs/data",
+      "displayType": "directories",
+      "isVisible": true,
+      "category": "DataNode",
+      "serviceName": "HDFS",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "dfs.datanode.data.dir.perm",
+      "displayName": "DataNode directories permission",
+      "description": "",
+      "defaultValue": "",
+      "displayType": "int",
+      "isVisible": false,
+      "category": "DataNode",
+      "serviceName": "HDFS"
+    },
+    {
+      "id": "site property",
+      "name": "dfs.replication",
+      "displayName": "Block replication",
+      "description": "Default block replication.",
+      "displayType": "int",
+      "defaultValue": "",
+      "isVisible": true,
+      "category": "Advanced",
+      "serviceName": "HDFS"
+    },
+    {
+      "id": "site property",
+      "name": "dfs.datanode.du.reserved",
+      "displayName": "Reserved space for HDFS",
+      "description": "Reserved space in bytes per volume. Always leave this much space free for non dfs use.",
+      "defaultValue": "1073741824",
+      "displayType": "int",
+      "unit": "bytes",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HDFS",
+      "index": 2
+    },
+    {
+      "id": "site property",
+      "name": "dfs.client.read.shortcircuit",
+      "displayName": "HDFS Short-circuit read",
+      "description": "This configuration parameter turns on short-circuit local reads.",
+      "defaultValue": "",
+      "displayType": "checkbox",
+      "category": "Advanced",
+      "serviceName": "HDFS"
+
+    },
+
+  /**********************************************YARN***************************************/
     {
       "id": "site property",
       "name": "yarn.acl.enable",
@@ -158,6 +299,8 @@ module.exports =
       "serviceName": "YARN",
       "category": "NodeManager"
     },
+
+  /**********************************************MAPREDUCE2***************************************/
     {
       "id": "site property",
       "name": "mapreduce.map.memory.mb",
@@ -203,6 +346,307 @@ module.exports =
       "category": "Advanced",
       "value": "",
       "defaultValue": ""
+    },
+
+  /**********************************************oozie-site***************************************/
+    {
+      "id": "site property",
+      "name": "oozie.db.schema.name",
+      "displayName": "Database Name",
+      "description": "Database name used for the Oozie",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "host",
+      "isVisible": true,
+      "isObserved": true,
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.username",
+      "displayName": "Database Username",
+      "description": "Database user name to use to connect to the database",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "host",
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.password",
+      "displayName": "Database Password",
+      "description": "Database password to use to connect to the database",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "password",
+      "isVisible": true,
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.driver",
+      "displayName": "JDBC driver class",
+      "defaultValue": "",
+      "value": "",     // the value is overwritten in code
+      "isVisible": false,
+      "description": "Database name used for the Oozie",
+      "category": "Advanced",
+      "serviceName": "OOZIE"
+    },
+    {
+      "id": "site property",
+      "name": "oozie.service.JPAService.jdbc.url",
+      "displayName": "Database URL",
+      "description": "The JDBC connection URL to the database",
+      "defaultValue": "",
+      "isOverridable": false,
+      "displayType": "advanced",
+      "category": "Oozie Server",
+      "serviceName": "OOZIE",
+      "index": 6
+    },
+
+  /**********************************************hive-site***************************************/
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionDriverName",
+      "displayName": "JDBC driver class",
+      "defaultValue": "",
+      "value": "",     // the value is overwritten in code
+      "isVisible": false,
+      "description": "Driver class name for a JDBC metastore",
+      "category": "Advanced",
+      "serviceName": "HIVE"
+    },
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionUserName",
+      "displayName": "Database Username",
+      "description": "Database user name to use to connect to the database",
+      "defaultValue": "hive",
+      "displayType": "host",
+      "isOverridable": false,
+      "category": "Hive Metastore",
+      "serviceName": "HIVE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionPassword",
+      "displayName": "Database Password",
+      "description": "Database password to use to connect to the metastore database",
+      "defaultValue": "",
+      "displayType": "password",
+      "isOverridable": false,
+      "isVisible": true,
+      "category": "Hive Metastore",
+      "serviceName": "HIVE",
+      "index": 6
+    },
+    {
+      "id": "site property",
+      "name": "javax.jdo.option.ConnectionURL",
+      "displayName": "Database URL",
+      "value": "",
+      "defaultValue": "", // set to a 'jdbc' to not include this in initial error count
+      "description": "The JDBC connection URL to the database",
+      "displayType": "advanced",
+      "isOverridable": false,
+      "isVisible": true,
+      "category": "Hive Metastore",
+      "serviceName": "HIVE",
+      "index": 7
+    },
+
+  /**********************************************hbase-site***************************************/
+    {
+      "id": "site property",
+      "name": "hbase.tmp.dir",
+      "displayName": "HBase local directory",
+      "description": "Temporary directory on the local filesystem",
+      "defaultDirectory": "/hadoop/hbase",
+      "defaultValue": "",
+      "displayType": "directory",
+      "isVisible": true,
+      "category": "Advanced",
+      "serviceName": "HBASE"
+
+    },
+    {
+      "id": "site property",
+      "name": "hbase.regionserver.global.memstore.upperLimit",
+      "displayName": "hbase.regionserver.global.memstore.upperLimit",
+      "description": "",
+      "defaultValue": "",
+      "displayType": "float",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.regionserver.global.memstore.lowerLimit",
+      "displayName": "hbase.regionserver.global.memstore.lowerLimit",
+      "defaultValue": "",
+      "description": "When memstores are being forced to flush to make room in\
+      memory, keep flushing until we hit this mark. Defaults to 35% of heap.\
+      This value equal to hbase.regionserver.global.memstore.upperLimit causes\
+      the minimum possible flushing to occur when updates are blocked due to\
+      memstore limiting.",
+      "displayType": "float",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hstore.blockingStoreFiles",
+      "displayName": "hstore blocking storefiles",
+      "description": "If more than this number of StoreFiles in any one Store (one StoreFile is written per flush of " +
+        "MemStore) then updates are blocked for this HRegion until a compaction is completed, or until " +
+        "hbase.hstore.blockingWaitTime has been exceeded.",
+      "defaultValue": "",
+      "isRequired": true,
+      "displayType": "int",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hstore.compactionThreshold",
+      "displayName": "HBase HStore compaction threshold",
+      "description": "If more than this number of HStoreFiles in any one HStore then a compaction is run to rewrite all HStoreFiles files as one.",
+      "defaultValue": "3",
+      "displayType": "int",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 0
+    },
+    {
+      "id": "site property",
+      "name": "hfile.block.cache.size",
+      "displayName": "HFile block cache size ",
+      "description": "Percentage of maximum heap (-Xmx setting) to allocate to block cache used by HFile/StoreFile. Set to 0 to disable but it's not recommended.",
+      "defaultValue": "0.40",
+      "displayType": "float",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 1
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.max.filesize",
+      "displayName": "Maximum HStoreFile Size",
+      "description": "If any one of a column families' HStoreFiles has grown to exceed this value, the hosting HRegion is split in two.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 2
+    },
+    {
+      "id": "site property",
+      "name": "hbase.regionserver.handler.count",
+      "displayName": "HBase RegionServer Handler",
+      "description": "Count of RPC Listener instances spun up on RegionServers",
+      "defaultValue": "60",
+      "displayType": "int",
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 2
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.majorcompaction",
+      "displayName": "HBase Region Major Compaction",
+      "description": "The time between major compactions of all HStoreFiles in a region. Set to 0 to disable automated major compactions.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "ms",
+      "isVisible": true,
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.memstore.block.multiplier",
+      "displayName": "HBase Region Block Multiplier",
+      "description": "Block updates if memstore has \"Multiplier * HBase Region Memstore Flush Size\" bytes. Useful preventing runaway memstore during spikes in update traffic",
+      "defaultValue": "",
+      "displayType": "int",
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.memstore.mslab.enabled",
+      "displayName": "hbase.hregion.memstore.mslab.enabled",
+      "description": "Enables the MemStore-Local Allocation Buffer,\
+      a feature which works to prevent heap fragmentation under\
+      heavy write loads. This can reduce the frequency of stop-the-world\
+      GC pauses on large heaps.",
+      "defaultValue": "",
+      "displayType": "checkbox",
+      "category": "Advanced",
+      "serviceName": "HBASE"
+    },
+    {
+      "id": "site property",
+      "name": "hbase.hregion.memstore.flush.size",
+      "displayName": "HBase Region Memstore Flush Size",
+      "description": "Memstore will be flushed to disk if size of the memstore exceeds this number of bytes.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "category": "RegionServer",
+      "serviceName": "HBASE",
+      "index": 5
+    },
+    {
+      "id": "site property",
+      "name": "hbase.client.scanner.caching",
+      "displayName": "HBase Client Scanner Caching",
+      "description": "Number of rows that will be fetched when calling next on a scanner if it is not served from \
+      (local, client) memory. Do not set this value such that the time between invocations is greater than the scanner timeout",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "rows",
+      "isVisible": true,
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 3
+    },
+    {
+      "id": "site property",
+      "name": "zookeeper.session.timeout",
+      "displayName": "Zookeeper timeout for HBase Session",
+      "description": "HBase passes this to the zk quorum as suggested maximum time for a session",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "ms",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 4
+    },
+    {
+      "id": "site property",
+      "name": "hbase.client.keyvalue.maxsize",
+      "displayName": "HBase Client Maximum key-value Size",
+      "description": "Specifies the combined maximum allowed size of a KeyValue instance. It should be set to a fraction of the maximum region size.",
+      "defaultValue": "",
+      "displayType": "int",
+      "unit": "bytes",
+      "category": "General",
+      "serviceName": "HBASE",
+      "index": 5
     }
   ]
 };

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/data/config_mapping.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/config_mapping.js b/ambari-web/app/data/config_mapping.js
index 9603dfb..8581b93 100644
--- a/ambari-web/app/data/config_mapping.js
+++ b/ambari-web/app/data/config_mapping.js
@@ -17,6 +17,7 @@
  */
 
 var configs = [
+  /**********************************************HDFS***************************************/
   {
     "name": "fs.default.name",
     "templateName": ["namenode_host"],
@@ -25,27 +26,6 @@ var configs = [
     "filename": "core-site.xml"
   },
   {
-    "name": "fs.checkpoint.dir",
-    "templateName": ["fs_checkpoint_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "core-site.xml"
-  },
-  {
-    "name": "fs.checkpoint.period",
-    "templateName": ["fs_checkpoint_period"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "core-site.xml"
-  },
-  {
-    "name": "fs.checkpoint.size",
-    "templateName": ["fs_checkpoint_size"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "core-site.xml"
-  },
-  {
     "name": "hadoop.proxyuser.<foreignKey[0]>.groups",
     "templateName": ["proxyuser_group"],
     "foreignKey": ["hive_user"],
@@ -94,50 +74,6 @@ var configs = [
     "isOverridable" : true
   },
   {
-    "name": "dfs.name.dir",
-    "templateName": ["dfs_name_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  /*
-  {
-    "name": "dfs.support.append",
-    "templateName": ["dfs_support_append"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  */
-  {
-    "name": "dfs.webhdfs.enabled",
-    "templateName": ["dfs_webhdfs_enabled"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "dfs.datanode.failed.volumes.tolerated",
-    "templateName": ["dfs_datanode_failed_volume_tolerated"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "dfs.block.local-path-access.user",
-    "templateName": ["dfs_block_local_path_access_user"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "dfs.data.dir",
-    "templateName": ["dfs_data_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
     "name": "dfs.hosts.exclude",
     "templateName": ["hadoop_conf_dir", "dfs_exclude"],
     "foreignKey": null,
@@ -152,13 +88,6 @@ var configs = [
     "filename": "hdfs-site.xml"
   },
   {
-    "name": "dfs.replication",
-    "templateName": ["dfs_replication"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
     "name": "dfs.datanode.address",
     "templateName": ["dfs_datanode_address"],
     "foreignKey": null,
@@ -180,13 +109,6 @@ var configs = [
     "filename": "hdfs-site.xml"
   },
   {
-    "name": "dfs.datanode.du.reserved",
-    "templateName": ["datanode_du_reserved"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
     "name": "dfs.secondary.http.address",
     "templateName": ["snamenode_host"],
     "foreignKey": null,
@@ -200,160 +122,8 @@ var configs = [
     "value": "<templateName[0]>:50470",
     "filename": "hdfs-site.xml"
   },
-  {
-    "name": "dfs.datanode.data.dir.perm",
-    "templateName": ["dfs_datanode_data_dir_perm"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "mapred.local.dir",
-    "templateName": ["mapred_local_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "oozie.base.url",
-    "templateName": ["oozieserver_host"],
-    "foreignKey": null,
-    "value": "http://<templateName[0]>:11000/oozie",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.create.db.schema",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "false",  // this is always false
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.db.schema.name",
-    "templateName": ['oozie_database_name'],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.driver",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "org.apache.derby.jdbc.EmbeddedDriver",  // this value is overwritten in code
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.username",
-    "templateName": ['oozie_metastore_user_name'],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.password",
-    "templateName": ['oozie_metastore_user_passwd'],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.url",
-    "templateName": ["oozie_jdbc_connection_url"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "javax.jdo.option.ConnectionDriverName",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "com.mysql.jdbc.Driver",  // this value is overwritten in code
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "javax.jdo.option.ConnectionUserName",
-    "templateName": ["hive_metastore_user_name"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "javax.jdo.option.ConnectionPassword",
-    "templateName": ["hive_metastore_user_passwd"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "javax.jdo.option.ConnectionURL",
-    "templateName": ["hive_jdbc_connection_url"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "hive.metastore.uris",
-    "templateName": ["hivemetastore_host"],
-    "foreignKey": null,
-    "value": "thrift://<templateName[0]>:9083",
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "mapred.jobtracker.taskScheduler",
-    "templateName": ["scheduler_name"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.tasktracker.map.tasks.maximum",
-    "templateName": ["mapred_map_tasks_max"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.tasktracker.reduce.tasks.maximum",
-    "templateName": ["mapred_red_tasks_max"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.cluster.reduce.memory.mb",
-    "templateName": ["mapred_cluster_red_mem_mb"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.job.map.memory.mb",
-    "templateName": ["mapred_job_map_mem_mb"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.cluster.max.map.memory.mb",
-    "templateName": ["mapred_cluster_max_map_mem_mb"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.cluster.max.reduce.memory.mb",
-    "templateName": ["mapred_cluster_max_red_mem_mb"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.job.reduce.memory.mb",
-    "templateName": ["mapred_job_red_mem_mb"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
+
+  /******************************************MAPREDUCE***************************************/
   {
     "name": "mapred.hosts",
     "templateName": ["hadoop_conf_dir", "mapred_hosts_include"],
@@ -369,13 +139,6 @@ var configs = [
     "filename": "mapred-site.xml"
   },
   {
-    "name": "mapred.healthChecker.script.path",
-    "templateName": ["mapred_jobstatus_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
     "name": "mapred.job.tracker.persist.jobstatus.dir",
     "templateName": ["hadoop_conf_dir"],
     "foreignKey": null,
@@ -390,34 +153,6 @@ var configs = [
     "filename": "mapred-site.xml"
   },
   {
-    "name": "mapred.cluster.map.memory.mb",
-    "templateName": ["mapred_cluster_map_mem_mb"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "io.sort.mb",
-    "templateName": ["io_sort_mb"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "io.sort.spill.percent",
-    "templateName": ["io_sort_spill_percent"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.system.dir",
-    "templateName": ["mapred_system_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
     "name": "mapred.job.tracker",
     "templateName": ["jobtracker_host"],
     "foreignKey": null,
@@ -432,190 +167,50 @@ var configs = [
     "filename": "mapred-site.xml"
   },
   {
-    "name": "mapred.userlog.retain.hours",
-    "templateName": ["mapreduce_userlog_retainhours"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapred.jobtracker.maxtasks.per.job",
-    "templateName": ["maxtasks_per_job"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "mapred-site.xml"
-  },
-  {
-    "name": "mapreduce.history.server.embedded",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "false",
-    "filename": "mapred-site.xml"
-  },
-  {
     "name": "mapreduce.history.server.http.address",
     "templateName": ["jobtracker_host"],
     "foreignKey": null,
     "value": "<templateName[0]>:51111",
     "filename": "mapred-site.xml"
   },
+
+  /**********************************************oozie-site***************************************/
   {
-    "name": "hbase.rootdir",
-    "templateName": ["namenode_host", "hbase_hdfs_root_dir"],
-    "foreignKey": null,
-    "value": "hdfs://<templateName[0]>:8020<templateName[1]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.tmp.dir",
-    "templateName": ["hbase_tmp_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  /*
-   {
-   "name": "hbase.master.info.bindAddress",
-   "templateName": ["hbasemaster.host"],
-   "foreignKey": null,
-   "value": "<templateName[0]>",
-   "filename": "hbase-site.xml"
-   },
-   */
-  {
-    "name": "hbase.regionserver.global.memstore.upperLimit",
-    "templateName": ["regionserver_memstore_upperlimit"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hstore.blockingStoreFiles",
-    "templateName": ["hstore_blockingstorefiles"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hstore.compactionThreshold",
-    "templateName": ["hstore_compactionthreshold"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hfile.block.cache.size",
-    "templateName": ["hfile_blockcache_size"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.max.filesize",
-    "templateName": ["hstorefile_maxsize"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.regionserver.handler.count",
-    "templateName": ["regionserver_handlers"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.majorcompaction",
-    "templateName": ["hregion_majorcompaction"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.regionserver.global.memstore.lowerLimit",
-    "templateName": ["regionserver_memstore_lowerlimit"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.memstore.block.multiplier",
-    "templateName": ["hregion_blockmultiplier"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.memstore.mslab.enabled",
-    "templateName": ["regionserver_memstore_lab"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.memstore.flush.size",
-    "templateName": ["hregion_memstoreflushsize"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.client.scanner.caching",
-    "templateName": ["client_scannercaching"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.client.scanner.caching",
-    "templateName": ["client_scannercaching"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.cluster.distributed",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "true",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.zookeeper.property.clientPort",
-    "templateName": [],
+    "name": "oozie.base.url",
+    "templateName": ["oozieserver_host"],
     "foreignKey": null,
-    "value": "2181",
-    "filename": "hbase-site.xml"
+    "value": "http://<templateName[0]>:11000/oozie",
+    "filename": "oozie-site.xml"
   },
+
+  /**********************************************hive-site***************************************/
+
   {
-    "name": "zookeeper.session.timeout",
-    "templateName": ["zookeeper_sessiontimeout"],
+    "name": "hive.metastore.uris",
+    "templateName": ["hivemetastore_host"],
     "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
+    "value": "thrift://<templateName[0]>:9083",
+    "filename": "hive-site.xml"
   },
+
+/**********************************************hbase-site***************************************/
   {
-    "name": "hbase.client.keyvalue.maxsize",
-    "templateName": ["hfile_max_keyvalue_size"],
+    "name": "hbase.rootdir",
+    "templateName": ["namenode_host", "hbase_hdfs_root_dir"],
     "foreignKey": null,
-    "value": "<templateName[0]>",
+    "value": "hdfs://<templateName[0]>:8020<templateName[1]>",
     "filename": "hbase-site.xml"
   },
   {
-    "name": "dfs.support.append",
-    "templateName": ["hdfs_support_append"],
+    "name": "hbase.zookeeper.quorum",
+    "templateName": ["zookeeperserver_hosts"],
     "foreignKey": null,
     "value": "<templateName[0]>",
     "filename": "hbase-site.xml"
   },
 
-  {
-    "name": "dfs.client.read.shortcircuit",
-    "templateName": ["hdfs_enable_shortcircuit_read"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
+/**********************************************webhcat-site***************************************/
+
   {
     "name": "templeton.zookeeper.hosts",
     "templateName": ["zookeeperserver_hosts"],
@@ -623,13 +218,8 @@ var configs = [
     "value": "<templateName[0]>",
     "filename": "webhcat-site.xml"
   },
-  {
-    "name": "hbase.zookeeper.quorum",
-    "templateName": ["zookeeperserver_hosts"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
+
+/**********************************************core-site for glusterfs***************************************/
   {
     "name": "fs.glusterfs.impl",
     "templateName": ["fs_glusterfs_impl"],

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/data/global_properties.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/global_properties.js b/ambari-web/app/data/global_properties.js
index 5b3e3fc..778af30 100644
--- a/ambari-web/app/data/global_properties.js
+++ b/ambari-web/app/data/global_properties.js
@@ -93,21 +93,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_name_dir",
-      "displayName": "NameNode directories",
-      "description": "NameNode directories for HDFS to store the file system image",
-      "defaultValue": "",
-      "defaultDirectory": "/hadoop/hdfs/namenode",
-      "displayType": "directories",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "NameNode",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
       "name": "namenode_heapsize",
       "displayName": "NameNode Java heap size",
       "description": "Initial and maximum Java heap size for NameNode (Java options -Xms and -Xmx).  This also applies to the Secondary NameNode.",
@@ -153,21 +138,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "fs_checkpoint_dir",
-      "displayName": "SecondaryNameNode Checkpoint directory",
-      "description": "Directory on the local filesystem where the Secondary NameNode should store the temporary images to merge",
-      "defaultValue": "",
-      "defaultDirectory": "/hadoop/hdfs/namesecondary",
-      "displayType": "directory",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "SNameNode",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
       "name": "datanode_hosts", //not in the schema. For UI purpose
       "displayName": "DataNode hosts",
       "value": "",
@@ -184,20 +154,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_data_dir",
-      "displayName": "DataNode directories",
-      "description": "DataNode directories for HDFS to store the data blocks",
-      "defaultValue": "",
-      "defaultDirectory": "/hadoop/hdfs/data",
-      "displayType": "directories",
-      "isVisible": true,
-      "domain": "datanode-global",
-      "serviceName": "HDFS",
-      "category": "DataNode",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
       "name": "dtnode_heapsize",
       "displayName": "DataNode maximum Java heap size",
       "description": "Maximum Java heap size for DataNode (Java option -Xmx)",
@@ -212,32 +168,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_datanode_failed_volume_tolerated",
-      "displayName": "DataNode volumes failure toleration",
-      "description": "The number of volumes that are allowed to fail before a DataNode stops offering service",
-      "defaultValue": "0",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "datanode-global",
-      "serviceName": "HDFS",
-      "category": "DataNode",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "dfs_webhdfs_enabled",
-      "displayName": "WebHDFS enabled",
-      "description": "Whether to enable WebHDFS feature",
-      "defaultValue": true,
-      "displayType": "checkbox",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "index": 0
-    },
-    {
-      "id": "puppet var",
       "name": "hadoop_heapsize",
       "displayName": "Hadoop maximum Java heap size",
       "description": "Maximum Java heap size for daemons such as Balancer (Java option -Xmx)",
@@ -251,47 +181,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "datanode_du_reserved",
-      "displayName": "Reserved space for HDFS",
-      "description": "Reserved space in bytes per volume. Always leave this much space free for non dfs use.",
-      "defaultValue": "0",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HDFS",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "fs_checkpoint_period",
-      "displayName": "HDFS Maximum Checkpoint Delay",
-      "description": "Maximum delay between two consecutive checkpoints for HDFS",
-      "defaultValue": "21600",
-      "displayType": "int",
-      "unit": "seconds",
-      "isVisible": true,
-      "domain": "global",
-      "filename": "core-site.xml",
-      "serviceName": "HDFS",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "fs_checkpoint_size",
-      "displayName": "HDFS Maximum Edit Log Size for Checkpointing",
-      "description": "Maximum size of the edits log file that forces an urgent checkpoint even if the maximum checkpoint delay is not reached",
-      "defaultValue": "67108864",
-      "displayType": "float",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "global",
-      "filename": "core-site.xml",
-      "serviceName": "HDFS",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
       "name": "hdfs_log_dir_prefix",
       "displayName": "Hadoop Log Dir Prefix",
       "description": "The parent directory for Hadoop log files.  The HDFS log directory will be ${hadoop_log_dir_prefix} / ${hdfs_user} and the MapReduce log directory will be ${hadoop_log_dir_prefix} / ${mapred_user}.",
@@ -373,46 +262,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "dfs_replication",
-      "displayName": "Block replication",
-      "description": "Default block replication.",
-      "displayType": "int",
-      "defaultValue": "3",
-      "isVisible": true,
-      "filename": "hdfs-site.xml",
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "dfs_block_local_path_access_user",
-      "displayName": "dfs.block.local-path-access.user",
-      "description": "the user who is allowed to perform short circuit reads",
-      "displayType": "advanced",
-      "defaultValue": "hbase",
-      "isVisible": true,
-      "filename": "hdfs-site.xml",
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "dfs_datanode_data_dir_perm",
-      "displayName": "dfs_datanode_data_dir_perm",
-      "description": "",
-      "defaultValue": "750",
-      "isReconfigurable": true,
-      "displayType": "int",
-      "isVisible": false,
-      "filename": "hdfs-site.xml",
-      "domain": "global",
-      "serviceName": "HDFS",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
       "name": "security_enabled",
       "displayName": "Hadoop Security",
       "description": "Enable hadoop security",
@@ -617,48 +466,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "mapred_local_dir",
-      "displayName": "MapReduce local directories",
-      "description": "Directories for MapReduce to store intermediate data files",
-      "defaultValue": "",
-      "defaultDirectory": "/hadoop/mapred",
-      "displayType": "directories",
-      "isReconfigurable": true,
-      "isVisible": true,
-      "domain": "tasktracker-global",
-      "serviceName": "MAPREDUCE",
-      "category": "TaskTracker",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_map_tasks_max",
-      "displayName": "Number of Map slots per node",
-      "description": "Number of slots that Map tasks that run simultaneously can occupy on a TaskTracker",
-      "defaultValue": "4",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "tasktracker-global",
-      "serviceName": "MAPREDUCE",
-      "category": "TaskTracker",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_red_tasks_max",
-      "displayName": "Number of Reduce slots per node",
-      "description": "Number of slots that Reduce tasks that run simultaneously can occupy on a TaskTracker.",
-      "defaultValue": "2",
-      "displayType": "int",
-      "isVisible": true,
-      "isRequired": false,
-      "domain": "tasktracker-global",
-      "serviceName": "MAPREDUCE",
-      "category": "TaskTracker",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
       "name": "mapred_child_java_opts_sz",
       "displayName": "Java options for MapReduce tasks",
       "description": "Java options for the TaskTracker child processes.",
@@ -673,146 +480,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "scheduler_name",
-      "displayName": "MapReduce Capacity Scheduler",
-      "description": "The scheduler to use for scheduling of MapReduce jobs",
-      "defaultValue": "org.apache.hadoop.mapred.CapacityTaskScheduler",
-      "displayType": "advanced",
-      "isOverridable": false,
-      "isVisible": true,
-      "serviceName": "MAPREDUCE",
-      "index": 0
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_cluster_map_mem_mb",
-      "displayName": "Cluster's Map slot size (virtual memory)",
-      "description": "The virtual memory size of a single Map slot in the MapReduce framework",
-      "defaultValue": "1536",
-      "displayType": "int",
-      "unit": "MB",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_cluster_red_mem_mb",
-      "displayName": "Cluster's Reduce slot size (virtual memory)",
-      "description": "The virtual memory size of a single Reduce slot in the MapReduce framework",
-      "defaultValue": "2048",
-      "displayType": "int",
-      "unit": "MB",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_cluster_max_map_mem_mb",
-      "displayName": "Upper limit on virtual memory for single Map task",
-      "description": "Upper limit on virtual memory size for a single Map task of any MapReduce job",
-      "defaultValue": "6144",
-      "displayType": "int",
-      "unit": "MB",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_cluster_max_red_mem_mb",
-      "displayName": "Upper limit on virtual memory for single Reduce task",
-      "description": "Upper limit on virtual memory size for a single Reduce task of any MapReduce job",
-      "defaultValue": "4096",
-      "displayType": "int",
-      "unit": "MB",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_job_map_mem_mb",
-      "displayName": "Default virtual memory for a job's map-task",
-      "description": "Virtual memory for single Map task",
-      "defaultValue": "1536",
-      "displayType": "int",
-      "unit": "MB",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
-      "name": "mapred_job_red_mem_mb",
-      "displayName": "Default virtual memory for a job's reduce-task",
-      "description": "Virtual memory for single Reduce task",
-      "defaultValue": "2048",
-      "displayType": "int",
-      "unit": "MB",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 6
-    },
-    {
-      "id": "puppet var",
-      "name": "io_sort_mb",
-      "displayName": "Map-side sort buffer memory",
-      "description": "The total amount of Map-side buffer memory to use while sorting files (Expert-only configuration)",
-      "defaultValue": "200",
-      "displayType": "int",
-      "unit": "MB",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 7
-    },
-    {
-      "id": "puppet var",
-      "name": "io_sort_spill_percent",
-      "displayName": "Limit on buffer",
-      "description": "Percentage of sort buffer used for record collection (Expert-only configuration)",
-      "defaultValue": "0.9",
-      "displayType": "float",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 8
-    },
-    {
-      "id": "puppet var",
-      "name": "mapreduce_userlog_retainhours",
-      "displayName": "Job log retention (hours)",
-      "description": "The maximum time, in hours, for which the user-logs are to be retained after the job completion.",
-      "defaultValue": "24",
-      "displayType": "int",
-      "unit": "hours",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 9
-    },
-    {
-      "id": "puppet var",
-      "name": "maxtasks_per_job",
-      "displayName": "Maximum number tasks for a Job",
-      "description": "Maximum number of tasks for a single Job",
-      "defaultValue": "-1",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "index": 10
-    },
-    {
-      "id": "puppet var",
       "name": "lzo_enabled",
       "displayName": "LZO compression",
       "description": "LZO compression enabled",
@@ -854,19 +521,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "mapred_system_dir",
-      "displayName": "MapReduce system directories",
-      "description": "",
-      "defaultValue": "/mapred/system",
-      "displayType": "directories",
-      "isReconfigurable": true,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "MAPREDUCE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
       "name": "mapred_hosts_exclude",
       "displayName": "Exclude hosts",
       "description": "Exclude entered hosts",
@@ -889,34 +543,6 @@ module.exports =
       "domain": "global",
       "category": "Advanced"
     },
-    {
-      "id": "puppet var",
-      "name": "mapred_jobstatus_dir",
-      "displayName": "Job Status directory",
-      "description": "Directory path to view job status",
-      "defaultValue": "file:////mapred/jobstatus",
-      "displayType": "advanced",
-      "isVisible": false,
-      "serviceName": "MAPREDUCE",
-      "domain": "global",
-      "category": "Advanced"
-    },
-    /*{
-      "id": "site property",
-      "name": "mapred.capacity-scheduler.maximum-system-jobs",
-      "displayName": "Max system jobs",
-      "displayType": "int",
-      "value": '3000',
-      "defaultValue": '3000',
-      "description": "Maximum number of jobs in the system which can be initialized, concurrently, by the " +
-        "CapacityScheduler. Individual queue limits on initialized jobs are directly proportional to their queue " +
-        "capacities.",
-      "isVisible": true,
-      "isRequired": true,
-      "serviceName": "MAPREDUCE",
-      "category": "CapacityScheduler",
-      "filename": "capacity-scheduler.xml"
-    },*/
   /**********************************************HBASE***************************************/
     {
       "id": "puppet var",
@@ -980,136 +606,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "regionserver_handlers",
-      "displayName": "HBase RegionServer Handler",
-      "description": "Count of RPC Listener instances spun up on RegionServers",
-      "defaultValue": "60",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "hregion_majorcompaction",
-      "displayName": "HBase Region Major Compaction",
-      "description": "The time between major compactions of all HStoreFiles in a region. Set to 0 to disable automated major compactions.",
-      "defaultValue": "86400000",
-      "displayType": "int",
-      "unit": "ms",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "hregion_blockmultiplier",
-      "displayName": "HBase Region Block Multiplier",
-      "description": "Block updates if memstore has \"Multiplier * HBase Region Memstore Flush Size\" bytes. Useful preventing runaway memstore during spikes in update traffic",
-      "defaultValue": "2",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
-      "name": "hregion_memstoreflushsize",
-      "displayName": "HBase Region Memstore Flush Size",
-      "description": "Memstore will be flushed to disk if size of the memstore exceeds this number of bytes.",
-      "defaultValue": "134217728",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "regionserver-global",
-      "serviceName": "HBASE",
-      "category": "RegionServer",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
-      "name": "hstore_compactionthreshold",
-      "displayName": "HBase HStore compaction threshold",
-      "description": "If more than this number of HStoreFiles in any one HStore then a compaction is run to rewrite all HStoreFiles files as one.",
-      "defaultValue": "3",
-      "displayType": "int",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 0
-    },
-    {
-      "id": "puppet var",
-      "name": "hfile_blockcache_size",
-      "displayName": "HFile block cache size ",
-      "description": "Percentage of maximum heap (-Xmx setting) to allocate to block cache used by HFile/StoreFile. Set to 0 to disable but it's not recommended.",
-      "defaultValue": "0.40",
-      "displayType": "float",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 1
-    },
-    {
-      "id": "puppet var",
-      "name": "hstorefile_maxsize",
-      "displayName": "Maximum HStoreFile Size",
-      "description": "If any one of a column families' HStoreFiles has grown to exceed this value, the hosting HRegion is split in two.",
-      "defaultValue": "10737418240",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 2
-    },
-    {
-      "id": "puppet var",
-      "name": "client_scannercaching",
-      "displayName": "HBase Client Scanner Caching",
-      "description": "Number of rows that will be fetched when calling next on a scanner if it is not served from (local, client) memory. Do not set this value such that the time between invocations is greater than the scanner timeout",
-      "defaultValue": "100",
-      "displayType": "int",
-      "unit": "rows",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 3
-    },
-    {
-      "id": "puppet var",
-      "name": "zookeeper_sessiontimeout",
-      "displayName": "Zookeeper timeout for HBase Session",
-      "description": "HBase passes this to the zk quorum as suggested maximum time for a session",
-      "defaultValue": "60000",
-      "displayType": "int",
-      "unit": "ms",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 4
-    },
-    {
-      "id": "puppet var",
-      "name": "hfile_max_keyvalue_size",
-      "displayName": "HBase Client Maximum key-value Size",
-      "description": "Specifies the combined maximum allowed size of a KeyValue instance. It should be set to a fraction of the maximum region size.",
-      "defaultValue": "10485760",
-      "displayType": "int",
-      "unit": "bytes",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
       "name": "hbase_log_dir",
       "displayName": "HBase Log Dir",
       "description": "Directory for HBase logs",
@@ -1149,98 +645,6 @@ module.exports =
       "serviceName": "HBASE",
       "category": "Advanced"
     },
-    {
-      "id": "puppet var",
-      "name": "hbase_tmp_dir",
-      "displayName": "HBase local directory",
-      "description": "Temporary directory on the local filesystem",
-      "defaultDirectory": "/hadoop/hbase",
-      "defaultValue": "",
-      "displayType": "directory",
-      "isVisible": true,
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "hdfs_enable_shortcircuit_read",
-      "displayName": "HDFS Short-circuit read",
-      "description": "",
-      "defaultValue": true,
-      "isRequired": false,
-      "displayType": "checkbox",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "hdfs_support_append",
-      "displayName": "HDFS append support",
-      "description": "HDFS append support",
-      "defaultValue": true,
-      "isRequired": false,
-      "displayType": "checkbox",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "hstore_blockingstorefiles",
-      "displayName": "hstore blocking storefiles",
-      "description": "If more than this number of StoreFiles in any one Store (one StoreFile is written per flush of " +
-        "MemStore) then updates are blocked for this HRegion until a compaction is completed, or until " +
-        "hbase.hstore.blockingWaitTime has been exceeded.",
-      "defaultValue": 7,
-      "isRequired": true,
-      "displayType": "init",
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "regionserver_memstore_lab",
-      "displayName": "regionserver_memstore_lab",
-      "description": "",
-      "defaultValue": true,
-      "isRequired": false,
-      "displayType": "checkbox",
-      "isVisible": false,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "regionserver_memstore_lowerlimit",
-      "displayName": "regionserver_memstore_lowerlimit",
-      "description": "",
-      "defaultValue": "0.38",
-      "isRequired": false,
-      "displayType": "float",
-      "isVisible": false,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
-    {
-      "id": "puppet var",
-      "name": "regionserver_memstore_upperlimit",
-      "displayName": "regionserver_memstore_upperlimit",
-      "description": "",
-      "defaultValue": "0.4",
-      "isRequired": true,
-      "displayType": "float",
-      "isVisible": false,
-      "domain": "global",
-      "serviceName": "HBASE",
-      "category": "Advanced"
-    },
   /**********************************************HIVE***************************************/
     {
       "id": "puppet var",
@@ -1421,36 +825,6 @@ module.exports =
     },
     {
       "id": "puppet var",
-      "name": "hive_metastore_user_name",
-      "displayName": "Database Username",
-      "description": "Database user name to use to connect to the database",
-      "defaultValue": "hive",
-      "isReconfigurable": true,
-      "displayType": "host",  // don't apply user validation
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HIVE",
-      "category": "Hive Metastore",
-      "index": 5
-    },
-    {
-      "id": "puppet var",
-      "name": "hive_metastore_user_passwd",
-      "displayName": "Database Password",
-      "description": "Database password to use to connect to the PostgreSQL database",
-      "defaultValue": "",
-      "isReconfigurable": true,
-      "displayType": "password",
-      "isOverridable": false,
-      "isVisible": true,
-      "domain": "global",
-      "serviceName": "HIVE",
-      "category": "Hive Metastore",
-      "index": 6
-    },
-    {
-      "id": "puppet var",
       "name": "hive_jdbc_connection_url",
       "displayName": "Database URL",
       "value": "",
@@ -1798,66 +1172,6 @@ module.exports =
           "category": "Oozie Server",
           "index": 3
       },
-      {
-          "id": "puppet var",
-          "name": "oozie_database_name",
-          "displayName": "Database Name",
-          "description": "Database name used for the Oozie",
-          "defaultValue": "oozie",
-          "isReconfigurable": true,
-          "isOverridable": false,
-          "displayType": "host",
-          "isVisible": true,
-          "isObserved": true,
-          //"domain": "global",
-          "serviceName": "OOZIE",
-          "category": "Oozie Server",
-          "index": 4
-      },
-      {
-          "id": "puppet var",
-          "name": "oozie_metastore_user_name",
-          "displayName": "Database Username",
-          "description": "Database user name to use to connect to the database",
-          "defaultValue": "oozie",
-          "isReconfigurable": true,
-          "isOverridable": false,
-          "displayType": "host",  // don't apply user validation
-          "isVisible": true,
-          //"domain": "global",
-          "serviceName": "OOZIE",
-          "category": "Oozie Server",
-          "index": 5
-      },
-      {
-          "id": "puppet var",
-          "name": "oozie_metastore_user_passwd",
-          "displayName": "Database Password",
-          "description": "Database password to use to connect to the database",
-          "defaultValue": "",
-          "isReconfigurable": true,
-          "isOverridable": false,
-          "displayType": "password",
-          "isVisible": true,
-          //"domain": "global",
-          "serviceName": "OOZIE",
-          "category": "Oozie Server",
-          "index": 6
-      },
-    {
-      "id": "puppet var",
-      "name": "oozie_jdbc_connection_url",
-      "displayName": "Database URL",
-      "description": "The JDBC connection URL to the database",
-      "defaultValue": "jdbc",  // set to a 'jdbc' to not include this in initial error count
-      "isReconfigurable": true,
-      "isOverridable": false,
-      "displayType": "advanced",
-      "isVisible": true,
-      "serviceName": "OOZIE",
-      "category": "Oozie Server",
-      "index": 7
-    },
     {
       "id": "puppet var",
       "name": "oozie_data_dir",


[4/4] git commit: AMBARI-3568: Clean up unnecessary globals from Ambari Web. (jaimin)

Posted by ja...@apache.org.
AMBARI-3568: Clean up unnecessary globals from Ambari Web. (jaimin)


Project: http://git-wip-us.apache.org/repos/asf/incubator-ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-ambari/commit/88f51325
Tree: http://git-wip-us.apache.org/repos/asf/incubator-ambari/tree/88f51325
Diff: http://git-wip-us.apache.org/repos/asf/incubator-ambari/diff/88f51325

Branch: refs/heads/trunk
Commit: 88f513259283798b8746a607354d345e6754890d
Parents: 63d5638
Author: Jaimin Jetly <ja...@hortonworks.com>
Authored: Tue Oct 22 15:55:54 2013 -0700
Committer: Jaimin Jetly <ja...@hortonworks.com>
Committed: Tue Oct 22 15:56:21 2013 -0700

----------------------------------------------------------------------
 .../puppet/modules/hdp-hive/manifests/params.pp |   4 +-
 .../services/HBASE/configuration/hbase-site.xml |  36 +-
 .../services/HDFS/configuration/core-site.xml   |   4 +-
 .../services/HDFS/configuration/hdfs-site.xml   |  18 +-
 .../services/HIVE/configuration/hive-site.xml   |   6 +-
 .../MAPREDUCE/configuration/mapred-site.xml     |  50 +-
 .../services/OOZIE/configuration/oozie-site.xml |   2 +-
 .../services/HBASE/configuration/hbase-site.xml |  30 +-
 .../services/HDFS/configuration/hdfs-site.xml   |  21 +-
 .../services/HIVE/configuration/hive-site.xml   |   6 +-
 .../services/OOZIE/configuration/oozie-site.xml |   4 +-
 .../services/HBASE/configuration/hbase-site.xml | 198 +++---
 .../services/HDFS/configuration/core-site.xml   | 274 ++++----
 .../services/HDFS/configuration/hdfs-site.xml   | 388 +++++------
 .../services/HIVE/configuration/hive-site.xml   |  14 +-
 .../services/OOZIE/configuration/oozie-site.xml | 388 +++++------
 .../services/HBASE/configuration/hbase-site.xml | 184 ++---
 .../services/HDFS/configuration/hdfs-site.xml   | 321 ++++-----
 .../services/HIVE/configuration/hive-site.xml   |  16 +-
 .../services/OOZIE/configuration/oozie-site.xml | 484 ++++++-------
 .../controllers/main/service/info/configs.js    |   3 +-
 .../app/controllers/wizard/step8_controller.js  |   5 -
 ambari-web/app/data/HDP2/config_mapping.js      | 276 +-------
 ambari-web/app/data/HDP2/global_properties.js   | 438 ------------
 ambari-web/app/data/HDP2/site_properties.js     | 444 ++++++++++++
 ambari-web/app/data/config_mapping.js           | 464 +------------
 ambari-web/app/data/global_properties.js        | 686 -------------------
 ambari-web/app/data/site_properties.js          | 624 +++++++++++++++++
 ambari-web/app/models/service_config.js         |  36 +-
 ambari-web/app/utils/config.js                  |   2 +-
 .../app/views/common/configs/services_config.js |   6 +-
 ambari-web/app/views/wizard/controls_view.js    |   6 +-
 32 files changed, 2368 insertions(+), 3070 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-agent/src/main/puppet/modules/hdp-hive/manifests/params.pp
----------------------------------------------------------------------
diff --git a/ambari-agent/src/main/puppet/modules/hdp-hive/manifests/params.pp b/ambari-agent/src/main/puppet/modules/hdp-hive/manifests/params.pp
index 3ec57ff..4fb7c41 100644
--- a/ambari-agent/src/main/puppet/modules/hdp-hive/manifests/params.pp
+++ b/ambari-agent/src/main/puppet/modules/hdp-hive/manifests/params.pp
@@ -22,8 +22,8 @@ class hdp-hive::params() inherits hdp::params
 {
 
   #TODO: will move to globals
-  $hive_metastore_user_name = hdp_default("hive_metastore_user_name","dbusername")
-  $hive_metastore_user_passwd = hdp_default("hive_metastore_user_passwd","dbpassword")
+  $hive_metastore_user_name = hdp_default("hive-site/javax.jdo.option.ConnectionUserName","dbusername")
+  $hive_metastore_user_passwd = hdp_default("hive-site/javax.jdo.option.ConnectionPassword","dbpassword")
   $hive_server_conf_dir = hdp_default("hive_server_conf_dir", "/etc/hive/conf.server")
 
   ### users

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HBASE/configuration/hbase-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HBASE/configuration/hbase-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HBASE/configuration/hbase-site.xml
index be9c023..e002d6b 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HBASE/configuration/hbase-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HBASE/configuration/hbase-site.xml
@@ -44,7 +44,7 @@
   </property>
   <property>
     <name>hbase.tmp.dir</name>
-    <value></value>
+    <value>/hadoop/hbase</value>
     <description>Temporary directory on the local filesystem.
     Change this setting to point to a location more permanent
     than '/tmp' (The '/tmp' directory is often cleared on
@@ -69,14 +69,14 @@
   </property>
   <property>
     <name>hbase.regionserver.global.memstore.upperLimit</name>
-    <value></value>
+    <value>0.4</value>
     <description>Maximum size of all memstores in a region server before new
       updates are blocked and flushes are forced. Defaults to 40% of heap
     </description>
   </property>
   <property>
     <name>hbase.regionserver.handler.count</name>
-    <value></value>
+    <value>60</value>
     <description>Count of RPC Listener instances spun up on RegionServers.
     Same property is used by the Master for count of master handlers.
     Default is 10.
@@ -84,8 +84,8 @@
   </property>
   <property>
     <name>hbase.hregion.majorcompaction</name>
-    <value></value>
-    <description>The time (in miliseconds) between 'major' compactions of all
+    <value>86400000</value>
+    <description>The time (in milliseconds) between 'major' compactions of all
     HStoreFiles in a region.  Default: 1 day.
     Set to 0 to disable automated major compactions.
     </description>
@@ -93,7 +93,7 @@
   
   <property>
     <name>hbase.regionserver.global.memstore.lowerLimit</name>
-    <value></value>
+    <value>0.38</value>
     <description>When memstores are being forced to flush to make room in
       memory, keep flushing until we hit this mark. Defaults to 35% of heap.
       This value equal to hbase.regionserver.global.memstore.upperLimit causes
@@ -103,7 +103,7 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.block.multiplier</name>
-    <value></value>
+    <value>2</value>
     <description>Block updates if memstore has hbase.hregion.memstore.block.multiplier
     time hbase.hregion.flush.size bytes.  Useful preventing
     runaway memstore during spikes in update traffic.  Without an
@@ -114,7 +114,7 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.flush.size</name>
-    <value></value>
+    <value>134217728</value>
     <description>
     Memstore will be flushed to disk if size of the memstore
     exceeds this number of bytes.  Value is checked by a thread that runs
@@ -123,7 +123,7 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.mslab.enabled</name>
-    <value></value>
+    <value>true</value>
     <description>
       Enables the MemStore-Local Allocation Buffer,
       a feature which works to prevent heap fragmentation under
@@ -133,7 +133,7 @@
   </property>
   <property>
     <name>hbase.hregion.max.filesize</name>
-    <value></value>
+    <value>10737418240</value>
     <description>
     Maximum HStoreFile size. If any one of a column families' HStoreFiles has
     grown to exceed this value, the hosting HRegion is split in two.
@@ -142,7 +142,7 @@
   </property>
   <property>
     <name>hbase.client.scanner.caching</name>
-    <value></value>
+    <value>100</value>
     <description>Number of rows that will be fetched when calling next
     on a scanner if it is not served from (local, client) memory. Higher
     caching values will enable faster scanners but will eat up more memory
@@ -153,7 +153,7 @@
   </property>
   <property>
     <name>zookeeper.session.timeout</name>
-    <value>30000</value>
+    <value>60000</value>
     <description>ZooKeeper session timeout.
       HBase passes this to the zk quorum as suggested maximum time for a
       session (This setting becomes zookeeper's 'maxSessionTimeout').  See
@@ -164,7 +164,7 @@
   </property>
   <property>
     <name>hbase.client.keyvalue.maxsize</name>
-    <value></value>
+    <value>10485760</value>
     <description>Specifies the combined maximum allowed size of a KeyValue
     instance. This is to set an upper boundary for a single entry saved in a
     storage file. Since they cannot be split it helps avoiding that a region
@@ -175,7 +175,7 @@
   </property>
   <property>
     <name>hbase.hstore.compactionThreshold</name>
-    <value></value>
+    <value>3</value>
     <description>
     If more than this number of HStoreFiles in any one HStore
     (one HStoreFile is written per flush of memstore) then a compaction
@@ -185,7 +185,7 @@
   </property>
   <property>
     <name>hbase.hstore.blockingStoreFiles</name>
-    <value></value>
+    <value>10</value>
     <description>
     If more than this number of StoreFiles in any one Store
     (one StoreFile is written per flush of MemStore) then updates are
@@ -195,7 +195,7 @@
   </property>
   <property>
     <name>hfile.block.cache.size</name>
-    <value></value>
+    <value>0.40</value>
     <description>
         Percentage of maximum heap (-Xmx setting) to allocate to block cache
         used by HFile/StoreFile. Default of 0.25 means allocate 25%.
@@ -319,7 +319,7 @@
 
   <property>
     <name>dfs.support.append</name>
-    <value></value>
+    <value>true</value>
     <description>Does HDFS allow appends to files?
     This is an hdfs config. set in here so the hdfs client will do append support.
     You must ensure that this config. is true serverside too when running hbase
@@ -329,7 +329,7 @@
 
   <property>
     <name>dfs.client.read.shortcircuit</name>
-    <value></value>
+    <value>true</value>
     <description>Enable/Disable short circuit read for your client.
     Hadoop servers should be configured to allow short circuit read
     for the hbase user for this to take effect

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml
index fe0bcba..8cf73f5 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml
@@ -74,7 +74,7 @@
 
   <property>
     <name>fs.checkpoint.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/namesecondary</value>
     <description>Determines where on the local filesystem the DFS secondary
         name node should store the temporary images to merge.
         If this is a comma-delimited list of directories then the image is
@@ -102,7 +102,7 @@
 
   <property>
     <name>fs.checkpoint.size</name>
-    <value>536870912</value>
+    <value>67108864</value>
     <description>The size of the current edit log (in bytes) that triggers
        a periodic checkpoint even if the fs.checkpoint.period hasn't expired.
   </description>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
index bd978a9..989b92d 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
@@ -27,7 +27,7 @@
   <property>
     <name>dfs.name.dir</name>
     <!-- cluster variant -->
-    <value></value>
+    <value>/hadoop/hdfs/namenode</value>
     <description>Determines where on the local filesystem the DFS name node
       should store the name table.  If this is a comma-delimited list
       of directories then the name table is replicated in all of the
@@ -37,14 +37,14 @@
 
   <property>
     <name>dfs.support.append</name>
-    <value></value>
+    <value>true</value>
     <description>to enable dfs append</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.webhdfs.enabled</name>
-    <value></value>
+    <value>true</value>
     <description>to enable webhdfs</description>
     <final>true</final>
   </property>
@@ -57,14 +57,14 @@
 
   <property>
     <name>dfs.datanode.failed.volumes.tolerated</name>
-    <value></value>
+    <value>0</value>
     <description>#of failed disks dn would tolerate</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.block.local-path-access.user</name>
-    <value></value>
+    <value>hbase</value>
     <description>the user who is allowed to perform short
     circuit reads.
     </description>
@@ -73,7 +73,7 @@
 
   <property>
     <name>dfs.data.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/data</value>
     <description>Determines where on the local filesystem an DFS data node
   should store its blocks.  If this is a comma-delimited
   list of directories, then data will be stored in all named
@@ -110,7 +110,7 @@
 
   <property>
     <name>dfs.replication</name>
-    <value></value>
+    <value>3</value>
     <description>Default block replication.
   </description>
   </property>
@@ -177,7 +177,7 @@ literal string "local" or a host:port for NDFS.</description>
 <property>
 <name>dfs.datanode.du.reserved</name>
 <!-- cluster variant -->
-<value></value>
+<value>1073741824</value>
 <description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
 </description>
 </property>
@@ -386,7 +386,7 @@ Kerberos principal name for the NameNode
 
   <property>
     <name>dfs.datanode.data.dir.perm</name>
-    <value></value>
+    <value>750</value>
 <description>The permissions that should be there on dfs.data.dir
 directories. The datanode will not come up if the permissions are
 different on existing dfs.data.dir directories. If the directories

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
index e5c834e..3baeee3 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
@@ -20,7 +20,7 @@ limitations under the License.
 <configuration>
   <property>
     <name>javax.jdo.option.ConnectionURL</name>
-    <value></value>
+    <value>jdbc</value>
     <description>JDBC connect string for a JDBC metastore</description>
   </property>
 
@@ -32,13 +32,13 @@ limitations under the License.
 
   <property>
     <name>javax.jdo.option.ConnectionUserName</name>
-    <value></value>
+    <value>hive</value>
     <description>username to use against metastore database</description>
   </property>
 
   <property>
     <name>javax.jdo.option.ConnectionPassword</name>
-    <value></value>
+    <value> </value>
     <description>password to use against metastore database</description>
   </property>
 

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
index df6ca71..fffcdb2 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
@@ -26,8 +26,10 @@
 
   <property>
     <name>io.sort.mb</name>
-    <value></value>
-    <description>No description</description>
+    <value>200</value>
+    <description>
+      The total amount of Map-side buffer memory to use while sorting files
+    </description>
   </property>
 
   <property>
@@ -38,8 +40,8 @@
 
   <property>
     <name>io.sort.spill.percent</name>
-    <value></value>
-    <description>No description</description>
+    <value>0.9</value>
+    <description>Percentage of sort buffer used for record collection</description>
   </property>
 
   <property>
@@ -73,7 +75,7 @@
   <property>
     <name>mapred.system.dir</name>
     <value>/mapred/system</value>
-    <description>No description</description>
+    <description>Path on the HDFS where where the MapReduce framework stores system files</description>
     <final>true</final>
   </property>
 
@@ -96,7 +98,7 @@
   <property>
     <!-- cluster specific -->
     <name>mapred.local.dir</name>
-    <value></value>
+    <value>/hadoop/mapred</value>
     <description>No description</description>
     <final>true</final>
   </property>
@@ -114,13 +116,13 @@
 
   <property>
     <name>mapred.tasktracker.map.tasks.maximum</name>
-    <value></value>
+    <value>4</value>
     <description>No description</description>
   </property>
 
   <property>
     <name>mapred.tasktracker.reduce.tasks.maximum</name>
-    <value></value>
+    <value>2</value>
     <description>No description</description>
   </property>
 
@@ -201,7 +203,7 @@
 
   <property>
     <name>mapred.jobtracker.taskScheduler</name>
-    <value></value>
+    <value>org.apache.hadoop.mapred.CapacityTaskScheduler</value>
   </property>
 
   <property>
@@ -277,31 +279,49 @@
   <property>
     <name>mapred.cluster.map.memory.mb</name>
     <value>1536</value>
+    <description>
+      The virtual memory size of a single Map slot in the MapReduce framework
+    </description>
   </property>
 
   <property>
     <name>mapred.cluster.reduce.memory.mb</name>
     <value>2048</value>
+    <description>
+    The virtual memory size of a single Reduce slot in the MapReduce framework
+    </description>
   </property>
 
   <property>
     <name>mapred.job.map.memory.mb</name>
     <value>1536</value>
+    <description>
+      Virtual memory for single Map task
+    </description>
   </property>
 
   <property>
     <name>mapred.job.reduce.memory.mb</name>
     <value>2048</value>
+    <description>
+      Virtual memory for single Reduce task
+    </description>
   </property>
 
   <property>
     <name>mapred.cluster.max.map.memory.mb</name>
     <value>6144</value>
+    <description>
+      Upper limit on virtual memory size for a single Map task of any MapReduce job
+    </description>
   </property>
 
   <property>
     <name>mapred.cluster.max.reduce.memory.mb</name>
     <value>4096</value>
+    <description>
+      Upper limit on virtual memory size for a single Reduce task of any MapReduce job
+    </description>
   </property>
 
 <property>
@@ -324,7 +344,10 @@
 
 <property>
   <name>mapred.healthChecker.script.path</name>
-  <value></value>
+  <value>file:////mapred/jobstatus</value>
+  <description>
+    Directory path to view job status
+  </description>
 </property>
 
 <property>
@@ -389,7 +412,7 @@
 
 <property>
   <name>mapred.jobtracker.maxtasks.per.job</name>
-  <value></value>
+  <value>-1</value>
   <final>true</final>
   <description>The maximum number of tasks for a single job.
   A value of -1 indicates that there is no maximum.  </description>
@@ -402,7 +425,10 @@
 
 <property>
   <name>mapred.userlog.retain.hours</name>
-  <value></value>
+  <value>24</value>
+  <description>
+    The maximum time, in hours, for which the user-logs are to be retained after the job completion.
+  </description>
 </property>
 
 <property>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-site.xml
index eeed0d8..57239c3 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/OOZIE/configuration/oozie-site.xml
@@ -210,7 +210,7 @@ org.apache.oozie.action.hadoop.DistcpActionExecutor</value>
 
     <property>
         <name>oozie.service.JPAService.jdbc.username</name>
-        <value>sa</value>
+        <value>oozie</value>
         <description>
             DB user name.
         </description>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/configuration/hbase-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/configuration/hbase-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/configuration/hbase-site.xml
index 4270410..375c7af 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/configuration/hbase-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HBASE/configuration/hbase-site.xml
@@ -44,7 +44,7 @@
   </property>
   <property>
     <name>hbase.tmp.dir</name>
-    <value></value>
+    <value>/hadoop/hbase</value>
     <description>Temporary directory on the local filesystem.
     Change this setting to point to a location more permanent
     than '/tmp' (The '/tmp' directory is often cleared on
@@ -69,14 +69,14 @@
   </property>
   <property>
     <name>hbase.regionserver.global.memstore.upperLimit</name>
-    <value></value>
+    <value>0.4</value>
     <description>Maximum size of all memstores in a region server before new
       updates are blocked and flushes are forced. Defaults to 40% of heap
     </description>
   </property>
   <property>
     <name>hbase.regionserver.handler.count</name>
-    <value></value>
+    <value>60</value>
     <description>Count of RPC Listener instances spun up on RegionServers.
     Same property is used by the Master for count of master handlers.
     Default is 10.
@@ -84,8 +84,8 @@
   </property>
   <property>
     <name>hbase.hregion.majorcompaction</name>
-    <value></value>
-    <description>The time (in miliseconds) between 'major' compactions of all
+    <value>86400000</value>
+    <description>The time (in milliseconds) between 'major' compactions of all
     HStoreFiles in a region.  Default: 1 day.
     Set to 0 to disable automated major compactions.
     </description>
@@ -93,7 +93,7 @@
   
   <property>
     <name>hbase.regionserver.global.memstore.lowerLimit</name>
-    <value></value>
+    <value>0.38</value>
     <description>When memstores are being forced to flush to make room in
       memory, keep flushing until we hit this mark. Defaults to 35% of heap.
       This value equal to hbase.regionserver.global.memstore.upperLimit causes
@@ -103,7 +103,7 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.block.multiplier</name>
-    <value></value>
+    <value>2</value>
     <description>Block updates if memstore has hbase.hregion.memstore.block.multiplier
     time hbase.hregion.flush.size bytes.  Useful preventing
     runaway memstore during spikes in update traffic.  Without an
@@ -114,7 +114,7 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.flush.size</name>
-    <value></value>
+    <value>134217728</value>
     <description>
     Memstore will be flushed to disk if size of the memstore
     exceeds this number of bytes.  Value is checked by a thread that runs
@@ -123,7 +123,7 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.mslab.enabled</name>
-    <value></value>
+    <value>true</value>
     <description>
       Enables the MemStore-Local Allocation Buffer,
       a feature which works to prevent heap fragmentation under
@@ -133,7 +133,7 @@
   </property>
   <property>
     <name>hbase.hregion.max.filesize</name>
-    <value></value>
+    <value>10737418240</value>
     <description>
     Maximum HStoreFile size. If any one of a column families' HStoreFiles has
     grown to exceed this value, the hosting HRegion is split in two.
@@ -142,7 +142,7 @@
   </property>
   <property>
     <name>hbase.client.scanner.caching</name>
-    <value></value>
+    <value>100</value>
     <description>Number of rows that will be fetched when calling next
     on a scanner if it is not served from (local, client) memory. Higher
     caching values will enable faster scanners but will eat up more memory
@@ -164,7 +164,7 @@
   </property>
   <property>
     <name>hbase.client.keyvalue.maxsize</name>
-    <value></value>
+    <value>10485760</value>
     <description>Specifies the combined maximum allowed size of a KeyValue
     instance. This is to set an upper boundary for a single entry saved in a
     storage file. Since they cannot be split it helps avoiding that a region
@@ -175,7 +175,7 @@
   </property>
   <property>
     <name>hbase.hstore.compactionThreshold</name>
-    <value></value>
+    <value>3</value>
     <description>
     If more than this number of HStoreFiles in any one HStore
     (one HStoreFile is written per flush of memstore) then a compaction
@@ -193,7 +193,7 @@
   
   <property>
     <name>hbase.hstore.blockingStoreFiles</name>
-    <value></value>
+    <value>10</value>
     <description>
     If more than this number of StoreFiles in any one Store
     (one StoreFile is written per flush of MemStore) then updates are
@@ -203,7 +203,7 @@
   </property>
   <property>
     <name>hfile.block.cache.size</name>
-    <value></value>
+    <value>0.40</value>
     <description>
         Percentage of maximum heap (-Xmx setting) to allocate to block cache
         used by HFile/StoreFile. Default of 0.25 means allocate 25%.

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
index 79e0151..df45e1b 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
@@ -27,7 +27,7 @@
   <property>
     <name>dfs.namenode.name.dir</name>
     <!-- cluster variant -->
-    <value></value>
+    <value>/hadoop/hdfs/namenode</value>
     <description>Determines where on the local filesystem the DFS name node
       should store the name table.  If this is a comma-delimited list
       of directories then the name table is replicated in all of the
@@ -45,20 +45,20 @@
   <property>
     <name>dfs.webhdfs.enabled</name>
     <value>true</value>
-    <description>to enable webhdfs</description>
+    <description>Whether to enable WebHDFS feature</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.datanode.failed.volumes.tolerated</name>
     <value>0</value>
-    <description>#of failed disks dn would tolerate</description>
+    <description> Number of failed disks a DataNode would tolerate before it stops offering service</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.datanode.data.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/data</value>
     <description>Determines where on the local filesystem an DFS data node
   should store its blocks.  If this is a comma-delimited
   list of directories, then data will be stored in all named
@@ -90,7 +90,7 @@
 
   <property>
     <name>dfs.namenode.checkpoint.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/namesecondary</value>
     <description>Determines where on the local filesystem the DFS secondary
       name node should store the temporary images to merge.
       If this is a comma-delimited list of directories then the image is
@@ -109,14 +109,23 @@
     </description>
   </property>
 
+
   <property>
     <name>dfs.namenode.checkpoint.period</name>
-    <value></value>
+    <value>21600</value>
     <description>The number of seconds between two periodic checkpoints.
     </description>
   </property>
 
   <property>
+    <name>fs.checkpoint.size</name>
+    <value>67108864</value>
+    <description>The size of the current edit log (in bytes) that triggers
+      a periodic checkpoint even if the maximum checkpoint delay is not reached
+    </description>
+  </property>
+
+  <property>
     <name>dfs.replication.max</name>
     <value>50</value>
     <description>Maximal block replication.

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
index 60ecfe7..053eccd 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
@@ -20,7 +20,7 @@ limitations under the License.
 <configuration>
   <property>
     <name>javax.jdo.option.ConnectionURL</name>
-    <value></value>
+    <value>jdbc</value>
     <description>JDBC connect string for a JDBC metastore</description>
   </property>
 
@@ -32,13 +32,13 @@ limitations under the License.
 
   <property>
     <name>javax.jdo.option.ConnectionUserName</name>
-    <value></value>
+    <value>hive</value>
     <description>username to use against metastore database</description>
   </property>
 
   <property>
     <name>javax.jdo.option.ConnectionPassword</name>
-    <value></value>
+    <value> </value>
     <description>password to use against metastore database</description>
   </property>
 

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-site.xml
index afd1e7f..bf4533f 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/OOZIE/configuration/oozie-site.xml
@@ -218,9 +218,9 @@
 
     <property>
         <name>oozie.service.JPAService.jdbc.username</name>
-        <value>sa</value>
+        <value>oozie</value>
         <description>
-            DB user name.
+          Database user name to use to connect to the database
         </description>
     </property>
 

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HBASE/configuration/hbase-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HBASE/configuration/hbase-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HBASE/configuration/hbase-site.xml
index be9c023..9d44de8 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HBASE/configuration/hbase-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HBASE/configuration/hbase-site.xml
@@ -24,13 +24,13 @@
     <name>hbase.rootdir</name>
     <value></value>
     <description>The directory shared by region servers and into
-    which HBase persists.  The URL should be 'fully-qualified'
-    to include the filesystem scheme.  For example, to specify the
-    HDFS directory '/hbase' where the HDFS instance's namenode is
-    running at namenode.example.org on port 9000, set this value to:
-    hdfs://namenode.example.org:9000/hbase.  By default HBase writes
-    into /tmp.  Change this configuration else all data will be lost
-    on machine restart.
+      which HBase persists.  The URL should be 'fully-qualified'
+      to include the filesystem scheme.  For example, to specify the
+      HDFS directory '/hbase' where the HDFS instance's namenode is
+      running at namenode.example.org on port 9000, set this value to:
+      hdfs://namenode.example.org:9000/hbase.  By default HBase writes
+      into /tmp.  Change this configuration else all data will be lost
+      on machine restart.
     </description>
   </property>
   <property>
@@ -44,11 +44,11 @@
   </property>
   <property>
     <name>hbase.tmp.dir</name>
-    <value></value>
+    <value>/hadoop/hbase</value>
     <description>Temporary directory on the local filesystem.
-    Change this setting to point to a location more permanent
-    than '/tmp' (The '/tmp' directory is often cleared on
-    machine restart).
+      Change this setting to point to a location more permanent
+      than '/tmp' (The '/tmp' directory is often cleared on
+      machine restart).
     </description>
   </property>
   <property>
@@ -69,31 +69,31 @@
   </property>
   <property>
     <name>hbase.regionserver.global.memstore.upperLimit</name>
-    <value></value>
+    <value>0.4</value>
     <description>Maximum size of all memstores in a region server before new
       updates are blocked and flushes are forced. Defaults to 40% of heap
     </description>
   </property>
   <property>
     <name>hbase.regionserver.handler.count</name>
-    <value></value>
+    <value>60</value>
     <description>Count of RPC Listener instances spun up on RegionServers.
-    Same property is used by the Master for count of master handlers.
-    Default is 10.
+      Same property is used by the Master for count of master handlers.
+      Default is 10.
     </description>
   </property>
   <property>
     <name>hbase.hregion.majorcompaction</name>
-    <value></value>
-    <description>The time (in miliseconds) between 'major' compactions of all
-    HStoreFiles in a region.  Default: 1 day.
-    Set to 0 to disable automated major compactions.
+    <value>86400000</value>
+    <description>The time (in milliseconds) between 'major' compactions of all
+      HStoreFiles in a region.  Default: 1 day.
+      Set to 0 to disable automated major compactions.
     </description>
   </property>
-  
+
   <property>
     <name>hbase.regionserver.global.memstore.lowerLimit</name>
-    <value></value>
+    <value>0.38</value>
     <description>When memstores are being forced to flush to make room in
       memory, keep flushing until we hit this mark. Defaults to 35% of heap.
       This value equal to hbase.regionserver.global.memstore.upperLimit causes
@@ -103,27 +103,27 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.block.multiplier</name>
-    <value></value>
+    <value>2</value>
     <description>Block updates if memstore has hbase.hregion.memstore.block.multiplier
-    time hbase.hregion.flush.size bytes.  Useful preventing
-    runaway memstore during spikes in update traffic.  Without an
-    upper-bound, memstore fills such that when it flushes the
-    resultant flush files take a long time to compact or split, or
-    worse, we OOME
+      time hbase.hregion.flush.size bytes.  Useful preventing
+      runaway memstore during spikes in update traffic.  Without an
+      upper-bound, memstore fills such that when it flushes the
+      resultant flush files take a long time to compact or split, or
+      worse, we OOME
     </description>
   </property>
   <property>
     <name>hbase.hregion.memstore.flush.size</name>
-    <value></value>
+    <value>134217728</value>
     <description>
-    Memstore will be flushed to disk if size of the memstore
-    exceeds this number of bytes.  Value is checked by a thread that runs
-    every hbase.server.thread.wakefrequency.
+      Memstore will be flushed to disk if size of the memstore
+      exceeds this number of bytes.  Value is checked by a thread that runs
+      every hbase.server.thread.wakefrequency.
     </description>
   </property>
   <property>
     <name>hbase.hregion.memstore.mslab.enabled</name>
-    <value></value>
+    <value>true</value>
     <description>
       Enables the MemStore-Local Allocation Buffer,
       a feature which works to prevent heap fragmentation under
@@ -133,27 +133,27 @@
   </property>
   <property>
     <name>hbase.hregion.max.filesize</name>
-    <value></value>
+    <value>10737418240</value>
     <description>
-    Maximum HStoreFile size. If any one of a column families' HStoreFiles has
-    grown to exceed this value, the hosting HRegion is split in two.
-    Default: 1G.
+      Maximum HStoreFile size. If any one of a column families' HStoreFiles has
+      grown to exceed this value, the hosting HRegion is split in two.
+      Default: 1G.
     </description>
   </property>
   <property>
     <name>hbase.client.scanner.caching</name>
-    <value></value>
+    <value>100</value>
     <description>Number of rows that will be fetched when calling next
-    on a scanner if it is not served from (local, client) memory. Higher
-    caching values will enable faster scanners but will eat up more memory
-    and some calls of next may take longer and longer times when the cache is empty.
-    Do not set this value such that the time between invocations is greater
-    than the scanner timeout; i.e. hbase.regionserver.lease.period
+      on a scanner if it is not served from (local, client) memory. Higher
+      caching values will enable faster scanners but will eat up more memory
+      and some calls of next may take longer and longer times when the cache is empty.
+      Do not set this value such that the time between invocations is greater
+      than the scanner timeout; i.e. hbase.regionserver.lease.period
     </description>
   </property>
   <property>
     <name>zookeeper.session.timeout</name>
-    <value>30000</value>
+    <value>60000</value>
     <description>ZooKeeper session timeout.
       HBase passes this to the zk quorum as suggested maximum time for a
       session (This setting becomes zookeeper's 'maxSessionTimeout').  See
@@ -164,42 +164,42 @@
   </property>
   <property>
     <name>hbase.client.keyvalue.maxsize</name>
-    <value></value>
+    <value>10485760</value>
     <description>Specifies the combined maximum allowed size of a KeyValue
-    instance. This is to set an upper boundary for a single entry saved in a
-    storage file. Since they cannot be split it helps avoiding that a region
-    cannot be split any further because the data is too large. It seems wise
-    to set this to a fraction of the maximum region size. Setting it to zero
-    or less disables the check.
+      instance. This is to set an upper boundary for a single entry saved in a
+      storage file. Since they cannot be split it helps avoiding that a region
+      cannot be split any further because the data is too large. It seems wise
+      to set this to a fraction of the maximum region size. Setting it to zero
+      or less disables the check.
     </description>
   </property>
   <property>
     <name>hbase.hstore.compactionThreshold</name>
-    <value></value>
+    <value>3</value>
     <description>
-    If more than this number of HStoreFiles in any one HStore
-    (one HStoreFile is written per flush of memstore) then a compaction
-    is run to rewrite all HStoreFiles files as one.  Larger numbers
-    put off compaction but when it runs, it takes longer to complete.
+      If more than this number of HStoreFiles in any one HStore
+      (one HStoreFile is written per flush of memstore) then a compaction
+      is run to rewrite all HStoreFiles files as one.  Larger numbers
+      put off compaction but when it runs, it takes longer to complete.
     </description>
   </property>
   <property>
     <name>hbase.hstore.blockingStoreFiles</name>
-    <value></value>
+    <value>10</value>
     <description>
-    If more than this number of StoreFiles in any one Store
-    (one StoreFile is written per flush of MemStore) then updates are
-    blocked for this HRegion until a compaction is completed, or
-    until hbase.hstore.blockingWaitTime has been exceeded.
+      If more than this number of StoreFiles in any one Store
+      (one StoreFile is written per flush of MemStore) then updates are
+      blocked for this HRegion until a compaction is completed, or
+      until hbase.hstore.blockingWaitTime has been exceeded.
     </description>
   </property>
   <property>
     <name>hfile.block.cache.size</name>
-    <value></value>
+    <value>0.40</value>
     <description>
-        Percentage of maximum heap (-Xmx setting) to allocate to block cache
-        used by HFile/StoreFile. Default of 0.25 means allocate 25%.
-        Set to 0 to disable but it's not recommended.
+      Percentage of maximum heap (-Xmx setting) to allocate to block cache
+      used by HFile/StoreFile. Default of 0.25 means allocate 25%.
+      Set to 0 to disable but it's not recommended.
     </description>
   </property>
 
@@ -210,35 +210,35 @@
     <name>hbase.master.keytab.file</name>
     <value></value>
     <description>Full path to the kerberos keytab file to use for logging in
-    the configured HMaster server principal.
+      the configured HMaster server principal.
     </description>
   </property>
   <property>
     <name>hbase.master.kerberos.principal</name>
     <value></value>
     <description>Ex. "hbase/_HOST@EXAMPLE.COM".  The kerberos principal name
-    that should be used to run the HMaster process.  The principal name should
-    be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the hostname
-    portion, it will be replaced with the actual hostname of the running
-    instance.
+      that should be used to run the HMaster process.  The principal name should
+      be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the hostname
+      portion, it will be replaced with the actual hostname of the running
+      instance.
     </description>
   </property>
   <property>
     <name>hbase.regionserver.keytab.file</name>
     <value></value>
     <description>Full path to the kerberos keytab file to use for logging in
-    the configured HRegionServer server principal.
+      the configured HRegionServer server principal.
     </description>
   </property>
   <property>
     <name>hbase.regionserver.kerberos.principal</name>
     <value></value>
     <description>Ex. "hbase/_HOST@EXAMPLE.COM".  The kerberos principal name
-    that should be used to run the HRegionServer process.  The principal name
-    should be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the
-    hostname portion, it will be replaced with the actual hostname of the
-    running instance.  An entry for this principal must exist in the file
-    specified in hbase.regionserver.keytab.file
+      that should be used to run the HRegionServer process.  The principal name
+      should be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the
+      hostname portion, it will be replaced with the actual hostname of the
+      running instance.  An entry for this principal must exist in the file
+      specified in hbase.regionserver.keytab.file
     </description>
   </property>
 
@@ -247,8 +247,8 @@
     <name>hbase.superuser</name>
     <value>hbase</value>
     <description>List of users or groups (comma-separated), who are allowed
-    full privileges, regardless of stored ACLs, across the cluster.
-    Only used when HBase security is enabled.
+      full privileges, regardless of stored ACLs, across the cluster.
+      Only used when HBase security is enabled.
     </description>
   </property>
 
@@ -273,10 +273,10 @@
     <name>hbase.coprocessor.region.classes</name>
     <value></value>
     <description>A comma-separated list of Coprocessors that are loaded by
-    default on all tables. For any override coprocessor method, these classes
-    will be called in order. After implementing your own Coprocessor, just put
-    it in HBase's classpath and add the fully qualified class name here.
-    A coprocessor can also be loaded on demand by setting HTableDescriptor.
+      default on all tables. For any override coprocessor method, these classes
+      will be called in order. After implementing your own Coprocessor, just put
+      it in HBase's classpath and add the fully qualified class name here.
+      A coprocessor can also be loaded on demand by setting HTableDescriptor.
     </description>
   </property>
 
@@ -296,7 +296,7 @@
     <name>hbase.zookeeper.property.clientPort</name>
     <value>2181</value>
     <description>Property from ZooKeeper's config zoo.cfg.
-    The port at which the clients will connect.
+      The port at which the clients will connect.
     </description>
   </property>
 
@@ -308,31 +308,31 @@
     <name>hbase.zookeeper.quorum</name>
     <value></value>
     <description>Comma separated list of servers in the ZooKeeper Quorum.
-    For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
-    By default this is set to localhost for local and pseudo-distributed modes
-    of operation. For a fully-distributed setup, this should be set to a full
-    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
-    this is the list of servers which we will start/stop ZooKeeper on.
+      For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
+      By default this is set to localhost for local and pseudo-distributed modes
+      of operation. For a fully-distributed setup, this should be set to a full
+      list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
+      this is the list of servers which we will start/stop ZooKeeper on.
     </description>
   </property>
   <!-- End of properties used to generate ZooKeeper host:port quorum list. -->
 
   <property>
     <name>dfs.support.append</name>
-    <value></value>
+    <value>true</value>
     <description>Does HDFS allow appends to files?
-    This is an hdfs config. set in here so the hdfs client will do append support.
-    You must ensure that this config. is true serverside too when running hbase
-    (You will have to restart your cluster after setting it).
+      This is an hdfs config. set in here so the hdfs client will do append support.
+      You must ensure that this config. is true serverside too when running hbase
+      (You will have to restart your cluster after setting it).
     </description>
   </property>
 
   <property>
     <name>dfs.client.read.shortcircuit</name>
-    <value></value>
+    <value>true</value>
     <description>Enable/Disable short circuit read for your client.
-    Hadoop servers should be configured to allow short circuit read
-    for the hbase user for this to take effect
+      Hadoop servers should be configured to allow short circuit read
+      for the hbase user for this to take effect
     </description>
   </property>
 
@@ -341,16 +341,16 @@
     <value></value>
     <description>Enable/disbale skipping the checksum check</description>
   </property>
-  
+
   <property>
     <name>hbase.zookeeper.useMulti</name>
     <value>true</value>
     <description>Instructs HBase to make use of ZooKeeper's multi-update functionality.
-    This allows certain ZooKeeper operations to complete more quickly and prevents some issues
-    with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).·
-    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+
-    and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will
-    not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).
+      This allows certain ZooKeeper operations to complete more quickly and prevents some issues
+      with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).·
+      IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+
+      and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will
+      not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).
     </description>
   </property>
   <property>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/core-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/core-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/core-site.xml
index fe0bcba..afa7338 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/core-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/core-site.xml
@@ -1,36 +1,36 @@
 <?xml version="1.0"?>
 <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
 
- <!--
-    Licensed to the Apache Software Foundation (ASF) under one or more
-    contributor license agreements.  See the NOTICE file distributed with
-    this work for additional information regarding copyright ownership.
-    The ASF licenses this file to You under the Apache License, Version 2.0
-    (the "License"); you may not use this file except in compliance with
-    the License.  You may obtain a copy of the License at
- 
-        http://www.apache.org/licenses/LICENSE-2.0
- 
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
- -->
- 
+<!--
+   Licensed to the Apache Software Foundation (ASF) under one or more
+   contributor license agreements.  See the NOTICE file distributed with
+   this work for additional information regarding copyright ownership.
+   The ASF licenses this file to You under the Apache License, Version 2.0
+   (the "License"); you may not use this file except in compliance with
+   the License.  You may obtain a copy of the License at
+
+       http://www.apache.org/licenses/LICENSE-2.0
+
+   Unless required by applicable law or agreed to in writing, software
+   distributed under the License is distributed on an "AS IS" BASIS,
+   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+   See the License for the specific language governing permissions and
+   limitations under the License.
+-->
+
 <!-- Put site-specific property overrides in this file. -->
 
 <configuration xmlns:xi="http://www.w3.org/2001/XInclude">
 
-<!-- i/o properties -->
+  <!-- i/o properties -->
 
   <property>
     <name>io.file.buffer.size</name>
     <value>131072</value>
     <description>The size of buffer for use in sequence files.
-  The size of this buffer should probably be a multiple of hardware
-  page size (4096 on Intel x86), and it determines how much data is
-  buffered during read and write operations.</description>
+      The size of this buffer should probably be a multiple of hardware
+      page size (4096 on Intel x86), and it determines how much data is
+      buffered during read and write operations.</description>
   </property>
 
   <property>
@@ -42,7 +42,7 @@
     <name>io.compression.codecs</name>
     <value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,com.hadoop.compression.lzo.LzoCodec,com.hadoop.compression.lzo.LzopCodec,org.apache.hadoop.io.compress.SnappyCodec</value>
     <description>A list of the compression codec classes that can be used
-                 for compression/decompression.</description>
+      for compression/decompression.</description>
   </property>
 
 
@@ -53,14 +53,14 @@
   </property>
 
 
-<!-- file system properties -->
+  <!-- file system properties -->
 
   <property>
     <name>fs.default.name</name>
     <!-- cluster variant -->
     <value></value>
     <description>The name of the default file system.  Either the
-  literal string "local" or a host:port for NDFS.</description>
+      literal string "local" or a host:port for NDFS.</description>
     <final>true</final>
   </property>
 
@@ -68,17 +68,17 @@
     <name>fs.trash.interval</name>
     <value>360</value>
     <description>Number of minutes between trash checkpoints.
-  If zero, the trash feature is disabled.
-  </description>
+      If zero, the trash feature is disabled.
+    </description>
   </property>
 
   <property>
     <name>fs.checkpoint.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/namesecondary</value>
     <description>Determines where on the local filesystem the DFS secondary
-        name node should store the temporary images to merge.
-        If this is a comma-delimited list of directories then the image is
-        replicated in all of the directories for redundancy.
+      name node should store the temporary images to merge.
+      If this is a comma-delimited list of directories then the image is
+      replicated in all of the directories for redundancy.
     </description>
   </property>
 
@@ -86,10 +86,10 @@
     <name>fs.checkpoint.edits.dir</name>
     <value>${fs.checkpoint.dir}</value>
     <description>Determines where on the local filesystem the DFS secondary
-        name node should store the temporary edits to merge.
-        If this is a comma-delimited list of directoires then teh edits is
-        replicated in all of the directoires for redundancy.
-        Default value is same as fs.checkpoint.dir
+      name node should store the temporary edits to merge.
+      If this is a comma-delimited list of directoires then teh edits is
+      replicated in all of the directoires for redundancy.
+      Default value is same as fs.checkpoint.dir
     </description>
   </property>
 
@@ -97,15 +97,15 @@
     <name>fs.checkpoint.period</name>
     <value>21600</value>
     <description>The number of seconds between two periodic checkpoints.
-  </description>
+    </description>
   </property>
 
   <property>
     <name>fs.checkpoint.size</name>
-    <value>536870912</value>
+    <value>67108864</value>
     <description>The size of the current edit log (in bytes) that triggers
-       a periodic checkpoint even if the fs.checkpoint.period hasn't expired.
-  </description>
+      a periodic checkpoint even if the fs.checkpoint.period hasn't expired.
+    </description>
   </property>
 
   <!-- ipc properties: copied from kryptonite configuration -->
@@ -113,16 +113,16 @@
     <name>ipc.client.idlethreshold</name>
     <value>8000</value>
     <description>Defines the threshold number of connections after which
-               connections will be inspected for idleness.
-  </description>
+      connections will be inspected for idleness.
+    </description>
   </property>
 
   <property>
     <name>ipc.client.connection.maxidletime</name>
     <value>30000</value>
     <description>The maximum time after which a client will bring down the
-               connection to the server.
-  </description>
+      connection to the server.
+    </description>
   </property>
 
   <property>
@@ -136,118 +136,118 @@
     <name>webinterface.private.actions</name>
     <value>false</value>
     <description> If set to true, the web interfaces of JT and NN may contain
-                actions, such as kill job, delete file, etc., that should
-                not be exposed to public. Enable this option if the interfaces
-                are only reachable by those who have the right authorization.
-  </description>
-  </property>
-
- <property>
-   <name>hadoop.security.authentication</name>
-   <value>simple</value>
-   <description>
-   Set the authentication for the cluster. Valid values are: simple or
-   kerberos.
-   </description>
- </property>
-<property>
-  <name>hadoop.security.authorization</name>
-  <value></value>
-  <description>
-     Enable authorization for different protocols.
-  </description>
-</property>
+      actions, such as kill job, delete file, etc., that should
+      not be exposed to public. Enable this option if the interfaces
+      are only reachable by those who have the right authorization.
+    </description>
+  </property>
+
+  <property>
+    <name>hadoop.security.authentication</name>
+    <value>simple</value>
+    <description>
+      Set the authentication for the cluster. Valid values are: simple or
+      kerberos.
+    </description>
+  </property>
+  <property>
+    <name>hadoop.security.authorization</name>
+    <value></value>
+    <description>
+      Enable authorization for different protocols.
+    </description>
+  </property>
 
   <property>
     <name>hadoop.security.auth_to_local</name>
     <value></value>
-<description>The mapping from kerberos principal names to local OS user names.
-  So the default rule is just "DEFAULT" which takes all principals in your default domain to their first component.
-  "omalley@APACHE.ORG" and "omalley/admin@APACHE.ORG" to "omalley", if your default domain is APACHE.ORG.
-The translations rules have 3 sections:
+    <description>The mapping from kerberos principal names to local OS user names.
+      So the default rule is just "DEFAULT" which takes all principals in your default domain to their first component.
+      "omalley@APACHE.ORG" and "omalley/admin@APACHE.ORG" to "omalley", if your default domain is APACHE.ORG.
+      The translations rules have 3 sections:
       base     filter    substitution
-The base consists of a number that represents the number of components in the principal name excluding the realm and the pattern for building the name from the sections of the principal name. The base uses $0 to mean the realm, $1 to mean the first component and $2 to mean the second component.
+      The base consists of a number that represents the number of components in the principal name excluding the realm and the pattern for building the name from the sections of the principal name. The base uses $0 to mean the realm, $1 to mean the first component and $2 to mean the second component.
 
-[1:$1@$0] translates "omalley@APACHE.ORG" to "omalley@APACHE.ORG"
-[2:$1] translates "omalley/admin@APACHE.ORG" to "omalley"
-[2:$1%$2] translates "omalley/admin@APACHE.ORG" to "omalley%admin"
+      [1:$1@$0] translates "omalley@APACHE.ORG" to "omalley@APACHE.ORG"
+      [2:$1] translates "omalley/admin@APACHE.ORG" to "omalley"
+      [2:$1%$2] translates "omalley/admin@APACHE.ORG" to "omalley%admin"
 
-The filter is a regex in parens that must the generated string for the rule to apply.
+      The filter is a regex in parens that must the generated string for the rule to apply.
 
-"(.*%admin)" will take any string that ends in "%admin"
-"(.*@ACME.COM)" will take any string that ends in "@ACME.COM"
+      "(.*%admin)" will take any string that ends in "%admin"
+      "(.*@ACME.COM)" will take any string that ends in "@ACME.COM"
 
-Finally, the substitution is a sed rule to translate a regex into a fixed string.
+      Finally, the substitution is a sed rule to translate a regex into a fixed string.
 
-"s/@ACME\.COM//" removes the first instance of "@ACME.COM".
-"s/@[A-Z]*\.COM//" removes the first instance of "@" followed by a name followed by ".COM".
-"s/X/Y/g" replaces all of the "X" in the name with "Y"
+      "s/@ACME\.COM//" removes the first instance of "@ACME.COM".
+      "s/@[A-Z]*\.COM//" removes the first instance of "@" followed by a name followed by ".COM".
+      "s/X/Y/g" replaces all of the "X" in the name with "Y"
 
-So, if your default realm was APACHE.ORG, but you also wanted to take all principals from ACME.COM that had a single component "joe@ACME.COM", you'd do:
+      So, if your default realm was APACHE.ORG, but you also wanted to take all principals from ACME.COM that had a single component "joe@ACME.COM", you'd do:
 
-RULE:[1:$1@$0](.@ACME.ORG)s/@.//
-DEFAULT
+      RULE:[1:$1@$0](.@ACME.ORG)s/@.//
+      DEFAULT
 
-To also translate the names with a second component, you'd make the rules:
+      To also translate the names with a second component, you'd make the rules:
 
-RULE:[1:$1@$0](.@ACME.ORG)s/@.//
-RULE:[2:$1@$0](.@ACME.ORG)s/@.//
-DEFAULT
+      RULE:[1:$1@$0](.@ACME.ORG)s/@.//
+      RULE:[2:$1@$0](.@ACME.ORG)s/@.//
+      DEFAULT
 
-If you want to treat all principals from APACHE.ORG with /admin as "admin", your rules would look like:
+      If you want to treat all principals from APACHE.ORG with /admin as "admin", your rules would look like:
 
-RULE[2:$1%$2@$0](.%admin@APACHE.ORG)s/./admin/
-DEFAULT
+      RULE[2:$1%$2@$0](.%admin@APACHE.ORG)s/./admin/
+      DEFAULT
     </description>
   </property>
 
-<!--
-<property>
-  <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").groups</name>
-  <value></value>
-  <description>
-     Proxy group for Hadoop.
-  </description>
-</property>
-
-<property>
-  <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").hosts</name>
-  <value></value>
-  <description>
-     Proxy host for Hadoop.
-  </description>
-</property>
-
-<property>
-  <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").groups</name>
-  <value></value>
-  <description>
-     Proxy group for Hadoop.
-  </description>
-</property>
-
-<property>
-  <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").hosts</name>
-  <value></value>
-  <description>
-     Proxy host for Hadoop.
-  </description>
-</property>
-
-<property>
-  <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").groups</name>
-  <value></value>
-  <description>
-    Proxy group for templeton.
-  </description>
-</property>
-
-<property>
-  <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").hosts</name>
-  <value></value>
-  <description>
-    Proxy host for templeton.
-  </description>
-</property>
--->
+  <!--
+  <property>
+    <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").groups</name>
+    <value></value>
+    <description>
+       Proxy group for Hadoop.
+    </description>
+  </property>
+
+  <property>
+    <name>hadoop.proxyuser.scope.function_hdp_user("hcat_user").hosts</name>
+    <value></value>
+    <description>
+       Proxy host for Hadoop.
+    </description>
+  </property>
+
+  <property>
+    <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").groups</name>
+    <value></value>
+    <description>
+       Proxy group for Hadoop.
+    </description>
+  </property>
+
+  <property>
+    <name>hadoop.proxyuser.scope.function_hdp_user("oozie_user").hosts</name>
+    <value></value>
+    <description>
+       Proxy host for Hadoop.
+    </description>
+  </property>
+
+  <property>
+    <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").groups</name>
+    <value></value>
+    <description>
+      Proxy group for templeton.
+    </description>
+  </property>
+
+  <property>
+    <name>hadoop.proxyuser.scope.function_hdp_user("templeton_user").hosts</name>
+    <value></value>
+    <description>
+      Proxy host for templeton.
+    </description>
+  </property>
+  -->
 </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
index bd978a9..5cb7e4a 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
@@ -22,12 +22,12 @@
 
 <configuration>
 
-<!-- file system properties -->
+  <!-- file system properties -->
 
   <property>
     <name>dfs.name.dir</name>
     <!-- cluster variant -->
-    <value></value>
+    <value>/hadoop/hdfs/namenode</value>
     <description>Determines where on the local filesystem the DFS name node
       should store the name table.  If this is a comma-delimited list
       of directories then the name table is replicated in all of the
@@ -37,19 +37,19 @@
 
   <property>
     <name>dfs.support.append</name>
-    <value></value>
+    <value>true</value>
     <description>to enable dfs append</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.webhdfs.enabled</name>
-    <value></value>
+    <value>true</value>
     <description>to enable webhdfs</description>
     <final>true</final>
   </property>
 
- <property>
+  <property>
     <name>dfs.datanode.socket.write.timeout</name>
     <value>0</value>
     <description>DFS Client write socket timeout</description>
@@ -57,29 +57,29 @@
 
   <property>
     <name>dfs.datanode.failed.volumes.tolerated</name>
-    <value></value>
+    <value>0</value>
     <description>#of failed disks dn would tolerate</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.block.local-path-access.user</name>
-    <value></value>
+    <value>hbase</value>
     <description>the user who is allowed to perform short
-    circuit reads.
+      circuit reads.
     </description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.data.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/data</value>
     <description>Determines where on the local filesystem an DFS data node
-  should store its blocks.  If this is a comma-delimited
-  list of directories, then data will be stored in all named
-  directories, typically on different devices.
-  Directories that do not exist are ignored.
-  </description>
+      should store its blocks.  If this is a comma-delimited
+      list of directories, then data will be stored in all named
+      directories, typically on different devices.
+      Directories that do not exist are ignored.
+    </description>
     <final>true</final>
   </property>
 
@@ -87,32 +87,32 @@
     <name>dfs.hosts.exclude</name>
     <value></value>
     <description>Names a file that contains a list of hosts that are
-    not permitted to connect to the namenode.  The full pathname of the
-    file must be specified.  If the value is empty, no hosts are
-    excluded.</description>
+      not permitted to connect to the namenode.  The full pathname of the
+      file must be specified.  If the value is empty, no hosts are
+      excluded.</description>
   </property>
 
   <property>
     <name>dfs.hosts</name>
     <value></value>
     <description>Names a file that contains a list of hosts that are
-    permitted to connect to the namenode. The full pathname of the file
-    must be specified.  If the value is empty, all hosts are
-    permitted.</description>
+      permitted to connect to the namenode. The full pathname of the file
+      must be specified.  If the value is empty, all hosts are
+      permitted.</description>
   </property>
 
   <property>
     <name>dfs.replication.max</name>
     <value>50</value>
     <description>Maximal block replication.
-  </description>
+    </description>
   </property>
 
   <property>
     <name>dfs.replication</name>
-    <value></value>
+    <value>3</value>
     <description>Default block replication.
-  </description>
+    </description>
   </property>
 
   <property>
@@ -125,21 +125,21 @@
     <name>dfs.safemode.threshold.pct</name>
     <value>1.0f</value>
     <description>
-        Specifies the percentage of blocks that should satisfy
-        the minimal replication requirement defined by dfs.replication.min.
-        Values less than or equal to 0 mean not to start in safe mode.
-        Values greater than 1 will make safe mode permanent.
-        </description>
+      Specifies the percentage of blocks that should satisfy
+      the minimal replication requirement defined by dfs.replication.min.
+      Values less than or equal to 0 mean not to start in safe mode.
+      Values greater than 1 will make safe mode permanent.
+    </description>
   </property>
 
   <property>
     <name>dfs.balance.bandwidthPerSec</name>
     <value>6250000</value>
     <description>
-        Specifies the maximum amount of bandwidth that each datanode
-        can utilize for the balancing purpose in term of
-        the number of bytes per second.
-  </description>
+      Specifies the maximum amount of bandwidth that each datanode
+      can utilize for the balancing purpose in term of
+      the number of bytes per second.
+    </description>
   </property>
 
   <property>
@@ -169,133 +169,133 @@
   <property>
     <name>dfs.http.address</name>
     <value></value>
-<description>The name of the default file system.  Either the
-literal string "local" or a host:port for NDFS.</description>
-<final>true</final>
-</property>
-
-<property>
-<name>dfs.datanode.du.reserved</name>
-<!-- cluster variant -->
-<value></value>
-<description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
-</description>
-</property>
-
-<property>
-<name>dfs.datanode.ipc.address</name>
-<value>0.0.0.0:8010</value>
-<description>
-The datanode ipc server address and port.
-If the port is 0 then the server will start on a free port.
-</description>
-</property>
-
-<property>
-<name>dfs.blockreport.initialDelay</name>
-<value>120</value>
-<description>Delay for first block report in seconds.</description>
-</property>
-
-<property>
-<name>dfs.datanode.du.pct</name>
-<value>0.85f</value>
-<description>When calculating remaining space, only use this percentage of the real available space
-</description>
-</property>
-
-<property>
-<name>dfs.namenode.handler.count</name>
-<value>40</value>
-<description>The number of server threads for the namenode.</description>
-</property>
-
-<property>
-<name>dfs.datanode.max.xcievers</name>
-<value>4096</value>
-<description>PRIVATE CONFIG VARIABLE</description>
-</property>
-
-<!-- Permissions configuration -->
-
-<property>
-<name>dfs.umaskmode</name>
-<value>077</value>
-<description>
-The octal umask used when creating files and directories.
-</description>
-</property>
-
-<property>
-<name>dfs.web.ugi</name>
-<!-- cluster variant -->
-<value>gopher,gopher</value>
-<description>The user account used by the web interface.
-Syntax: USERNAME,GROUP1,GROUP2, ...
-</description>
-</property>
-
-<property>
-<name>dfs.permissions</name>
-<value>true</value>
-<description>
-If "true", enable permission checking in HDFS.
-If "false", permission checking is turned off,
-but all other behavior is unchanged.
-Switching from one parameter value to the other does not change the mode,
-owner or group of files or directories.
-</description>
-</property>
-
-<property>
-<name>dfs.permissions.supergroup</name>
-<value>hdfs</value>
-<description>The name of the group of super-users.</description>
-</property>
-
-<property>
-<name>dfs.namenode.handler.count</name>
-<value>100</value>
-<description>Added to grow Queue size so that more client connections are allowed</description>
-</property>
-
-<property>
-<name>ipc.server.max.response.size</name>
-<value>5242880</value>
-</property>
-<property>
-<name>dfs.block.access.token.enable</name>
-<value>true</value>
-<description>
-If "true", access tokens are used as capabilities for accessing datanodes.
-If "false", no access tokens are checked on accessing datanodes.
-</description>
-</property>
-
-<property>
-<name>dfs.namenode.kerberos.principal</name>
-<value></value>
-<description>
-Kerberos principal name for the NameNode
-</description>
-</property>
-
-<property>
-<name>dfs.secondary.namenode.kerberos.principal</name>
-<value></value>
+    <description>The name of the default file system.  Either the
+      literal string "local" or a host:port for NDFS.</description>
+    <final>true</final>
+  </property>
+
+  <property>
+    <name>dfs.datanode.du.reserved</name>
+    <!-- cluster variant -->
+    <value>1073741824</value>
+    <description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.datanode.ipc.address</name>
+    <value>0.0.0.0:8010</value>
     <description>
-        Kerberos principal name for the secondary NameNode.
+      The datanode ipc server address and port.
+      If the port is 0 then the server will start on a free port.
     </description>
   </property>
 
+  <property>
+    <name>dfs.blockreport.initialDelay</name>
+    <value>120</value>
+    <description>Delay for first block report in seconds.</description>
+  </property>
+
+  <property>
+    <name>dfs.datanode.du.pct</name>
+    <value>0.85f</value>
+    <description>When calculating remaining space, only use this percentage of the real available space
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.namenode.handler.count</name>
+    <value>40</value>
+    <description>The number of server threads for the namenode.</description>
+  </property>
+
+  <property>
+    <name>dfs.datanode.max.xcievers</name>
+    <value>4096</value>
+    <description>PRIVATE CONFIG VARIABLE</description>
+  </property>
+
+  <!-- Permissions configuration -->
+
+  <property>
+    <name>dfs.umaskmode</name>
+    <value>077</value>
+    <description>
+      The octal umask used when creating files and directories.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.web.ugi</name>
+    <!-- cluster variant -->
+    <value>gopher,gopher</value>
+    <description>The user account used by the web interface.
+      Syntax: USERNAME,GROUP1,GROUP2, ...
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.permissions</name>
+    <value>true</value>
+    <description>
+      If "true", enable permission checking in HDFS.
+      If "false", permission checking is turned off,
+      but all other behavior is unchanged.
+      Switching from one parameter value to the other does not change the mode,
+      owner or group of files or directories.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.permissions.supergroup</name>
+    <value>hdfs</value>
+    <description>The name of the group of super-users.</description>
+  </property>
 
-<!--
-  This is KRB DOMAIN specific. The FQDN of the namenode has to be mentioned.
--->
+  <property>
+    <name>dfs.namenode.handler.count</name>
+    <value>100</value>
+    <description>Added to grow Queue size so that more client connections are allowed</description>
+  </property>
+
+  <property>
+    <name>ipc.server.max.response.size</name>
+    <value>5242880</value>
+  </property>
+  <property>
+    <name>dfs.block.access.token.enable</name>
+    <value>true</value>
+    <description>
+      If "true", access tokens are used as capabilities for accessing datanodes.
+      If "false", no access tokens are checked on accessing datanodes.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.namenode.kerberos.principal</name>
+    <value></value>
+    <description>
+      Kerberos principal name for the NameNode
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.secondary.namenode.kerberos.principal</name>
+    <value></value>
+    <description>
+      Kerberos principal name for the secondary NameNode.
+    </description>
+  </property>
+
+
+  <!--
+    This is KRB DOMAIN specific. The FQDN of the namenode has to be mentioned.
+  -->
   <property>
     <name>dfs.namenode.kerberos.https.principal</name>
     <value></value>
-     <description>The Kerberos principal for the host that the NameNode runs on.</description>
+    <description>The Kerberos principal for the host that the NameNode runs on.</description>
 
   </property>
 
@@ -341,84 +341,84 @@ Kerberos principal name for the NameNode
   <property>
     <name>dfs.datanode.kerberos.principal</name>
     <value></value>
- <description>
-        The Kerberos principal that the DataNode runs as. "_HOST" is replaced by the real host name.
+    <description>
+      The Kerberos principal that the DataNode runs as. "_HOST" is replaced by the real host name.
     </description>
   </property>
 
   <property>
     <name>dfs.namenode.keytab.file</name>
     <value></value>
- <description>
-        Combined keytab file containing the namenode service and host principals.
+    <description>
+      Combined keytab file containing the namenode service and host principals.
     </description>
   </property>
 
   <property>
     <name>dfs.secondary.namenode.keytab.file</name>
     <value></value>
-  <description>
-        Combined keytab file containing the namenode service and host principals.
+    <description>
+      Combined keytab file containing the namenode service and host principals.
     </description>
   </property>
 
   <property>
     <name>dfs.datanode.keytab.file</name>
     <value></value>
- <description>
-        The filename of the keytab file for the DataNode.
+    <description>
+      The filename of the keytab file for the DataNode.
     </description>
   </property>
 
   <property>
     <name>dfs.https.port</name>
     <value>50470</value>
- <description>The https port where namenode binds</description>
+    <description>The https port where namenode binds</description>
 
   </property>
 
   <property>
     <name>dfs.https.address</name>
     <value></value>
-  <description>The https address where namenode binds</description>
+    <description>The https address where namenode binds</description>
 
   </property>
 
   <property>
     <name>dfs.datanode.data.dir.perm</name>
-    <value></value>
-<description>The permissions that should be there on dfs.data.dir
-directories. The datanode will not come up if the permissions are
-different on existing dfs.data.dir directories. If the directories
-don't exist, they will be created with this permission.</description>
-  </property>
-
-  <property>
-  <name>dfs.access.time.precision</name>
-  <value>0</value>
-  <description>The access time for HDFS file is precise upto this value.
-               The default value is 1 hour. Setting a value of 0 disables
-               access times for HDFS.
-  </description>
-</property>
-
-<property>
- <name>dfs.cluster.administrators</name>
- <value> hdfs</value>
- <description>ACL for who all can view the default servlets in the HDFS</description>
-</property>
-
-<property>
-  <name>ipc.server.read.threadpool.size</name>
-  <value>5</value>
-  <description></description>
-</property>
-
-<property>
-  <name>dfs.datanode.failed.volumes.tolerated</name>
-  <value>0</value>
-  <description>Number of failed disks datanode would tolerate</description>
-</property>
+    <value>750</value>
+    <description>The permissions that should be there on dfs.data.dir
+      directories. The datanode will not come up if the permissions are
+      different on existing dfs.data.dir directories. If the directories
+      don't exist, they will be created with this permission.</description>
+  </property>
+
+  <property>
+    <name>dfs.access.time.precision</name>
+    <value>0</value>
+    <description>The access time for HDFS file is precise upto this value.
+      The default value is 1 hour. Setting a value of 0 disables
+      access times for HDFS.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.cluster.administrators</name>
+    <value> hdfs</value>
+    <description>ACL for who all can view the default servlets in the HDFS</description>
+  </property>
+
+  <property>
+    <name>ipc.server.read.threadpool.size</name>
+    <value>5</value>
+    <description></description>
+  </property>
+
+  <property>
+    <name>dfs.datanode.failed.volumes.tolerated</name>
+    <value>0</value>
+    <description>Number of failed disks datanode would tolerate</description>
+  </property>
 
   <property>
     <name>dfs.namenode.avoid.read.stale.datanode</name>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
index e5c834e..7ee89b9 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
@@ -20,7 +20,7 @@ limitations under the License.
 <configuration>
   <property>
     <name>javax.jdo.option.ConnectionURL</name>
-    <value></value>
+    <value>jdbc</value>
     <description>JDBC connect string for a JDBC metastore</description>
   </property>
 
@@ -32,13 +32,13 @@ limitations under the License.
 
   <property>
     <name>javax.jdo.option.ConnectionUserName</name>
-    <value></value>
+    <value>hive</value>
     <description>username to use against metastore database</description>
   </property>
 
   <property>
     <name>javax.jdo.option.ConnectionPassword</name>
-    <value></value>
+    <value> </value>
     <description>password to use against metastore database</description>
   </property>
 
@@ -52,21 +52,21 @@ limitations under the License.
     <name>hive.metastore.sasl.enabled</name>
     <value></value>
     <description>If true, the metastore thrift interface will be secured with SASL.
-     Clients must authenticate with Kerberos.</description>
+      Clients must authenticate with Kerberos.</description>
   </property>
 
   <property>
     <name>hive.metastore.kerberos.keytab.file</name>
     <value></value>
     <description>The path to the Kerberos Keytab file containing the metastore
-     thrift server's service principal.</description>
+      thrift server's service principal.</description>
   </property>
 
   <property>
     <name>hive.metastore.kerberos.principal</name>
     <value></value>
     <description>The service principal for the metastore thrift server. The special
-    string _HOST will be replaced automatically with the correct host name.</description>
+      string _HOST will be replaced automatically with the correct host name.</description>
   </property>
 
   <property>
@@ -109,7 +109,7 @@ limitations under the License.
     <name>hive.security.authorization.manager</name>
     <value>org.apache.hcatalog.security.HdfsAuthorizationProvider</value>
     <description>the hive client authorization manager class name.
-    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  </description>
+      The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  </description>
   </property>
 
   <property>


[3/4] AMBARI-3568: Clean up unnecessary globals from Ambari Web. (jaimin)

Posted by ja...@apache.org.
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/OOZIE/configuration/oozie-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/OOZIE/configuration/oozie-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/OOZIE/configuration/oozie-site.xml
index eeed0d8..1f83735 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/OOZIE/configuration/oozie-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/OOZIE/configuration/oozie-site.xml
@@ -15,223 +15,223 @@
   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
   See the License for the specific language governing permissions and
   limitations under the License.
--->     
+-->
 
 <configuration>
 
-<!--
-    Refer to the oozie-default.xml file for the complete list of
-    Oozie configuration properties and their default values.
--->
+  <!--
+      Refer to the oozie-default.xml file for the complete list of
+      Oozie configuration properties and their default values.
+  -->
   <property>
     <name>oozie.base.url</name>
     <value>http://localhost:11000/oozie</value>
     <description>Base Oozie URL.</description>
-   </property>
+  </property>
 
   <property>
     <name>oozie.system.id</name>
     <value>oozie-${user.name}</value>
     <description>
-    The Oozie system ID.
-    </description>
-   </property>
-
-   <property>
-     <name>oozie.systemmode</name>
-     <value>NORMAL</value>
-     <description>
-     System mode for  Oozie at startup.
-     </description>
-   </property>
-
-   <property>
-     <name>oozie.service.AuthorizationService.authorization.enabled</name>
-     <value>true</value>
-     <description>
-     Specifies whether security (user name/admin role) is enabled or not.
-     If disabled any user can manage Oozie system and manage any job.
-     </description>
-   </property>
-
-   <property>
-     <name>oozie.service.PurgeService.older.than</name>
-     <value>30</value>
-     <description>
-     Jobs older than this value, in days, will be purged by the PurgeService.
-     </description>
-   </property>
-
-   <property>
-     <name>oozie.service.PurgeService.purge.interval</name>
-     <value>3600</value>
-     <description>
-     Interval at which the purge service will run, in seconds.
-     </description>
-   </property>
-
-   <property>
-     <name>oozie.service.CallableQueueService.queue.size</name>
-     <value>1000</value>
-     <description>Max callable queue size</description>
-   </property>
-
-   <property>
-     <name>oozie.service.CallableQueueService.threads</name>
-     <value>10</value>
-     <description>Number of threads used for executing callables</description>
-   </property>
-
-   <property>
-     <name>oozie.service.CallableQueueService.callable.concurrency</name>
-     <value>3</value>
-     <description>
-     Maximum concurrency for a given callable type.
-     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).
-     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).
-     All commands that use action executors (action-start, action-end, action-kill and action-check) use
-     the action type as the callable type.
-     </description>
-   </property>
-
-   <property>
-     <name>oozie.service.coord.normal.default.timeout</name>
-     <value>120</value>
-     <description>Default timeout for a coordinator action input check (in minutes) for normal job.
+      The Oozie system ID.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.systemmode</name>
+    <value>NORMAL</value>
+    <description>
+      System mode for  Oozie at startup.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.AuthorizationService.authorization.enabled</name>
+    <value>true</value>
+    <description>
+      Specifies whether security (user name/admin role) is enabled or not.
+      If disabled any user can manage Oozie system and manage any job.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.PurgeService.older.than</name>
+    <value>30</value>
+    <description>
+      Jobs older than this value, in days, will be purged by the PurgeService.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.PurgeService.purge.interval</name>
+    <value>3600</value>
+    <description>
+      Interval at which the purge service will run, in seconds.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.CallableQueueService.queue.size</name>
+    <value>1000</value>
+    <description>Max callable queue size</description>
+  </property>
+
+  <property>
+    <name>oozie.service.CallableQueueService.threads</name>
+    <value>10</value>
+    <description>Number of threads used for executing callables</description>
+  </property>
+
+  <property>
+    <name>oozie.service.CallableQueueService.callable.concurrency</name>
+    <value>3</value>
+    <description>
+      Maximum concurrency for a given callable type.
+      Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).
+      Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).
+      All commands that use action executors (action-start, action-end, action-kill and action-check) use
+      the action type as the callable type.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.coord.normal.default.timeout</name>
+    <value>120</value>
+    <description>Default timeout for a coordinator action input check (in minutes) for normal job.
       -1 means infinite timeout</description>
-   </property>
+  </property>
 
-   <property>
-     <name>oozie.db.schema.name</name>
-     <value>oozie</value>
-     <description>
+  <property>
+    <name>oozie.db.schema.name</name>
+    <value>oozie</value>
+    <description>
       Oozie DataBase Name
-     </description>
-   </property>
+    </description>
+  </property>
 
-    <property>
-      <name>oozie.service.HadoopAccessorService.jobTracker.whitelist</name>
-      <value> </value>
-      <description>
+  <property>
+    <name>oozie.service.HadoopAccessorService.jobTracker.whitelist</name>
+    <value> </value>
+    <description>
       Whitelisted job tracker for Oozie service.
-      </description>
-    </property>
-   
-    <property>
-      <name>oozie.authentication.type</name>
-      <value>simple</value>
-      <description>
-      </description>
-    </property>
-   
-    <property>
-      <name>oozie.service.HadoopAccessorService.nameNode.whitelist</name>
-      <value> </value>
-      <description>
-      </description>
-    </property>
-
-    <property>
-      <name>oozie.service.WorkflowAppService.system.libpath</name>
-      <value>/user/${user.name}/share/lib</value>
-      <description>
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.authentication.type</name>
+    <value>simple</value>
+    <description>
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.HadoopAccessorService.nameNode.whitelist</name>
+    <value> </value>
+    <description>
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.WorkflowAppService.system.libpath</name>
+    <value>/user/${user.name}/share/lib</value>
+    <description>
       System library path to use for workflow applications.
       This path is added to workflow application if their job properties sets
       the property 'oozie.use.system.libpath' to true.
-      </description>
-    </property>
+    </description>
+  </property>
 
-    <property>
-      <name>use.system.libpath.for.mapreduce.and.pig.jobs</name>
-      <value>false</value>
-      <description>
+  <property>
+    <name>use.system.libpath.for.mapreduce.and.pig.jobs</name>
+    <value>false</value>
+    <description>
       If set to true, submissions of MapReduce and Pig jobs will include
       automatically the system library path, thus not requiring users to
       specify where the Pig JAR files are. Instead, the ones from the system
       library path are used.
-      </description>
-    </property>
-    <property>
-      <name>oozie.authentication.kerberos.name.rules</name>
-      <value>DEFAULT</value>
-      <description>The mapping from kerberos principal names to local OS user names.</description>
-    </property>
-    <property>
-      <name>oozie.service.HadoopAccessorService.hadoop.configurations</name>
-      <value>*=/etc/hadoop/conf</value>
-      <description>
-          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of
-          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is
-          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains
-          the relevant Hadoop *-site.xml files. If the path is relative is looked within
-          the Oozie configuration directory; though the path can be absolute (i.e. to point
-          to Hadoop client conf/ directories in the local filesystem.
-      </description>
-    </property>
-    <property>
-        <name>oozie.service.ActionService.executor.ext.classes</name>
-        <value>org.apache.oozie.action.email.EmailActionExecutor,
-org.apache.oozie.action.hadoop.HiveActionExecutor,
-org.apache.oozie.action.hadoop.ShellActionExecutor,
-org.apache.oozie.action.hadoop.SqoopActionExecutor,
-org.apache.oozie.action.hadoop.DistcpActionExecutor</value>
-    </property>
-
-    <property>
-        <name>oozie.service.SchemaService.wf.ext.schemas</name>
-        <value>shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd</value>
-    </property>
-    <property>
-        <name>oozie.service.JPAService.create.db.schema</name>
-        <value>false</value>
-        <description>
-            Creates Oozie DB.
-
-            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.
-            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.
-        </description>
-    </property>
-
-    <property>
-        <name>oozie.service.JPAService.jdbc.driver</name>
-        <value>org.apache.derby.jdbc.EmbeddedDriver</value>
-        <description>
-            JDBC driver class.
-        </description>
-    </property>
-
-    <property>
-        <name>oozie.service.JPAService.jdbc.url</name>
-        <value>jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true</value>
-        <description>
-            JDBC URL.
-        </description>
-    </property>
-
-    <property>
-        <name>oozie.service.JPAService.jdbc.username</name>
-        <value>sa</value>
-        <description>
-            DB user name.
-        </description>
-    </property>
-
-    <property>
-        <name>oozie.service.JPAService.jdbc.password</name>
-        <value> </value>
-        <description>
-            DB user password.
-
-            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,
-                       if empty Configuration assumes it is NULL.
-        </description>
-    </property>
-
-    <property>
-        <name>oozie.service.JPAService.pool.max.active.conn</name>
-        <value>10</value>
-        <description>
-             Max number of connections.
-        </description>
-    </property>
+    </description>
+  </property>
+  <property>
+    <name>oozie.authentication.kerberos.name.rules</name>
+    <value>DEFAULT</value>
+    <description>The mapping from kerberos principal names to local OS user names.</description>
+  </property>
+  <property>
+    <name>oozie.service.HadoopAccessorService.hadoop.configurations</name>
+    <value>*=/etc/hadoop/conf</value>
+    <description>
+      Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of
+      the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is
+      used when there is no exact match for an authority. The HADOOP_CONF_DIR contains
+      the relevant Hadoop *-site.xml files. If the path is relative is looked within
+      the Oozie configuration directory; though the path can be absolute (i.e. to point
+      to Hadoop client conf/ directories in the local filesystem.
+    </description>
+  </property>
+  <property>
+    <name>oozie.service.ActionService.executor.ext.classes</name>
+    <value>org.apache.oozie.action.email.EmailActionExecutor,
+      org.apache.oozie.action.hadoop.HiveActionExecutor,
+      org.apache.oozie.action.hadoop.ShellActionExecutor,
+      org.apache.oozie.action.hadoop.SqoopActionExecutor,
+      org.apache.oozie.action.hadoop.DistcpActionExecutor</value>
+  </property>
+
+  <property>
+    <name>oozie.service.SchemaService.wf.ext.schemas</name>
+    <value>shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd</value>
+  </property>
+  <property>
+    <name>oozie.service.JPAService.create.db.schema</name>
+    <value>false</value>
+    <description>
+      Creates Oozie DB.
+
+      If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.
+      If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.JPAService.jdbc.driver</name>
+    <value>org.apache.derby.jdbc.EmbeddedDriver</value>
+    <description>
+      JDBC driver class.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.JPAService.jdbc.url</name>
+    <value>jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true</value>
+    <description>
+      JDBC URL.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.JPAService.jdbc.username</name>
+    <value>oozie</value>
+    <description>
+      DB user name.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.JPAService.jdbc.password</name>
+    <value> </value>
+    <description>
+      DB user password.
+
+      IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,
+      if empty Configuration assumes it is NULL.
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.JPAService.pool.max.active.conn</name>
+    <value>10</value>
+    <description>
+      Max number of connections.
+    </description>
+  </property>
 </configuration>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HBASE/configuration/hbase-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HBASE/configuration/hbase-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HBASE/configuration/hbase-site.xml
index 4270410..f5fd2f3 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HBASE/configuration/hbase-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HBASE/configuration/hbase-site.xml
@@ -24,13 +24,13 @@
     <name>hbase.rootdir</name>
     <value></value>
     <description>The directory shared by region servers and into
-    which HBase persists.  The URL should be 'fully-qualified'
-    to include the filesystem scheme.  For example, to specify the
-    HDFS directory '/hbase' where the HDFS instance's namenode is
-    running at namenode.example.org on port 9000, set this value to:
-    hdfs://namenode.example.org:9000/hbase.  By default HBase writes
-    into /tmp.  Change this configuration else all data will be lost
-    on machine restart.
+      which HBase persists.  The URL should be 'fully-qualified'
+      to include the filesystem scheme.  For example, to specify the
+      HDFS directory '/hbase' where the HDFS instance's namenode is
+      running at namenode.example.org on port 9000, set this value to:
+      hdfs://namenode.example.org:9000/hbase.  By default HBase writes
+      into /tmp.  Change this configuration else all data will be lost
+      on machine restart.
     </description>
   </property>
   <property>
@@ -44,11 +44,11 @@
   </property>
   <property>
     <name>hbase.tmp.dir</name>
-    <value></value>
+    <value>/hadoop/hbase</value>
     <description>Temporary directory on the local filesystem.
-    Change this setting to point to a location more permanent
-    than '/tmp' (The '/tmp' directory is often cleared on
-    machine restart).
+      Change this setting to point to a location more permanent
+      than '/tmp' (The '/tmp' directory is often cleared on
+      machine restart).
     </description>
   </property>
   <property>
@@ -69,31 +69,31 @@
   </property>
   <property>
     <name>hbase.regionserver.global.memstore.upperLimit</name>
-    <value></value>
+    <value>0.4</value>
     <description>Maximum size of all memstores in a region server before new
       updates are blocked and flushes are forced. Defaults to 40% of heap
     </description>
   </property>
   <property>
     <name>hbase.regionserver.handler.count</name>
-    <value></value>
+    <value>60</value>
     <description>Count of RPC Listener instances spun up on RegionServers.
-    Same property is used by the Master for count of master handlers.
-    Default is 10.
+      Same property is used by the Master for count of master handlers.
+      Default is 10.
     </description>
   </property>
   <property>
     <name>hbase.hregion.majorcompaction</name>
-    <value></value>
-    <description>The time (in miliseconds) between 'major' compactions of all
-    HStoreFiles in a region.  Default: 1 day.
-    Set to 0 to disable automated major compactions.
+    <value>86400000</value>
+    <description>The time (in milliseconds) between 'major' compactions of all
+      HStoreFiles in a region.  Default: 1 day.
+      Set to 0 to disable automated major compactions.
     </description>
   </property>
-  
+
   <property>
     <name>hbase.regionserver.global.memstore.lowerLimit</name>
-    <value></value>
+    <value>0.38</value>
     <description>When memstores are being forced to flush to make room in
       memory, keep flushing until we hit this mark. Defaults to 35% of heap.
       This value equal to hbase.regionserver.global.memstore.upperLimit causes
@@ -103,27 +103,27 @@
   </property>
   <property>
     <name>hbase.hregion.memstore.block.multiplier</name>
-    <value></value>
+    <value>2</value>
     <description>Block updates if memstore has hbase.hregion.memstore.block.multiplier
-    time hbase.hregion.flush.size bytes.  Useful preventing
-    runaway memstore during spikes in update traffic.  Without an
-    upper-bound, memstore fills such that when it flushes the
-    resultant flush files take a long time to compact or split, or
-    worse, we OOME
+      time hbase.hregion.flush.size bytes.  Useful preventing
+      runaway memstore during spikes in update traffic.  Without an
+      upper-bound, memstore fills such that when it flushes the
+      resultant flush files take a long time to compact or split, or
+      worse, we OOME
     </description>
   </property>
   <property>
     <name>hbase.hregion.memstore.flush.size</name>
-    <value></value>
+    <value>134217728</value>
     <description>
-    Memstore will be flushed to disk if size of the memstore
-    exceeds this number of bytes.  Value is checked by a thread that runs
-    every hbase.server.thread.wakefrequency.
+      Memstore will be flushed to disk if size of the memstore
+      exceeds this number of bytes.  Value is checked by a thread that runs
+      every hbase.server.thread.wakefrequency.
     </description>
   </property>
   <property>
     <name>hbase.hregion.memstore.mslab.enabled</name>
-    <value></value>
+    <value>true</value>
     <description>
       Enables the MemStore-Local Allocation Buffer,
       a feature which works to prevent heap fragmentation under
@@ -133,22 +133,22 @@
   </property>
   <property>
     <name>hbase.hregion.max.filesize</name>
-    <value></value>
+    <value>10737418240</value>
     <description>
-    Maximum HStoreFile size. If any one of a column families' HStoreFiles has
-    grown to exceed this value, the hosting HRegion is split in two.
-    Default: 1G.
+      Maximum HStoreFile size. If any one of a column families' HStoreFiles has
+      grown to exceed this value, the hosting HRegion is split in two.
+      Default: 1G.
     </description>
   </property>
   <property>
     <name>hbase.client.scanner.caching</name>
-    <value></value>
+    <value>100</value>
     <description>Number of rows that will be fetched when calling next
-    on a scanner if it is not served from (local, client) memory. Higher
-    caching values will enable faster scanners but will eat up more memory
-    and some calls of next may take longer and longer times when the cache is empty.
-    Do not set this value such that the time between invocations is greater
-    than the scanner timeout; i.e. hbase.regionserver.lease.period
+      on a scanner if it is not served from (local, client) memory. Higher
+      caching values will enable faster scanners but will eat up more memory
+      and some calls of next may take longer and longer times when the cache is empty.
+      Do not set this value such that the time between invocations is greater
+      than the scanner timeout; i.e. hbase.regionserver.lease.period
     </description>
   </property>
   <property>
@@ -164,50 +164,50 @@
   </property>
   <property>
     <name>hbase.client.keyvalue.maxsize</name>
-    <value></value>
+    <value>10485760</value>
     <description>Specifies the combined maximum allowed size of a KeyValue
-    instance. This is to set an upper boundary for a single entry saved in a
-    storage file. Since they cannot be split it helps avoiding that a region
-    cannot be split any further because the data is too large. It seems wise
-    to set this to a fraction of the maximum region size. Setting it to zero
-    or less disables the check.
+      instance. This is to set an upper boundary for a single entry saved in a
+      storage file. Since they cannot be split it helps avoiding that a region
+      cannot be split any further because the data is too large. It seems wise
+      to set this to a fraction of the maximum region size. Setting it to zero
+      or less disables the check.
     </description>
   </property>
   <property>
     <name>hbase.hstore.compactionThreshold</name>
-    <value></value>
+    <value>3</value>
     <description>
-    If more than this number of HStoreFiles in any one HStore
-    (one HStoreFile is written per flush of memstore) then a compaction
-    is run to rewrite all HStoreFiles files as one.  Larger numbers
-    put off compaction but when it runs, it takes longer to complete.
+      If more than this number of HStoreFiles in any one HStore
+      (one HStoreFile is written per flush of memstore) then a compaction
+      is run to rewrite all HStoreFiles files as one.  Larger numbers
+      put off compaction but when it runs, it takes longer to complete.
     </description>
   </property>
   <property>
     <name>hbase.hstore.flush.retries.number</name>
     <value>120</value>
     <description>
-    The number of times the region flush operation will be retried.
+      The number of times the region flush operation will be retried.
     </description>
   </property>
-  
+
   <property>
     <name>hbase.hstore.blockingStoreFiles</name>
-    <value></value>
+    <value>10</value>
     <description>
-    If more than this number of StoreFiles in any one Store
-    (one StoreFile is written per flush of MemStore) then updates are
-    blocked for this HRegion until a compaction is completed, or
-    until hbase.hstore.blockingWaitTime has been exceeded.
+      If more than this number of StoreFiles in any one Store
+      (one StoreFile is written per flush of MemStore) then updates are
+      blocked for this HRegion until a compaction is completed, or
+      until hbase.hstore.blockingWaitTime has been exceeded.
     </description>
   </property>
   <property>
     <name>hfile.block.cache.size</name>
-    <value></value>
+    <value>0.40</value>
     <description>
-        Percentage of maximum heap (-Xmx setting) to allocate to block cache
-        used by HFile/StoreFile. Default of 0.25 means allocate 25%.
-        Set to 0 to disable but it's not recommended.
+      Percentage of maximum heap (-Xmx setting) to allocate to block cache
+      used by HFile/StoreFile. Default of 0.25 means allocate 25%.
+      Set to 0 to disable but it's not recommended.
     </description>
   </property>
 
@@ -218,35 +218,35 @@
     <name>hbase.master.keytab.file</name>
     <value></value>
     <description>Full path to the kerberos keytab file to use for logging in
-    the configured HMaster server principal.
+      the configured HMaster server principal.
     </description>
   </property>
   <property>
     <name>hbase.master.kerberos.principal</name>
     <value></value>
     <description>Ex. "hbase/_HOST@EXAMPLE.COM".  The kerberos principal name
-    that should be used to run the HMaster process.  The principal name should
-    be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the hostname
-    portion, it will be replaced with the actual hostname of the running
-    instance.
+      that should be used to run the HMaster process.  The principal name should
+      be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the hostname
+      portion, it will be replaced with the actual hostname of the running
+      instance.
     </description>
   </property>
   <property>
     <name>hbase.regionserver.keytab.file</name>
     <value></value>
     <description>Full path to the kerberos keytab file to use for logging in
-    the configured HRegionServer server principal.
+      the configured HRegionServer server principal.
     </description>
   </property>
   <property>
     <name>hbase.regionserver.kerberos.principal</name>
     <value></value>
     <description>Ex. "hbase/_HOST@EXAMPLE.COM".  The kerberos principal name
-    that should be used to run the HRegionServer process.  The principal name
-    should be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the
-    hostname portion, it will be replaced with the actual hostname of the
-    running instance.  An entry for this principal must exist in the file
-    specified in hbase.regionserver.keytab.file
+      that should be used to run the HRegionServer process.  The principal name
+      should be in the form: user/hostname@DOMAIN.  If "_HOST" is used as the
+      hostname portion, it will be replaced with the actual hostname of the
+      running instance.  An entry for this principal must exist in the file
+      specified in hbase.regionserver.keytab.file
     </description>
   </property>
 
@@ -255,8 +255,8 @@
     <name>hbase.superuser</name>
     <value>hbase</value>
     <description>List of users or groups (comma-separated), who are allowed
-    full privileges, regardless of stored ACLs, across the cluster.
-    Only used when HBase security is enabled.
+      full privileges, regardless of stored ACLs, across the cluster.
+      Only used when HBase security is enabled.
     </description>
   </property>
 
@@ -276,10 +276,10 @@
     <name>hbase.coprocessor.region.classes</name>
     <value></value>
     <description>A comma-separated list of Coprocessors that are loaded by
-    default on all tables. For any override coprocessor method, these classes
-    will be called in order. After implementing your own Coprocessor, just put
-    it in HBase's classpath and add the fully qualified class name here.
-    A coprocessor can also be loaded on demand by setting HTableDescriptor.
+      default on all tables. For any override coprocessor method, these classes
+      will be called in order. After implementing your own Coprocessor, just put
+      it in HBase's classpath and add the fully qualified class name here.
+      A coprocessor can also be loaded on demand by setting HTableDescriptor.
     </description>
   </property>
 
@@ -299,7 +299,7 @@
     <name>hbase.zookeeper.property.clientPort</name>
     <value>2181</value>
     <description>Property from ZooKeeper's config zoo.cfg.
-    The port at which the clients will connect.
+      The port at which the clients will connect.
     </description>
   </property>
 
@@ -311,11 +311,11 @@
     <name>hbase.zookeeper.quorum</name>
     <value></value>
     <description>Comma separated list of servers in the ZooKeeper Quorum.
-    For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
-    By default this is set to localhost for local and pseudo-distributed modes
-    of operation. For a fully-distributed setup, this should be set to a full
-    list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
-    this is the list of servers which we will start/stop ZooKeeper on.
+      For example, "host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
+      By default this is set to localhost for local and pseudo-distributed modes
+      of operation. For a fully-distributed setup, this should be set to a full
+      list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in hbase-env.sh
+      this is the list of servers which we will start/stop ZooKeeper on.
     </description>
   </property>
   <!-- End of properties used to generate ZooKeeper host:port quorum list. -->
@@ -324,11 +324,11 @@
     <name>hbase.zookeeper.useMulti</name>
     <value>true</value>
     <description>Instructs HBase to make use of ZooKeeper's multi-update functionality.
-    This allows certain ZooKeeper operations to complete more quickly and prevents some issues
-    with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).·
-    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+
-    and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will
-    not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).
+      This allows certain ZooKeeper operations to complete more quickly and prevents some issues
+      with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).·
+      IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+
+      and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will
+      not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).
     </description>
   </property>
   <property>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
index 79e0151..b8e4772 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
@@ -22,12 +22,12 @@
 
 <configuration>
 
-<!-- file system properties -->
+  <!-- file system properties -->
 
   <property>
     <name>dfs.namenode.name.dir</name>
     <!-- cluster variant -->
-    <value></value>
+    <value>/hadoop/hdfs/namenode</value>
     <description>Determines where on the local filesystem the DFS name node
       should store the name table.  If this is a comma-delimited list
       of directories then the name table is replicated in all of the
@@ -45,26 +45,26 @@
   <property>
     <name>dfs.webhdfs.enabled</name>
     <value>true</value>
-    <description>to enable webhdfs</description>
+    <description>Whether to enable WebHDFS feature</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.datanode.failed.volumes.tolerated</name>
     <value>0</value>
-    <description>#of failed disks dn would tolerate</description>
+    <description> Number of failed disks a DataNode would tolerate before it stops offering service</description>
     <final>true</final>
   </property>
 
   <property>
     <name>dfs.datanode.data.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/data</value>
     <description>Determines where on the local filesystem an DFS data node
-  should store its blocks.  If this is a comma-delimited
-  list of directories, then data will be stored in all named
-  directories, typically on different devices.
-  Directories that do not exist are ignored.
-  </description>
+      should store its blocks.  If this is a comma-delimited
+      list of directories, then data will be stored in all named
+      directories, typically on different devices.
+      Directories that do not exist are ignored.
+    </description>
     <final>true</final>
   </property>
 
@@ -72,25 +72,25 @@
     <name>dfs.hosts.exclude</name>
     <value></value>
     <description>Names a file that contains a list of hosts that are
-    not permitted to connect to the namenode.  The full pathname of the
-    file must be specified.  If the value is empty, no hosts are
-    excluded.</description>
+      not permitted to connect to the namenode.  The full pathname of the
+      file must be specified.  If the value is empty, no hosts are
+      excluded.</description>
   </property>
 
-<!--
-  <property>
-    <name>dfs.hosts</name>
-    <value></value>
-    <description>Names a file that contains a list of hosts that are
-    permitted to connect to the namenode. The full pathname of the file
-    must be specified.  If the value is empty, all hosts are
-    permitted.</description>
-  </property>
--->
+  <!--
+    <property>
+      <name>dfs.hosts</name>
+      <value></value>
+      <description>Names a file that contains a list of hosts that are
+      permitted to connect to the namenode. The full pathname of the file
+      must be specified.  If the value is empty, all hosts are
+      permitted.</description>
+    </property>
+  -->
 
   <property>
     <name>dfs.namenode.checkpoint.dir</name>
-    <value></value>
+    <value>/hadoop/hdfs/namesecondary</value>
     <description>Determines where on the local filesystem the DFS secondary
       name node should store the temporary images to merge.
       If this is a comma-delimited list of directories then the image is
@@ -109,25 +109,34 @@
     </description>
   </property>
 
+
   <property>
     <name>dfs.namenode.checkpoint.period</name>
-    <value></value>
+    <value>21600</value>
     <description>The number of seconds between two periodic checkpoints.
     </description>
   </property>
 
   <property>
+    <name>fs.checkpoint.size</name>
+    <value>67108864</value>
+    <description>The size of the current edit log (in bytes) that triggers
+      a periodic checkpoint even if the maximum checkpoint delay is not reached
+    </description>
+  </property>
+
+  <property>
     <name>dfs.replication.max</name>
     <value>50</value>
     <description>Maximal block replication.
-  </description>
+    </description>
   </property>
 
   <property>
     <name>dfs.replication</name>
     <value>3</value>
     <description>Default block replication.
-  </description>
+    </description>
   </property>
 
   <property>
@@ -146,21 +155,21 @@
     <name>dfs.namenode.safemode.threshold-pct</name>
     <value>1.0f</value>
     <description>
-        Specifies the percentage of blocks that should satisfy
-        the minimal replication requirement defined by dfs.namenode.replication.min.
-        Values less than or equal to 0 mean not to start in safe mode.
-        Values greater than 1 will make safe mode permanent.
-        </description>
+      Specifies the percentage of blocks that should satisfy
+      the minimal replication requirement defined by dfs.namenode.replication.min.
+      Values less than or equal to 0 mean not to start in safe mode.
+      Values greater than 1 will make safe mode permanent.
+    </description>
   </property>
 
   <property>
     <name>dfs.datanode.balance.bandwidthPerSec</name>
     <value>6250000</value>
     <description>
-        Specifies the maximum amount of bandwidth that each datanode
-        can utilize for the balancing purpose in term of
-        the number of bytes per second.
-  </description>
+      Specifies the maximum amount of bandwidth that each datanode
+      can utilize for the balancing purpose in term of
+      the number of bytes per second.
+    </description>
   </property>
 
   <property>
@@ -190,113 +199,113 @@
   <property>
     <name>dfs.namenode.http-address</name>
     <value></value>
-<description>The name of the default file system.  Either the
-literal string "local" or a host:port for NDFS.</description>
-<final>true</final>
-</property>
-
-<property>
-<name>dfs.datanode.du.reserved</name>
-<!-- cluster variant -->
-<value>1073741824</value>
-<description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
-</description>
-</property>
-
-<property>
-<name>dfs.datanode.ipc.address</name>
-<value>0.0.0.0:8010</value>
-<description>
-The datanode ipc server address and port.
-If the port is 0 then the server will start on a free port.
-</description>
-</property>
-
-<property>
-<name>dfs.blockreport.initialDelay</name>
-<value>120</value>
-<description>Delay for first block report in seconds.</description>
-</property>
-
-<property>
-<name>dfs.namenode.handler.count</name>
-<value>40</value>
-<description>The number of server threads for the namenode.</description>
-</property>
-
-<property>
-<name>dfs.datanode.max.transfer.threads</name>
-<value>1024</value>
-<description>PRIVATE CONFIG VARIABLE</description>
-</property>
-
-<!-- Permissions configuration -->
-
-<property>
-<name>fs.permissions.umask-mode</name>
-<value>022</value>
-<description>
-The octal umask used when creating files and directories.
-</description>
-</property>
-
-<property>
-<name>dfs.permissions.enabled</name>
-<value>true</value>
-<description>
-If "true", enable permission checking in HDFS.
-If "false", permission checking is turned off,
-but all other behavior is unchanged.
-Switching from one parameter value to the other does not change the mode,
-owner or group of files or directories.
-</description>
-</property>
-
-<property>
-<name>dfs.permissions.superusergroup</name>
-<value>hdfs</value>
-<description>The name of the group of super-users.</description>
-</property>
-
-<property>
-<name>dfs.namenode.handler.count</name>
-<value>100</value>
-<description>Added to grow Queue size so that more client connections are allowed</description>
-</property>
-
-<property>
-<name>dfs.block.access.token.enable</name>
-<value>true</value>
-<description>
-If "true", access tokens are used as capabilities for accessing datanodes.
-If "false", no access tokens are checked on accessing datanodes.
-</description>
-</property>
-
-<property>
-<name>dfs.namenode.kerberos.principal</name>
-<value></value>
-<description>
-Kerberos principal name for the NameNode
-</description>
-</property>
-
-<property>
-<name>dfs.secondary.namenode.kerberos.principal</name>
-<value></value>
+    <description>The name of the default file system.  Either the
+      literal string "local" or a host:port for NDFS.</description>
+    <final>true</final>
+  </property>
+
+  <property>
+    <name>dfs.datanode.du.reserved</name>
+    <!-- cluster variant -->
+    <value>1073741824</value>
+    <description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.datanode.ipc.address</name>
+    <value>0.0.0.0:8010</value>
     <description>
-        Kerberos principal name for the secondary NameNode.
+      The datanode ipc server address and port.
+      If the port is 0 then the server will start on a free port.
     </description>
   </property>
 
+  <property>
+    <name>dfs.blockreport.initialDelay</name>
+    <value>120</value>
+    <description>Delay for first block report in seconds.</description>
+  </property>
+
+  <property>
+    <name>dfs.namenode.handler.count</name>
+    <value>40</value>
+    <description>The number of server threads for the namenode.</description>
+  </property>
+
+  <property>
+    <name>dfs.datanode.max.transfer.threads</name>
+    <value>1024</value>
+    <description>PRIVATE CONFIG VARIABLE</description>
+  </property>
+
+  <!-- Permissions configuration -->
+
+  <property>
+    <name>fs.permissions.umask-mode</name>
+    <value>022</value>
+    <description>
+      The octal umask used when creating files and directories.
+    </description>
+  </property>
 
-<!--
-  This is KRB DOMAIN specific. The FQDN of the namenode has to be mentioned.
--->
+  <property>
+    <name>dfs.permissions.enabled</name>
+    <value>true</value>
+    <description>
+      If "true", enable permission checking in HDFS.
+      If "false", permission checking is turned off,
+      but all other behavior is unchanged.
+      Switching from one parameter value to the other does not change the mode,
+      owner or group of files or directories.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.permissions.superusergroup</name>
+    <value>hdfs</value>
+    <description>The name of the group of super-users.</description>
+  </property>
+
+  <property>
+    <name>dfs.namenode.handler.count</name>
+    <value>100</value>
+    <description>Added to grow Queue size so that more client connections are allowed</description>
+  </property>
+
+  <property>
+    <name>dfs.block.access.token.enable</name>
+    <value>true</value>
+    <description>
+      If "true", access tokens are used as capabilities for accessing datanodes.
+      If "false", no access tokens are checked on accessing datanodes.
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.namenode.kerberos.principal</name>
+    <value></value>
+    <description>
+      Kerberos principal name for the NameNode
+    </description>
+  </property>
+
+  <property>
+    <name>dfs.secondary.namenode.kerberos.principal</name>
+    <value></value>
+    <description>
+      Kerberos principal name for the secondary NameNode.
+    </description>
+  </property>
+
+
+  <!--
+    This is KRB DOMAIN specific. The FQDN of the namenode has to be mentioned.
+  -->
   <property>
     <name>dfs.namenode.kerberos.https.principal</name>
     <value></value>
-     <description>The Kerberos principal for the host that the NameNode runs on.</description>
+    <description>The Kerberos principal for the host that the NameNode runs on.</description>
 
   </property>
 
@@ -336,64 +345,64 @@ Kerberos principal name for the NameNode
   <property>
     <name>dfs.datanode.kerberos.principal</name>
     <value></value>
- <description>
-        The Kerberos principal that the DataNode runs as. "_HOST" is replaced by the real host name.
+    <description>
+      The Kerberos principal that the DataNode runs as. "_HOST" is replaced by the real host name.
     </description>
   </property>
 
   <property>
     <name>dfs.namenode.keytab.file</name>
     <value></value>
- <description>
-        Combined keytab file containing the namenode service and host principals.
+    <description>
+      Combined keytab file containing the namenode service and host principals.
     </description>
   </property>
 
   <property>
     <name>dfs.secondary.namenode.keytab.file</name>
     <value></value>
-  <description>
-        Combined keytab file containing the namenode service and host principals.
+    <description>
+      Combined keytab file containing the namenode service and host principals.
     </description>
   </property>
 
   <property>
     <name>dfs.datanode.keytab.file</name>
     <value></value>
- <description>
-        The filename of the keytab file for the DataNode.
+    <description>
+      The filename of the keytab file for the DataNode.
     </description>
   </property>
 
   <property>
     <name>dfs.namenode.https-address</name>
     <value></value>
-  <description>The https address where namenode binds</description>
+    <description>The https address where namenode binds</description>
 
   </property>
 
   <property>
     <name>dfs.datanode.data.dir.perm</name>
     <value>750</value>
-<description>The permissions that should be there on dfs.datanode.data.dir
-directories. The datanode will not come up if the permissions are
-different on existing dfs.datanode.data.dir directories. If the directories
-don't exist, they will be created with this permission.</description>
+    <description>The permissions that should be there on dfs.datanode.data.dir
+      directories. The datanode will not come up if the permissions are
+      different on existing dfs.datanode.data.dir directories. If the directories
+      don't exist, they will be created with this permission.</description>
   </property>
 
   <property>
     <name>dfs.namenode.accesstime.precision</name>
     <value>0</value>
     <description>The access time for HDFS file is precise upto this value.
-                 The default value is 1 hour. Setting a value of 0 disables
-                 access times for HDFS.
+      The default value is 1 hour. Setting a value of 0 disables
+      access times for HDFS.
     </description>
   </property>
 
   <property>
-   <name>dfs.cluster.administrators</name>
-   <value> hdfs</value>
-   <description>ACL for who all can view the default servlets in the HDFS</description>
+    <name>dfs.cluster.administrators</name>
+    <value> hdfs</value>
+    <description>ACL for who all can view the default servlets in the HDFS</description>
   </property>
 
   <property>
@@ -426,14 +435,14 @@ don't exist, they will be created with this permission.</description>
     <value>30000</value>
     <description>Datanode is stale after not getting a heartbeat in this interval in ms</description>
   </property>
-  
+
   <property>
     <name>dfs.journalnode.http-address</name>
     <value>0.0.0.0:8480</value>
     <description>The address and port the JournalNode web UI listens on.
-     If the port is 0 then the server will start on a free port. </description>
+      If the port is 0 then the server will start on a free port. </description>
   </property>
-  
+
   <property>
     <name>dfs.journalnode.edits.dir</name>
     <value>/grid/0/hdfs/journal</value>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
index 60ecfe7..1b5fa9c 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
@@ -20,7 +20,7 @@ limitations under the License.
 <configuration>
   <property>
     <name>javax.jdo.option.ConnectionURL</name>
-    <value></value>
+    <value>jdbc</value>
     <description>JDBC connect string for a JDBC metastore</description>
   </property>
 
@@ -32,13 +32,13 @@ limitations under the License.
 
   <property>
     <name>javax.jdo.option.ConnectionUserName</name>
-    <value></value>
+    <value>hive</value>
     <description>username to use against metastore database</description>
   </property>
 
   <property>
     <name>javax.jdo.option.ConnectionPassword</name>
-    <value></value>
+    <value> </value>
     <description>password to use against metastore database</description>
   </property>
 
@@ -52,21 +52,21 @@ limitations under the License.
     <name>hive.metastore.sasl.enabled</name>
     <value></value>
     <description>If true, the metastore thrift interface will be secured with SASL.
-     Clients must authenticate with Kerberos.</description>
+      Clients must authenticate with Kerberos.</description>
   </property>
 
   <property>
     <name>hive.metastore.kerberos.keytab.file</name>
     <value></value>
     <description>The path to the Kerberos Keytab file containing the metastore
-     thrift server's service principal.</description>
+      thrift server's service principal.</description>
   </property>
 
   <property>
     <name>hive.metastore.kerberos.principal</name>
     <value></value>
     <description>The service principal for the metastore thrift server. The special
-    string _HOST will be replaced automatically with the correct host name.</description>
+      string _HOST will be replaced automatically with the correct host name.</description>
   </property>
 
   <property>
@@ -109,7 +109,7 @@ limitations under the License.
     <name>hive.security.authorization.manager</name>
     <value>org.apache.hadoop.hive.ql.security.authorization.DefaultHiveAuthorizationProvider</value>
     <description>the hive client authorization manager class name.
-    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  </description>
+      The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  </description>
   </property>
 
   <property>
@@ -241,7 +241,7 @@ limitations under the License.
     <name>hive.optimize.index.filter</name>
     <value>true</value>
     <description>
-    Whether to enable automatic use of indexes
+      Whether to enable automatic use of indexes
     </description>
   </property>
 

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/OOZIE/configuration/oozie-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/OOZIE/configuration/oozie-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/OOZIE/configuration/oozie-site.xml
index afd1e7f..28529ff 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/OOZIE/configuration/oozie-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/OOZIE/configuration/oozie-site.xml
@@ -15,299 +15,299 @@
   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
   See the License for the specific language governing permissions and
   limitations under the License.
--->     
+-->
 
 <configuration>
 
-<!--
-    Refer to the oozie-default.xml file for the complete list of
-    Oozie configuration properties and their default values.
--->
+  <!--
+      Refer to the oozie-default.xml file for the complete list of
+      Oozie configuration properties and their default values.
+  -->
   <property>
     <name>oozie.base.url</name>
     <value>http://localhost:11000/oozie</value>
     <description>Base Oozie URL.</description>
-   </property>
+  </property>
 
   <property>
     <name>oozie.system.id</name>
     <value>oozie-${user.name}</value>
     <description>
-    The Oozie system ID.
+      The Oozie system ID.
     </description>
-   </property>
+  </property>
 
-   <property>
-     <name>oozie.systemmode</name>
-     <value>NORMAL</value>
-     <description>
-     System mode for  Oozie at startup.
-     </description>
-   </property>
+  <property>
+    <name>oozie.systemmode</name>
+    <value>NORMAL</value>
+    <description>
+      System mode for  Oozie at startup.
+    </description>
+  </property>
 
-   <property>
-     <name>oozie.service.AuthorizationService.security.enabled</name>
-     <value>true</value>
-     <description>
-     Specifies whether security (user name/admin role) is enabled or not.
-     If disabled any user can manage Oozie system and manage any job.
-     </description>
-   </property>
+  <property>
+    <name>oozie.service.AuthorizationService.security.enabled</name>
+    <value>true</value>
+    <description>
+      Specifies whether security (user name/admin role) is enabled or not.
+      If disabled any user can manage Oozie system and manage any job.
+    </description>
+  </property>
 
-   <property>
-     <name>oozie.service.PurgeService.older.than</name>
-     <value>30</value>
-     <description>
-     Jobs older than this value, in days, will be purged by the PurgeService.
-     </description>
-   </property>
+  <property>
+    <name>oozie.service.PurgeService.older.than</name>
+    <value>30</value>
+    <description>
+      Jobs older than this value, in days, will be purged by the PurgeService.
+    </description>
+  </property>
 
-   <property>
-     <name>oozie.service.PurgeService.purge.interval</name>
-     <value>3600</value>
-     <description>
-     Interval at which the purge service will run, in seconds.
-     </description>
-   </property>
+  <property>
+    <name>oozie.service.PurgeService.purge.interval</name>
+    <value>3600</value>
+    <description>
+      Interval at which the purge service will run, in seconds.
+    </description>
+  </property>
 
-   <property>
-     <name>oozie.service.CallableQueueService.queue.size</name>
-     <value>1000</value>
-     <description>Max callable queue size</description>
-   </property>
+  <property>
+    <name>oozie.service.CallableQueueService.queue.size</name>
+    <value>1000</value>
+    <description>Max callable queue size</description>
+  </property>
 
-   <property>
-     <name>oozie.service.CallableQueueService.threads</name>
-     <value>10</value>
-     <description>Number of threads used for executing callables</description>
-   </property>
+  <property>
+    <name>oozie.service.CallableQueueService.threads</name>
+    <value>10</value>
+    <description>Number of threads used for executing callables</description>
+  </property>
 
-   <property>
-     <name>oozie.service.CallableQueueService.callable.concurrency</name>
-     <value>3</value>
-     <description>
-     Maximum concurrency for a given callable type.
-     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).
-     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).
-     All commands that use action executors (action-start, action-end, action-kill and action-check) use
-     the action type as the callable type.
-     </description>
-   </property>
+  <property>
+    <name>oozie.service.CallableQueueService.callable.concurrency</name>
+    <value>3</value>
+    <description>
+      Maximum concurrency for a given callable type.
+      Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).
+      Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).
+      All commands that use action executors (action-start, action-end, action-kill and action-check) use
+      the action type as the callable type.
+    </description>
+  </property>
 
-   <property>
-     <name>oozie.service.coord.normal.default.timeout</name>
-     <value>120</value>
-     <description>Default timeout for a coordinator action input check (in minutes) for normal job.
+  <property>
+    <name>oozie.service.coord.normal.default.timeout</name>
+    <value>120</value>
+    <description>Default timeout for a coordinator action input check (in minutes) for normal job.
       -1 means infinite timeout</description>
-   </property>
+  </property>
 
-   <property>
-     <name>oozie.db.schema.name</name>
-     <value>oozie</value>
-     <description>
+  <property>
+    <name>oozie.db.schema.name</name>
+    <value>oozie</value>
+    <description>
       Oozie DataBase Name
-     </description>
-   </property>
+    </description>
+  </property>
 
-    <property>
-      <name>oozie.service.HadoopAccessorService.jobTracker.whitelist</name>
-      <value> </value>
-      <description>
+  <property>
+    <name>oozie.service.HadoopAccessorService.jobTracker.whitelist</name>
+    <value> </value>
+    <description>
       Whitelisted job tracker for Oozie service.
-      </description>
-    </property>
-   
-    <property>
-      <name>oozie.authentication.type</name>
-      <value>simple</value>
-      <description>
-      </description>
-    </property>
-   
-    <property>
-      <name>oozie.service.HadoopAccessorService.nameNode.whitelist</name>
-      <value> </value>
-      <description>
-      </description>
-    </property>
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.authentication.type</name>
+    <value>simple</value>
+    <description>
+    </description>
+  </property>
 
-    <property>
-      <name>oozie.service.WorkflowAppService.system.libpath</name>
-      <value>/user/${user.name}/share/lib</value>
-      <description>
+  <property>
+    <name>oozie.service.HadoopAccessorService.nameNode.whitelist</name>
+    <value> </value>
+    <description>
+    </description>
+  </property>
+
+  <property>
+    <name>oozie.service.WorkflowAppService.system.libpath</name>
+    <value>/user/${user.name}/share/lib</value>
+    <description>
       System library path to use for workflow applications.
       This path is added to workflow application if their job properties sets
       the property 'oozie.use.system.libpath' to true.
-      </description>
-    </property>
+    </description>
+  </property>
 
-    <property>
-      <name>use.system.libpath.for.mapreduce.and.pig.jobs</name>
-      <value>false</value>
-      <description>
+  <property>
+    <name>use.system.libpath.for.mapreduce.and.pig.jobs</name>
+    <value>false</value>
+    <description>
       If set to true, submissions of MapReduce and Pig jobs will include
       automatically the system library path, thus not requiring users to
       specify where the Pig JAR files are. Instead, the ones from the system
       library path are used.
-      </description>
-    </property>
-    <property>
-      <name>oozie.authentication.kerberos.name.rules</name>
-      <value>
-        RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/
-        RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/
-        RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/
-        RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/
-        DEFAULT
-        </value>
-      <description>The mapping from kerberos principal names to local OS user names.</description>
-    </property>
-    <property>
-      <name>oozie.service.HadoopAccessorService.hadoop.configurations</name>
-      <value>*=/etc/hadoop/conf</value>
-      <description>
-          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of
-          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is
-          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains
-          the relevant Hadoop *-site.xml files. If the path is relative is looked within
-          the Oozie configuration directory; though the path can be absolute (i.e. to point
-          to Hadoop client conf/ directories in the local filesystem.
-      </description>
-    </property>
-    <property>
-        <name>oozie.service.ActionService.executor.ext.classes</name>
-        <value>
-            org.apache.oozie.action.email.EmailActionExecutor,
-            org.apache.oozie.action.hadoop.HiveActionExecutor,
-            org.apache.oozie.action.hadoop.ShellActionExecutor,
-            org.apache.oozie.action.hadoop.SqoopActionExecutor,
-            org.apache.oozie.action.hadoop.DistcpActionExecutor
-        </value>
-    </property>
+    </description>
+  </property>
+  <property>
+    <name>oozie.authentication.kerberos.name.rules</name>
+    <value>
+      RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/
+      RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/
+      RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/
+      RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/
+      DEFAULT
+    </value>
+    <description>The mapping from kerberos principal names to local OS user names.</description>
+  </property>
+  <property>
+    <name>oozie.service.HadoopAccessorService.hadoop.configurations</name>
+    <value>*=/etc/hadoop/conf</value>
+    <description>
+      Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of
+      the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is
+      used when there is no exact match for an authority. The HADOOP_CONF_DIR contains
+      the relevant Hadoop *-site.xml files. If the path is relative is looked within
+      the Oozie configuration directory; though the path can be absolute (i.e. to point
+      to Hadoop client conf/ directories in the local filesystem.
+    </description>
+  </property>
+  <property>
+    <name>oozie.service.ActionService.executor.ext.classes</name>
+    <value>
+      org.apache.oozie.action.email.EmailActionExecutor,
+      org.apache.oozie.action.hadoop.HiveActionExecutor,
+      org.apache.oozie.action.hadoop.ShellActionExecutor,
+      org.apache.oozie.action.hadoop.SqoopActionExecutor,
+      org.apache.oozie.action.hadoop.DistcpActionExecutor
+    </value>
+  </property>
 
-    <property>
-        <name>oozie.service.SchemaService.wf.ext.schemas</name>
-        <value>shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd,shell-action-0.2.xsd,oozie-sla-0.1.xsd,oozie-sla-0.2.xsd</value>
-    </property>
-    <property>
-        <name>oozie.service.JPAService.create.db.schema</name>
-        <value>false</value>
-        <description>
-            Creates Oozie DB.
+  <property>
+    <name>oozie.service.SchemaService.wf.ext.schemas</name>
+    <value>shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd,shell-action-0.2.xsd,oozie-sla-0.1.xsd,oozie-sla-0.2.xsd</value>
+  </property>
+  <property>
+    <name>oozie.service.JPAService.create.db.schema</name>
+    <value>false</value>
+    <description>
+      Creates Oozie DB.
 
-            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.
-            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.
-        </description>
-    </property>
+      If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.
+      If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.
+    </description>
+  </property>
 
-    <property>
-        <name>oozie.service.JPAService.jdbc.driver</name>
-        <value>org.apache.derby.jdbc.EmbeddedDriver</value>
-        <description>
-            JDBC driver class.
-        </description>
-    </property>
+  <property>
+    <name>oozie.service.JPAService.jdbc.driver</name>
+    <value>org.apache.derby.jdbc.EmbeddedDriver</value>
+    <description>
+      JDBC driver class.
+    </description>
+  </property>
 
-    <property>
-        <name>oozie.service.JPAService.jdbc.url</name>
-        <value>jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true</value>
-        <description>
-            JDBC URL.
-        </description>
-    </property>
+  <property>
+    <name>oozie.service.JPAService.jdbc.url</name>
+    <value>jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true</value>
+    <description>
+      JDBC URL.
+    </description>
+  </property>
 
-    <property>
-        <name>oozie.service.JPAService.jdbc.username</name>
-        <value>sa</value>
-        <description>
-            DB user name.
-        </description>
-    </property>
+  <property>
+    <name>oozie.service.JPAService.jdbc.username</name>
+    <value>oozie</value>
+    <description>
+      Database user name to use to connect to the database
+    </description>
+  </property>
 
-    <property>
-        <name>oozie.service.JPAService.jdbc.password</name>
-        <value> </value>
-        <description>
-            DB user password.
+  <property>
+    <name>oozie.service.JPAService.jdbc.password</name>
+    <value> </value>
+    <description>
+      DB user password.
 
-            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,
-                       if empty Configuration assumes it is NULL.
-        </description>
-    </property>
+      IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,
+      if empty Configuration assumes it is NULL.
+    </description>
+  </property>
 
-    <property>
-        <name>oozie.service.JPAService.pool.max.active.conn</name>
-        <value>10</value>
-        <description>
-             Max number of connections.
-        </description>
-    </property>
+  <property>
+    <name>oozie.service.JPAService.pool.max.active.conn</name>
+    <value>10</value>
+    <description>
+      Max number of connections.
+    </description>
+  </property>
 
-    <property>
-      <name>oozie.services</name>
-      <value>
-        org.apache.oozie.service.SchedulerService,
-        org.apache.oozie.service.InstrumentationService,
-        org.apache.oozie.service.CallableQueueService,
-        org.apache.oozie.service.UUIDService,
-        org.apache.oozie.service.ELService,
-        org.apache.oozie.service.AuthorizationService,
-        org.apache.oozie.service.UserGroupInformationService,
-        org.apache.oozie.service.HadoopAccessorService,
-        org.apache.oozie.service.URIHandlerService,
-        org.apache.oozie.service.MemoryLocksService,
-        org.apache.oozie.service.DagXLogInfoService,
-        org.apache.oozie.service.SchemaService,
-        org.apache.oozie.service.LiteWorkflowAppService,
-        org.apache.oozie.service.JPAService,
-        org.apache.oozie.service.StoreService,
-        org.apache.oozie.service.CoordinatorStoreService,
-        org.apache.oozie.service.SLAStoreService,
-        org.apache.oozie.service.DBLiteWorkflowStoreService,
-        org.apache.oozie.service.CallbackService,
-        org.apache.oozie.service.ActionService,
-        org.apache.oozie.service.ActionCheckerService,
-        org.apache.oozie.service.RecoveryService,
-        org.apache.oozie.service.PurgeService,
-        org.apache.oozie.service.CoordinatorEngineService,
-        org.apache.oozie.service.BundleEngineService,
-        org.apache.oozie.service.DagEngineService,
-        org.apache.oozie.service.CoordMaterializeTriggerService,
-        org.apache.oozie.service.StatusTransitService,
-        org.apache.oozie.service.PauseTransitService,
-        org.apache.oozie.service.GroupsService,
-        org.apache.oozie.service.ProxyUserService
-      </value>
-      <description>List of Oozie services</description>
-    </property>
-    <property>
-      <name>oozie.service.URIHandlerService.uri.handlers</name>
-      <value>org.apache.oozie.dependency.FSURIHandler,org.apache.oozie.dependency.HCatURIHandler</value>
-      <description>
-        Enlist the different uri handlers supported for data availability checks.
-      </description>
-    </property>
-    <property>
+  <property>
+    <name>oozie.services</name>
+    <value>
+      org.apache.oozie.service.SchedulerService,
+      org.apache.oozie.service.InstrumentationService,
+      org.apache.oozie.service.CallableQueueService,
+      org.apache.oozie.service.UUIDService,
+      org.apache.oozie.service.ELService,
+      org.apache.oozie.service.AuthorizationService,
+      org.apache.oozie.service.UserGroupInformationService,
+      org.apache.oozie.service.HadoopAccessorService,
+      org.apache.oozie.service.URIHandlerService,
+      org.apache.oozie.service.MemoryLocksService,
+      org.apache.oozie.service.DagXLogInfoService,
+      org.apache.oozie.service.SchemaService,
+      org.apache.oozie.service.LiteWorkflowAppService,
+      org.apache.oozie.service.JPAService,
+      org.apache.oozie.service.StoreService,
+      org.apache.oozie.service.CoordinatorStoreService,
+      org.apache.oozie.service.SLAStoreService,
+      org.apache.oozie.service.DBLiteWorkflowStoreService,
+      org.apache.oozie.service.CallbackService,
+      org.apache.oozie.service.ActionService,
+      org.apache.oozie.service.ActionCheckerService,
+      org.apache.oozie.service.RecoveryService,
+      org.apache.oozie.service.PurgeService,
+      org.apache.oozie.service.CoordinatorEngineService,
+      org.apache.oozie.service.BundleEngineService,
+      org.apache.oozie.service.DagEngineService,
+      org.apache.oozie.service.CoordMaterializeTriggerService,
+      org.apache.oozie.service.StatusTransitService,
+      org.apache.oozie.service.PauseTransitService,
+      org.apache.oozie.service.GroupsService,
+      org.apache.oozie.service.ProxyUserService
+    </value>
+    <description>List of Oozie services</description>
+  </property>
+  <property>
+    <name>oozie.service.URIHandlerService.uri.handlers</name>
+    <value>org.apache.oozie.dependency.FSURIHandler,org.apache.oozie.dependency.HCatURIHandler</value>
+    <description>
+      Enlist the different uri handlers supported for data availability checks.
+    </description>
+  </property>
+  <property>
     <name>oozie.services.ext</name>
     <value>org.apache.oozie.service.PartitionDependencyManagerService,org.apache.oozie.service.HCatAccessorService</value>
     <description>
-       To add/replace services defined in 'oozie.services' with custom implementations.
-       Class names must be separated by commas.
+      To add/replace services defined in 'oozie.services' with custom implementations.
+      Class names must be separated by commas.
     </description>
-    </property>
-    <property>
+  </property>
+  <property>
     <name>oozie.service.coord.push.check.requeue.interval</name>
     <value>30000</value>
     <description>
-        Command re-queue interval for push dependencies (in millisecond).
+      Command re-queue interval for push dependencies (in millisecond).
+    </description>
+  </property>
+  <property>
+    <name>oozie.credentials.credentialclasses</name>
+    <value>hcat=org.apache.oozie.action.hadoop.HCatCredentials</value>
+    <description>
+      Credential Class to be used for HCat.
     </description>
-    </property>
-    <property>
-      <name>oozie.credentials.credentialclasses</name>
-      <value>hcat=org.apache.oozie.action.hadoop.HCatCredentials</value>
-      <description>
-        Credential Class to be used for HCat.
-      </description>
-    </property>
+  </property>
 
 </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/controllers/main/service/info/configs.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/main/service/info/configs.js b/ambari-web/app/controllers/main/service/info/configs.js
index e96635d..5dc2251 100644
--- a/ambari-web/app/controllers/main/service/info/configs.js
+++ b/ambari-web/app/controllers/main/service/info/configs.js
@@ -1495,11 +1495,10 @@ App.MainServiceInfoConfigsController = Em.Controller.extend({
   getOozieSiteObj: function (siteObj) {
     var jdbcUrl = siteObj.findProperty('name', 'oozie.service.JPAService.jdbc.url');
     var jdbcDriver = siteObj.findProperty('name', 'oozie.service.JPAService.jdbc.driver');
-
+    var oozieDbName = siteObj.findProperty('name', 'oozie.db.schema.name');
     var oozieDb = this.get('globalConfigs').findProperty('name', 'oozie_database').value;
     // oozieHost is undefined if the database is derby
     var oozieHost = (oozieDb == 'New Derby Database') ? '' : this.get('globalConfigs').findProperty('name', 'oozie_hostname').value;
-    var oozieDbName = this.get('globalConfigs').findProperty('name', 'oozie_database_name').value;
     var defaultJdbcUrl;
 
     switch (oozieDb) {

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/controllers/wizard/step8_controller.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/wizard/step8_controller.js b/ambari-web/app/controllers/wizard/step8_controller.js
index 78a9151..a1539b1 100644
--- a/ambari-web/app/controllers/wizard/step8_controller.js
+++ b/ambari-web/app/controllers/wizard/step8_controller.js
@@ -1546,8 +1546,6 @@ App.WizardStep8Controller = Em.Controller.extend({
     var globals = this.get('content.serviceConfigProperties').filterProperty('id', 'puppet var');
     if (globals.someProperty('name', 'oozie_database')) {
       var oozieDb = globals.findProperty('name', 'oozie_database');
-      var oozieHost = globals.findProperty('name', 'oozie_hostname').value;
-      var oozieDbName = globals.findProperty('name', 'oozie_database_name').value;
       if (oozieDb.value === 'New Derby Database') {
         // oozieProperties["oozie.service.JPAService.jdbc.url"] = "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true";
         oozieProperties["oozie.service.JPAService.jdbc.driver"] = "org.apache.derby.jdbc.EmbeddedDriver";
@@ -1558,9 +1556,6 @@ App.WizardStep8Controller = Em.Controller.extend({
         // oozieProperties["oozie.service.JPAService.jdbc.url"] = "jdbc:oracle:thin:@//" + oozieHost + ":1521/" + oozieDbName;
         oozieProperties["oozie.service.JPAService.jdbc.driver"] = "oracle.jdbc.driver.OracleDriver";
       }
-      //set oozie username and password
-      oozieProperties["oozie.service.JPAService.jdbc.username"] = globals.findProperty('name', 'oozie_metastore_user_name').value;
-      oozieProperties["oozie.service.JPAService.jdbc.password"] = globals.findProperty('name', 'oozie_metastore_user_passwd').value;
     }
     return {type: 'oozie-site', tag: 'version1', properties: oozieProperties};
   },

http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/88f51325/ambari-web/app/data/HDP2/config_mapping.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/config_mapping.js b/ambari-web/app/data/HDP2/config_mapping.js
index 9e824cb..0e1c7ed 100644
--- a/ambari-web/app/data/HDP2/config_mapping.js
+++ b/ambari-web/app/data/HDP2/config_mapping.js
@@ -30,27 +30,6 @@ var configs = [
     "filename": "core-site.xml"
   },
   {
-    "name": "dfs.namenode.checkpoint.dir",
-    "templateName": ["dfs_namenode_checkpoint_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "dfs.namenode.checkpoint.period",
-    "templateName": ["dfs_namenode_checkpoint_period"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "fs.checkpoint.size",
-    "templateName": ["fs_checkpoint_size"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "core-site.xml"
-  },
-  {
     "name": "hadoop.proxyuser.<foreignKey[0]>.groups",
     "templateName": ["proxyuser_group"],
     "foreignKey": ["hive_user"],
@@ -100,34 +79,6 @@ var configs = [
   },
 /**********************************************hdfs-site***************************************/
   {
-    "name": "dfs.namenode.name.dir",
-    "templateName": ["dfs_namenode_name_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "dfs.webhdfs.enabled",
-    "templateName": ["dfs_webhdfs_enabled"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "dfs.datanode.failed.volumes.tolerated",
-    "templateName": ["dfs_datanode_failed_volume_tolerated"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
-    "name": "dfs.datanode.data.dir",
-    "templateName": ["dfs_datanode_data_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
     "name": "dfs.hosts.exclude",
     "templateName": ["hadoop_conf_dir", "dfs_exclude"],
     "foreignKey": null,
@@ -135,13 +86,6 @@ var configs = [
     "filename": "hdfs-site.xml"
   },
   {
-    "name": "dfs.replication",
-    "templateName": ["dfs_replication"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
     "name": "dfs.datanode.address",
     "templateName": ["dfs_datanode_address"],
     "foreignKey": null,
@@ -163,14 +107,6 @@ var configs = [
     "filename": "hdfs-site.xml"
   },
   {
-    "name": "dfs.datanode.du.reserved",
-    "templateName": ["datanode_du_reserved"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-
-  {
     "name": "dfs.namenode.secondary.http-address",
     "templateName": ["snamenode_host"],
     "foreignKey": null,
@@ -184,13 +120,7 @@ var configs = [
     "value": "<templateName[0]>:50470",
     "filename": "hdfs-site.xml"
   },
-  {
-    "name": "dfs.datanode.data.dir.perm",
-    "templateName": ["dfs_datanode_data_dir_perm"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
+
 /**********************************************oozie-site***************************************/
   {
     "name": "oozie.base.url",
@@ -199,78 +129,9 @@ var configs = [
     "value": "http://<templateName[0]>:11000/oozie",
     "filename": "oozie-site.xml"
   },
-  {
-    "name": "oozie.service.JPAService.create.db.schema",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "false",  // this is always false
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.db.schema.name",
-    "templateName": ['oozie_database_name'],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.driver",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "org.apache.derby.jdbc.EmbeddedDriver",  // this value is overwritten in code
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.username",
-    "templateName": ['oozie_metastore_user_name'],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.password",
-    "templateName": ['oozie_metastore_user_passwd'],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
-  {
-    "name": "oozie.service.JPAService.jdbc.url",
-    "templateName": ["oozie_jdbc_connection_url"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "oozie-site.xml"
-  },
+
 /**********************************************hive-site***************************************/
   {
-    "name": "javax.jdo.option.ConnectionDriverName",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "com.mysql.jdbc.Driver",  // this value is overwritten in code
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "javax.jdo.option.ConnectionUserName",
-    "templateName": ["hive_metastore_user_name"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "javax.jdo.option.ConnectionPassword",
-    "templateName": ["hive_metastore_user_passwd"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hive-site.xml"
-  },
-  {
-    "name": "javax.jdo.option.ConnectionURL",
-    "templateName": ["hive_jdbc_connection_url"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hive-site.xml"
-  },
-  {
     "name": "hive.metastore.uris",
     "templateName": ["hivemetastore_host"],
     "foreignKey": null,
@@ -356,139 +217,6 @@ var configs = [
     "filename": "hbase-site.xml"
   },
   {
-    "name": "hbase.tmp.dir",
-    "templateName": ["hbase_tmp_dir"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.regionserver.global.memstore.upperLimit",
-    "templateName": ["regionserver_memstore_upperlimit"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hstore.blockingStoreFiles",
-    "templateName": ["hstore_blockingstorefiles"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hstore.compactionThreshold",
-    "templateName": ["hstore_compactionthreshold"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hfile.block.cache.size",
-    "templateName": ["hfile_blockcache_size"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.max.filesize",
-    "templateName": ["hstorefile_maxsize"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.regionserver.handler.count",
-    "templateName": ["regionserver_handlers"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.majorcompaction",
-    "templateName": ["hregion_majorcompaction"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.regionserver.global.memstore.lowerLimit",
-    "templateName": ["regionserver_memstore_lowerlimit"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.memstore.block.multiplier",
-    "templateName": ["hregion_blockmultiplier"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.memstore.mslab.enabled",
-    "templateName": ["regionserver_memstore_lab"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.hregion.memstore.flush.size",
-    "templateName": ["hregion_memstoreflushsize"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.client.scanner.caching",
-    "templateName": ["client_scannercaching"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.client.scanner.caching",
-    "templateName": ["client_scannercaching"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.cluster.distributed",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "true",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.zookeeper.property.clientPort",
-    "templateName": [],
-    "foreignKey": null,
-    "value": "2181",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "zookeeper.session.timeout",
-    "templateName": ["zookeeper_sessiontimeout"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "hbase.client.keyvalue.maxsize",
-    "templateName": ["hfile_max_keyvalue_size"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hbase-site.xml"
-  },
-  {
-    "name": "dfs.client.read.shortcircuit",
-    "templateName": ["hdfs_enable_shortcircuit_read"],
-    "foreignKey": null,
-    "value": "<templateName[0]>",
-    "filename": "hdfs-site.xml"
-  },
-  {
     "name": "hbase.zookeeper.quorum",
     "templateName": ["zookeeperserver_hosts"],
     "foreignKey": null,