You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by yu...@apache.org on 2013/11/15 04:52:59 UTC
[2/2] git commit: Revert "AMBARI-3755: Config Refactor: Installer
wizard should not save global configuration not used by agent. (jaimin)"
Revert "AMBARI-3755: Config Refactor: Installer wizard should not save global configuration not used by agent. (jaimin)"
This reverts commit fb60f21dac3e98a7925cdac76fc4225118cec5cb.
Project: http://git-wip-us.apache.org/repos/asf/incubator-ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-ambari/commit/05be921e
Tree: http://git-wip-us.apache.org/repos/asf/incubator-ambari/tree/05be921e
Diff: http://git-wip-us.apache.org/repos/asf/incubator-ambari/diff/05be921e
Branch: refs/heads/trunk
Commit: 05be921e3eee9f229a7954a75210731eacfa6c34
Parents: fb60f21
Author: Yusaku Sako <yu...@hortonworks.com>
Authored: Thu Nov 14 19:52:57 2013 -0800
Committer: Yusaku Sako <yu...@hortonworks.com>
Committed: Thu Nov 14 19:52:57 2013 -0800
----------------------------------------------------------------------
.../services/HDFS/configuration/hdfs-site.xml | 386 +++++++-------
.../services/HIVE/configuration/hive-site.xml | 14 +-
.../MAPREDUCE/configuration/mapred-site.xml | 505 +++++++++---------
.../services/HDFS/configuration/hdfs-site.xml | 14 +-
.../services/HIVE/configuration/hive-site.xml | 6 -
.../MAPREDUCE/configuration/mapred-site.xml | 11 +-
.../services/HDFS/configuration/hdfs-site.xml | 14 +-
.../services/HIVE/configuration/hive-site.xml | 7 -
.../services/HDFS/configuration/hdfs-site.xml | 14 +-
.../services/HIVE/configuration/hive-site.xml | 6 -
.../MAPREDUCE/configuration/mapred-site.xml | 15 +-
.../services/HDFS/configuration/hdfs-site.xml | 14 +-
.../services/HIVE/configuration/hive-site.xml | 6 -
.../MAPREDUCE/configuration/mapred-site.xml | 11 +-
.../services/HDFS/configuration/hdfs-site.xml | 14 +-
.../services/HIVE/configuration/hive-site.xml | 7 -
.../main/admin/security/add/step4.js | 6 +
.../controllers/main/admin/security/disable.js | 8 +-
.../controllers/main/service/info/configs.js | 2 +-
ambari-web/app/controllers/wizard.js | 1 -
.../app/controllers/wizard/step8_controller.js | 15 +-
ambari-web/app/data/HDP2/config_mapping.js | 15 +
ambari-web/app/data/HDP2/global_properties.js | 219 ++++++--
ambari-web/app/data/HDP2/secure_mapping.js | 8 +-
ambari-web/app/data/HDP2/site_properties.js | 14 -
ambari-web/app/data/config_mapping.js | 23 +
ambari-web/app/data/global_properties.js | 526 ++++++++++++++++++-
ambari-web/app/data/secure_mapping.js | 8 +-
ambari-web/app/data/site_properties.js | 26 -
ambari-web/app/models/service_config.js | 1 -
ambari-web/app/utils/config.js | 170 +++---
ambari-web/app/utils/helper.js | 4 -
ambari-web/app/views/wizard/controls_view.js | 2 +-
33 files changed, 1282 insertions(+), 810 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
index e5c8229..1a0b933 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/hdfs-site.xml
@@ -22,7 +22,7 @@
<configuration>
- <!-- file system properties -->
+<!-- file system properties -->
<property>
<name>dfs.name.dir</name>
@@ -49,7 +49,7 @@
<final>true</final>
</property>
- <property>
+ <property>
<name>dfs.datanode.socket.write.timeout</name>
<value>0</value>
<description>DFS Client write socket timeout</description>
@@ -66,7 +66,7 @@
<name>dfs.block.local-path-access.user</name>
<value>hbase</value>
<description>the user who is allowed to perform short
- circuit reads.
+ circuit reads.
</description>
<final>true</final>
</property>
@@ -75,11 +75,11 @@
<name>dfs.data.dir</name>
<value>/hadoop/hdfs/data</value>
<description>Determines where on the local filesystem an DFS data node
- should store its blocks. If this is a comma-delimited
- list of directories, then data will be stored in all named
- directories, typically on different devices.
- Directories that do not exist are ignored.
- </description>
+ should store its blocks. If this is a comma-delimited
+ list of directories, then data will be stored in all named
+ directories, typically on different devices.
+ Directories that do not exist are ignored.
+ </description>
<final>true</final>
</property>
@@ -87,32 +87,32 @@
<name>dfs.hosts.exclude</name>
<value>/etc/hadoop/conf/dfs.exclude</value>
<description>Names a file that contains a list of hosts that are
- not permitted to connect to the namenode. The full pathname of the
- file must be specified. If the value is empty, no hosts are
- excluded.</description>
+ not permitted to connect to the namenode. The full pathname of the
+ file must be specified. If the value is empty, no hosts are
+ excluded.</description>
</property>
<property>
<name>dfs.hosts</name>
<value>/etc/hadoop/conf/dfs.include</value>
<description>Names a file that contains a list of hosts that are
- permitted to connect to the namenode. The full pathname of the file
- must be specified. If the value is empty, all hosts are
- permitted.</description>
+ permitted to connect to the namenode. The full pathname of the file
+ must be specified. If the value is empty, all hosts are
+ permitted.</description>
</property>
<property>
<name>dfs.replication.max</name>
<value>50</value>
<description>Maximal block replication.
- </description>
+ </description>
</property>
<property>
<name>dfs.replication</name>
<value>3</value>
<description>Default block replication.
- </description>
+ </description>
</property>
<property>
@@ -125,21 +125,21 @@
<name>dfs.safemode.threshold.pct</name>
<value>1.0f</value>
<description>
- Specifies the percentage of blocks that should satisfy
- the minimal replication requirement defined by dfs.replication.min.
- Values less than or equal to 0 mean not to start in safe mode.
- Values greater than 1 will make safe mode permanent.
- </description>
+ Specifies the percentage of blocks that should satisfy
+ the minimal replication requirement defined by dfs.replication.min.
+ Values less than or equal to 0 mean not to start in safe mode.
+ Values greater than 1 will make safe mode permanent.
+ </description>
</property>
<property>
<name>dfs.balance.bandwidthPerSec</name>
<value>6250000</value>
<description>
- Specifies the maximum amount of bandwidth that each datanode
- can utilize for the balancing purpose in term of
- the number of bytes per second.
- </description>
+ Specifies the maximum amount of bandwidth that each datanode
+ can utilize for the balancing purpose in term of
+ the number of bytes per second.
+ </description>
</property>
<property>
@@ -151,23 +151,13 @@
</property>
<property>
- <name>ambari.dfs.datanode.port</name>
- <value>50010</value>
- </property>
-
- <property>
<name>dfs.datanode.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.port}</value>
- </property>
-
- <property>
- <name>ambari.dfs.datanode.http.port</name>
- <value>50075</value>
+ <value></value>
</property>
<property>
<name>dfs.datanode.http.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.http.port}</value>
+ <value></value>
</property>
<property>
@@ -179,133 +169,133 @@
<property>
<name>dfs.http.address</name>
<value>localhost:50070</value>
- <description>The name of the default file system. Either the
- literal string "local" or a host:port for NDFS.</description>
- <final>true</final>
- </property>
-
- <property>
- <name>dfs.datanode.du.reserved</name>
- <!-- cluster variant -->
- <value>1073741824</value>
- <description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
- </description>
- </property>
-
- <property>
- <name>dfs.datanode.ipc.address</name>
- <value>0.0.0.0:8010</value>
- <description>
- The datanode ipc server address and port.
- If the port is 0 then the server will start on a free port.
- </description>
- </property>
-
- <property>
- <name>dfs.blockreport.initialDelay</name>
- <value>120</value>
- <description>Delay for first block report in seconds.</description>
- </property>
-
- <property>
- <name>dfs.datanode.du.pct</name>
- <value>0.85f</value>
- <description>When calculating remaining space, only use this percentage of the real available space
- </description>
- </property>
-
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>40</value>
- <description>The number of server threads for the namenode.</description>
- </property>
-
- <property>
- <name>dfs.datanode.max.xcievers</name>
- <value>4096</value>
- <description>PRIVATE CONFIG VARIABLE</description>
- </property>
-
- <!-- Permissions configuration -->
-
- <property>
- <name>dfs.umaskmode</name>
- <value>077</value>
- <description>
- The octal umask used when creating files and directories.
- </description>
- </property>
-
- <property>
- <name>dfs.web.ugi</name>
- <!-- cluster variant -->
- <value>gopher,gopher</value>
- <description>The user account used by the web interface.
- Syntax: USERNAME,GROUP1,GROUP2, ...
- </description>
- </property>
-
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- <description>
- If "true", enable permission checking in HDFS.
- If "false", permission checking is turned off,
- but all other behavior is unchanged.
- Switching from one parameter value to the other does not change the mode,
- owner or group of files or directories.
- </description>
- </property>
-
- <property>
- <name>dfs.permissions.supergroup</name>
- <value>hdfs</value>
- <description>The name of the group of super-users.</description>
- </property>
-
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>100</value>
- <description>Added to grow Queue size so that more client connections are allowed</description>
- </property>
-
- <property>
- <name>ipc.server.max.response.size</name>
- <value>5242880</value>
- </property>
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
+<description>The name of the default file system. Either the
+literal string "local" or a host:port for NDFS.</description>
+<final>true</final>
+</property>
+
+<property>
+<name>dfs.datanode.du.reserved</name>
+<!-- cluster variant -->
+<value>1073741824</value>
+<description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
+</description>
+</property>
+
+<property>
+<name>dfs.datanode.ipc.address</name>
+<value>0.0.0.0:8010</value>
+<description>
+The datanode ipc server address and port.
+If the port is 0 then the server will start on a free port.
+</description>
+</property>
+
+<property>
+<name>dfs.blockreport.initialDelay</name>
+<value>120</value>
+<description>Delay for first block report in seconds.</description>
+</property>
+
+<property>
+<name>dfs.datanode.du.pct</name>
+<value>0.85f</value>
+<description>When calculating remaining space, only use this percentage of the real available space
+</description>
+</property>
+
+<property>
+<name>dfs.namenode.handler.count</name>
+<value>40</value>
+<description>The number of server threads for the namenode.</description>
+</property>
+
+<property>
+<name>dfs.datanode.max.xcievers</name>
+<value>4096</value>
+<description>PRIVATE CONFIG VARIABLE</description>
+</property>
+
+<!-- Permissions configuration -->
+
+<property>
+<name>dfs.umaskmode</name>
+<value>077</value>
+<description>
+The octal umask used when creating files and directories.
+</description>
+</property>
+
+<property>
+<name>dfs.web.ugi</name>
+<!-- cluster variant -->
+<value>gopher,gopher</value>
+<description>The user account used by the web interface.
+Syntax: USERNAME,GROUP1,GROUP2, ...
+</description>
+</property>
+
+<property>
+<name>dfs.permissions</name>
+<value>true</value>
+<description>
+If "true", enable permission checking in HDFS.
+If "false", permission checking is turned off,
+but all other behavior is unchanged.
+Switching from one parameter value to the other does not change the mode,
+owner or group of files or directories.
+</description>
+</property>
+
+<property>
+<name>dfs.permissions.supergroup</name>
+<value>hdfs</value>
+<description>The name of the group of super-users.</description>
+</property>
+
+<property>
+<name>dfs.namenode.handler.count</name>
+<value>100</value>
+<description>Added to grow Queue size so that more client connections are allowed</description>
+</property>
+
+<property>
+<name>ipc.server.max.response.size</name>
+<value>5242880</value>
+</property>
+<property>
+<name>dfs.block.access.token.enable</name>
+<value>true</value>
+<description>
+If "true", access tokens are used as capabilities for accessing datanodes.
+If "false", no access tokens are checked on accessing datanodes.
+</description>
+</property>
+
+<property>
+<name>dfs.namenode.kerberos.principal</name>
+<value></value>
+<description>
+Kerberos principal name for the NameNode
+</description>
+</property>
+
+<property>
+<name>dfs.secondary.namenode.kerberos.principal</name>
+<value></value>
<description>
- If "true", access tokens are used as capabilities for accessing datanodes.
- If "false", no access tokens are checked on accessing datanodes.
+ Kerberos principal name for the secondary NameNode.
</description>
</property>
- <property>
- <name>dfs.namenode.kerberos.principal</name>
- <value></value>
- <description>
- Kerberos principal name for the NameNode
- </description>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.principal</name>
- <value></value>
- <description>
- Kerberos principal name for the secondary NameNode.
- </description>
- </property>
-
-
- <!--
- This is KRB DOMAIN specific. The FQDN of the namenode has to be mentioned.
- -->
+<!--
+ This is KRB DOMAIN specific. The FQDN of the namenode has to be mentioned.
+-->
<property>
<name>dfs.namenode.kerberos.https.principal</name>
<value></value>
- <description>The Kerberos principal for the host that the NameNode runs on.</description>
+ <description>The Kerberos principal for the host that the NameNode runs on.</description>
</property>
@@ -351,84 +341,84 @@
<property>
<name>dfs.datanode.kerberos.principal</name>
<value></value>
- <description>
- The Kerberos principal that the DataNode runs as. "_HOST" is replaced by the real host name.
+ <description>
+ The Kerberos principal that the DataNode runs as. "_HOST" is replaced by the real host name.
</description>
</property>
<property>
<name>dfs.namenode.keytab.file</name>
<value></value>
- <description>
- Combined keytab file containing the namenode service and host principals.
+ <description>
+ Combined keytab file containing the namenode service and host principals.
</description>
</property>
<property>
<name>dfs.secondary.namenode.keytab.file</name>
<value></value>
- <description>
- Combined keytab file containing the namenode service and host principals.
+ <description>
+ Combined keytab file containing the namenode service and host principals.
</description>
</property>
<property>
<name>dfs.datanode.keytab.file</name>
<value></value>
- <description>
- The filename of the keytab file for the DataNode.
+ <description>
+ The filename of the keytab file for the DataNode.
</description>
</property>
<property>
<name>dfs.https.port</name>
<value>50470</value>
- <description>The https port where namenode binds</description>
+ <description>The https port where namenode binds</description>
</property>
<property>
<name>dfs.https.address</name>
<value>localhost:50470</value>
- <description>The https address where namenode binds</description>
+ <description>The https address where namenode binds</description>
</property>
<property>
<name>dfs.datanode.data.dir.perm</name>
<value>750</value>
- <description>The permissions that should be there on dfs.data.dir
- directories. The datanode will not come up if the permissions are
- different on existing dfs.data.dir directories. If the directories
- don't exist, they will be created with this permission.</description>
- </property>
-
- <property>
- <name>dfs.access.time.precision</name>
- <value>0</value>
- <description>The access time for HDFS file is precise upto this value.
- The default value is 1 hour. Setting a value of 0 disables
- access times for HDFS.
- </description>
- </property>
-
- <property>
- <name>dfs.cluster.administrators</name>
- <value> hdfs</value>
- <description>ACL for who all can view the default servlets in the HDFS</description>
- </property>
-
- <property>
- <name>ipc.server.read.threadpool.size</name>
- <value>5</value>
- <description></description>
- </property>
-
- <property>
- <name>dfs.datanode.failed.volumes.tolerated</name>
- <value>0</value>
- <description>Number of failed disks datanode would tolerate</description>
- </property>
+<description>The permissions that should be there on dfs.data.dir
+directories. The datanode will not come up if the permissions are
+different on existing dfs.data.dir directories. If the directories
+don't exist, they will be created with this permission.</description>
+ </property>
+
+ <property>
+ <name>dfs.access.time.precision</name>
+ <value>0</value>
+ <description>The access time for HDFS file is precise upto this value.
+ The default value is 1 hour. Setting a value of 0 disables
+ access times for HDFS.
+ </description>
+</property>
+
+<property>
+ <name>dfs.cluster.administrators</name>
+ <value> hdfs</value>
+ <description>ACL for who all can view the default servlets in the HDFS</description>
+</property>
+
+<property>
+ <name>ipc.server.read.threadpool.size</name>
+ <value>5</value>
+ <description></description>
+</property>
+
+<property>
+ <name>dfs.datanode.failed.volumes.tolerated</name>
+ <value>0</value>
+ <description>Number of failed disks datanode would tolerate</description>
+</property>
<property>
<name>dfs.namenode.avoid.read.stale.datanode</name>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
index 29ed54e..9d0b168 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HIVE/configuration/hive-site.xml
@@ -19,12 +19,6 @@ limitations under the License.
<configuration>
<property>
- <name>ambari.hive.db.schema.name</name>
- <value>hive</value>
- <description>Database name used as the Hive Metastore</description>
- </property>
-
- <property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc</value>
<description>JDBC connect string for a JDBC metastore</description>
@@ -58,21 +52,21 @@ limitations under the License.
<name>hive.metastore.sasl.enabled</name>
<value></value>
<description>If true, the metastore thrift interface will be secured with SASL.
- Clients must authenticate with Kerberos.</description>
+ Clients must authenticate with Kerberos.</description>
</property>
<property>
<name>hive.metastore.kerberos.keytab.file</name>
<value></value>
<description>The path to the Kerberos Keytab file containing the metastore
- thrift server's service principal.</description>
+ thrift server's service principal.</description>
</property>
<property>
<name>hive.metastore.kerberos.principal</name>
<value></value>
<description>The service principal for the metastore thrift server. The special
- string _HOST will be replaced automatically with the correct host name.</description>
+ string _HOST will be replaced automatically with the correct host name.</description>
</property>
<property>
@@ -115,7 +109,7 @@ limitations under the License.
<name>hive.security.authorization.manager</name>
<value>org.apache.hcatalog.security.HdfsAuthorizationProvider</value>
<description>the hive client authorization manager class name.
- The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider. </description>
+ The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider. </description>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
index f386cf6..9ef7eaa 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
@@ -22,7 +22,7 @@
<configuration xmlns:xi="http://www.w3.org/2001/XInclude">
- <!-- i/o properties -->
+<!-- i/o properties -->
<property>
<name>io.sort.mb</name>
@@ -50,25 +50,25 @@
<description>No description</description>
</property>
- <!-- map/reduce properties -->
+<!-- map/reduce properties -->
- <property>
- <name>mapred.tasktracker.tasks.sleeptime-before-sigkill</name>
- <value>250</value>
- <description>Normally, this is the amount of time before killing
- processes, and the recommended-default is 5.000 seconds - a value of
- 5000 here. In this case, we are using it solely to blast tasks before
- killing them, and killing them very quickly (1/4 second) to guarantee
- that we do not leave VMs around for later jobs.
- </description>
- </property>
+<property>
+ <name>mapred.tasktracker.tasks.sleeptime-before-sigkill</name>
+ <value>250</value>
+ <description>Normally, this is the amount of time before killing
+ processes, and the recommended-default is 5.000 seconds - a value of
+ 5000 here. In this case, we are using it solely to blast tasks before
+ killing them, and killing them very quickly (1/4 second) to guarantee
+ that we do not leave VMs around for later jobs.
+ </description>
+</property>
<property>
<name>mapred.job.tracker.handler.count</name>
<value>50</value>
<description>
- The number of server threads for the JobTracker. This should be roughly
- 4% of the number of tasktracker nodes.
+ The number of server threads for the JobTracker. This should be roughly
+ 4% of the number of tasktracker nodes.
</description>
</property>
@@ -91,10 +91,11 @@
<name>mapred.job.tracker.http.address</name>
<!-- cluster variant -->
<value>localhost:50030</value>
- <description>JobTracker host and http port address</description>
+ <description>Http address for JobTracker</description>
<final>true</final>
</property>
+
<property>
<!-- cluster specific -->
<name>mapred.local.dir</name>
@@ -104,8 +105,8 @@
</property>
<property>
- <name>mapreduce.cluster.administrators</name>
- <value> hadoop</value>
+ <name>mapreduce.cluster.administrators</name>
+ <value> hadoop</value>
</property>
<property>
@@ -135,14 +136,14 @@
<name>mapred.map.tasks.speculative.execution</name>
<value>false</value>
<description>If true, then multiple instances of some map tasks
- may be executed in parallel.</description>
+ may be executed in parallel.</description>
</property>
<property>
<name>mapred.reduce.tasks.speculative.execution</name>
<value>false</value>
<description>If true, then multiple instances of some reduce tasks
- may be executed in parallel.</description>
+ may be executed in parallel.</description>
</property>
<property>
@@ -154,29 +155,29 @@
<name>mapred.inmem.merge.threshold</name>
<value>1000</value>
<description>The threshold, in terms of the number of files
- for the in-memory merge process. When we accumulate threshold number of files
- we initiate the in-memory merge and spill to disk. A value of 0 or less than
- 0 indicates we want to DON'T have any threshold and instead depend only on
- the ramfs's memory consumption to trigger the merge.
- </description>
+ for the in-memory merge process. When we accumulate threshold number of files
+ we initiate the in-memory merge and spill to disk. A value of 0 or less than
+ 0 indicates we want to DON'T have any threshold and instead depend only on
+ the ramfs's memory consumption to trigger the merge.
+ </description>
</property>
<property>
<name>mapred.job.shuffle.merge.percent</name>
<value>0.66</value>
<description>The usage threshold at which an in-memory merge will be
- initiated, expressed as a percentage of the total memory allocated to
- storing in-memory map outputs, as defined by
- mapred.job.shuffle.input.buffer.percent.
- </description>
+ initiated, expressed as a percentage of the total memory allocated to
+ storing in-memory map outputs, as defined by
+ mapred.job.shuffle.input.buffer.percent.
+ </description>
</property>
<property>
<name>mapred.job.shuffle.input.buffer.percent</name>
<value>0.7</value>
<description>The percentage of memory to be allocated from the maximum heap
- size to storing map outputs during the shuffle.
- </description>
+ size to storing map outputs during the shuffle.
+ </description>
</property>
<property>
@@ -187,13 +188,13 @@
</description>
</property>
- <property>
- <name>mapred.output.compression.type</name>
- <value>BLOCK</value>
- <description>If the job outputs are to compressed as SequenceFiles, how should
- they be compressed? Should be one of NONE, RECORD or BLOCK.
- </description>
- </property>
+<property>
+ <name>mapred.output.compression.type</name>
+ <value>BLOCK</value>
+ <description>If the job outputs are to compressed as SequenceFiles, how should
+ they be compressed? Should be one of NONE, RECORD or BLOCK.
+ </description>
+</property>
<property>
@@ -210,7 +211,7 @@
<name>mapred.jobtracker.restart.recover</name>
<value>false</value>
<description>"true" to enable (job) recovery upon restart,
- "false" to start afresh
+ "false" to start afresh
</description>
</property>
@@ -218,20 +219,20 @@
<name>mapred.job.reduce.input.buffer.percent</name>
<value>0.0</value>
<description>The percentage of memory- relative to the maximum heap size- to
- retain map outputs during the reduce. When the shuffle is concluded, any
- remaining map outputs in memory must consume less than this threshold before
- the reduce can begin.
- </description>
+ retain map outputs during the reduce. When the shuffle is concluded, any
+ remaining map outputs in memory must consume less than this threshold before
+ the reduce can begin.
+ </description>
</property>
- <property>
- <name>mapreduce.reduce.input.limit</name>
- <value>10737418240</value>
- <description>The limit on the input size of the reduce. (This value
- is 10 Gb.) If the estimated input size of the reduce is greater than
- this value, job is failed. A value of -1 means that there is no limit
- set. </description>
- </property>
+ <property>
+ <name>mapreduce.reduce.input.limit</name>
+ <value>10737418240</value>
+ <description>The limit on the input size of the reduce. (This value
+ is 10 Gb.) If the estimated input size of the reduce is greater than
+ this value, job is failed. A value of -1 means that there is no limit
+ set. </description>
+</property>
<!-- copied from kryptonite configuration -->
@@ -245,9 +246,9 @@
<name>mapred.task.timeout</name>
<value>600000</value>
<description>The number of milliseconds before a task will be
- terminated if it neither reads an input, writes an output, nor
- updates its status string.
- </description>
+ terminated if it neither reads an input, writes an output, nor
+ updates its status string.
+ </description>
</property>
<property>
@@ -259,9 +260,9 @@
<property>
<name>mapred.task.tracker.task-controller</name>
<value>org.apache.hadoop.mapred.DefaultTaskController</value>
- <description>
- TaskController which is used to launch and manage task execution.
- </description>
+ <description>
+ TaskController which is used to launch and manage task execution.
+ </description>
</property>
<property>
@@ -270,16 +271,10 @@
</property>
<property>
- <name>ambari.mapred.child.java.opts.memory</name>
- <value>768</value>
-
- <description>Java options Memory for the TaskTracker child processes</description>
- </property>
-
- <property>
<name>mapred.child.java.opts</name>
- <value>-server -Xmx${ambari.mapred.child.java.opts.memory}m -Djava.net.preferIPv4Stack=true</value>
- <description>Java options for the TaskTracker child processes</description>
+ <value></value>
+
+ <description>No description</description>
</property>
<property>
@@ -294,7 +289,7 @@
<name>mapred.cluster.reduce.memory.mb</name>
<value>2048</value>
<description>
- The virtual memory size of a single Reduce slot in the MapReduce framework
+ The virtual memory size of a single Reduce slot in the MapReduce framework
</description>
</property>
@@ -340,137 +335,137 @@
</description>
</property>
- <property>
- <name>mapred.hosts.exclude</name>
- <value>/etc/hadoop/conf/mapred.exclude</value>
- <description>
- Names a file that contains the list of hosts that
- should be excluded by the jobtracker. If the value is empty, no
- hosts are excluded.
- </description>
- </property>
-
- <property>
- <name>mapred.max.tracker.blacklists</name>
- <value>16</value>
- <description>
- if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted
- </description>
- </property>
-
- <property>
- <name>mapred.healthChecker.script.path</name>
- <value>file:////mapred/jobstatus</value>
- <description>
- Directory path to view job status
- </description>
- </property>
-
- <property>
- <name>mapred.healthChecker.interval</name>
- <value>135000</value>
- </property>
-
- <property>
- <name>mapred.healthChecker.script.timeout</name>
- <value>60000</value>
- </property>
-
- <property>
- <name>mapred.job.tracker.persist.jobstatus.active</name>
- <value>false</value>
- <description>Indicates if persistency of job status information is
- active or not.
- </description>
- </property>
-
- <property>
- <name>mapred.job.tracker.persist.jobstatus.hours</name>
- <value>1</value>
- <description>The number of hours job status information is persisted in DFS.
- The job status information will be available after it drops of the memory
- queue and between jobtracker restarts. With a zero value the job status
- information is not persisted at all in DFS.
- </description>
- </property>
-
- <property>
- <name>mapred.job.tracker.persist.jobstatus.dir</name>
- <value>/etc/hadoop/conf/health_check</value>
- <description>The directory where the job status information is persisted
- in a file system to be available after it drops of the memory queue and
- between jobtracker restarts.
- </description>
- </property>
-
- <property>
- <name>mapred.jobtracker.retirejob.check</name>
- <value>10000</value>
- </property>
-
- <property>
- <name>mapred.jobtracker.retirejob.interval</name>
- <value>21600000</value>
- </property>
-
- <property>
- <name>mapred.job.tracker.history.completed.location</name>
- <value>/mapred/history/done</value>
- <description>No description</description>
- </property>
-
- <property>
- <name>mapred.task.maxvmem</name>
- <value></value>
- <final>true</final>
- <description>No description</description>
- </property>
-
- <property>
- <name>mapred.jobtracker.maxtasks.per.job</name>
- <value>-1</value>
- <final>true</final>
- <description>The maximum number of tasks for a single job.
- A value of -1 indicates that there is no maximum. </description>
- </property>
-
- <property>
- <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name>
- <value>false</value>
- </property>
-
- <property>
- <name>mapred.userlog.retain.hours</name>
- <value>24</value>
- <description>
- The maximum time, in hours, for which the user-logs are to be retained after the job completion.
- </description>
- </property>
-
- <property>
- <name>mapred.job.reuse.jvm.num.tasks</name>
- <value>1</value>
- <description>
- How many tasks to run per jvm. If set to -1, there is no limit
- </description>
- <final>true</final>
- </property>
-
- <property>
- <name>mapreduce.jobtracker.kerberos.principal</name>
- <value></value>
- <description>
+<property>
+ <name>mapred.hosts.exclude</name>
+ <value>/etc/hadoop/conf/mapred.exclude</value>
+ <description>
+ Names a file that contains the list of hosts that
+ should be excluded by the jobtracker. If the value is empty, no
+ hosts are excluded.
+ </description>
+</property>
+
+<property>
+ <name>mapred.max.tracker.blacklists</name>
+ <value>16</value>
+ <description>
+ if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted
+ </description>
+</property>
+
+<property>
+ <name>mapred.healthChecker.script.path</name>
+ <value>file:////mapred/jobstatus</value>
+ <description>
+ Directory path to view job status
+ </description>
+</property>
+
+<property>
+ <name>mapred.healthChecker.interval</name>
+ <value>135000</value>
+</property>
+
+<property>
+ <name>mapred.healthChecker.script.timeout</name>
+ <value>60000</value>
+</property>
+
+<property>
+ <name>mapred.job.tracker.persist.jobstatus.active</name>
+ <value>false</value>
+ <description>Indicates if persistency of job status information is
+ active or not.
+ </description>
+</property>
+
+<property>
+ <name>mapred.job.tracker.persist.jobstatus.hours</name>
+ <value>1</value>
+ <description>The number of hours job status information is persisted in DFS.
+ The job status information will be available after it drops of the memory
+ queue and between jobtracker restarts. With a zero value the job status
+ information is not persisted at all in DFS.
+ </description>
+</property>
+
+<property>
+ <name>mapred.job.tracker.persist.jobstatus.dir</name>
+ <value>/etc/hadoop/conf/health_check</value>
+ <description>The directory where the job status information is persisted
+ in a file system to be available after it drops of the memory queue and
+ between jobtracker restarts.
+ </description>
+</property>
+
+<property>
+ <name>mapred.jobtracker.retirejob.check</name>
+ <value>10000</value>
+</property>
+
+<property>
+ <name>mapred.jobtracker.retirejob.interval</name>
+ <value>21600000</value>
+</property>
+
+<property>
+ <name>mapred.job.tracker.history.completed.location</name>
+ <value>/mapred/history/done</value>
+ <description>No description</description>
+</property>
+
+<property>
+ <name>mapred.task.maxvmem</name>
+ <value></value>
+ <final>true</final>
+ <description>No description</description>
+</property>
+
+<property>
+ <name>mapred.jobtracker.maxtasks.per.job</name>
+ <value>-1</value>
+ <final>true</final>
+ <description>The maximum number of tasks for a single job.
+ A value of -1 indicates that there is no maximum. </description>
+</property>
+
+<property>
+ <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name>
+ <value>false</value>
+</property>
+
+<property>
+ <name>mapred.userlog.retain.hours</name>
+ <value>24</value>
+ <description>
+ The maximum time, in hours, for which the user-logs are to be retained after the job completion.
+ </description>
+</property>
+
+<property>
+ <name>mapred.job.reuse.jvm.num.tasks</name>
+ <value>1</value>
+ <description>
+ How many tasks to run per jvm. If set to -1, there is no limit
+ </description>
+ <final>true</final>
+</property>
+
+<property>
+ <name>mapreduce.jobtracker.kerberos.principal</name>
+ <value></value>
+ <description>
JT user name key.
- </description>
- </property>
+ </description>
+</property>
- <property>
- <name>mapreduce.tasktracker.kerberos.principal</name>
- <value></value>
- <description>
- tt user name key. "_HOST" is replaced by the host name of the task tracker.
- </description>
- </property>
+<property>
+ <name>mapreduce.tasktracker.kerberos.principal</name>
+ <value></value>
+ <description>
+ tt user name key. "_HOST" is replaced by the host name of the task tracker.
+ </description>
+</property>
<property>
@@ -480,54 +475,54 @@
</property>
- <property>
- <name>mapreduce.jobtracker.keytab.file</name>
- <value></value>
- <description>
- The keytab for the jobtracker principal.
- </description>
+ <property>
+ <name>mapreduce.jobtracker.keytab.file</name>
+ <value></value>
+ <description>
+ The keytab for the jobtracker principal.
+ </description>
- </property>
+</property>
- <property>
- <name>mapreduce.tasktracker.keytab.file</name>
- <value></value>
+ <property>
+ <name>mapreduce.tasktracker.keytab.file</name>
+ <value></value>
<description>The filename of the keytab for the task tracker</description>
- </property>
+ </property>
- <property>
- <name>mapred.task.tracker.http.address</name>
- <value></value>
- <description>Http address for task tracker.</description>
- </property>
+ <property>
+ <name>mapred.task.tracker.http.address</name>
+ <value></value>
+ <description>Http address for task tracker.</description>
+ </property>
- <property>
- <name>mapreduce.jobtracker.staging.root.dir</name>
- <value>/user</value>
- <description>The Path prefix for where the staging directories should be placed. The next level is always the user's
- name. It is a path in the default file system.</description>
- </property>
+ <property>
+ <name>mapreduce.jobtracker.staging.root.dir</name>
+ <value>/user</value>
+ <description>The Path prefix for where the staging directories should be placed. The next level is always the user's
+ name. It is a path in the default file system.</description>
+ </property>
- <property>
- <name>mapreduce.tasktracker.group</name>
- <value>hadoop</value>
- <description>The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.</description>
+ <property>
+ <name>mapreduce.tasktracker.group</name>
+ <value>hadoop</value>
+ <description>The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.</description>
- </property>
+ </property>
<property>
<name>mapreduce.jobtracker.split.metainfo.maxsize</name>
<value>50000000</value>
<final>true</final>
- <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during
- initialize.
- </description>
+ <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during
+ initialize.
+ </description>
</property>
<property>
<name>mapreduce.history.server.embedded</name>
<value>false</value>
<description>Should job history server be embedded within Job tracker
- process</description>
+process</description>
<final>true</final>
</property>
@@ -542,38 +537,38 @@
<property>
<name>mapreduce.jobhistory.kerberos.principal</name>
<!-- cluster variant -->
- <value></value>
+ <value></value>
<description>Job history user name key. (must map to same user as JT
- user)</description>
+user)</description>
</property>
- <property>
- <name>mapreduce.jobhistory.keytab.file</name>
+ <property>
+ <name>mapreduce.jobhistory.keytab.file</name>
<!-- cluster variant -->
- <value></value>
- <description>The keytab for the job history server principal.</description>
- </property>
-
- <property>
- <name>mapred.jobtracker.blacklist.fault-timeout-window</name>
- <value>180</value>
- <description>
- 3-hour sliding window (value is in minutes)
- </description>
- </property>
-
- <property>
- <name>mapred.jobtracker.blacklist.fault-bucket-width</name>
- <value>15</value>
- <description>
- 15-minute bucket size (value is in minutes)
- </description>
- </property>
-
- <property>
- <name>mapred.queue.names</name>
- <value>default</value>
- <description> Comma separated list of queues configured for this jobtracker.</description>
- </property>
+ <value></value>
+ <description>The keytab for the job history server principal.</description>
+ </property>
+
+<property>
+ <name>mapred.jobtracker.blacklist.fault-timeout-window</name>
+ <value>180</value>
+ <description>
+ 3-hour sliding window (value is in minutes)
+ </description>
+</property>
+
+<property>
+ <name>mapred.jobtracker.blacklist.fault-bucket-width</name>
+ <value>15</value>
+ <description>
+ 15-minute bucket size (value is in minutes)
+ </description>
+</property>
+
+<property>
+ <name>mapred.queue.names</name>
+ <value>default</value>
+ <description> Comma separated list of queues configured for this jobtracker.</description>
+</property>
</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HDFS/configuration/hdfs-site.xml
index 758ea6d..1a0b933 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HDFS/configuration/hdfs-site.xml
@@ -151,23 +151,13 @@
</property>
<property>
- <name>ambari.dfs.datanode.port</name>
- <value>50010</value>
- </property>
-
- <property>
<name>dfs.datanode.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.port}</value>
- </property>
-
- <property>
- <name>ambari.dfs.datanode.http.port</name>
- <value>50075</value>
+ <value></value>
</property>
<property>
<name>dfs.datanode.http.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.http.port}</value>
+ <value></value>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HIVE/configuration/hive-site.xml
index 24de30b..9d0b168 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/HIVE/configuration/hive-site.xml
@@ -19,12 +19,6 @@ limitations under the License.
<configuration>
<property>
- <name>ambari.hive.db.schema.name</name>
- <value>hive</value>
- <description>Database name used as the Hive Metastore</description>
- </property>
-
- <property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc</value>
<description>JDBC connect string for a JDBC metastore</description>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
index 732be1b..567c1e1 100644
--- a/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
@@ -270,17 +270,10 @@
</property>
<property>
- <name>ambari.mapred.child.java.opts.memory</name>
- <value>768</value>
-
- <description>Java options Memory for the TaskTracker child processes</description>
- </property>
-
- <property>
<name>mapred.child.java.opts</name>
- <value>-server -Xmx${ambari.mapred.child.java.opts.memory}m -Djava.net.preferIPv4Stack=true</value>
+ <value></value>
- <description>Java options for the TaskTracker child processes</description>
+ <description>No description</description>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
index 19fd655..bd3b12a 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/hdfs-site.xml
@@ -181,23 +181,13 @@
</property>
<property>
- <name>ambari.dfs.datanode.port</name>
- <value>50010</value>
- </property>
-
- <property>
<name>dfs.datanode.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.port}</value>
- </property>
-
- <property>
- <name>ambari.dfs.datanode.http.port</name>
- <value>50075</value>
+ <value>0.0.0.0:50010</value>
</property>
<property>
<name>dfs.datanode.http.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.http.port}</value>
+ <value>0.0.0.0:50075</value>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
index b3ed5f5..7fb8969 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HIVE/configuration/hive-site.xml
@@ -18,13 +18,6 @@ limitations under the License.
-->
<configuration>
-
- <property>
- <name>ambari.hive.db.schema.name</name>
- <value>hive</value>
- <description>Database name used as the Hive Metastore</description>
- </property>
-
<property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc</value>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
index e5c8229..15666e0 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HDFS/configuration/hdfs-site.xml
@@ -151,23 +151,13 @@
</property>
<property>
- <name>ambari.dfs.datanode.port</name>
- <value>50010</value>
- </property>
-
- <property>
<name>dfs.datanode.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.port}</value>
- </property>
-
- <property>
- <name>ambari.dfs.datanode.http.port</name>
- <value>50075</value>
+ <value></value>
</property>
<property>
<name>dfs.datanode.http.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.http.port}</value>
+ <value></value>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
index 29ed54e..1337fa4 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/HIVE/configuration/hive-site.xml
@@ -19,12 +19,6 @@ limitations under the License.
<configuration>
<property>
- <name>ambari.hive.db.schema.name</name>
- <value>hive</value>
- <description>Database name used as the Hive Metastore</description>
- </property>
-
- <property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc</value>
<description>JDBC connect string for a JDBC metastore</description>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
index f386cf6..743ca6a 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.2/services/MAPREDUCE/configuration/mapred-site.xml
@@ -91,10 +91,11 @@
<name>mapred.job.tracker.http.address</name>
<!-- cluster variant -->
<value>localhost:50030</value>
- <description>JobTracker host and http port address</description>
+ <description>Http address for JobTracker</description>
<final>true</final>
</property>
+
<property>
<!-- cluster specific -->
<name>mapred.local.dir</name>
@@ -270,16 +271,10 @@
</property>
<property>
- <name>ambari.mapred.child.java.opts.memory</name>
- <value>768</value>
-
- <description>Java options Memory for the TaskTracker child processes</description>
- </property>
-
- <property>
<name>mapred.child.java.opts</name>
- <value>-server -Xmx${ambari.mapred.child.java.opts.memory}m -Djava.net.preferIPv4Stack=true</value>
- <description>Java options for the TaskTracker child processes</description>
+ <value></value>
+
+ <description>No description</description>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HDFS/configuration/hdfs-site.xml
index e5c8229..15666e0 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HDFS/configuration/hdfs-site.xml
@@ -151,23 +151,13 @@
</property>
<property>
- <name>ambari.dfs.datanode.port</name>
- <value>50010</value>
- </property>
-
- <property>
<name>dfs.datanode.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.port}</value>
- </property>
-
- <property>
- <name>ambari.dfs.datanode.http.port</name>
- <value>50075</value>
+ <value></value>
</property>
<property>
<name>dfs.datanode.http.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.http.port}</value>
+ <value></value>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HIVE/configuration/hive-site.xml
index 29ed54e..1337fa4 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/HIVE/configuration/hive-site.xml
@@ -19,12 +19,6 @@ limitations under the License.
<configuration>
<property>
- <name>ambari.hive.db.schema.name</name>
- <value>hive</value>
- <description>Database name used as the Hive Metastore</description>
- </property>
-
- <property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc</value>
<description>JDBC connect string for a JDBC metastore</description>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
index a1ea9de..56eeff5 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/1.3.3/services/MAPREDUCE/configuration/mapred-site.xml
@@ -270,17 +270,10 @@
</property>
<property>
- <name>ambari.mapred.child.java.opts.memory</name>
- <value>768</value>
-
- <description>Java options Memory for the TaskTracker child processes</description>
- </property>
-
- <property>
<name>mapred.child.java.opts</name>
- <value>-server -Xmx${ambari.mapred.child.java.opts.memory}m -Djava.net.preferIPv4Stack=true</value>
+ <value></value>
- <description>Java options for the TaskTracker child processes</description>
+ <description>No description</description>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
index 6694866..d1f271e 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HDFS/configuration/hdfs-site.xml
@@ -181,23 +181,13 @@
</property>
<property>
- <name>ambari.dfs.datanode.port</name>
- <value>50010</value>
- </property>
-
- <property>
<name>dfs.datanode.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.port}</value>
- </property>
-
- <property>
- <name>ambari.dfs.datanode.http.port</name>
- <value>50075</value>
+ <value>0.0.0.0:50010</value>
</property>
<property>
<name>dfs.datanode.http.address</name>
- <value>0.0.0.0:${ambari.dfs.datanode.http.port}</value>
+ <value>0.0.0.0:50075</value>
</property>
<property>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
index 17571f6..2a3f5d4 100644
--- a/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
+++ b/ambari-server/src/main/resources/stacks/HDPLocal/2.0.6/services/HIVE/configuration/hive-site.xml
@@ -18,13 +18,6 @@ limitations under the License.
-->
<configuration>
-
- <property>
- <name>ambari.hive.db.schema.name</name>
- <value>hive</value>
- <description>Database name used as the Hive Metastore</description>
- </property>
-
<property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc</value>
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/controllers/main/admin/security/add/step4.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/main/admin/security/add/step4.js b/ambari-web/app/controllers/main/admin/security/add/step4.js
index 479e228..d62fd1d 100644
--- a/ambari-web/app/controllers/main/admin/security/add/step4.js
+++ b/ambari-web/app/controllers/main/admin/security/add/step4.js
@@ -418,6 +418,12 @@ App.MainAdminSecurityAddStep4Controller = Em.Controller.extend({
case 'security_enabled':
_property.value = 'true';
break;
+ case 'dfs_datanode_address':
+ _property.value = '1019';
+ break;
+ case 'dfs_datanode_http_address':
+ _property.value = '1022';
+ break;
}
}, this);
},
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/controllers/main/admin/security/disable.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/main/admin/security/disable.js b/ambari-web/app/controllers/main/admin/security/disable.js
index 618a822..d15265d 100644
--- a/ambari-web/app/controllers/main/admin/security/disable.js
+++ b/ambari-web/app/controllers/main/admin/security/disable.js
@@ -387,11 +387,11 @@ App.MainAdminSecurityDisableController = Em.Controller.extend({
var configName = _config.name;
if (configName in _serviceConfigTags.configs) {
switch (configName) {
- case 'ambari.dfs.datanode.port':
- _serviceConfigTags.configs[configName] = '50010';
+ case 'dfs.datanode.address':
+ _serviceConfigTags.configs[configName] = '0.0.0.0:50010';
break;
- case 'ambari.dfs.datanode.http.port':
- _serviceConfigTags.configs[configName] = '50075';
+ case 'dfs.datanode.http.address':
+ _serviceConfigTags.configs[configName] = '0.0.0.0:50075';
break;
case 'mapred.task.tracker.task-controller':
_serviceConfigTags.configs[configName] = 'org.apache.hadoop.mapred.DefaultTaskController';
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/controllers/main/service/info/configs.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/main/service/info/configs.js b/ambari-web/app/controllers/main/service/info/configs.js
index 6330c6f..8d1bbb5 100644
--- a/ambari-web/app/controllers/main/service/info/configs.js
+++ b/ambari-web/app/controllers/main/service/info/configs.js
@@ -1378,7 +1378,7 @@ App.MainServiceInfoConfigsController = Em.Controller.extend({
var globalSiteProperties = {};
this.get('globalConfigs').forEach(function (_globalSiteObj) {
// do not pass any globalConfigs whose name ends with _host or _hosts
- if (_globalSiteObj.isRequiredByAgent === true) {
+ if (!/_hosts?$/.test(_globalSiteObj.name)) {
// append "m" to JVM memory options except for hadoop_heapsize
if (/_heapsize|_newsize|_maxnewsize$/.test(_globalSiteObj.name) && !heapsizeException.contains(_globalSiteObj.name)) {
_globalSiteObj.value += "m";
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/controllers/wizard.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/wizard.js b/ambari-web/app/controllers/wizard.js
index b2774fc..c22158e 100644
--- a/ambari-web/app/controllers/wizard.js
+++ b/ambari-web/app/controllers/wizard.js
@@ -736,7 +736,6 @@ App.WizardController = Em.Controller.extend({
domain: _configProperties.get('domain'),
filename: _configProperties.get('filename'),
displayType: _configProperties.get('displayType'),
- isRequiredByAgent: _configProperties.get('isRequiredByAgent'),
overrides: overridesArray
};
serviceConfigProperties.push(configProperty);
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/controllers/wizard/step8_controller.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/controllers/wizard/step8_controller.js b/ambari-web/app/controllers/wizard/step8_controller.js
index 956d4fb..ca17bb2 100644
--- a/ambari-web/app/controllers/wizard/step8_controller.js
+++ b/ambari-web/app/controllers/wizard/step8_controller.js
@@ -1404,16 +1404,18 @@ App.WizardStep8Controller = Em.Controller.extend({
globalSiteObj.forEach(function (_globalSiteObj) {
var heapsizeException = ['hadoop_heapsize','yarn_heapsize','nodemanager_heapsize','resourcemanager_heapsize'];
// do not pass any globals whose name ends with _host or _hosts
- if (_globalSiteObj.isRequiredByAgent === true) {
+ if (!/_hosts?$/.test(_globalSiteObj.name)) {
// append "m" to JVM memory options except for hadoop_heapsize
if (/_heapsize|_newsize|_maxnewsize$/.test(_globalSiteObj.name) && !heapsizeException.contains(_globalSiteObj.name)) {
globalSiteProperties[_globalSiteObj.name] = _globalSiteObj.value + "m";
} else {
globalSiteProperties[_globalSiteObj.name] = App.config.escapeXMLCharacters(_globalSiteObj.value);
}
- if (_globalSiteObj.name == 'java64_home') {
- globalSiteProperties['java64_home'] = this.get('content.installOptions.javaHome');
- }
+ console.log("STEP8: name of the global property is: " + _globalSiteObj.name);
+ console.log("STEP8: value of the global property is: " + _globalSiteObj.value);
+ }
+ if (_globalSiteObj.name == 'java64_home') {
+ globalSiteProperties['java64_home'] = this.get('content.installOptions.javaHome');
}
this._recordHostOverrideFromObj(_globalSiteObj, 'global', 'version1', this);
}, this);
@@ -1570,11 +1572,16 @@ App.WizardStep8Controller = Em.Controller.extend({
var globals = this.get('content.serviceConfigProperties').filterProperty('id', 'puppet var');
if (globals.someProperty('name', 'hive_database')) {
var hiveDb = globals.findProperty('name', 'hive_database');
+ var hiveHost = globals.findProperty('name', 'hive_hostname').value;
+ var hiveDbName = globals.findProperty('name', 'hive_database_name').value;
if (hiveDb.value === 'New MySQL Database') {
+ // hiveProperties["javax.jdo.option.ConnectionURL"] = "jdbc:mysql://"+ hiveHost + "/" + hiveDbName + "?createDatabaseIfNotExist=true";
hiveProperties["javax.jdo.option.ConnectionDriverName"] = "com.mysql.jdbc.Driver";
} else if (hiveDb.value === 'Existing MySQL Database'){
+ // hiveProperties["javax.jdo.option.ConnectionURL"] = "jdbc:mysql://"+ hiveHost + "/" + hiveDbName + "?createDatabaseIfNotExist=true";
hiveProperties["javax.jdo.option.ConnectionDriverName"] = "com.mysql.jdbc.Driver";
} else { //existing oracle database
+ // hiveProperties["javax.jdo.option.ConnectionURL"] = "jdbc:oracle:thin:@//"+ hiveHost + ":1521/" + hiveDbName;
hiveProperties["javax.jdo.option.ConnectionDriverName"] = "oracle.jdbc.driver.OracleDriver";
}
}
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/data/HDP2/config_mapping.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/config_mapping.js b/ambari-web/app/data/HDP2/config_mapping.js
index 742d505..7f284be 100644
--- a/ambari-web/app/data/HDP2/config_mapping.js
+++ b/ambari-web/app/data/HDP2/config_mapping.js
@@ -67,6 +67,21 @@ var configs = [
"filename": "core-site.xml",
"isOverridable": true
},
+/**********************************************hdfs-site***************************************/
+ {
+ "name": "dfs.datanode.address",
+ "templateName": ["dfs_datanode_address"],
+ "foreignKey": null,
+ "value": "0.0.0.0:<templateName[0]>",
+ "filename": "hdfs-site.xml"
+ },
+ {
+ "name": "dfs.datanode.http.address",
+ "templateName": ["dfs_datanode_http_address"],
+ "foreignKey": null,
+ "value": "0.0.0.0:<templateName[0]>",
+ "filename": "hdfs-site.xml"
+ },
/**********************************************hbase-site***************************************/
{
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/data/HDP2/global_properties.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/global_properties.js b/ambari-web/app/data/HDP2/global_properties.js
index f1d26a1..c3f2f29 100644
--- a/ambari-web/app/data/HDP2/global_properties.js
+++ b/ambari-web/app/data/HDP2/global_properties.js
@@ -47,10 +47,6 @@
* If this is unspecified, true is assumed.
* E.g., true, false
*
- * isRequiredByAgent:
- * Whether the config property is required by agent or not.
- * If value is true then it will be persisted in global configuration
- *
* displayType:
* How the config property is to be rendered for user input.
* If this is left unspecified, "string" is assumed
@@ -90,7 +86,6 @@ module.exports =
"displayType": "masterHosts",
"isOverridable": false,
"isVisible": true,
- "isRequiredByAgent": false,
"domain": "global",
"serviceName": "HDFS",
"category": "NameNode",
@@ -106,7 +101,6 @@ module.exports =
"unit": "MB",
"isOverridable": false,
"isVisible": true,
- "isRequiredByAgent": true,
"domain": "global",
"serviceName": "HDFS",
"category": "NameNode",
@@ -122,7 +116,6 @@ module.exports =
"unit": "MB",
"isOverridable": false,
"isVisible": true,
- "isRequiredByAgent": true,
"domain": "global",
"serviceName": "HDFS",
"category": "NameNode",
@@ -137,7 +130,6 @@ module.exports =
"description": "The host that has been assigned to run SecondaryNameNode",
"displayType": "masterHost",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "HDFS",
@@ -155,7 +147,6 @@ module.exports =
"isRequired": false,
"isOverridable": false,
"isVisible": true,
- "isRequiredByAgent": false,
"domain": "datanode-global",
"serviceName": "HDFS",
"category": "DataNode",
@@ -170,7 +161,6 @@ module.exports =
"displayType": "int",
"unit": "MB",
"isVisible": true,
- "isRequiredByAgent": true,
"domain": "datanode-global",
"serviceName": "HDFS",
"category": "DataNode",
@@ -184,7 +174,6 @@ module.exports =
"defaultValue": "1024",
"displayType": "int",
"unit": "MB",
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "HDFS",
@@ -199,7 +188,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "HDFS",
@@ -214,7 +202,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "HDFS",
@@ -229,7 +216,6 @@ module.exports =
"displayType": "int",
"unit": "MB",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": false,
"domain": "global",
"serviceName": "HDFS",
@@ -244,7 +230,6 @@ module.exports =
"isRequired": false,
"displayType": "checkbox",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": false,
"domain": "global",
"serviceName": "HDFS",
@@ -252,6 +237,20 @@ module.exports =
},
{
"id": "puppet var",
+ "name": "namenode_formatted_mark_dir",
+ "displayName": "Hadoop formatted mark directory",
+ "description": "",
+ "defaultValue": "/var/run/hadoop/hdfs/namenode/formatted/",
+ "isRequired": false,
+ "isReconfigurable": false,
+ "displayType": "directory",
+ "isVisible": false,
+ "domain": "global",
+ "serviceName": "HDFS",
+ "category": "NameNode"
+ },
+ {
+ "id": "puppet var",
"name": "hcat_conf_dir",
"displayName": "HCat conf directory",
"description": "",
@@ -259,7 +258,6 @@ module.exports =
"isRequired": false,
"isReconfigurable": false,
"displayType": "directory",
- "isRequiredByAgent": true,
"isVisible": false,
"domain": "global",
"serviceName": "HDFS",
@@ -267,13 +265,25 @@ module.exports =
},
{
"id": "puppet var",
+ "name": "hdfs_enable_shortcircuit_read",
+ "displayName": "HDFS Short-circuit read",
+ "description": "",
+ "defaultValue": true,
+ "isRequired": false,
+ "displayType": "checkbox",
+ "isVisible": true,
+ "domain": "global",
+ "serviceName": "HDFS",
+ "category": "Advanced"
+ },
+ {
+ "id": "puppet var",
"name": "lzo_enabled",
"displayName": "LZO compression",
"description": "LZO compression enabled",
"defaultValue": true,
"displayType": "checkbox",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": false,
"domain": "global",
"serviceName": "HDFS"
@@ -288,7 +298,6 @@ module.exports =
"defaultValue": "",
"isOverridable": false,
"displayType": "masterHost",
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "MAPREDUCE2",
@@ -302,7 +311,6 @@ module.exports =
"description": "",
"defaultValue": "/var/log/hadoop-mapreduce",
"displayType": "directory",
- "isRequiredByAgent": true,
"isVisible": true,
"isReconfigurable": false,
"domain": "global",
@@ -316,7 +324,6 @@ module.exports =
"description": "",
"defaultValue": "/var/run/hadoop-mapreduce",
"displayType": "directory",
- "isRequiredByAgent": true,
"isVisible": true,
"isReconfigurable": false,
"domain": "global",
@@ -334,7 +341,6 @@ module.exports =
"isOverridable": false,
"displayType": "int",
"unit": "MB",
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "YARN",
@@ -349,7 +355,6 @@ module.exports =
"defaultValue": "",
"isOverridable": false,
"displayType": "masterHost",
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "YARN",
@@ -365,7 +370,6 @@ module.exports =
"isOverridable": false,
"displayType": "int",
"unit": "MB",
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "YARN",
@@ -380,7 +384,6 @@ module.exports =
"defaultValue": "",
"isOverridable": false,
"displayType": "slaveHosts",
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "YARN",
@@ -396,7 +399,6 @@ module.exports =
"isOverridable": false,
"displayType": "int",
"unit": "MB",
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "YARN",
@@ -411,7 +413,6 @@ module.exports =
"defaultValue": "/var/log/hadoop-yarn",
"displayType": "directory",
"isReconfigurable": false,
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "YARN",
@@ -424,7 +425,6 @@ module.exports =
"description": "",
"defaultValue": "/var/run/hadoop-yarn",
"displayType": "directory",
- "isRequiredByAgent": true,
"isVisible": true,
"isReconfigurable": false,
"domain": "global",
@@ -441,7 +441,6 @@ module.exports =
"description": "The host that has been assigned to run HBase Master",
"displayType": "masterHosts",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "HBASE",
@@ -457,7 +456,6 @@ module.exports =
"displayType": "int",
"unit": "MB",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "HBASE",
@@ -473,7 +471,6 @@ module.exports =
"description": "The hosts that have been assigned to run RegionServer",
"displayType": "slaveHosts",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": true,
"isRequired": false,
"domain": "regionserver-global",
@@ -489,7 +486,6 @@ module.exports =
"defaultValue": "1024",
"displayType": "int",
"unit": "MB",
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "regionserver-global",
"serviceName": "HBASE",
@@ -505,7 +501,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "HBASE",
@@ -520,7 +515,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isOverridable": false,
- "isRequiredByAgent": true,
"isVisible": true,
"domain": "global",
"serviceName": "HBASE",
@@ -549,7 +543,6 @@ module.exports =
"description": "The host that has been assigned to run Hive Metastore",
"displayType": "masterHost",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": true,
"serviceName": "HIVE",
"category": "Hive Metastore",
@@ -662,7 +655,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "host",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": false,
"isObserved": true,
"domain": "global",
@@ -679,7 +671,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "host",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": false,
"isObserved": true,
"domain": "global",
@@ -696,7 +687,6 @@ module.exports =
"description": "Host on which the database will be created by Ambari",
"isReconfigurable": false,
"displayType": "masterHost",
- "isRequiredByAgent": false,
"isOverridable": false,
"isVisible": false,
"domain": "global",
@@ -706,6 +696,22 @@ module.exports =
},
{
"id": "puppet var",
+ "name": "hive_database_name",
+ "displayName": "Database Name",
+ "description": "Database name used as the Hive Metastore",
+ "defaultValue": "hive",
+ "isReconfigurable": true,
+ "displayType": "host",
+ "isOverridable": false,
+ "isVisible": true,
+ "isObserved": true,
+ "domain": "global",
+ "serviceName": "HIVE",
+ "category": "Hive Metastore",
+ "index": 4
+ },
+ {
+ "id": "puppet var",
"name": "hive_metastore_port",
"displayName": "Hive metastore port",
"description": "",
@@ -713,7 +719,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "int",
"isOverridable": false,
- "isRequiredByAgent": false, // Make this to true when we expose the property on ui by making "isVisible": true
"isVisible": false,
"domain": "global",
"serviceName": "HIVE",
@@ -728,7 +733,19 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isVisible": false,
- "isRequiredByAgent": false, // Make this to true when we expose the property on ui by making "isVisible": true
+ "domain": "global",
+ "serviceName": "HIVE",
+ "category": "Advanced"
+ },
+ {
+ "id": "puppet var",
+ "name": "hive_conf_dir",
+ "displayName": "Hive conf directory",
+ "description": "",
+ "defaultValue": "/etc/hive/conf",
+ "isReconfigurable": false,
+ "displayType": "directory",
+ "isVisible": false,
"domain": "global",
"serviceName": "HIVE",
"category": "Advanced"
@@ -742,7 +759,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isVisible": false,
- "isRequiredByAgent": false, // Make this to true when we expose the property on ui by making "isVisible": true
"domain": "global",
"serviceName": "HIVE",
"category": "Advanced"
@@ -777,6 +793,19 @@ module.exports =
},
{
"id": "puppet var",
+ "name": "mysql_connector_url",
+ "displayName": "MySQL connector url",
+ "description": "",
+ "defaultValue": "${download_url}/mysql-connector-java-5.1.18.zip",
+ "isReconfigurable": false,
+ "displayType": "directory",
+ "isVisible": false,
+ "domain": "global",
+ "serviceName": "HIVE",
+ "category": "Advanced"
+ },
+ {
+ "id": "puppet var",
"name": "hive_aux_jars_path",
"displayName": "Hive auxilary jar path",
"description": "",
@@ -799,7 +828,6 @@ module.exports =
"displayType": "masterHost",
"isOverridable": false,
"isVisible": true,
- "isRequiredByAgent": false,
"domain": "global",
"serviceName": "WEBHCAT",
"category": "WebHCat Server"
@@ -843,7 +871,6 @@ module.exports =
"displayType": "masterHost",
"isOverridable": false,
"isVisible": true,
- "isRequiredByAgent": false,
"domain": "global",
"serviceName": "OOZIE",
"category": "Oozie Server",
@@ -1028,7 +1055,6 @@ module.exports =
"isOverridable": false,
"displayType": "masterHost",
"isVisible": false,
- "isRequiredByAgent": false,
//"domain": "global",
"serviceName": "OOZIE",
"category": "Oozie Server"
@@ -1113,7 +1139,6 @@ module.exports =
"description": "The host that has been assigned to run ZooKeeper Server",
"displayType": "masterHosts",
"isVisible": true,
- "isRequiredByAgent": false,
"isOverridable": false,
"isRequired": false,
"serviceName": "ZOOKEEPER",
@@ -1222,6 +1247,19 @@ module.exports =
"category": "Advanced",
"index": 1
},
+ {
+ "id": "puppet var",
+ "name": "zk_pid_file",
+ "displayName": "ZooKeeper PID File",
+ "description": "",
+ "defaultValue": "/var/run/zookeeper/zookeeper_server.pid",
+ "isReconfigurable": false,
+ "displayType": "directory",
+ "isVisible": false,
+ "domain": "global",
+ "serviceName": "ZOOKEEPER",
+ "category": "Advanced"
+ },
/**********************************************HUE***************************************/
{
"id": "puppet var",
@@ -1232,7 +1270,6 @@ module.exports =
"description": "The host that has been assigned to run Hue Server",
"displayType": "masterHost",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "HUE",
@@ -1247,7 +1284,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "HUE",
@@ -1262,7 +1298,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isOverridable": false,
- "isRequiredByAgent": false,
"isVisible": true,
"domain": "global",
"serviceName": "HUE",
@@ -1278,7 +1313,6 @@ module.exports =
"isReconfigurable": false,
"displayType": "directory",
"isVisible": false,
- "isRequiredByAgent": false,
"domain": "global",
"serviceName": "GANGLIA",
"category": "Advanced"
@@ -1293,7 +1327,6 @@ module.exports =
"isRequired": false,
"displayType": "directory",
"isVisible": false,
- "isRequiredByAgent": false,
"domain": "global",
"serviceName": "MISC",
"category": "General",
@@ -1317,6 +1350,64 @@ module.exports =
},
{
"id": "puppet var",
+ "name": "dfs_datanode_address",
+ "displayName": "dfs_datanode_address",
+ "description": "",
+ "defaultValue": "50010",
+ "isReconfigurable": true,
+ "displayType": "int",
+ "isVisible": true,
+ "filename": "hdfs-site.xml",
+ "domain": "global",
+ "serviceName": "MISC",
+ "category": "Advanced",
+ "belongsToService":[]
+ },
+ {
+ "id": "puppet var",
+ "name": "dfs_datanode_http_address",
+ "displayName": "dfs_datanode_http_address",
+ "description": "",
+ "defaultValue": "50075",
+ "isReconfigurable": true,
+ "displayType": "int",
+ "isVisible": true,
+ "filename": "hdfs-site.xml",
+ "domain": "global",
+ "serviceName": "MISC",
+ "category": "Advanced",
+ "belongsToService":[]
+ },
+ {
+ "id": "puppet var",
+ "name": "gpl_artifacts_download_url",
+ "displayName": "gpl artifact download url",
+ "description": "",
+ "defaultValue": "",
+ "isReconfigurable": false,
+ "displayType": "advanced",
+ "isVisible": false,
+ "domain": "global",
+ "serviceName": "MISC",
+ "category": "General",
+ "belongsToService":[]
+ },
+ {
+ "id": "puppet var",
+ "name": "apache_artifacts_download_url",
+ "displayName": "apache artifact download url",
+ "description": "",
+ "defaultValue": "",
+ "isReconfigurable": false,
+ "displayType": "advanced",
+ "isVisible": false,
+ "domain": "global",
+ "serviceName": "MISC",
+ "category": "General",
+ "belongsToService":[]
+ },
+ {
+ "id": "puppet var",
"name": "ganglia_runtime_dir",
"displayName": "Ganglia runtime directory",
"description": "",
@@ -1345,6 +1436,36 @@ module.exports =
},
{
"id": "puppet var",
+ "name": "run_dir",
+ "displayName": "Hadoop run directory",
+ "description": "",
+ "defaultValue": "/var/run/hadoop",
+ "isRequired": false,
+ "isReconfigurable": false,
+ "displayType": "directory",
+ "isVisible": false,
+ "domain": "global",
+ "serviceName": "MISC",
+ "category": "Advanced",
+ "belongsToService":[]
+ },
+ {
+ "id": "puppet var",
+ "name": "hadoop_conf_dir",
+ "displayName": "Hadoop conf directory",
+ "description": "",
+ "defaultValue": "/etc/hadoop/conf",
+ "isRequired": false,
+ "isReconfigurable": false,
+ "displayType": "directory",
+ "isVisible": false,
+ "domain": "global",
+ "serviceName": "MISC",
+ "category": "Advanced",
+ "belongsToService":[]
+ },
+ {
+ "id": "puppet var",
"name": "hdfs_user",
"displayName": "HDFS User",
"description": "User to run HDFS as",
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/data/HDP2/secure_mapping.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/secure_mapping.js b/ambari-web/app/data/HDP2/secure_mapping.js
index 47be31c..421cf06 100644
--- a/ambari-web/app/data/HDP2/secure_mapping.js
+++ b/ambari-web/app/data/HDP2/secure_mapping.js
@@ -146,18 +146,18 @@ module.exports = [
"serviceName": "HDFS"
},
{
- "name": "ambari.dfs.datanode.port",
+ "name": "dfs.datanode.address",
"templateName": ["dfs_datanode_address"],
"foreignKey": null,
- "value": "<templateName[0]>",
+ "value": "0.0.0.0:<templateName[0]>",
"filename": "hdfs-site.xml",
"serviceName": "HDFS"
},
{
- "name": "ambari.dfs.datanode.http.port",
+ "name": "dfs.datanode.http.address",
"templateName": ["dfs_datanode_http_address"],
"foreignKey": null,
- "value": "<templateName[0]>",
+ "value": "0.0.0.0:<templateName[0]>",
"filename": "hdfs-site.xml",
"serviceName": "HDFS"
},
http://git-wip-us.apache.org/repos/asf/incubator-ambari/blob/05be921e/ambari-web/app/data/HDP2/site_properties.js
----------------------------------------------------------------------
diff --git a/ambari-web/app/data/HDP2/site_properties.js b/ambari-web/app/data/HDP2/site_properties.js
index 95ea021..8697c5f 100644
--- a/ambari-web/app/data/HDP2/site_properties.js
+++ b/ambari-web/app/data/HDP2/site_properties.js
@@ -485,20 +485,6 @@ module.exports =
"serviceName": "HIVE",
"index": 7
},
- {
- "id": "site property",
- "name": "ambari.hive.db.schema.name",
- "displayName": "Database Name",
- "description": "Database name used as the Hive Metastore",
- "defaultValue": "",
- "isReconfigurable": true,
- "displayType": "host",
- "isOverridable": false,
- "isObserved": true,
- "serviceName": "HIVE",
- "category": "Hive Metastore",
- "index": 4
- },
/**********************************************hbase-site***************************************/
{