You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by yu...@apache.org on 2013/05/11 19:25:53 UTC
svn commit: r1481380 [3/4] - in /incubator/ambari/branches/branch-1.2.3: ./
ambari-agent/ ambari-agent/src/main/puppet/modules/hdp-hadoop/manifests/
ambari-agent/src/main/puppet/modules/hdp-hadoop/templates/
ambari-agent/src/main/puppet/modules/hdp-nag...
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java Sat May 11 17:25:52 2013
@@ -33,7 +33,7 @@ import org.springframework.security.ldap
import java.util.Collection;
/**
- * Provides authorities population for LDAP user from LDAP catalog
+ * Provides authorities population for LDAP user from local DB
*/
public class AmbariLdapAuthoritiesPopulator implements LdapAuthoritiesPopulator {
private static final Logger log = LoggerFactory.getLogger(AmbariLdapAuthoritiesPopulator.class);
@@ -43,8 +43,6 @@ public class AmbariLdapAuthoritiesPopula
UserDAO userDAO;
RoleDAO roleDAO;
- private static final String AMBARI_ADMIN_LDAP_ATTRIBUTE_KEY = "ambari_admin";
-
@Inject
public AmbariLdapAuthoritiesPopulator(Configuration configuration, AuthorizationHelper authorizationHelper,
UserDAO userDAO, RoleDAO roleDAO) {
@@ -70,80 +68,29 @@ public class AmbariLdapAuthoritiesPopula
newUser.setLdapUser(true);
newUser.setUserName(username);
- //Adding a default "user" role
- addRole(newUser, configuration.getConfigsMap().
- get(Configuration.USER_ROLE_NAME_KEY));
- }
+ String roleName = (configuration.getConfigsMap().get(Configuration.USER_ROLE_NAME_KEY));
+ log.info("Using default role name " + roleName);
- user = userDAO.findLdapUserByName(username);
+ RoleEntity role = roleDAO.findByName(roleName);
- //Adding an "admin" user role if user is a member of ambari administrators
- // LDAP group
- Boolean isAdmin =
- (Boolean) userData.getObjectAttribute(AMBARI_ADMIN_LDAP_ATTRIBUTE_KEY);
- if ((isAdmin != null) && isAdmin) {
- log.info("Adding admin role to LDAP user " + username);
- addRole(user, configuration.getConfigsMap().
- get(Configuration.ADMIN_ROLE_NAME_KEY));
- } else {
- removeRole(user, configuration.getConfigsMap().
- get(Configuration.ADMIN_ROLE_NAME_KEY));
+ if (role == null) {
+ log.info("Role " + roleName + " not present in local DB - creating");
+ role = new RoleEntity();
+ role.setRoleName(roleName);
+ roleDAO.create(role);
+ role = roleDAO.findByName(role.getRoleName());
+ }
+
+ userDAO.create(newUser);
+
+ user = userDAO.findLdapUserByName(newUser.getUserName());
+
+ user.getRoleEntities().add(role);
+ role.getUserEntities().add(user);
+ roleDAO.merge(role);
+ userDAO.merge(user);
}
- user = userDAO.findLdapUserByName(username);
return authorizationHelper.convertRolesToAuthorities(user.getRoleEntities());
}
-
- /**
- * Adds role to user's role entities
- * Adds user to roleName's user entities
- *
- * @param user - the user entity to be modified
- * @param roleName - the role to add to user's roleEntities
- */
- private void addRole(UserEntity user, String roleName) {
- log.info("Using default role name " + roleName);
-
- RoleEntity roleEntity = roleDAO.findByName(roleName);
-
- if (roleEntity == null) {
- log.info("Role " + roleName + " not present in local DB - creating");
- roleEntity = new RoleEntity();
- roleEntity.setRoleName(roleName);
- roleDAO.create(roleEntity);
- roleEntity = roleDAO.findByName(roleEntity.getRoleName());
- }
-
- UserEntity userEntity = userDAO.findLdapUserByName(user.getUserName());
- if (userEntity == null) {
- userDAO.create(user);
- userEntity = userDAO.findLdapUserByName(user.getUserName());
- }
-
- if (!userEntity.getRoleEntities().contains(roleEntity)) {
- userEntity.getRoleEntities().add(roleEntity);
- roleEntity.getUserEntities().add(userEntity);
- roleDAO.merge(roleEntity);
- userDAO.merge(userEntity);
- }
- }
-
- /**
- * Remove role "roleName" from user "user"
- * @param user
- * @param roleName
- */
- private void removeRole(UserEntity user, String roleName) {
- UserEntity userEntity = userDAO.findByPK(user.getUserId());
- RoleEntity roleEntity = roleDAO.findByName(roleName);
-
- if (userEntity.getRoleEntities().contains(roleEntity)) {
- log.info("Removing admin role from LDAP user " + user.getUserName());
- userEntity.getRoleEntities().remove(roleEntity);
- roleEntity.getUserEntities().remove(userEntity);
- userDAO.merge(userEntity);
- roleDAO.merge(roleEntity);
- }
-
- }
}
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java Sat May 11 17:25:52 2013
@@ -205,11 +205,6 @@ public class Users {
public synchronized void addRoleToUser(User user, String role)
throws AmbariException {
- if (userDAO.findLdapUserByName(user.getUserName()) != null) {
- LOG.warn("Trying to add a role to the LDAP user"
- + ", user=" + user.getUserName());
- throw new AmbariException("Roles are not editable for LDAP users");
- }
UserEntity userEntity = userDAO.findByPK(user.getUserId());
if (userEntity == null) {
@@ -239,11 +234,6 @@ public class Users {
public synchronized void removeRoleFromUser(User user, String role)
throws AmbariException {
- if (userDAO.findLdapUserByName(user.getUserName()) != null) {
- LOG.warn("Trying to add a role to the LDAP user"
- + ", user=" + user.getUserName());
- throw new AmbariException("Roles are not editable for LDAP users");
- }
UserEntity userEntity = userDAO.findByPK(user.getUserId());
if (userEntity == null) {
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java Sat May 11 17:25:52 2013
@@ -86,9 +86,6 @@ public class StageUtils {
componentToClusterInfoKeyMap.put("NAMENODE", "namenode_host");
componentToClusterInfoKeyMap.put("JOBTRACKER", "jtnode_host");
componentToClusterInfoKeyMap.put("SNAMENODE", "snamenode_host");
- componentToClusterInfoKeyMap.put("RESOURCEMANAGER", "rm_host");
- componentToClusterInfoKeyMap.put("NODEMANAGER", "nm_hosts");
- componentToClusterInfoKeyMap.put("HISTORYSERVER", "hs_host");
componentToClusterInfoKeyMap.put("ZOOKEEPER_SERVER", "zookeeper_hosts");
componentToClusterInfoKeyMap.put("HBASE_MASTER", "hbase_master_hosts");
componentToClusterInfoKeyMap.put("HBASE_REGIONSERVER", "hbase_rs_hosts");
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py Sat May 11 17:25:52 2013
@@ -144,11 +144,6 @@ OS_TYPE_PROPERTY = "server.os_type"
JDK_DOWNLOAD_CMD = "curl --create-dirs -o {0} {1}"
JDK_DOWNLOAD_SIZE_CMD = "curl -I {0}"
-#JCE Policy files
-JCE_POLICY_FILENAME = "jce_policy-6.zip"
-JCE_DOWNLOAD_CMD = "curl -o {0} {1}"
-JCE_MIN_FILESIZE = 5000
-
def configure_pg_hba_ambaridb_users():
args = optparse.Values()
configure_postgres_username_password(args)
@@ -569,10 +564,6 @@ def download_jdk(args):
if args.java_home and os.path.exists(args.java_home):
print_warning_msg("JAVA_HOME " + args.java_home
+ " must be valid on ALL hosts")
- print_warning_msg("Please make sure the JCE Unlimited Strength "
- "Jurisdiction Policy Files 6, "
- "are dwonloaded on all "
- "hosts")
write_property(JAVA_HOME_PROPERTY, args.java_home)
return 0
@@ -656,75 +647,8 @@ def download_jdk(args):
format(JDK_INSTALL_DIR, jdk_version)
write_property(JAVA_HOME_PROPERTY, "{0}/{1}".
format(JDK_INSTALL_DIR, jdk_version))
- jce_download = download_jce_policy(properties, ok)
- if (jce_download == -1):
- print "JCE Policy files are required for secure HDP setup. Please ensure " \
- " all hosts have the JCE unlimited strength policy 6, files."
return 0
-def download_jce_policy(properties, accpeted_bcl):
- try:
- jce_url = properties['jce_policy.url']
- resources_dir = properties['resources.dir']
- except (KeyError), e:
- print 'Property ' + str(e) + ' is not defined in properties file'
- return -1
- dest_file = resources_dir + os.sep + JCE_POLICY_FILENAME
- if not os.path.exists(dest_file):
- print 'Downloading JCE Policy archive from ' + jce_url + ' to ' + dest_file
- try:
- size_command = JDK_DOWNLOAD_SIZE_CMD.format(jce_url);
- #Get Header from url,to get file size then
- retcode, out, err = run_os_command(size_command)
- if out.find("Content-Length") == -1:
- print "Request header doesn't contain Content-Length";
- return -1
- start_with = int(out.find("Content-Length") + len("Content-Length") + 2)
- end_with = out.find("\r\n", start_with)
- src_size = int(out[start_with:end_with])
- print_info_msg('JCE zip distribution size is ' + str(src_size) + 'bytes')
- file_exists = os.path.isfile(dest_file)
- file_size = -1
- if file_exists:
- file_size = os.stat(dest_file).st_size
- if file_exists and file_size == src_size:
- print_info_msg("File already exists")
- else:
- #BCL license before download
- jce_download_cmd = JCE_DOWNLOAD_CMD.format(dest_file, jce_url)
- print_info_msg("JCE download cmd: " + jce_download_cmd)
- if (accpeted_bcl == True):
- retcode, out, err = run_os_command(jce_download_cmd)
- if retcode == 0:
- print 'Successfully downloaded JCE Policy archive to ' + dest_file
- else:
- return -1
- else:
- ok = get_YN_input("To download the JCE Policy archive you must "
- "accept the license terms found at "
- "http://www.oracle.com/technetwork/java/javase"
- "/terms/license/index.html"
- "Not accepting might result in failure when "
- "setting up HDP security. \nDo you accept the "
- "Oracle Binary Code License Agreement [y/n] (y)? ", True)
- if (ok == True):
- retcode, out, err = run_os_command(jce_download_cmd)
- if retcode == 0:
- print 'Successfully downloaded JCE Policy archive to ' + dest_file
- else:
- return -1
- except Exception, e:
- print_error_msg('Failed to download JCE Policy archive: ' + str(e))
- return -1
- downloaded_size = os.stat(dest_file).st_size
- if downloaded_size != src_size or downloaded_size < JCE_MIN_FILESIZE:
- print_error_msg('Size of downloaded JCE Policy archive is '
- + str(downloaded_size) + ' bytes, it is probably \
- damaged or incomplete')
- return -1
- else:
- print "JCE Policy archive already exists, using " + dest_file
-
class RetCodeException(Exception): pass
def install_jdk(dest_file):
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml Sat May 11 17:25:52 2013
@@ -412,14 +412,4 @@ don't exist, they will be created with t
<description>Number of failed disks datanode would tolerate</description>
</property>
- <property>
- <name>dfs.namenode.check.stale.datanode</name>
- <value>true</value>
- <description>
- With this setting, the datanodes that have not replied to the heartbeat
- for more than 30s (i.e. in a stale state) are used for reads only if all
- other remote replicas have failed.
- </description>
- </property>
-
</configuration>
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml Sat May 11 17:25:52 2013
@@ -37,21 +37,6 @@
<description>Secondary NameNode.</description>
</property>
<property>
- <name>rm_host</name>
- <value></value>
- <description>Resource Manager.</description>
- </property>
- <property>
- <name>nm_hosts</name>
- <value></value>
- <description>List of Node Manager Hosts.</description>
- </property>
- <property>
- <name>hs_host</name>
- <value></value>
- <description>History Server.</description>
- </property>
- <property>
<name>fs_checkpoint_dir</name>
<value>/hadoop/hdfs/namesecondary</value>
<description>Secondary NameNode checkpoint dir.</description>
Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,112 @@
+<!--
+ Licensed to the Apache Software Foundation (ASF) under one or more
+ contributor license agreements. See the NOTICE file distributed with
+ this work for additional information regarding copyright ownership.
+ The ASF licenses this file to You under the Apache License, Version 2.0
+ (the "License"); you may not use this file except in compliance with
+ the License. You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+-->
+
+<configuration>
+
+ <property>
+ <name>yarn.scheduler.capacity.maximum-applications</name>
+ <value>10000</value>
+ <description>
+ Maximum number of applications that can be pending and running.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
+ <value>0.1</value>
+ <description>
+ Maximum percent of resources in the cluster which can be used to run
+ application masters i.e. controls number of concurrent running
+ applications.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.queues</name>
+ <value>default</value>
+ <description>
+ The queues at the this level (root is the root queue).
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.capacity</name>
+ <value>100</value>
+ <description>
+ The total capacity as a percentage out of 100 for this queue.
+ If it has child queues then this includes their capacity as well.
+ The child queues capacity should add up to their parent queue's capacity
+ or less.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.default.capacity</name>
+ <value>100</value>
+ <description>Default queue target capacity.</description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
+ <value>1</value>
+ <description>
+ Default queue user limit a percentage from 0.0 to 1.0.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
+ <value>100</value>
+ <description>
+ The maximum capacity of the default queue.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.default.state</name>
+ <value>RUNNING</value>
+ <description>
+ The state of the default queue. State can be one of RUNNING or STOPPED.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.default.acl_submit_jobs</name>
+ <value>*</value>
+ <description>
+ The ACL of who can submit jobs to the default queue.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.default.acl_administer_jobs</name>
+ <value>*</value>
+ <description>
+ The ACL of who can administer jobs on the default queue.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.capacity.root.acl_administer_queues</name>
+ <value>*</value>
+ <description>
+ The ACL for who can administer this queue i.e. change sub-queue
+ allocations.
+ </description>
+ </property>
+
+</configuration>
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml Sat May 11 17:25:52 2013
@@ -528,22 +528,4 @@ user)</description>
<description> Comma separated list of queues configured for this jobtracker.</description>
</property>
-<property>
- <name>mapreduce.shuffle.port</name>
- <value>8081</value>
- <description>Default port that the ShuffleHandler will run on. ShuffleHandler is a service run at the NodeManager to facilitate transfers of intermediate Map outputs to requesting Reducers.</description>
-</property>
-
-<property>
- <name>mapreduce.jobhistory.intermediate-done-dir</name>
- <value>/mr-history/tmp</value>
- <description>Directory where history files are written by MapReduce jobs.</description>
-</property>
-
-<property>
- <name>mapreduce.jobhistory.done-dir</name>
- <value>/mr-history/done</value>
- <description>Directory where history files are managed by the MR JobHistory Server.</description>
-</property>
-
</configuration>
Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,172 @@
+<?xml version="1.0"?>
+<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
+<!--
+ Licensed to the Apache Software Foundation (ASF) under one or more
+ contributor license agreements. See the NOTICE file distributed with
+ this work for additional information regarding copyright ownership.
+ The ASF licenses this file to You under the Apache License, Version 2.0
+ (the "License"); you may not use this file except in compliance with
+ the License. You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+-->
+
+<!-- Put site-specific property overrides in this file. -->
+
+<configuration xmlns:xi="http://www.w3.org/2001/XInclude">
+
+<!-- ResourceManager -->
+
+ <property>
+ <name>yarn.resourcemanager.resource-tracker.address</name>
+ <value>TODO-RMNODE-HOSTNAME:8025</value>
+ </property>
+
+ <property>
+ <name>yarn.resourcemanager.scheduler.address</name>
+ <value>TODO-RMNODE-HOSTNAME:8030</value>
+ </property>
+
+ <property>
+ <name>yarn.resourcemanager.address</name>
+ <value>TODO-RMNODE-HOSTNAME:8050</value>
+ </property>
+
+ <property>
+ <name>yarn.resourcemanager.admin.address</name>
+ <value>TODO-RMNODE-HOSTNAME:8141</value>
+ </property>
+
+ <property>
+ <name>yarn.resourcemanager.scheduler.class</name>
+ <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.minimum-allocation-mb</name>
+ <value>1024</value>
+ </property>
+
+ <property>
+ <name>yarn.scheduler.maximum-allocation-mb</name>
+ <value>8192</value>
+ </property>
+
+<!-- NodeManager -->
+
+ <property>
+ <name>yarn.nodemanager.address</name>
+ <value>0.0.0.0:45454</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.local-dirs</name>
+ <value>TODO-YARN-LOCAL-DIR</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.resource.memory-mb</name>
+ <value>8192</value>
+ <description>Amount of physical memory, in MB, that can be allocated
+ for containers.</description>
+ </property>
+
+ <property>
+ <name>yarn.application.classpath</name>
+ <value>/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*</value>
+ <description>Classpath for typical applications.</description>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.vmem-pmem-ratio</name>
+ <value>2.1</value>
+ <description>Ratio between virtual memory to physical memory when
+ setting memory limits for containers. Container allocations are
+ expressed in terms of physical memory, and virtual memory usage
+ is allowed to exceed this allocation by this ratio.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.container-executor.class</name>
+ <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value>
+ <description>ContainerExecutor for launching containers</description>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.aux-services</name>
+ <value>mapreduce.shuffle</value>
+ <description>Auxilliary services of NodeManager</description>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.aux-services.\1.class</name>
+ <value>org.apache.hadoop.mapred.ShuffleHandler</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.log-dirs</name>
+ <value>TODO-YARN-LOG-DIR</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.container-monitor.interval-ms</name>
+ <value>3000</value>
+ <description>The interval, in milliseconds, for which the node manager
+ waits between two cycles of monitoring its containers' memory usage.
+ </description>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.health-checker.script.path</name>
+ <value>/etc/hadoop/conf/health_check</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.health-checker.interval-ms</name>
+ <value>135000</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.health-checker.script.timeout-ms</name>
+ <value>60000</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.log.retain-second</name>
+ <value>604800</value>
+ </property>
+
+ <property>
+ <name>yarn.log-aggregation-enable</name>
+ <value>true</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.remote-app-log-dir</name>
+ <value>/app-logs</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.remote-app-log-dir-suffix</name>
+ <value>logs</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.log-aggregation.compression-type</name>
+ <value>gz</value>
+ </property>
+
+ <property>
+ <name>yarn.nodemanager.delete.debug-delay-sec</name>
+ <value>36000</value>
+ </property>
+
+
+</configuration>
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml Sat May 11 17:25:52 2013
@@ -17,12 +17,12 @@
-->
<metainfo>
<user>root</user>
- <comment>Tez is the next generation Hadoop Query Processing framework written on top of YARN</comment>
+ <comment>This is comment for TEZ service</comment>
<version>0.1.0.22-1</version>
<components>
<component>
- <name>TEZ_CLIENT</name>
+ <name>TEZ</name>
<category>CLIENT</category>
</component>
</components>
Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,39 @@
+<?xml version="1.0"?>
+<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
+
+<!--
+ Licensed to the Apache Software Foundation (ASF) under one or more
+ contributor license agreements. See the NOTICE file distributed with
+ this work for additional information regarding copyright ownership.
+ The ASF licenses this file to You under the Apache License, Version 2.0
+ (the "License"); you may not use this file except in compliance with
+ the License. You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+-->
+
+<!-- mapred-queue-acls.xml -->
+<configuration>
+
+
+<!-- queue default -->
+
+ <property>
+ <name>mapred.queue.default.acl-submit-job</name>
+ <value>*</value>
+ </property>
+
+ <property>
+ <name>mapred.queue.default.acl-administer-jobs</name>
+ <value>*</value>
+ </property>
+
+ <!-- END ACLs -->
+
+</configuration>
Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,531 @@
+<?xml version="1.0"?>
+<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
+
+<!--
+ Licensed to the Apache Software Foundation (ASF) under one or more
+ contributor license agreements. See the NOTICE file distributed with
+ this work for additional information regarding copyright ownership.
+ The ASF licenses this file to You under the Apache License, Version 2.0
+ (the "License"); you may not use this file except in compliance with
+ the License. You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+-->
+
+<!-- Put site-specific property overrides in this file. -->
+
+<configuration xmlns:xi="http://www.w3.org/2001/XInclude">
+
+<!-- i/o properties -->
+
+ <property>
+ <name>io.sort.mb</name>
+ <value></value>
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>io.sort.record.percent</name>
+ <value>.2</value>
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>io.sort.spill.percent</name>
+ <value></value>
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>io.sort.factor</name>
+ <value>100</value>
+ <description>No description</description>
+ </property>
+
+<!-- map/reduce properties -->
+
+<property>
+ <name>mapred.tasktracker.tasks.sleeptime-before-sigkill</name>
+ <value>250</value>
+ <description>Normally, this is the amount of time before killing
+ processes, and the recommended-default is 5.000 seconds - a value of
+ 5000 here. In this case, we are using it solely to blast tasks before
+ killing them, and killing them very quickly (1/4 second) to guarantee
+ that we do not leave VMs around for later jobs.
+ </description>
+</property>
+
+ <property>
+ <name>mapred.job.tracker.handler.count</name>
+ <value>50</value>
+ <description>
+ The number of server threads for the JobTracker. This should be roughly
+ 4% of the number of tasktracker nodes.
+ </description>
+ </property>
+
+ <property>
+ <name>mapred.system.dir</name>
+ <value>/mapred/system</value>
+ <description>No description</description>
+ <final>true</final>
+ </property>
+
+ <property>
+ <name>mapred.job.tracker</name>
+ <!-- cluster variant -->
+ <value></value>
+ <description>No description</description>
+ <final>true</final>
+ </property>
+
+ <property>
+ <name>mapred.job.tracker.http.address</name>
+ <!-- cluster variant -->
+ <value></value>
+ <description>No description</description>
+ <final>true</final>
+ </property>
+
+ <property>
+ <!-- cluster specific -->
+ <name>mapred.local.dir</name>
+ <value></value>
+ <description>No description</description>
+ <final>true</final>
+ </property>
+
+ <property>
+ <name>mapreduce.cluster.administrators</name>
+ <value> hadoop</value>
+ </property>
+
+ <property>
+ <name>mapred.reduce.parallel.copies</name>
+ <value>30</value>
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>mapred.tasktracker.map.tasks.maximum</name>
+ <value></value>
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>mapred.tasktracker.reduce.tasks.maximum</name>
+ <value></value>
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>tasktracker.http.threads</name>
+ <value>50</value>
+ </property>
+
+ <property>
+ <name>mapred.map.tasks.speculative.execution</name>
+ <value>false</value>
+ <description>If true, then multiple instances of some map tasks
+ may be executed in parallel.</description>
+ </property>
+
+ <property>
+ <name>mapred.reduce.tasks.speculative.execution</name>
+ <value>false</value>
+ <description>If true, then multiple instances of some reduce tasks
+ may be executed in parallel.</description>
+ </property>
+
+ <property>
+ <name>mapred.reduce.slowstart.completed.maps</name>
+ <value>0.05</value>
+ </property>
+
+ <property>
+ <name>mapred.inmem.merge.threshold</name>
+ <value>1000</value>
+ <description>The threshold, in terms of the number of files
+ for the in-memory merge process. When we accumulate threshold number of files
+ we initiate the in-memory merge and spill to disk. A value of 0 or less than
+ 0 indicates we want to DON'T have any threshold and instead depend only on
+ the ramfs's memory consumption to trigger the merge.
+ </description>
+ </property>
+
+ <property>
+ <name>mapred.job.shuffle.merge.percent</name>
+ <value>0.66</value>
+ <description>The usage threshold at which an in-memory merge will be
+ initiated, expressed as a percentage of the total memory allocated to
+ storing in-memory map outputs, as defined by
+ mapred.job.shuffle.input.buffer.percent.
+ </description>
+ </property>
+
+ <property>
+ <name>mapred.job.shuffle.input.buffer.percent</name>
+ <value>0.7</value>
+ <description>The percentage of memory to be allocated from the maximum heap
+ size to storing map outputs during the shuffle.
+ </description>
+ </property>
+
+ <property>
+ <name>mapred.map.output.compression.codec</name>
+ <value></value>
+ <description>If the map outputs are compressed, how should they be
+ compressed
+ </description>
+ </property>
+
+<property>
+ <name>mapred.output.compression.type</name>
+ <value>BLOCK</value>
+ <description>If the job outputs are to compressed as SequenceFiles, how should
+ they be compressed? Should be one of NONE, RECORD or BLOCK.
+ </description>
+</property>
+
+
+ <property>
+ <name>mapred.jobtracker.completeuserjobs.maximum</name>
+ <value>5</value>
+ </property>
+
+ <property>
+ <name>mapred.jobtracker.taskScheduler</name>
+ <value></value>
+ </property>
+
+ <property>
+ <name>mapred.jobtracker.restart.recover</name>
+ <value>false</value>
+ <description>"true" to enable (job) recovery upon restart,
+ "false" to start afresh
+ </description>
+ </property>
+
+ <property>
+ <name>mapred.job.reduce.input.buffer.percent</name>
+ <value>0.0</value>
+ <description>The percentage of memory- relative to the maximum heap size- to
+ retain map outputs during the reduce. When the shuffle is concluded, any
+ remaining map outputs in memory must consume less than this threshold before
+ the reduce can begin.
+ </description>
+ </property>
+
+ <property>
+ <name>mapreduce.reduce.input.limit</name>
+ <value>10737418240</value>
+ <description>The limit on the input size of the reduce. (This value
+ is 10 Gb.) If the estimated input size of the reduce is greater than
+ this value, job is failed. A value of -1 means that there is no limit
+ set. </description>
+</property>
+
+
+ <!-- copied from kryptonite configuration -->
+ <property>
+ <name>mapred.compress.map.output</name>
+ <value></value>
+ </property>
+
+
+ <property>
+ <name>mapred.task.timeout</name>
+ <value>600000</value>
+ <description>The number of milliseconds before a task will be
+ terminated if it neither reads an input, writes an output, nor
+ updates its status string.
+ </description>
+ </property>
+
+ <property>
+ <name>jetty.connector</name>
+ <value>org.mortbay.jetty.nio.SelectChannelConnector</value>
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>mapred.task.tracker.task-controller</name>
+ <value></value>
+ <description>
+ TaskController which is used to launch and manage task execution.
+ </description>
+ </property>
+
+ <property>
+ <name>mapred.child.root.logger</name>
+ <value>INFO,TLA</value>
+ </property>
+
+ <property>
+ <name>mapred.child.java.opts</name>
+ <value></value>
+
+ <description>No description</description>
+ </property>
+
+ <property>
+ <name>mapred.cluster.map.memory.mb</name>
+ <value></value>
+ </property>
+
+ <property>
+ <name>mapred.cluster.reduce.memory.mb</name>
+ <value></value>
+ </property>
+
+ <property>
+ <name>mapred.job.map.memory.mb</name>
+ <value></value>
+ </property>
+
+ <property>
+ <name>mapred.job.reduce.memory.mb</name>
+ <value></value>
+ </property>
+
+ <property>
+ <name>mapred.cluster.max.map.memory.mb</name>
+ <value></value>
+ </property>
+
+ <property>
+ <name>mapred.cluster.max.reduce.memory.mb</name>
+ <value></value>
+ </property>
+
+<property>
+ <name>mapred.hosts</name>
+ <value></value>
+</property>
+
+<property>
+ <name>mapred.hosts.exclude</name>
+ <value></value>
+</property>
+
+<property>
+ <name>mapred.max.tracker.blacklists</name>
+ <value>16</value>
+ <description>
+ if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted
+ </description>
+</property>
+
+<property>
+ <name>mapred.healthChecker.script.path</name>
+ <value></value>
+</property>
+
+<property>
+ <name>mapred.healthChecker.interval</name>
+ <value>135000</value>
+</property>
+
+<property>
+ <name>mapred.healthChecker.script.timeout</name>
+ <value>60000</value>
+</property>
+
+<property>
+ <name>mapred.job.tracker.persist.jobstatus.active</name>
+ <value>false</value>
+ <description>Indicates if persistency of job status information is
+ active or not.
+ </description>
+</property>
+
+<property>
+ <name>mapred.job.tracker.persist.jobstatus.hours</name>
+ <value>1</value>
+ <description>The number of hours job status information is persisted in DFS.
+ The job status information will be available after it drops of the memory
+ queue and between jobtracker restarts. With a zero value the job status
+ information is not persisted at all in DFS.
+ </description>
+</property>
+
+<property>
+ <name>mapred.job.tracker.persist.jobstatus.dir</name>
+ <value></value>
+ <description>The directory where the job status information is persisted
+ in a file system to be available after it drops of the memory queue and
+ between jobtracker restarts.
+ </description>
+</property>
+
+<property>
+ <name>mapred.jobtracker.retirejob.check</name>
+ <value>10000</value>
+</property>
+
+<property>
+ <name>mapred.jobtracker.retirejob.interval</name>
+ <value>0</value>
+</property>
+
+<property>
+ <name>mapred.job.tracker.history.completed.location</name>
+ <value>/mapred/history/done</value>
+ <description>No description</description>
+</property>
+
+<property>
+ <name>mapred.task.maxvmem</name>
+ <value></value>
+ <final>true</final>
+ <description>No description</description>
+</property>
+
+<property>
+ <name>mapred.jobtracker.maxtasks.per.job</name>
+ <value></value>
+ <final>true</final>
+ <description>The maximum number of tasks for a single job.
+ A value of -1 indicates that there is no maximum. </description>
+</property>
+
+<property>
+ <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name>
+ <value>false</value>
+</property>
+
+<property>
+ <name>mapred.userlog.retain.hours</name>
+ <value></value>
+</property>
+
+<property>
+ <name>mapred.job.reuse.jvm.num.tasks</name>
+ <value>1</value>
+ <description>
+ How many tasks to run per jvm. If set to -1, there is no limit
+ </description>
+ <final>true</final>
+</property>
+
+<property>
+ <name>mapreduce.jobtracker.kerberos.principal</name>
+ <value></value>
+ <description>
+ JT user name key.
+ </description>
+</property>
+
+<property>
+ <name>mapreduce.tasktracker.kerberos.principal</name>
+ <value></value>
+ <description>
+ tt user name key. "_HOST" is replaced by the host name of the task tracker.
+ </description>
+</property>
+
+
+ <property>
+ <name>hadoop.job.history.user.location</name>
+ <value>none</value>
+ <final>true</final>
+ </property>
+
+
+ <property>
+ <name>mapreduce.jobtracker.keytab.file</name>
+ <value></value>
+ <description>
+ The keytab for the jobtracker principal.
+ </description>
+
+</property>
+
+ <property>
+ <name>mapreduce.tasktracker.keytab.file</name>
+ <value></value>
+ <description>The filename of the keytab for the task tracker</description>
+ </property>
+
+ <property>
+ <name>mapreduce.jobtracker.staging.root.dir</name>
+ <value>/user</value>
+ <description>The Path prefix for where the staging directories should be placed. The next level is always the user's
+ name. It is a path in the default file system.</description>
+ </property>
+
+ <property>
+ <name>mapreduce.tasktracker.group</name>
+ <value>hadoop</value>
+ <description>The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.</description>
+
+ </property>
+
+ <property>
+ <name>mapreduce.jobtracker.split.metainfo.maxsize</name>
+ <value>50000000</value>
+ <final>true</final>
+ <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during
+ initialize.
+ </description>
+ </property>
+ <property>
+ <name>mapreduce.history.server.embedded</name>
+ <value>false</value>
+ <description>Should job history server be embedded within Job tracker
+process</description>
+ <final>true</final>
+ </property>
+
+ <property>
+ <name>mapreduce.history.server.http.address</name>
+ <!-- cluster variant -->
+ <value></value>
+ <description>Http address of the history server</description>
+ <final>true</final>
+ </property>
+
+ <property>
+ <name>mapreduce.jobhistory.kerberos.principal</name>
+ <!-- cluster variant -->
+ <value></value>
+ <description>Job history user name key. (must map to same user as JT
+user)</description>
+ </property>
+
+ <property>
+ <name>mapreduce.jobhistory.keytab.file</name>
+ <!-- cluster variant -->
+ <value></value>
+ <description>The keytab for the job history server principal.</description>
+ </property>
+
+<property>
+ <name>mapred.jobtracker.blacklist.fault-timeout-window</name>
+ <value>180</value>
+ <description>
+ 3-hour sliding window (value is in minutes)
+ </description>
+</property>
+
+<property>
+ <name>mapred.jobtracker.blacklist.fault-bucket-width</name>
+ <value>15</value>
+ <description>
+ 15-minute bucket size (value is in minutes)
+ </description>
+</property>
+
+<property>
+ <name>mapred.queue.names</name>
+ <value>default</value>
+ <description> Comma separated list of queues configured for this jobtracker.</description>
+</property>
+
+</configuration>
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml Sat May 11 17:25:52 2013
@@ -160,7 +160,7 @@
</property>
<property>
<name>zookeeper.session.timeout</name>
- <value>60000</value>
+ <value></value>
<description>ZooKeeper session timeout.
HBase passes this to the zk quorum as suggested maximum time for a
session (This setting becomes zookeeper's 'maxSessionTimeout'). See
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml Sat May 11 17:25:52 2013
@@ -412,14 +412,4 @@ don't exist, they will be created with t
<description>Number of failed disks datanode would tolerate</description>
</property>
- <property>
- <name>dfs.namenode.check.stale.datanode</name>
- <value>true</value>
- <description>
- With this setting, the datanodes that have not replied to the heartbeat
- for more than 30s (i.e. in a stale state) are used for reads only if all
- other remote replicas have failed.
- </description>
- </property>
-
</configuration>
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java Sat May 11 17:25:52 2013
@@ -18,7 +18,6 @@
package org.apache.ambari.server.agent;
import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
import static org.mockito.Matchers.eq;
@@ -167,99 +166,6 @@ public class TestHeartbeatMonitor {
assertTrue("HeartbeatMonitor should not generate StatusCommands for host2 because it has no services", cmds.isEmpty());
}
- @Test
- public void testNoStatusCommandForClientComponents() throws Exception {
- Clusters clusters = injector.getInstance(Clusters.class);
- clusters.addHost(hostname1);
- clusters.getHost(hostname1).setOsType("centos6");
- clusters.getHost(hostname1).persist();
- clusters.addHost(hostname2);
- clusters.getHost(hostname2).setOsType("centos6");
- clusters.getHost(hostname2).persist();
- clusters.addCluster(clusterName);
- Cluster cluster = clusters.getCluster(clusterName);
- cluster.setDesiredStackVersion(new StackId("HDP-0.1"));
- Set<String> hostNames = new HashSet<String>() {{
- add(hostname1);
- add(hostname2);
- }};
-
- ConfigFactory configFactory = injector.getInstance(ConfigFactory.class);
- Config config = configFactory.createNew(cluster, "global",
- new HashMap<String, String>() {{
- put("a", "b");
- }});
- config.setVersionTag("version1");
- cluster.addConfig(config);
- cluster.addDesiredConfig(config);
-
-
- clusters.mapHostsToCluster(hostNames, clusterName);
- Service hdfs = cluster.addService(serviceName);
- hdfs.persist();
- hdfs.addServiceComponent(Role.DATANODE.name()).persist();
- hdfs.getServiceComponent(Role.DATANODE.name()).addServiceComponentHost
- (hostname1).persist();
- hdfs.addServiceComponent(Role.NAMENODE.name()).persist();
- hdfs.getServiceComponent(Role.NAMENODE.name()).addServiceComponentHost
- (hostname1).persist();
- hdfs.addServiceComponent(Role.SECONDARY_NAMENODE.name()).persist();
- hdfs.getServiceComponent(Role.SECONDARY_NAMENODE.name()).
- addServiceComponentHost(hostname1).persist();
- hdfs.addServiceComponent(Role.HDFS_CLIENT.name()).persist();
- hdfs.getServiceComponent(Role.HDFS_CLIENT.name()).addServiceComponentHost
- (hostname1).persist();
- hdfs.getServiceComponent(Role.HDFS_CLIENT.name()).addServiceComponentHost
- (hostname2).persist();
-
- ActionQueue aq = new ActionQueue();
- ActionManager am = mock(ActionManager.class);
- HeartbeatMonitor hm = new HeartbeatMonitor(clusters, aq, am,
- heartbeatMonitorWakeupIntervalMS);
- HeartBeatHandler handler = new HeartBeatHandler(clusters, aq, am, injector);
- Register reg = new Register();
- reg.setHostname(hostname1);
- reg.setResponseId(12);
- reg.setTimestamp(System.currentTimeMillis() - 300);
- reg.setAgentVersion(ambariMetaInfo.getServerVersion());
- HostInfo hi = new HostInfo();
- hi.setOS("Centos5");
- reg.setHardwareProfile(hi);
- handler.handleRegistration(reg);
-
- HeartBeat hb = new HeartBeat();
- hb.setHostname(hostname1);
- hb.setNodeStatus(new HostStatus(HostStatus.Status.HEALTHY, "cool"));
- hb.setTimestamp(System.currentTimeMillis());
- hb.setResponseId(12);
- handler.handleHeartBeat(hb);
-
- List<StatusCommand> cmds = hm.generateStatusCommands(hostname1);
- assertTrue("HeartbeatMonitor should generate StatusCommands for host1",
- cmds.size() == 3);
- assertEquals("HDFS", cmds.get(0).getServiceName());
- boolean containsDATANODEStatus = false;
- boolean containsNAMENODEStatus = false;
- boolean containsSECONDARY_NAMENODEStatus = false;
- boolean containsHDFS_CLIENTStatus = false;
- for (StatusCommand cmd : cmds) {
- containsDATANODEStatus |= cmd.getComponentName().equals("DATANODE");
- containsNAMENODEStatus |= cmd.getComponentName().equals("NAMENODE");
- containsSECONDARY_NAMENODEStatus |= cmd.getComponentName().
- equals("SECONDARY_NAMENODE");
- containsHDFS_CLIENTStatus |= cmd.getComponentName().equals
- ("HDFS_CLIENT");
- assertTrue(cmd.getConfigurations().size() > 0);
- }
- assertTrue(containsDATANODEStatus);
- assertTrue(containsNAMENODEStatus);
- assertTrue(containsSECONDARY_NAMENODEStatus);
- assertFalse(containsHDFS_CLIENTStatus);
-
- cmds = hm.generateStatusCommands(hostname2);
- assertTrue("HeartbeatMonitor should not generate StatusCommands for host2" +
- " because it has only client components", cmds.isEmpty());
- }
@Test
public void testHeartbeatStateCommandsEnqueueing() throws AmbariException, InterruptedException,
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java Sat May 11 17:25:52 2013
@@ -36,7 +36,6 @@ import org.apache.ambari.server.AmbariEx
import org.apache.ambari.server.ClusterNotFoundException;
import org.apache.ambari.server.DuplicateResourceException;
import org.apache.ambari.server.ParentObjectNotFoundException;
-import org.apache.ambari.server.ObjectNotFoundException;
import org.apache.ambari.server.Role;
import org.apache.ambari.server.RoleCommand;
import org.apache.ambari.server.StackAccessException;
@@ -75,9 +74,7 @@ import org.apache.ambari.server.state.sv
import org.apache.ambari.server.utils.StageUtils;
import org.junit.After;
import org.junit.Before;
-import org.junit.Rule;
import org.junit.Test;
-import org.junit.rules.ExpectedException;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import com.google.inject.Guice;
@@ -122,9 +119,6 @@ public class AmbariManagementControllerT
private Users users;
private EntityManager entityManager;
- @Rule
- public ExpectedException expectedException = ExpectedException.none();
-
@Before
public void setup() throws Exception {
injector = Guice.createInjector(new InMemoryDefaultTestModule());
@@ -6048,128 +6042,4 @@ public class AmbariManagementControllerT
}
}
- @Test
- public void testGetTasksByRequestId() throws AmbariException {
- final long requestId1 = 1;
- final long requestId2 = 2;
- final String clusterName = "c1";
- final String hostName1 = "h1";
- final String context = "Test invocation";
-
- clusters.addCluster(clusterName);
- clusters.getCluster(clusterName).setDesiredStackVersion(new StackId("HDP-0.1"));
- clusters.addHost(hostName1);
- clusters.getHost("h1").setOsType("centos5");
- clusters.getHost(hostName1).persist();
-
- clusters.mapHostsToCluster(new HashSet<String>(){
- {add(hostName1);}}, clusterName);
-
-
- List<Stage> stages = new ArrayList<Stage>();
- stages.add(new Stage(requestId1, "/a1", clusterName, context));
- stages.get(0).setStageId(1);
- stages.get(0).addHostRoleExecutionCommand(hostName1, Role.HBASE_MASTER,
- RoleCommand.START,
- new ServiceComponentHostStartEvent(Role.HBASE_MASTER.toString(),
- hostName1, System.currentTimeMillis(),
- new HashMap<String, String>()),
- clusterName, "HBASE");
-
- stages.add(new Stage(requestId1, "/a2", clusterName, context));
- stages.get(1).setStageId(2);
- stages.get(1).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
- RoleCommand.START,
- new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
- hostName1, System.currentTimeMillis(),
- new HashMap<String, String>()), clusterName, "HBASE");
-
- stages.add(new Stage(requestId1, "/a3", clusterName, context));
- stages.get(2).setStageId(3);
- stages.get(2).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
- RoleCommand.START,
- new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
- hostName1, System.currentTimeMillis(),
- new HashMap<String, String>()), clusterName, "HBASE");
-
-
- stages.add(new Stage(requestId2, "/a4", clusterName, context));
- stages.get(3).setStageId(4);
- stages.get(3).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
- RoleCommand.START,
- new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
- hostName1, System.currentTimeMillis(),
- new HashMap<String, String>()), clusterName, "HBASE");
-
- stages.add(new Stage(requestId2, "/a5", clusterName, context));
- stages.get(4).setStageId(5);
- stages.get(4).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
- RoleCommand.START,
- new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
- hostName1, System.currentTimeMillis(),
- new HashMap<String, String>()), clusterName, "HBASE");
-
- actionDB.persistActions(stages);
-
- Set<TaskStatusRequest> taskStatusRequests;
- Set<TaskStatusResponse> taskStatusResponses;
-
- //check count of tasks by requestId1
- taskStatusRequests = new HashSet<TaskStatusRequest>(){
- {
- add(new TaskStatusRequest(requestId1, null));
- }
- };
- taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
- assertEquals(3, taskStatusResponses.size());
-
- //check a taskId that requested by requestId1 and task id
- taskStatusRequests = new HashSet<TaskStatusRequest>(){
- {
- add(new TaskStatusRequest(requestId1, 2L));
- }
- };
- taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
- assertEquals(1, taskStatusResponses.size());
- assertEquals(2L, taskStatusResponses.iterator().next().getTaskId());
-
- //check count of tasks by requestId2
- taskStatusRequests = new HashSet<TaskStatusRequest>(){
- {
- add(new TaskStatusRequest(requestId2, null));
- }
- };
- taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
- assertEquals(2, taskStatusResponses.size());
-
- //check a taskId that requested by requestId2 and task id
- taskStatusRequests = new HashSet<TaskStatusRequest>(){
- {
- add(new TaskStatusRequest(requestId2, 5L));
- }
- };
- taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
- assertEquals(5L, taskStatusResponses.iterator().next().getTaskId());
-
- //verify that task from second request (requestId2) does not present in first request (requestId1)
- taskStatusRequests = new HashSet<TaskStatusRequest>(){
- {
- add(new TaskStatusRequest(requestId1, 5L));
- }
- };
- expectedException.expect(ObjectNotFoundException.class);
- expectedException.expectMessage("Task resource doesn't exist.");
- controller.getTaskStatus(taskStatusRequests);
-
- //verify that task from first request (requestId1) does not present in second request (requestId2)
- taskStatusRequests = new HashSet<TaskStatusRequest>(){
- {
- add(new TaskStatusRequest(requestId2, 2L));
- }
- };
- expectedException.expect(ObjectNotFoundException.class);
- expectedException.expectMessage("Task resource doesn't exist.");
- controller.getTaskStatus(taskStatusRequests);
- }
-
}
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java Sat May 11 17:25:52 2013
@@ -97,38 +97,6 @@ public class AmbariLdapAuthenticationPro
Assert.assertTrue(auth == null);
}
- @Test
- public void testLdapAdminGroupToRolesMapping() throws Exception {
-
- Authentication authentication;
-
- authentication =
- new UsernamePasswordAuthenticationToken("allowedAdmin", "password");
- Authentication result = authenticationProvider.authenticate(authentication);
- assertTrue(result.isAuthenticated());
-
- UserEntity allowedAdminEntity = userDAO.findLdapUserByName("allowedAdmin");
-
- authentication =
- new UsernamePasswordAuthenticationToken("allowedUser", "password");
- authenticationProvider.authenticate(authentication);
- UserEntity allowedUserEntity = userDAO.findLdapUserByName("allowedUser");
-
-
- RoleEntity adminRole = roleDAO.findByName(
- configuration.getConfigsMap().get(Configuration.ADMIN_ROLE_NAME_KEY));
- RoleEntity userRole = roleDAO.findByName(
- configuration.getConfigsMap().get(Configuration.USER_ROLE_NAME_KEY));
-
-
- assertTrue(allowedAdminEntity.getRoleEntities().contains(userRole));
- assertTrue(allowedAdminEntity.getRoleEntities().contains(adminRole));
-
- assertTrue(allowedUserEntity.getRoleEntities().contains(userRole));
- assertFalse(allowedUserEntity.getRoleEntities().contains(adminRole));
-
-
- }
@AfterClass
public static void afterClass() {
Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh Sat May 11 17:25:52 2013
@@ -21,18 +21,18 @@ curl -i -X POST http://localhost:8080/ap
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HBASE
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/GANGLIA
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/NAGIOS
-
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "core-site", "tag": "version1", "properties" : { "fs.default.name" : "localhost:8020"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "core-site", "tag": "version2", "properties" : { "fs.default.name" : "localhost:8020"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "core-site", "tag": "version1"}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "hdfs-site", "tag": "version1", "properties" : { "dfs.datanode.data.dir.perm" : "750"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "global", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "mapred-site", "tag": "version1", "properties" : { "mapred.job.tracker" : "localhost:50300", "mapreduce.history.server.embedded": "false", "mapreduce.history.server.http.address": "localhost:51111"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "hbase-site", "tag": "version1", "properties" : { "hbase.rootdir" : "hdfs://localhost:8020/apps/hbase/", "hbase.cluster.distributed" : "true", "hbase.zookeeper.quorum": "localhost", "zookeeper.session.timeout": "60000" }}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "hbase-env", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "nagios-global", "tag": "version2", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password", "nagios_contact": "a\u0040b.c" }}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "nagios-global", "tag": "version1", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password" }}}}' http://localhost:8080/api/v1/clusters/c1
-
+curl -i -X POST -d '{"type": "core-site", "tag": "version1", "properties" : { "fs.default.name" : "localhost:8020"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "hdfs-site", "tag": "version1", "properties" : { "dfs.datanode.data.dir.perm" : "750"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "global", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "mapred-site", "tag": "version1", "properties" : { "mapred.job.tracker" : "localhost:50300", "mapreduce.history.server.embedded": "false", "mapreduce.history.server.http.address": "localhost:51111"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "hbase-site", "tag": "version1", "properties" : { "hbase.rootdir" : "hdfs://localhost:8020/apps/hbase/", "hbase.cluster.distributed" : "true", "hbase.zookeeper.quorum": "localhost", "zookeeper.session.timeout": "60000" }}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "hbase-env", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "nagios-global", "tag": "version2", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password", "nagios_contact": "a\u0040b.c" }}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "nagios-global", "tag": "version1", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password" }}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X PUT -d '{"config": {"core-site": "version1", "hdfs-site": "version1", "global" : "version1" }}' http://localhost:8080/api/v1/clusters/c1/services/HDFS
+curl -i -X PUT -d '{"config": {"core-site": "version1", "mapred-site": "version1"}}' http://localhost:8080/api/v1/clusters/c1/services/MAPREDUCE
+curl -i -X PUT -d '{"config": {"hbase-site": "version1", "hbase-env": "version1"}}' http://localhost:8080/api/v1/clusters/c1/services/HBASE
+curl -i -X PUT -d '{"config": {"nagios-global": "version2" }}' http://localhost:8080/api/v1/clusters/c1/services/NAGIOS
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HDFS/components/NAMENODE
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HDFS/components/SECONDARY_NAMENODE
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HDFS/components/DATANODE
@@ -60,7 +60,6 @@ curl -i -X POST http://localhost:8080/ap
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST/host_components/HBASE_REGIONSERVER
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST/host_components/HBASE_CLIENT
curl -i -X POST http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST/host_components/NAGIOS_SERVER
-curl -i -X PUT -d '{"Hosts": {"desired_config": {"type": "core-site", "tag": "version3", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password" }}}}' http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST
curl -i -X PUT -d '{"ServiceInfo": {"state" : "INSTALLED"}}' http://localhost:8080/api/v1/clusters/c1/services?state=INIT
#curl -i -X PUT -d '{"ServiceInfo": {"state" : "STARTED"}}' http://localhost:8080/api/v1/clusters/c1/services?state=INSTALLED
# http://localhost:8080/api/v1/clusters/c1/requests/2
Modified: incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js Sat May 11 17:25:52 2013
@@ -24,8 +24,8 @@ App.skipBootstrap = false;
App.alwaysGoToInstaller = false;
App.testEnableSecurity = true; // By default enable security is tested; turning it false tests disable security
App.apiPrefix = '/api/v1';
-App.defaultStackVersion = 'HDP-1.3.0';
-App.defaultLocalStackVersion = 'HDPLocal-1.3.0';
+App.defaultStackVersion = 'HDP-1.2.1';
+App.defaultLocalStackVersion = 'HDPLocal-1.2.1';
App.defaultJavaHome = '/usr/jdk/jdk1.6.0_31';
App.timeout = 180000; // default AJAX timeout
App.maxRetries = 3; // max number of retries for certain AJAX calls
@@ -50,7 +50,7 @@ App.supports = {
hiveOozieExtraDatabases: false,
multipleHBaseMasters: false,
addMasters: false,
- customizeSmokeTestUser: true,
+ customizeSmokeTestUser: false,
hue: false,
ldapGroupMapping: false
};
Modified: incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js Sat May 11 17:25:52 2013
@@ -39,10 +39,11 @@ App.MainAdminUserController = Em.Control
return;
}
+ ;
App.ModalPopup.show({
header:Em.I18n.t('admin.users.delete.header').format(event.context.get('userName')),
- body:Em.I18n.t('question.sure').format(''),
+ body:Em.I18n.t('question.sure'),
primary:Em.I18n.t('yes'),
secondary:Em.I18n.t('no'),
Modified: incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml Sat May 11 17:25:52 2013
@@ -19,7 +19,7 @@
<parent>
<groupId>org.apache.ambari</groupId>
<artifactId>ambari-project</artifactId>
- <version>1.3.0-SNAPSHOT</version>
+ <version>1.2.3-SNAPSHOT</version>
<relativePath>../ambari-project</relativePath>
</parent>
<modelVersion>4.0.0</modelVersion>
@@ -27,7 +27,7 @@
<artifactId>ambari-web</artifactId>
<packaging>pom</packaging>
<name>Ambari Web</name>
- <version>1.3.0-SNAPSHOT</version>
+ <version>1.2.3-SNAPSHOT</version>
<description>Ambari Web</description>
<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
Modified: incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php (original)
+++ incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php Sat May 11 17:25:52 2013
@@ -381,11 +381,11 @@ function hdp_mon_generate_response( $res
$pieces[0] = "HIVE";
break;
case "ZKSERVERS":
- $pieces[0] = "ZOOKEEPER";
+ $pieces[0] = "ZOOKEEPER";
break;
case "AMBARI":
- $pieces[0] = "AMBARI";
- break;
+ $pieces[0] = "AMBARI";
+ break;
case "NAGIOS":
case "HDFS":
case "MAPREDUCE":
Modified: incubator/ambari/branches/branch-1.2.3/docs/pom.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/pom.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/pom.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/pom.xml Sat May 11 17:25:52 2013
@@ -28,7 +28,7 @@
<modelVersion>4.0.0</modelVersion>
<groupId>org.apache.ambari</groupId>
- <version>1.2.2-SNAPSHOT</version>
+ <version>1.2.0-SNAPSHOT</version>
<artifactId>ambari</artifactId>
<packaging>pom</packaging>
@@ -118,18 +118,6 @@
<role>PMC</role>
</roles>
</developer>
- <developer>
- <id>billie</id>
- <name>Billie Rinaldi</name>
- <email>billie@apache.org</email>
- <timezone>-8</timezone>
- <roles>
- <role>Committer</role>
- </roles>
- <organization>
- Hortonworks
- </organization>
- </developer>
<developer>
<id>ddas</id>
<name>Devaraj Das</name>
@@ -237,19 +225,7 @@
<organization>
Hortonworks
</organization>
- </developer>
- <developer>
- <id>ncole</id>
- <name>Nate Cole</name>
- <email>ncole@apache.org</email>
- <timezone>-8</timezone>
- <roles>
- <role>Committer</role>
- </roles>
- <organization>
- Hortonworks
- </organization>
- </developer>
+ </developer>
<developer>
<id>ramya</id>
<name>Ramya Sunil</name>
@@ -275,30 +251,6 @@
</organization>
</developer>
<developer>
- <id>smohanty</id>
- <name>Sumit Mohanty</name>
- <email>smohanty@apache.org</email>
- <timezone>-8</timezone>
- <roles>
- <role>Committer</role>
- </roles>
- <organization>
- Hortonworks
- </organization>
- </developer>
- <developer>
- <id>swagle</id>
- <name>Siddharth Wagle</name>
- <email>swagle@apache.org</email>
- <timezone>-8</timezone>
- <roles>
- <role>Committer</role>
- </roles>
- <organization>
- Hortonworks
- </organization>
- </developer>
- <developer>
<id>srimanth</id>
<name>Srimanth Gunturi</name>
<email>srimanth@apache.org</email>
Modified: incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt Sat May 11 17:25:52 2013
@@ -15,20 +15,26 @@
~~
Introduction
- The Apache Ambari project is aimed at making Hadoop management simpler by developing software for provisioning, managing, and monitoring Apache Hadoop clusters.
- Ambari provides an intuitive, easy-to-use Hadoop management web UI backed by its RESTful APIs.
+ Apache Ambari is a web-based tool for provisioning, managing, and monitoring Apache Hadoop clusters. The set of
+ Hadoop components that are currently supported by Ambari includes:
- The set of Hadoop components that are currently supported by Ambari includes:
+ * {{{http://hadoop.apache.org/docs/hdfs} Apache Hadoop - HDFS}}
- {{{http://hadoop.apache.org/docs/hdfs} HDFS}},
- {{{http://hadoop.apache.org/docs/mapreduce} MapReduce}}
- {{{http://hive.apache.org} Hive}},
- {{{http://incubator.apache.org/hcatalog} HCatalog}},
- {{{http://hbase.apache.org} HBase}},
- {{{http://zookeeper.apache.org} ZooKeeper}},
- {{{http://incubator.apache.org/oozie/} Oozie}},
- {{{http://pig.apache.org} Pig}},
- {{{http://sqoop.apache.org} Sqoop}}
+ * {{{http://hadoop.apache.org/docs/mapreduce} Apache Hadoop - MapReduce}}
+
+ * {{{http://hive.apache.org} Apache Hive}}
+
+ * {{{http://incubator.apache.org/hcatalog} Apache HCatalog}}
+
+ * {{{http://hbase.apache.org} Apache HBase}}
+
+ * {{{http://zookeeper.apache.org} Apache Zookeeper}}
+
+ * {{{http://incubator.apache.org/oozie/} Apache Oozie}}
+
+ * {{{http://pig.apache.org} Apache Pig}}
+
+ * {{{http://sqoop.apache.org} Apache Sqoop}}
[]
@@ -36,7 +42,7 @@ Introduction
* Provision a Hadoop Cluster
- * Ambari provides a step-by-step wizard for installing Hadoop services across any number of hosts.
+ * Ambari provides an easy-to-use, step-by-step wizard for installing Hadoop services across any number of hosts.
* Ambari handles configuration of Hadoop services for the cluster.
@@ -55,31 +61,52 @@ Introduction
* Ambari leverages {{{http://ganglia.sourceforge.net/} Ganglia}} for metrics collection.
* Ambari leverages {{{http://www.nagios.org/} Nagios}} for system alerting and will send emails when your attention is needed (e.g., a node goes down, remaining disk space is low, etc).
+
+ []
+
+Ambari Source
+
+ Follow the {{{./1.2.0/installing-hadoop-using-ambari/content/index.html} installation guide for 1.2.0 (stable)}} or check out the work going on in {{{./whats-new.html} trunk}} for the upcoming 1.2.1 release.
+
+Roadmap
+
+ * Support for additional Operating Systems
+
+ * Ambari currently supports 64-bit RHEL/CentOS 5 + 6 and SLES 11
[]
- Ambari enables Application Developers and System Integrators to:
+ * RESTful API for integration
+
+ * Ambari will expose a unified, RESTful API to enable third-party applications to integrate
+ Hadoop cluster management and monitoring capabilities (1.2.0)
+
+ []
- * Easily integrate Hadoop provisioning, management, and monitoring capabilities to their own applications with the {{{https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md} Ambari REST APIs}}.
+ * Granular configurations
-Getting Started with Ambari
+ * Ambari currently applies service configurations at the cluster-level. For more
+ flexibility, Ambari will allow for configurations in a more granular manner, such as
+ applying a set of configurations to a specific group of hosts.
- Follow the {{{./1.2.2/installing-hadoop-using-ambari/content/index.html} installation guide for Ambari 1.2.2}}.
+ []
- Note: Ambari currently supports 64-bit RHEL/CentOS 5 + 6 and SLES 11.
+ * Security
-Get Involved
+ * Installation of secure Hadoop clusters (Kerberos-based)
- Visit the {{{https://cwiki.apache.org/confluence/display/AMBARI/Ambari} Ambari Wiki}} for design documents, roadmap, development guidelines, etc.
+ * Role-based user authentication, authorization, and auditing
- The first {{{http://www.meetup.com/Apache-Ambari-User-Group} Ambari User Group Meetup}} took place on April 2 in Palo Alto, California, USA. {{{http://www.meetup.com/Apache-Ambari-User-Group/events/109316812/} See the slides and WebEx session from the Meetup}}.
+ * Support for LDAP and Active Directory (1.2.0)
-What's New?
+ []
- Check out the work going on for the {{{./whats-new.html} upcoming 1.3.0 release}}.
+ * Visualization
-Disclaimer
+ * Interactive visualization of the current state of the cluster for a number of key metrics (1.2.0)
- Apache Ambari is an effort undergoing incubation at The Apache Software Foundation (ASF) sponsored by the Apache Incubator PMC. Incubation is required of all newly accepted projects until a further review indicates that the infrastructure, communications, and decision making process have stabilized in a manner consistent with other successful ASF projects. While incubation status is not necessarily a reflection of the completeness or stability of the code, it does indicate that the project has yet to be fully endorsed by the ASF.
+ * Interactive visualization of historical states of the cluster for a number of key metrics
+ * Interactive visualization of Pig, Hive, and MapReduce jobs (1.2.0)
+ []
Modified: incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt Sat May 11 17:25:52 2013
@@ -14,9 +14,28 @@
~~ limitations under the License.
~~
-What's New in Ambari 1.3.0?
+What's New?
- * Check back later for the new features in 1.3.0.
+ Thereâs a lot of activity occurring in trunk. Weâve had a great response from the
+ community and received a lot of input into the direction. To date, we have added
+ many new features. To highlight a few...
+
+ * Added more options on provisioning so you can better target your cluster
+ setup for your needs
+
+ * More service controls, including the ability to run ad hoc smoke tests
+
+ * Exposed a RESTful API for cluster metrics (with management and
+ provisioning to follow)
+
+ * Moved to a JavaScript front-end for Ambari Web and refreshed the entire UI
+ experience
+
+ * Used that RESTful API and added more charts and heatmaps to the dashboard
+
+ * Reduced the number of dependencies on the Ambari Server + Ambari Agent
+
+ * Added support for SLES 11
Getting Ambari Source
@@ -28,14 +47,14 @@ $ svn checkout http://svn.apache.org/rep
JIRA
- You can follow all the work going on by watching the Ambari JIRA project. You can see the
- JIRA details for Ambari 1.3.0 here:
+ You can follow all the work going on by watching the Ambari JIRA project. The next release of Apache Ambari will be version 1.2.1. You can see the
+ JIRA details for this release here:
- {{{https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.3.0%22%20AND%20project%20%3D%20AMBARI} https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.3.0%22%20AND%20project%20%3D%20AMBARI}}
+ {{{https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.2.1%22%20AND%20project%20%3D%20AMBARI} https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.2.1%22%20AND%20project%20%3D%20AMBARI}}
User Guide
- Take a look at {{{http://incubator.apache.org/ambari/1.2.2/installing-hadoop-using-ambari/content/index.html} how to install a Hadoop cluster using Ambari 1.2.2}}.
+ {{{http://incubator.apache.org/ambari/1.2.1/installing-hadoop-using-ambari/content/index.html} See how to install a Hadoop cluster using Ambari 1.2.1}}
Stay Tuned
Modified: incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml Sat May 11 17:25:52 2013
@@ -28,9 +28,7 @@
<custom>
<fluidoSkin>
<topBarEnabled>true</topBarEnabled>
- <sideBarEnabled>true</sideBarEnabled>
- <leftColumnClass>span2</leftColumnClass>
- <bodyColumnClass>span10</bodyColumnClass>
+ <sideBarEnabled>false</sideBarEnabled>
</fluidoSkin>
</custom>
<bannerLeft>
@@ -61,62 +59,38 @@
})();
</script>
<!-- End of Google analytics -->
- <style>
- .well.sidebar-nav {
- background-color: #fff;
- }
- a.externalLink[href^=http] {
- background-image: none;
- padding-right: 0;
- }
- body.topBarEnabled {
- padding-top: 40px;
- }
- #leftColumn .nav-list .active a {
- background-color: #a0a0a0;
- }
- .nav-list .active a:hover {
- background-color: #a0a0a0;
- }
- </style>
</head>
<links>
<item name="JIRA" href="https://issues.apache.org/jira/browse/AMBARI" />
<item name="SVN" href="https://svn.apache.org/repos/asf/incubator/ambari/" />
- <item name="Wiki" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari" />
+ <item name="WIKI" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari" />
</links>
<breadcrumbs>
- <item name="Incubator" href="../"/>
- <item name="Ambari" href="http://incubator.apache.org/ambari/"/>
+ <item name="Ambari" href="index.html"/>
</breadcrumbs>
<menu name="Ambari">
- <item name="Overview" href="index.html"/>
<item name="What's New?" href="whats-new.html"/>
+ <item name="About" href="index.html"/>
+ <item name="Wiki" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari"/>
<item name="Project Team" href="team-list.html"/>
- <item name="IRC Channel" href="irc.html"/>
<item name="Mailing Lists" href="mail-lists.html"/>
<item name="Issue Tracking" href="issue-tracking.html"/>
- <item name="User Group" href="http://www.meetup.com/Apache-Ambari-User-Group"/>
- <item name="Project License" href="http://www.apache.org/licenses/"/>
+ <item name="Project License" href="license.html"/>
</menu>
<menu name="Releases">
- <item name="1.2.2" href="http://www.apache.org/dist/incubator/ambari/ambari-1.2.2/"/>
- <item name="1.2.1" href="http://www.apache.org/dist/incubator/ambari/ambari-1.2.1/"/>
+ <item name="1.2.1 (coming soon)" href="whats-new.html"/>
<item name="1.2.0" href="http://www.apache.org/dist/incubator/ambari/ambari-1.2.0/"/>
<item name="0.9" href="http://www.apache.org/dist/incubator/ambari/ambari-0.9-incubating/"/>
</menu>
<menu name="Documentation">
- <item name="Install Guide for 1.2.2" href="1.2.2/installing-hadoop-using-ambari/content/index.html"/>
<item name="Install Guide for 1.2.1" href="1.2.1/installing-hadoop-using-ambari/content/index.html"/>
<item name="Install Guide for 1.2.0" href="1.2.0/installing-hadoop-using-ambari/content/index.html"/>
<item name="Install Guide for 0.9" href="install-0.9.html"/>
- <item name="API Reference" href="https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md"/>
- <item name="Wiki" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari"/>
</menu>
<footer>