You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by yu...@apache.org on 2013/05/11 19:25:53 UTC

svn commit: r1481380 [3/4] - in /incubator/ambari/branches/branch-1.2.3: ./ ambari-agent/ ambari-agent/src/main/puppet/modules/hdp-hadoop/manifests/ ambari-agent/src/main/puppet/modules/hdp-hadoop/templates/ ambari-agent/src/main/puppet/modules/hdp-nag...

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthoritiesPopulator.java Sat May 11 17:25:52 2013
@@ -33,7 +33,7 @@ import org.springframework.security.ldap
 import java.util.Collection;
 
 /**
- * Provides authorities population for LDAP user from LDAP catalog
+ * Provides authorities population for LDAP user from local DB
  */
 public class AmbariLdapAuthoritiesPopulator implements LdapAuthoritiesPopulator {
   private static final Logger log = LoggerFactory.getLogger(AmbariLdapAuthoritiesPopulator.class);
@@ -43,8 +43,6 @@ public class AmbariLdapAuthoritiesPopula
   UserDAO userDAO;
   RoleDAO roleDAO;
 
-  private static final String AMBARI_ADMIN_LDAP_ATTRIBUTE_KEY = "ambari_admin";
-
   @Inject
   public AmbariLdapAuthoritiesPopulator(Configuration configuration, AuthorizationHelper authorizationHelper,
                                         UserDAO userDAO, RoleDAO roleDAO) {
@@ -70,80 +68,29 @@ public class AmbariLdapAuthoritiesPopula
       newUser.setLdapUser(true);
       newUser.setUserName(username);
 
-      //Adding a default "user" role
-      addRole(newUser, configuration.getConfigsMap().
-          get(Configuration.USER_ROLE_NAME_KEY));
-    }
+      String roleName = (configuration.getConfigsMap().get(Configuration.USER_ROLE_NAME_KEY));
+      log.info("Using default role name " + roleName);
 
-    user = userDAO.findLdapUserByName(username);
+      RoleEntity role = roleDAO.findByName(roleName);
 
-    //Adding an "admin" user role if user is a member of ambari administrators
-    // LDAP group
-    Boolean isAdmin =
-        (Boolean) userData.getObjectAttribute(AMBARI_ADMIN_LDAP_ATTRIBUTE_KEY);
-    if ((isAdmin != null) && isAdmin) {
-      log.info("Adding admin role to LDAP user " + username);
-      addRole(user, configuration.getConfigsMap().
-          get(Configuration.ADMIN_ROLE_NAME_KEY));
-    } else {
-      removeRole(user, configuration.getConfigsMap().
-          get(Configuration.ADMIN_ROLE_NAME_KEY));
+      if (role == null) {
+        log.info("Role " + roleName + " not present in local DB - creating");
+        role = new RoleEntity();
+        role.setRoleName(roleName);
+        roleDAO.create(role);
+        role = roleDAO.findByName(role.getRoleName());
+      }
+
+      userDAO.create(newUser);
+
+      user = userDAO.findLdapUserByName(newUser.getUserName());
+
+      user.getRoleEntities().add(role);
+      role.getUserEntities().add(user);
+      roleDAO.merge(role);
+      userDAO.merge(user);
     }
 
-    user = userDAO.findLdapUserByName(username);
     return authorizationHelper.convertRolesToAuthorities(user.getRoleEntities());
   }
-
-  /**
-   * Adds role to user's role entities
-   * Adds user to roleName's user entities
-   *
-   * @param user - the user entity to be modified
-   * @param roleName - the role to add to user's roleEntities
-   */
-  private void addRole(UserEntity user, String roleName) {
-    log.info("Using default role name " + roleName);
-
-    RoleEntity roleEntity = roleDAO.findByName(roleName);
-
-    if (roleEntity == null) {
-      log.info("Role " + roleName + " not present in local DB - creating");
-      roleEntity = new RoleEntity();
-      roleEntity.setRoleName(roleName);
-      roleDAO.create(roleEntity);
-      roleEntity = roleDAO.findByName(roleEntity.getRoleName());
-    }
-
-    UserEntity userEntity = userDAO.findLdapUserByName(user.getUserName());
-    if (userEntity == null) {
-      userDAO.create(user);
-      userEntity = userDAO.findLdapUserByName(user.getUserName());
-    }
-
-    if (!userEntity.getRoleEntities().contains(roleEntity)) {
-      userEntity.getRoleEntities().add(roleEntity);
-      roleEntity.getUserEntities().add(userEntity);
-      roleDAO.merge(roleEntity);
-      userDAO.merge(userEntity);
-    }
-  }
-
-  /**
-   * Remove role "roleName" from user "user"
-   * @param user
-   * @param roleName
-   */
-  private void removeRole(UserEntity user, String roleName) {
-    UserEntity userEntity = userDAO.findByPK(user.getUserId());
-    RoleEntity roleEntity = roleDAO.findByName(roleName);
-
-    if (userEntity.getRoleEntities().contains(roleEntity)) {
-      log.info("Removing admin role from LDAP user " + user.getUserName());
-      userEntity.getRoleEntities().remove(roleEntity);
-      roleEntity.getUserEntities().remove(userEntity);
-      userDAO.merge(userEntity);
-      roleDAO.merge(roleEntity);
-    }
-
-  }
 }

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/security/authorization/Users.java Sat May 11 17:25:52 2013
@@ -205,11 +205,6 @@ public class Users {
   public synchronized void addRoleToUser(User user, String role)
       throws AmbariException {
 
-    if (userDAO.findLdapUserByName(user.getUserName()) != null) {
-      LOG.warn("Trying to add a role to the LDAP user"
-          + ", user=" + user.getUserName());
-      throw new AmbariException("Roles are not editable for LDAP users");
-    }
 
     UserEntity userEntity = userDAO.findByPK(user.getUserId());
     if (userEntity == null) {
@@ -239,11 +234,6 @@ public class Users {
   public synchronized void removeRoleFromUser(User user, String role)
       throws AmbariException {
 
-    if (userDAO.findLdapUserByName(user.getUserName()) != null) {
-      LOG.warn("Trying to add a role to the LDAP user"
-          + ", user=" + user.getUserName());
-      throw new AmbariException("Roles are not editable for LDAP users");
-    }
 
     UserEntity userEntity = userDAO.findByPK(user.getUserId());
     if (userEntity == null) {

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/java/org/apache/ambari/server/utils/StageUtils.java Sat May 11 17:25:52 2013
@@ -86,9 +86,6 @@ public class StageUtils {
     componentToClusterInfoKeyMap.put("NAMENODE", "namenode_host");
     componentToClusterInfoKeyMap.put("JOBTRACKER", "jtnode_host");
     componentToClusterInfoKeyMap.put("SNAMENODE", "snamenode_host");
-    componentToClusterInfoKeyMap.put("RESOURCEMANAGER", "rm_host");
-    componentToClusterInfoKeyMap.put("NODEMANAGER", "nm_hosts");
-    componentToClusterInfoKeyMap.put("HISTORYSERVER", "hs_host");
     componentToClusterInfoKeyMap.put("ZOOKEEPER_SERVER", "zookeeper_hosts");
     componentToClusterInfoKeyMap.put("HBASE_MASTER", "hbase_master_hosts");
     componentToClusterInfoKeyMap.put("HBASE_REGIONSERVER", "hbase_rs_hosts");

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/python/ambari-server.py Sat May 11 17:25:52 2013
@@ -144,11 +144,6 @@ OS_TYPE_PROPERTY = "server.os_type"
 JDK_DOWNLOAD_CMD = "curl --create-dirs -o {0} {1}"
 JDK_DOWNLOAD_SIZE_CMD = "curl -I {0}"
 
-#JCE Policy files
-JCE_POLICY_FILENAME = "jce_policy-6.zip"
-JCE_DOWNLOAD_CMD = "curl -o {0} {1}"
-JCE_MIN_FILESIZE = 5000
-
 def configure_pg_hba_ambaridb_users():
   args = optparse.Values()
   configure_postgres_username_password(args)
@@ -569,10 +564,6 @@ def download_jdk(args):
   if args.java_home and os.path.exists(args.java_home):
     print_warning_msg("JAVA_HOME " + args.java_home
                     + " must be valid on ALL hosts")
-    print_warning_msg("Please make sure the JCE Unlimited Strength "
-                      "Jurisdiction Policy Files 6, "
-                      "are dwonloaded on all "
-                      "hosts")
     write_property(JAVA_HOME_PROPERTY, args.java_home)
     return 0
 
@@ -656,75 +647,8 @@ def download_jdk(args):
       format(JDK_INSTALL_DIR, jdk_version)
   write_property(JAVA_HOME_PROPERTY, "{0}/{1}".
       format(JDK_INSTALL_DIR, jdk_version))
-  jce_download = download_jce_policy(properties, ok)
-  if (jce_download == -1):
-    print "JCE Policy files are required for secure HDP setup. Please ensure " \
-          " all hosts have the JCE unlimited strength policy 6, files."
   return 0
 
-def download_jce_policy(properties, accpeted_bcl):
-  try:
-    jce_url = properties['jce_policy.url']
-    resources_dir = properties['resources.dir']
-  except (KeyError), e:
-    print 'Property ' + str(e) + ' is not defined in properties file'
-    return -1
-  dest_file = resources_dir + os.sep + JCE_POLICY_FILENAME
-  if not os.path.exists(dest_file):
-    print 'Downloading JCE Policy archive from ' + jce_url + ' to ' + dest_file
-    try:
-      size_command = JDK_DOWNLOAD_SIZE_CMD.format(jce_url);
-      #Get Header from url,to get file size then
-      retcode, out, err = run_os_command(size_command)
-      if out.find("Content-Length") == -1:
-        print "Request header doesn't contain Content-Length";
-        return -1
-      start_with = int(out.find("Content-Length") + len("Content-Length") + 2)
-      end_with = out.find("\r\n", start_with)
-      src_size = int(out[start_with:end_with])
-      print_info_msg('JCE zip distribution size is ' + str(src_size) + 'bytes')
-      file_exists = os.path.isfile(dest_file)
-      file_size = -1
-      if file_exists:
-        file_size = os.stat(dest_file).st_size
-      if file_exists and file_size == src_size:
-        print_info_msg("File already exists")
-      else:
-        #BCL license before download
-        jce_download_cmd = JCE_DOWNLOAD_CMD.format(dest_file, jce_url)
-        print_info_msg("JCE download cmd: " + jce_download_cmd)
-        if (accpeted_bcl == True):
-          retcode, out, err = run_os_command(jce_download_cmd)
-          if retcode == 0:
-            print 'Successfully downloaded JCE Policy archive to ' + dest_file
-          else:
-            return -1
-        else:
-          ok = get_YN_input("To download the JCE Policy archive you must "
-                            "accept the license terms found at "
-                            "http://www.oracle.com/technetwork/java/javase"
-                            "/terms/license/index.html"
-                            "Not accepting might result in failure when "
-                            "setting up HDP security. \nDo you accept the "
-                            "Oracle Binary Code License Agreement [y/n] (y)? ", True)
-          if (ok == True):
-            retcode, out, err = run_os_command(jce_download_cmd)
-            if retcode == 0:
-              print 'Successfully downloaded JCE Policy archive to ' + dest_file
-          else:
-            return -1
-    except Exception, e:
-      print_error_msg('Failed to download JCE Policy archive: ' + str(e))
-      return -1
-    downloaded_size = os.stat(dest_file).st_size
-    if downloaded_size != src_size or downloaded_size < JCE_MIN_FILESIZE:
-      print_error_msg('Size of downloaded JCE Policy archive is '
-                      + str(downloaded_size) + ' bytes, it is probably \
-                    damaged or incomplete')
-      return -1
-  else:
-    print "JCE Policy archive already exists, using " + dest_file
-
 class RetCodeException(Exception): pass
 
 def install_jdk(dest_file):

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/1.3.0/services/HDFS/configuration/hdfs-site.xml Sat May 11 17:25:52 2013
@@ -412,14 +412,4 @@ don't exist, they will be created with t
   <description>Number of failed disks datanode would tolerate</description>
 </property>
 
-  <property>
-    <name>dfs.namenode.check.stale.datanode</name>
-    <value>true</value>
-    <description>
-      With this setting, the datanodes that have not replied to the heartbeat
-      for more than 30s (i.e. in a stale state) are used for reads only if all
-      other remote replicas have failed.
-    </description>
-  </property>
-
 </configuration>

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/HDFS/configuration/global.xml Sat May 11 17:25:52 2013
@@ -37,21 +37,6 @@
     <description>Secondary NameNode.</description>
   </property>
   <property>
-    <name>rm_host</name>
-    <value></value>
-    <description>Resource Manager.</description>
-  </property>
-  <property>
-    <name>nm_hosts</name>
-    <value></value>
-    <description>List of Node Manager Hosts.</description>
-  </property>
-  <property>
-    <name>hs_host</name>
-    <value></value>
-    <description>History Server.</description>
-  </property>
-  <property>
     <name>fs_checkpoint_dir</name>
     <value>/hadoop/hdfs/namesecondary</value>
     <description>Secondary NameNode checkpoint dir.</description>

Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/capacity-scheduler.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,112 @@
+<!--
+   Licensed to the Apache Software Foundation (ASF) under one or more
+   contributor license agreements.  See the NOTICE file distributed with
+   this work for additional information regarding copyright ownership.
+   The ASF licenses this file to You under the Apache License, Version 2.0
+   (the "License"); you may not use this file except in compliance with
+   the License.  You may obtain a copy of the License at
+
+       http://www.apache.org/licenses/LICENSE-2.0
+
+   Unless required by applicable law or agreed to in writing, software
+   distributed under the License is distributed on an "AS IS" BASIS,
+   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+   See the License for the specific language governing permissions and
+   limitations under the License.
+-->
+
+<configuration>
+
+  <property>
+    <name>yarn.scheduler.capacity.maximum-applications</name>
+    <value>10000</value>
+    <description>
+      Maximum number of applications that can be pending and running.
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
+    <value>0.1</value>
+    <description>
+      Maximum percent of resources in the cluster which can be used to run 
+      application masters i.e. controls number of concurrent running
+      applications.
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.queues</name>
+    <value>default</value>
+    <description>
+      The queues at the this level (root is the root queue).
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.capacity</name>
+    <value>100</value>
+    <description>
+      The total capacity as a percentage out of 100 for this queue.
+      If it has child queues then this includes their capacity as well.
+      The child queues capacity should add up to their parent queue's capacity
+      or less.
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.default.capacity</name>
+    <value>100</value>
+    <description>Default queue target capacity.</description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
+    <value>1</value>
+    <description>
+      Default queue user limit a percentage from 0.0 to 1.0.
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
+    <value>100</value>
+    <description>
+      The maximum capacity of the default queue. 
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.default.state</name>
+    <value>RUNNING</value>
+    <description>
+      The state of the default queue. State can be one of RUNNING or STOPPED.
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.default.acl_submit_jobs</name>
+    <value>*</value>
+    <description>
+      The ACL of who can submit jobs to the default queue.
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.default.acl_administer_jobs</name>
+    <value>*</value>
+    <description>
+      The ACL of who can administer jobs on the default queue.
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.capacity.root.acl_administer_queues</name>
+    <value>*</value>
+    <description>
+      The ACL for who can administer this queue i.e. change sub-queue 
+      allocations.
+    </description>
+  </property>
+
+</configuration>

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/mapred-site.xml Sat May 11 17:25:52 2013
@@ -528,22 +528,4 @@ user)</description>
   <description> Comma separated list of queues configured for this jobtracker.</description>
 </property>
 
-<property>
-  <name>mapreduce.shuffle.port</name>
-  <value>8081</value>
-  <description>Default port that the ShuffleHandler will run on. ShuffleHandler is a service run at the NodeManager to facilitate transfers of intermediate Map outputs to requesting Reducers.</description>
-</property>
-
-<property>
-  <name>mapreduce.jobhistory.intermediate-done-dir</name>
-  <value>/mr-history/tmp</value>
-  <description>Directory where history files are written by MapReduce jobs.</description>
-</property>
-
-<property>
-  <name>mapreduce.jobhistory.done-dir</name>
-  <value>/mr-history/done</value>
-  <description>Directory where history files are managed by the MR JobHistory Server.</description>
-</property>
-
 </configuration>

Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/MAPREDUCEv2/configuration/yarn-site.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,172 @@
+<?xml version="1.0"?>
+<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
+<!--
+   Licensed to the Apache Software Foundation (ASF) under one or more
+   contributor license agreements.  See the NOTICE file distributed with
+   this work for additional information regarding copyright ownership.
+   The ASF licenses this file to You under the Apache License, Version 2.0
+   (the "License"); you may not use this file except in compliance with
+   the License.  You may obtain a copy of the License at
+
+       http://www.apache.org/licenses/LICENSE-2.0
+
+   Unless required by applicable law or agreed to in writing, software
+   distributed under the License is distributed on an "AS IS" BASIS,
+   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+   See the License for the specific language governing permissions and
+   limitations under the License.
+-->
+
+<!-- Put site-specific property overrides in this file. -->
+
+<configuration xmlns:xi="http://www.w3.org/2001/XInclude">
+
+<!-- ResourceManager -->
+
+  <property>
+    <name>yarn.resourcemanager.resource-tracker.address</name>
+    <value>TODO-RMNODE-HOSTNAME:8025</value>
+  </property>
+
+  <property>
+    <name>yarn.resourcemanager.scheduler.address</name>
+    <value>TODO-RMNODE-HOSTNAME:8030</value>
+  </property>
+  
+  <property>
+    <name>yarn.resourcemanager.address</name>
+    <value>TODO-RMNODE-HOSTNAME:8050</value>
+  </property>
+
+	<property>
+    <name>yarn.resourcemanager.admin.address</name>
+    <value>TODO-RMNODE-HOSTNAME:8141</value>
+	</property>
+
+  <property>
+   <name>yarn.resourcemanager.scheduler.class</name>
+   <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.minimum-allocation-mb</name>
+    <value>1024</value>
+  </property>
+
+  <property>
+    <name>yarn.scheduler.maximum-allocation-mb</name>
+    <value>8192</value>
+  </property>
+
+<!-- NodeManager -->
+
+  <property>
+    <name>yarn.nodemanager.address</name>
+    <value>0.0.0.0:45454</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.local-dirs</name>
+    <value>TODO-YARN-LOCAL-DIR</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.resource.memory-mb</name>
+    <value>8192</value>
+    <description>Amount of physical memory, in MB, that can be allocated
+      for containers.</description>
+  </property>
+
+  <property>
+    <name>yarn.application.classpath</name>
+    <value>/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*</value>
+  <description>Classpath for typical applications.</description>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.vmem-pmem-ratio</name>
+    <value>2.1</value>
+    <description>Ratio between virtual memory to physical memory when
+    setting memory limits for containers. Container allocations are
+    expressed in terms of physical memory, and virtual memory usage
+    is allowed to exceed this allocation by this ratio.
+    </description>
+  </property>
+  
+  <property>
+    <name>yarn.nodemanager.container-executor.class</name>
+    <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value>
+    <description>ContainerExecutor for launching containers</description>
+  </property>
+ 
+  <property>
+    <name>yarn.nodemanager.aux-services</name>
+    <value>mapreduce.shuffle</value>
+    <description>Auxilliary services of NodeManager</description>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.aux-services.\1.class</name>
+    <value>org.apache.hadoop.mapred.ShuffleHandler</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.log-dirs</name>
+    <value>TODO-YARN-LOG-DIR</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.container-monitor.interval-ms</name>
+    <value>3000</value>
+    <description>The interval, in milliseconds, for which the node manager
+    waits  between two cycles of monitoring its containers' memory usage. 
+    </description>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.health-checker.script.path</name>
+    <value>/etc/hadoop/conf/health_check</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.health-checker.interval-ms</name>
+    <value>135000</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.health-checker.script.timeout-ms</name>
+    <value>60000</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.log.retain-second</name>
+    <value>604800</value>
+  </property>
+
+  <property>
+    <name>yarn.log-aggregation-enable</name>
+    <value>true</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.remote-app-log-dir</name>
+    <value>/app-logs</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.remote-app-log-dir-suffix</name>
+    <value>logs</value>
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.log-aggregation.compression-type</name>
+    <value>gz</value> 
+  </property>
+
+  <property>
+    <name>yarn.nodemanager.delete.debug-delay-sec</name>
+    <value>36000</value>
+  </property>
+
+
+</configuration>

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/TEZ/metainfo.xml Sat May 11 17:25:52 2013
@@ -17,12 +17,12 @@
 -->
 <metainfo>
     <user>root</user>
-    <comment>Tez is the next generation Hadoop Query Processing framework written on top of YARN</comment>
+    <comment>This is comment for TEZ service</comment>
     <version>0.1.0.22-1</version>
 
     <components>
         <component>
-            <name>TEZ_CLIENT</name>
+            <name>TEZ</name>
             <category>CLIENT</category>
         </component>
     </components>

Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-queue-acls.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,39 @@
+<?xml version="1.0"?>
+<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
+
+<!--
+   Licensed to the Apache Software Foundation (ASF) under one or more
+   contributor license agreements.  See the NOTICE file distributed with
+   this work for additional information regarding copyright ownership.
+   The ASF licenses this file to You under the Apache License, Version 2.0
+   (the "License"); you may not use this file except in compliance with
+   the License.  You may obtain a copy of the License at
+
+       http://www.apache.org/licenses/LICENSE-2.0
+
+   Unless required by applicable law or agreed to in writing, software
+   distributed under the License is distributed on an "AS IS" BASIS,
+   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+   See the License for the specific language governing permissions and
+   limitations under the License.
+-->
+
+<!-- mapred-queue-acls.xml -->
+<configuration>
+
+
+<!-- queue default -->
+
+  <property>
+    <name>mapred.queue.default.acl-submit-job</name>
+    <value>*</value>
+  </property>
+
+  <property>
+    <name>mapred.queue.default.acl-administer-jobs</name>
+    <value>*</value>
+  </property>
+
+  <!-- END ACLs -->
+
+</configuration>

Added: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml?rev=1481380&view=auto
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml (added)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDP/2.0.1/services/YARN/configuration/mapred-site.xml Sat May 11 17:25:52 2013
@@ -0,0 +1,531 @@
+<?xml version="1.0"?>
+<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
+
+<!--
+   Licensed to the Apache Software Foundation (ASF) under one or more
+   contributor license agreements.  See the NOTICE file distributed with
+   this work for additional information regarding copyright ownership.
+   The ASF licenses this file to You under the Apache License, Version 2.0
+   (the "License"); you may not use this file except in compliance with
+   the License.  You may obtain a copy of the License at
+
+       http://www.apache.org/licenses/LICENSE-2.0
+
+   Unless required by applicable law or agreed to in writing, software
+   distributed under the License is distributed on an "AS IS" BASIS,
+   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+   See the License for the specific language governing permissions and
+   limitations under the License.
+-->
+
+<!-- Put site-specific property overrides in this file. -->
+
+<configuration xmlns:xi="http://www.w3.org/2001/XInclude">
+
+<!-- i/o properties -->
+
+  <property>
+    <name>io.sort.mb</name>
+    <value></value>
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>io.sort.record.percent</name>
+    <value>.2</value>
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>io.sort.spill.percent</name>
+    <value></value>
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>io.sort.factor</name>
+    <value>100</value>
+    <description>No description</description>
+  </property>
+
+<!-- map/reduce properties -->
+
+<property>
+  <name>mapred.tasktracker.tasks.sleeptime-before-sigkill</name>
+  <value>250</value>
+  <description>Normally, this is the amount of time before killing
+  processes, and the recommended-default is 5.000 seconds - a value of
+  5000 here.  In this case, we are using it solely to blast tasks before
+  killing them, and killing them very quickly (1/4 second) to guarantee
+  that we do not leave VMs around for later jobs.
+  </description>
+</property>
+
+  <property>
+    <name>mapred.job.tracker.handler.count</name>
+    <value>50</value>
+    <description>
+    The number of server threads for the JobTracker. This should be roughly
+    4% of the number of tasktracker nodes.
+    </description>
+  </property>
+
+  <property>
+    <name>mapred.system.dir</name>
+    <value>/mapred/system</value>
+    <description>No description</description>
+    <final>true</final>
+  </property>
+
+  <property>
+    <name>mapred.job.tracker</name>
+    <!-- cluster variant -->
+    <value></value>
+    <description>No description</description>
+    <final>true</final>
+  </property>
+
+  <property>
+    <name>mapred.job.tracker.http.address</name>
+    <!-- cluster variant -->
+    <value></value>
+    <description>No description</description>
+    <final>true</final>
+  </property>
+
+  <property>
+    <!-- cluster specific -->
+    <name>mapred.local.dir</name>
+    <value></value>
+    <description>No description</description>
+    <final>true</final>
+  </property>
+
+  <property>
+  <name>mapreduce.cluster.administrators</name>
+  <value> hadoop</value>
+  </property>
+
+  <property>
+    <name>mapred.reduce.parallel.copies</name>
+    <value>30</value>
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>mapred.tasktracker.map.tasks.maximum</name>
+    <value></value>
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>mapred.tasktracker.reduce.tasks.maximum</name>
+    <value></value>
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>tasktracker.http.threads</name>
+    <value>50</value>
+  </property>
+
+  <property>
+    <name>mapred.map.tasks.speculative.execution</name>
+    <value>false</value>
+    <description>If true, then multiple instances of some map tasks
+               may be executed in parallel.</description>
+  </property>
+
+  <property>
+    <name>mapred.reduce.tasks.speculative.execution</name>
+    <value>false</value>
+    <description>If true, then multiple instances of some reduce tasks
+               may be executed in parallel.</description>
+  </property>
+
+  <property>
+    <name>mapred.reduce.slowstart.completed.maps</name>
+    <value>0.05</value>
+  </property>
+
+  <property>
+    <name>mapred.inmem.merge.threshold</name>
+    <value>1000</value>
+    <description>The threshold, in terms of the number of files
+  for the in-memory merge process. When we accumulate threshold number of files
+  we initiate the in-memory merge and spill to disk. A value of 0 or less than
+  0 indicates we want to DON'T have any threshold and instead depend only on
+  the ramfs's memory consumption to trigger the merge.
+  </description>
+  </property>
+
+  <property>
+    <name>mapred.job.shuffle.merge.percent</name>
+    <value>0.66</value>
+    <description>The usage threshold at which an in-memory merge will be
+  initiated, expressed as a percentage of the total memory allocated to
+  storing in-memory map outputs, as defined by
+  mapred.job.shuffle.input.buffer.percent.
+  </description>
+  </property>
+
+  <property>
+    <name>mapred.job.shuffle.input.buffer.percent</name>
+    <value>0.7</value>
+    <description>The percentage of memory to be allocated from the maximum heap
+  size to storing map outputs during the shuffle.
+  </description>
+  </property>
+
+  <property>
+    <name>mapred.map.output.compression.codec</name>
+    <value></value>
+    <description>If the map outputs are compressed, how should they be
+      compressed
+    </description>
+  </property>
+
+<property>
+  <name>mapred.output.compression.type</name>
+  <value>BLOCK</value>
+  <description>If the job outputs are to compressed as SequenceFiles, how should
+               they be compressed? Should be one of NONE, RECORD or BLOCK.
+  </description>
+</property>
+
+
+  <property>
+    <name>mapred.jobtracker.completeuserjobs.maximum</name>
+    <value>5</value>
+  </property>
+
+  <property>
+    <name>mapred.jobtracker.taskScheduler</name>
+    <value></value>
+  </property>
+
+  <property>
+    <name>mapred.jobtracker.restart.recover</name>
+    <value>false</value>
+    <description>"true" to enable (job) recovery upon restart,
+               "false" to start afresh
+    </description>
+  </property>
+
+  <property>
+    <name>mapred.job.reduce.input.buffer.percent</name>
+    <value>0.0</value>
+    <description>The percentage of memory- relative to the maximum heap size- to
+  retain map outputs during the reduce. When the shuffle is concluded, any
+  remaining map outputs in memory must consume less than this threshold before
+  the reduce can begin.
+  </description>
+  </property>
+
+ <property>
+  <name>mapreduce.reduce.input.limit</name>
+  <value>10737418240</value>
+  <description>The limit on the input size of the reduce. (This value
+  is 10 Gb.)  If the estimated input size of the reduce is greater than
+  this value, job is failed. A value of -1 means that there is no limit
+  set. </description>
+</property>
+
+
+  <!-- copied from kryptonite configuration -->
+  <property>
+    <name>mapred.compress.map.output</name>
+    <value></value>
+  </property>
+
+
+  <property>
+    <name>mapred.task.timeout</name>
+    <value>600000</value>
+    <description>The number of milliseconds before a task will be
+  terminated if it neither reads an input, writes an output, nor
+  updates its status string.
+  </description>
+  </property>
+
+  <property>
+    <name>jetty.connector</name>
+    <value>org.mortbay.jetty.nio.SelectChannelConnector</value>
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>mapred.task.tracker.task-controller</name>
+    <value></value>
+   <description>
+     TaskController which is used to launch and manage task execution.
+  </description>
+  </property>
+
+  <property>
+    <name>mapred.child.root.logger</name>
+    <value>INFO,TLA</value>
+  </property>
+
+  <property>
+    <name>mapred.child.java.opts</name>
+    <value></value>
+
+    <description>No description</description>
+  </property>
+
+  <property>
+    <name>mapred.cluster.map.memory.mb</name>
+    <value></value>
+  </property>
+
+  <property>
+    <name>mapred.cluster.reduce.memory.mb</name>
+    <value></value>
+  </property>
+
+  <property>
+    <name>mapred.job.map.memory.mb</name>
+    <value></value>
+  </property>
+
+  <property>
+    <name>mapred.job.reduce.memory.mb</name>
+    <value></value>
+  </property>
+
+  <property>
+    <name>mapred.cluster.max.map.memory.mb</name>
+    <value></value>
+  </property>
+
+  <property>
+    <name>mapred.cluster.max.reduce.memory.mb</name>
+    <value></value>
+  </property>
+
+<property>
+  <name>mapred.hosts</name>
+  <value></value>
+</property>
+
+<property>
+  <name>mapred.hosts.exclude</name>
+  <value></value>
+</property>
+
+<property>
+  <name>mapred.max.tracker.blacklists</name>
+  <value>16</value>
+  <description>
+    if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted
+  </description>
+</property>
+
+<property>
+  <name>mapred.healthChecker.script.path</name>
+  <value></value>
+</property>
+
+<property>
+  <name>mapred.healthChecker.interval</name>
+  <value>135000</value>
+</property>
+
+<property>
+  <name>mapred.healthChecker.script.timeout</name>
+  <value>60000</value>
+</property>
+
+<property>
+  <name>mapred.job.tracker.persist.jobstatus.active</name>
+  <value>false</value>
+  <description>Indicates if persistency of job status information is
+  active or not.
+  </description>
+</property>
+
+<property>
+  <name>mapred.job.tracker.persist.jobstatus.hours</name>
+  <value>1</value>
+  <description>The number of hours job status information is persisted in DFS.
+    The job status information will be available after it drops of the memory
+    queue and between jobtracker restarts. With a zero value the job status
+    information is not persisted at all in DFS.
+  </description>
+</property>
+
+<property>
+  <name>mapred.job.tracker.persist.jobstatus.dir</name>
+  <value></value>
+  <description>The directory where the job status information is persisted
+   in a file system to be available after it drops of the memory queue and
+   between jobtracker restarts.
+  </description>
+</property>
+
+<property>
+  <name>mapred.jobtracker.retirejob.check</name>
+  <value>10000</value>
+</property>
+
+<property>
+  <name>mapred.jobtracker.retirejob.interval</name>
+  <value>0</value>
+</property>
+
+<property>
+  <name>mapred.job.tracker.history.completed.location</name>
+  <value>/mapred/history/done</value>
+  <description>No description</description>
+</property>
+
+<property>
+  <name>mapred.task.maxvmem</name>
+  <value></value>
+  <final>true</final>
+   <description>No description</description>
+</property>
+
+<property>
+  <name>mapred.jobtracker.maxtasks.per.job</name>
+  <value></value>
+  <final>true</final>
+  <description>The maximum number of tasks for a single job.
+  A value of -1 indicates that there is no maximum.  </description>
+</property>
+
+<property>
+  <name>mapreduce.fileoutputcommitter.marksuccessfuljobs</name>
+  <value>false</value>
+</property>
+
+<property>
+  <name>mapred.userlog.retain.hours</name>
+  <value></value>
+</property>
+
+<property>
+  <name>mapred.job.reuse.jvm.num.tasks</name>
+  <value>1</value>
+  <description>
+    How many tasks to run per jvm. If set to -1, there is no limit
+  </description>
+  <final>true</final>
+</property>
+
+<property>
+  <name>mapreduce.jobtracker.kerberos.principal</name>
+  <value></value>
+  <description>
+      JT user name key.
+ </description>
+</property>
+
+<property>
+  <name>mapreduce.tasktracker.kerberos.principal</name>
+   <value></value>
+  <description>
+       tt user name key. "_HOST" is replaced by the host name of the task tracker.
+   </description>
+</property>
+
+
+  <property>
+    <name>hadoop.job.history.user.location</name>
+    <value>none</value>
+    <final>true</final>
+  </property>
+
+
+ <property>
+   <name>mapreduce.jobtracker.keytab.file</name>
+   <value></value>
+   <description>
+       The keytab for the jobtracker principal.
+   </description>
+
+</property>
+
+ <property>
+   <name>mapreduce.tasktracker.keytab.file</name>
+   <value></value>
+    <description>The filename of the keytab for the task tracker</description>
+ </property>
+
+ <property>
+   <name>mapreduce.jobtracker.staging.root.dir</name>
+   <value>/user</value>
+ <description>The Path prefix for where the staging directories should be placed. The next level is always the user's
+   name. It is a path in the default file system.</description>
+ </property>
+
+ <property>
+      <name>mapreduce.tasktracker.group</name>
+      <value>hadoop</value>
+      <description>The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.</description>
+
+ </property>
+
+  <property>
+    <name>mapreduce.jobtracker.split.metainfo.maxsize</name>
+    <value>50000000</value>
+    <final>true</final>
+     <description>If the size of the split metainfo file is larger than this, the JobTracker will fail the job during
+    initialize.
+   </description>
+  </property>
+  <property>
+    <name>mapreduce.history.server.embedded</name>
+    <value>false</value>
+    <description>Should job history server be embedded within Job tracker
+process</description>
+    <final>true</final>
+  </property>
+
+  <property>
+    <name>mapreduce.history.server.http.address</name>
+    <!-- cluster variant -->
+    <value></value>
+    <description>Http address of the history server</description>
+    <final>true</final>
+  </property>
+
+  <property>
+    <name>mapreduce.jobhistory.kerberos.principal</name>
+    <!-- cluster variant -->
+  <value></value>
+    <description>Job history user name key. (must map to same user as JT
+user)</description>
+  </property>
+
+ <property>
+   <name>mapreduce.jobhistory.keytab.file</name>
+    <!-- cluster variant -->
+   <value></value>
+   <description>The keytab for the job history server principal.</description>
+ </property>
+
+<property>
+  <name>mapred.jobtracker.blacklist.fault-timeout-window</name>
+  <value>180</value>
+  <description>
+    3-hour sliding window (value is in minutes)
+  </description>
+</property>
+
+<property>
+  <name>mapred.jobtracker.blacklist.fault-bucket-width</name>
+  <value>15</value>
+  <description>
+    15-minute bucket size (value is in minutes)
+  </description>
+</property>
+
+<property>
+  <name>mapred.queue.names</name>
+  <value>default</value>
+  <description> Comma separated list of queues configured for this jobtracker.</description>
+</property>
+
+</configuration>

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HBASE/configuration/hbase-site.xml Sat May 11 17:25:52 2013
@@ -160,7 +160,7 @@
   </property>
   <property>
     <name>zookeeper.session.timeout</name>
-    <value>60000</value>
+    <value></value>
     <description>ZooKeeper session timeout.
       HBase passes this to the zk quorum as suggested maximum time for a
       session (This setting becomes zookeeper's 'maxSessionTimeout').  See

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/main/resources/stacks/HDPLocal/1.3.0/services/HDFS/configuration/hdfs-site.xml Sat May 11 17:25:52 2013
@@ -412,14 +412,4 @@ don't exist, they will be created with t
   <description>Number of failed disks datanode would tolerate</description>
 </property>
 
-  <property>
-    <name>dfs.namenode.check.stale.datanode</name>
-    <value>true</value>
-    <description>
-      With this setting, the datanodes that have not replied to the heartbeat
-      for more than 30s (i.e. in a stale state) are used for reads only if all
-      other remote replicas have failed.
-    </description>
-  </property>
-
 </configuration>

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatMonitor.java Sat May 11 17:25:52 2013
@@ -18,7 +18,6 @@
 package org.apache.ambari.server.agent;
 
 import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
 import static org.junit.Assert.assertTrue;
 import static org.junit.Assert.fail;
 import static org.mockito.Matchers.eq;
@@ -167,99 +166,6 @@ public class TestHeartbeatMonitor {
     assertTrue("HeartbeatMonitor should not generate StatusCommands for host2 because it has no services", cmds.isEmpty());
   }
 
-  @Test
-  public void testNoStatusCommandForClientComponents() throws Exception {
-    Clusters clusters = injector.getInstance(Clusters.class);
-    clusters.addHost(hostname1);
-    clusters.getHost(hostname1).setOsType("centos6");
-    clusters.getHost(hostname1).persist();
-    clusters.addHost(hostname2);
-    clusters.getHost(hostname2).setOsType("centos6");
-    clusters.getHost(hostname2).persist();
-    clusters.addCluster(clusterName);
-    Cluster cluster = clusters.getCluster(clusterName);
-    cluster.setDesiredStackVersion(new StackId("HDP-0.1"));
-    Set<String> hostNames = new HashSet<String>() {{
-      add(hostname1);
-      add(hostname2);
-    }};
-
-    ConfigFactory configFactory = injector.getInstance(ConfigFactory.class);
-    Config config = configFactory.createNew(cluster, "global",
-      new HashMap<String, String>() {{
-        put("a", "b");
-      }});
-    config.setVersionTag("version1");
-    cluster.addConfig(config);
-    cluster.addDesiredConfig(config);
-
-
-    clusters.mapHostsToCluster(hostNames, clusterName);
-    Service hdfs = cluster.addService(serviceName);
-    hdfs.persist();
-    hdfs.addServiceComponent(Role.DATANODE.name()).persist();
-    hdfs.getServiceComponent(Role.DATANODE.name()).addServiceComponentHost
-      (hostname1).persist();
-    hdfs.addServiceComponent(Role.NAMENODE.name()).persist();
-    hdfs.getServiceComponent(Role.NAMENODE.name()).addServiceComponentHost
-      (hostname1).persist();
-    hdfs.addServiceComponent(Role.SECONDARY_NAMENODE.name()).persist();
-    hdfs.getServiceComponent(Role.SECONDARY_NAMENODE.name()).
-      addServiceComponentHost(hostname1).persist();
-    hdfs.addServiceComponent(Role.HDFS_CLIENT.name()).persist();
-    hdfs.getServiceComponent(Role.HDFS_CLIENT.name()).addServiceComponentHost
-      (hostname1).persist();
-    hdfs.getServiceComponent(Role.HDFS_CLIENT.name()).addServiceComponentHost
-      (hostname2).persist();
-
-    ActionQueue aq = new ActionQueue();
-    ActionManager am = mock(ActionManager.class);
-    HeartbeatMonitor hm = new HeartbeatMonitor(clusters, aq, am,
-      heartbeatMonitorWakeupIntervalMS);
-    HeartBeatHandler handler = new HeartBeatHandler(clusters, aq, am, injector);
-    Register reg = new Register();
-    reg.setHostname(hostname1);
-    reg.setResponseId(12);
-    reg.setTimestamp(System.currentTimeMillis() - 300);
-    reg.setAgentVersion(ambariMetaInfo.getServerVersion());
-    HostInfo hi = new HostInfo();
-    hi.setOS("Centos5");
-    reg.setHardwareProfile(hi);
-    handler.handleRegistration(reg);
-
-    HeartBeat hb = new HeartBeat();
-    hb.setHostname(hostname1);
-    hb.setNodeStatus(new HostStatus(HostStatus.Status.HEALTHY, "cool"));
-    hb.setTimestamp(System.currentTimeMillis());
-    hb.setResponseId(12);
-    handler.handleHeartBeat(hb);
-
-    List<StatusCommand> cmds = hm.generateStatusCommands(hostname1);
-    assertTrue("HeartbeatMonitor should generate StatusCommands for host1",
-      cmds.size() == 3);
-    assertEquals("HDFS", cmds.get(0).getServiceName());
-    boolean containsDATANODEStatus = false;
-    boolean containsNAMENODEStatus = false;
-    boolean containsSECONDARY_NAMENODEStatus = false;
-    boolean containsHDFS_CLIENTStatus = false;
-    for (StatusCommand cmd : cmds) {
-      containsDATANODEStatus |= cmd.getComponentName().equals("DATANODE");
-      containsNAMENODEStatus |= cmd.getComponentName().equals("NAMENODE");
-      containsSECONDARY_NAMENODEStatus |= cmd.getComponentName().
-        equals("SECONDARY_NAMENODE");
-      containsHDFS_CLIENTStatus |= cmd.getComponentName().equals
-        ("HDFS_CLIENT");
-      assertTrue(cmd.getConfigurations().size() > 0);
-    }
-    assertTrue(containsDATANODEStatus);
-    assertTrue(containsNAMENODEStatus);
-    assertTrue(containsSECONDARY_NAMENODEStatus);
-    assertFalse(containsHDFS_CLIENTStatus);
-
-    cmds = hm.generateStatusCommands(hostname2);
-    assertTrue("HeartbeatMonitor should not generate StatusCommands for host2" +
-      " because it has only client components", cmds.isEmpty());
-  }
 
   @Test
   public void testHeartbeatStateCommandsEnqueueing() throws AmbariException, InterruptedException,

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariManagementControllerTest.java Sat May 11 17:25:52 2013
@@ -36,7 +36,6 @@ import org.apache.ambari.server.AmbariEx
 import org.apache.ambari.server.ClusterNotFoundException;
 import org.apache.ambari.server.DuplicateResourceException;
 import org.apache.ambari.server.ParentObjectNotFoundException;
-import org.apache.ambari.server.ObjectNotFoundException;
 import org.apache.ambari.server.Role;
 import org.apache.ambari.server.RoleCommand;
 import org.apache.ambari.server.StackAccessException;
@@ -75,9 +74,7 @@ import org.apache.ambari.server.state.sv
 import org.apache.ambari.server.utils.StageUtils;
 import org.junit.After;
 import org.junit.Before;
-import org.junit.Rule;
 import org.junit.Test;
-import org.junit.rules.ExpectedException;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 import com.google.inject.Guice;
@@ -122,9 +119,6 @@ public class AmbariManagementControllerT
   private Users users;
   private EntityManager entityManager;
 
-  @Rule
-  public ExpectedException expectedException = ExpectedException.none();
-
   @Before
   public void setup() throws Exception {
     injector = Guice.createInjector(new InMemoryDefaultTestModule());
@@ -6048,128 +6042,4 @@ public class AmbariManagementControllerT
     }
   }
 
-  @Test
-  public void testGetTasksByRequestId() throws AmbariException {
-    final long requestId1 = 1;
-    final long requestId2 = 2;
-    final String clusterName = "c1";
-    final String hostName1 = "h1";
-    final String context = "Test invocation";
-
-    clusters.addCluster(clusterName);
-    clusters.getCluster(clusterName).setDesiredStackVersion(new StackId("HDP-0.1"));
-    clusters.addHost(hostName1);
-    clusters.getHost("h1").setOsType("centos5");
-    clusters.getHost(hostName1).persist();
-
-    clusters.mapHostsToCluster(new HashSet<String>(){
-      {add(hostName1);}}, clusterName);
-
-
-    List<Stage> stages = new ArrayList<Stage>();
-    stages.add(new Stage(requestId1, "/a1", clusterName, context));
-    stages.get(0).setStageId(1);
-    stages.get(0).addHostRoleExecutionCommand(hostName1, Role.HBASE_MASTER,
-            RoleCommand.START,
-            new ServiceComponentHostStartEvent(Role.HBASE_MASTER.toString(),
-                    hostName1, System.currentTimeMillis(),
-                    new HashMap<String, String>()),
-            clusterName, "HBASE");
-
-    stages.add(new Stage(requestId1, "/a2", clusterName, context));
-    stages.get(1).setStageId(2);
-    stages.get(1).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
-            RoleCommand.START,
-            new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
-                    hostName1, System.currentTimeMillis(),
-                    new HashMap<String, String>()), clusterName, "HBASE");
-
-    stages.add(new Stage(requestId1, "/a3", clusterName, context));
-    stages.get(2).setStageId(3);
-    stages.get(2).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
-            RoleCommand.START,
-            new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
-                    hostName1, System.currentTimeMillis(),
-                    new HashMap<String, String>()), clusterName, "HBASE");
-
-
-    stages.add(new Stage(requestId2, "/a4", clusterName, context));
-    stages.get(3).setStageId(4);
-    stages.get(3).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
-            RoleCommand.START,
-            new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
-                    hostName1, System.currentTimeMillis(),
-                    new HashMap<String, String>()), clusterName, "HBASE");
-
-    stages.add(new Stage(requestId2, "/a5", clusterName, context));
-    stages.get(4).setStageId(5);
-    stages.get(4).addHostRoleExecutionCommand(hostName1, Role.HBASE_CLIENT,
-            RoleCommand.START,
-            new ServiceComponentHostStartEvent(Role.HBASE_CLIENT.toString(),
-                    hostName1, System.currentTimeMillis(),
-                    new HashMap<String, String>()), clusterName, "HBASE");
-
-    actionDB.persistActions(stages);
-
-    Set<TaskStatusRequest> taskStatusRequests;
-    Set<TaskStatusResponse> taskStatusResponses;
-
-    //check count of tasks by requestId1
-    taskStatusRequests = new HashSet<TaskStatusRequest>(){
-      {
-        add(new TaskStatusRequest(requestId1, null));
-      }
-    };
-    taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
-    assertEquals(3, taskStatusResponses.size());
-
-    //check a taskId that requested by requestId1 and task id
-    taskStatusRequests = new HashSet<TaskStatusRequest>(){
-      {
-        add(new TaskStatusRequest(requestId1, 2L));
-      }
-    };
-    taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
-    assertEquals(1, taskStatusResponses.size());
-    assertEquals(2L, taskStatusResponses.iterator().next().getTaskId());
-
-    //check count of tasks by requestId2
-    taskStatusRequests = new HashSet<TaskStatusRequest>(){
-      {
-        add(new TaskStatusRequest(requestId2, null));
-      }
-    };
-    taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
-    assertEquals(2, taskStatusResponses.size());
-
-    //check a taskId that requested by requestId2 and task id
-    taskStatusRequests = new HashSet<TaskStatusRequest>(){
-      {
-        add(new TaskStatusRequest(requestId2, 5L));
-      }
-    };
-    taskStatusResponses = controller.getTaskStatus(taskStatusRequests);
-    assertEquals(5L, taskStatusResponses.iterator().next().getTaskId());
-
-    //verify that task from second request (requestId2) does not present in first request (requestId1)
-    taskStatusRequests = new HashSet<TaskStatusRequest>(){
-      {
-        add(new TaskStatusRequest(requestId1, 5L));
-      }
-    };
-    expectedException.expect(ObjectNotFoundException.class);
-    expectedException.expectMessage("Task resource doesn't exist.");
-    controller.getTaskStatus(taskStatusRequests);
-
-    //verify that task from first request (requestId1) does not present in second request (requestId2)
-    taskStatusRequests = new HashSet<TaskStatusRequest>(){
-      {
-        add(new TaskStatusRequest(requestId2, 2L));
-      }
-    };
-    expectedException.expect(ObjectNotFoundException.class);
-    expectedException.expectMessage("Task resource doesn't exist.");
-    controller.getTaskStatus(taskStatusRequests);
-  }
-
 }

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/java/org/apache/ambari/server/security/authorization/AmbariLdapAuthenticationProviderTest.java Sat May 11 17:25:52 2013
@@ -97,38 +97,6 @@ public class AmbariLdapAuthenticationPro
     Assert.assertTrue(auth == null);
   }
 
-  @Test
-  public void testLdapAdminGroupToRolesMapping() throws Exception {
-
-    Authentication authentication;
-
-    authentication =
-        new UsernamePasswordAuthenticationToken("allowedAdmin", "password");
-    Authentication result = authenticationProvider.authenticate(authentication);
-    assertTrue(result.isAuthenticated());
-
-    UserEntity allowedAdminEntity = userDAO.findLdapUserByName("allowedAdmin");
-
-    authentication =
-        new UsernamePasswordAuthenticationToken("allowedUser", "password");
-    authenticationProvider.authenticate(authentication);
-    UserEntity allowedUserEntity = userDAO.findLdapUserByName("allowedUser");
-
-
-    RoleEntity adminRole = roleDAO.findByName(
-        configuration.getConfigsMap().get(Configuration.ADMIN_ROLE_NAME_KEY));
-    RoleEntity userRole = roleDAO.findByName(
-        configuration.getConfigsMap().get(Configuration.USER_ROLE_NAME_KEY));
-
-
-    assertTrue(allowedAdminEntity.getRoleEntities().contains(userRole));
-    assertTrue(allowedAdminEntity.getRoleEntities().contains(adminRole));
-
-    assertTrue(allowedUserEntity.getRoleEntities().contains(userRole));
-    assertFalse(allowedUserEntity.getRoleEntities().contains(adminRole));
-
-
-  }
 
   @AfterClass
   public static void afterClass() {

Modified: incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-server/src/test/resources/test_api.sh Sat May 11 17:25:52 2013
@@ -21,18 +21,18 @@ curl -i -X POST http://localhost:8080/ap
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HBASE
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/GANGLIA
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/NAGIOS
-
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "core-site", "tag": "version1", "properties" : { "fs.default.name" : "localhost:8020"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "core-site", "tag": "version2", "properties" : { "fs.default.name" : "localhost:8020"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "core-site", "tag": "version1"}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "hdfs-site", "tag": "version1", "properties" : { "dfs.datanode.data.dir.perm" : "750"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "global", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "mapred-site", "tag": "version1", "properties" : { "mapred.job.tracker" : "localhost:50300", "mapreduce.history.server.embedded": "false", "mapreduce.history.server.http.address": "localhost:51111"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "hbase-site", "tag": "version1", "properties" : { "hbase.rootdir" : "hdfs://localhost:8020/apps/hbase/", "hbase.cluster.distributed" : "true", "hbase.zookeeper.quorum": "localhost", "zookeeper.session.timeout": "60000" }}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "hbase-env", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "nagios-global", "tag": "version2", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password", "nagios_contact": "a\u0040b.c" }}}}' http://localhost:8080/api/v1/clusters/c1
-curl -i -X PUT -d '{"Clusters": {"desired_config": {"type": "nagios-global", "tag": "version1", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password"  }}}}' http://localhost:8080/api/v1/clusters/c1
-
+curl -i -X POST -d '{"type": "core-site", "tag": "version1", "properties" : { "fs.default.name" : "localhost:8020"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "hdfs-site", "tag": "version1", "properties" : { "dfs.datanode.data.dir.perm" : "750"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "global", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "mapred-site", "tag": "version1", "properties" : { "mapred.job.tracker" : "localhost:50300", "mapreduce.history.server.embedded": "false", "mapreduce.history.server.http.address": "localhost:51111"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "hbase-site", "tag": "version1", "properties" : { "hbase.rootdir" : "hdfs://localhost:8020/apps/hbase/", "hbase.cluster.distributed" : "true", "hbase.zookeeper.quorum": "localhost", "zookeeper.session.timeout": "60000" }}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "hbase-env", "tag": "version1", "properties" : { "hbase_hdfs_root_dir" : "/apps/hbase/"}}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "nagios-global", "tag": "version2", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password", "nagios_contact": "a\u0040b.c" }}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X POST -d '{"type": "nagios-global", "tag": "version1", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password"  }}' http://localhost:8080/api/v1/clusters/c1/configurations
+curl -i -X PUT -d '{"config": {"core-site": "version1", "hdfs-site": "version1", "global" : "version1" }}'  http://localhost:8080/api/v1/clusters/c1/services/HDFS
+curl -i -X PUT -d '{"config": {"core-site": "version1", "mapred-site": "version1"}}'  http://localhost:8080/api/v1/clusters/c1/services/MAPREDUCE
+curl -i -X PUT -d '{"config": {"hbase-site": "version1", "hbase-env": "version1"}}'  http://localhost:8080/api/v1/clusters/c1/services/HBASE
+curl -i -X PUT -d '{"config": {"nagios-global": "version2" }}'  http://localhost:8080/api/v1/clusters/c1/services/NAGIOS
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HDFS/components/NAMENODE
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HDFS/components/SECONDARY_NAMENODE
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/services/HDFS/components/DATANODE
@@ -60,7 +60,6 @@ curl -i -X POST http://localhost:8080/ap
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST/host_components/HBASE_REGIONSERVER
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST/host_components/HBASE_CLIENT
 curl -i -X POST http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST/host_components/NAGIOS_SERVER
-curl -i -X PUT -d '{"Hosts": {"desired_config": {"type": "core-site", "tag": "version3", "properties" : { "nagios_web_login" : "nagiosadmin", "nagios_web_password" : "password"  }}}}' http://localhost:8080/api/v1/clusters/c1/hosts/$AGENT_HOST
 curl -i -X PUT  -d '{"ServiceInfo": {"state" : "INSTALLED"}}'   http://localhost:8080/api/v1/clusters/c1/services?state=INIT
 #curl -i -X PUT  -d '{"ServiceInfo": {"state" : "STARTED"}}'   http://localhost:8080/api/v1/clusters/c1/services?state=INSTALLED
 # http://localhost:8080/api/v1/clusters/c1/requests/2

Modified: incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-web/app/config.js Sat May 11 17:25:52 2013
@@ -24,8 +24,8 @@ App.skipBootstrap = false;
 App.alwaysGoToInstaller = false;
 App.testEnableSecurity = true; // By default enable security is tested; turning it false tests disable security
 App.apiPrefix = '/api/v1';
-App.defaultStackVersion = 'HDP-1.3.0';
-App.defaultLocalStackVersion = 'HDPLocal-1.3.0';
+App.defaultStackVersion = 'HDP-1.2.1';
+App.defaultLocalStackVersion = 'HDPLocal-1.2.1';
 App.defaultJavaHome = '/usr/jdk/jdk1.6.0_31';
 App.timeout = 180000; // default AJAX timeout
 App.maxRetries = 3; // max number of retries for certain AJAX calls
@@ -50,7 +50,7 @@ App.supports = {
   hiveOozieExtraDatabases: false,
   multipleHBaseMasters: false,
   addMasters: false,
-  customizeSmokeTestUser: true,
+  customizeSmokeTestUser: false,
   hue: false,
   ldapGroupMapping: false
 };

Modified: incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-web/app/controllers/main/admin/user.js Sat May 11 17:25:52 2013
@@ -39,10 +39,11 @@ App.MainAdminUserController = Em.Control
 
       return;
     }
+    ;
 
     App.ModalPopup.show({
       header:Em.I18n.t('admin.users.delete.header').format(event.context.get('userName')),
-      body:Em.I18n.t('question.sure').format(''),
+      body:Em.I18n.t('question.sure'),
       primary:Em.I18n.t('yes'),
       secondary:Em.I18n.t('no'),
 

Modified: incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/ambari-web/pom.xml Sat May 11 17:25:52 2013
@@ -19,7 +19,7 @@
   <parent>
     <groupId>org.apache.ambari</groupId>
     <artifactId>ambari-project</artifactId>
-    <version>1.3.0-SNAPSHOT</version>
+    <version>1.2.3-SNAPSHOT</version>
     <relativePath>../ambari-project</relativePath>
   </parent>
   <modelVersion>4.0.0</modelVersion>
@@ -27,7 +27,7 @@
   <artifactId>ambari-web</artifactId>
   <packaging>pom</packaging>
   <name>Ambari Web</name>
-  <version>1.3.0-SNAPSHOT</version>
+  <version>1.2.3-SNAPSHOT</version>
   <description>Ambari Web</description>
   <properties>
     <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>

Modified: incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php (original)
+++ incubator/ambari/branches/branch-1.2.3/contrib/addons/src/addOns/nagios/scripts/nagios_alerts.php Sat May 11 17:25:52 2013
@@ -381,11 +381,11 @@ function hdp_mon_generate_response( $res
         $pieces[0] = "HIVE";
         break;
       case "ZKSERVERS":
-	    $pieces[0] = "ZOOKEEPER";
+	      $pieces[0] = "ZOOKEEPER";
         break;
       case "AMBARI":
-	    $pieces[0] = "AMBARI";
-      break;      
+	      $pieces[0] = "AMBARI";
+        break;
       case "NAGIOS":
       case "HDFS":
       case "MAPREDUCE":

Modified: incubator/ambari/branches/branch-1.2.3/docs/pom.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/pom.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/pom.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/pom.xml Sat May 11 17:25:52 2013
@@ -28,7 +28,7 @@
     <modelVersion>4.0.0</modelVersion>
 
     <groupId>org.apache.ambari</groupId>
-    <version>1.2.2-SNAPSHOT</version>
+    <version>1.2.0-SNAPSHOT</version>
     <artifactId>ambari</artifactId>
     <packaging>pom</packaging>
 
@@ -118,18 +118,6 @@
                 <role>PMC</role>
             </roles>
         </developer>
-        <developer>
-            <id>billie</id>
-            <name>Billie Rinaldi</name>
-            <email>billie@apache.org</email>
-            <timezone>-8</timezone>
-            <roles>
-                <role>Committer</role>
-            </roles>
-            <organization>
-                Hortonworks
-            </organization>
-        </developer>
         <developer>            
             <id>ddas</id>
             <name>Devaraj Das</name>
@@ -237,19 +225,7 @@
             <organization>
                 Hortonworks                
             </organization>            
-        </developer>
-        <developer>
-            <id>ncole</id>
-            <name>Nate Cole</name>
-            <email>ncole@apache.org</email>
-            <timezone>-8</timezone>
-            <roles>
-                <role>Committer</role>
-            </roles>
-            <organization>
-                Hortonworks
-            </organization>
-        </developer>
+        </developer>              
         <developer>
             <id>ramya</id>
             <name>Ramya Sunil</name>
@@ -275,30 +251,6 @@
             </organization>            
         </developer>
         <developer>
-            <id>smohanty</id>
-            <name>Sumit Mohanty</name>
-            <email>smohanty@apache.org</email>
-            <timezone>-8</timezone>
-            <roles>
-                <role>Committer</role>
-            </roles>
-            <organization>
-                Hortonworks
-            </organization>
-        </developer>
-        <developer>
-            <id>swagle</id>
-            <name>Siddharth Wagle</name>
-            <email>swagle@apache.org</email>
-            <timezone>-8</timezone>
-            <roles>
-                <role>Committer</role>
-            </roles>
-            <organization>
-                Hortonworks
-            </organization>
-        </developer>
-        <developer>
             <id>srimanth</id>
             <name>Srimanth Gunturi</name>
             <email>srimanth@apache.org</email>

Modified: incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/index.apt Sat May 11 17:25:52 2013
@@ -15,20 +15,26 @@
 ~~
 Introduction
 
-  The Apache Ambari project is aimed at making Hadoop management simpler by developing software for provisioning, managing, and monitoring Apache Hadoop clusters.
-  Ambari provides an intuitive, easy-to-use Hadoop management web UI backed by its RESTful APIs.
+  Apache Ambari is a web-based tool for provisioning, managing, and monitoring Apache Hadoop clusters. The set of
+  Hadoop components that are currently supported by Ambari includes:
 
-  The set of Hadoop components that are currently supported by Ambari includes:
+  * {{{http://hadoop.apache.org/docs/hdfs} Apache Hadoop - HDFS}}
 
-  {{{http://hadoop.apache.org/docs/hdfs} HDFS}},
-  {{{http://hadoop.apache.org/docs/mapreduce} MapReduce}}
-  {{{http://hive.apache.org} Hive}},
-  {{{http://incubator.apache.org/hcatalog} HCatalog}},
-  {{{http://hbase.apache.org} HBase}},
-  {{{http://zookeeper.apache.org} ZooKeeper}},
-  {{{http://incubator.apache.org/oozie/} Oozie}},
-  {{{http://pig.apache.org} Pig}},
-  {{{http://sqoop.apache.org} Sqoop}}
+  * {{{http://hadoop.apache.org/docs/mapreduce} Apache Hadoop - MapReduce}}
+
+  * {{{http://hive.apache.org} Apache Hive}}
+
+  * {{{http://incubator.apache.org/hcatalog} Apache HCatalog}}
+
+  * {{{http://hbase.apache.org} Apache HBase}}
+
+  * {{{http://zookeeper.apache.org} Apache Zookeeper}}
+
+  * {{{http://incubator.apache.org/oozie/} Apache Oozie}}
+
+  * {{{http://pig.apache.org} Apache Pig}}
+
+  * {{{http://sqoop.apache.org} Apache Sqoop}}
 
   []
 
@@ -36,7 +42,7 @@ Introduction
 
   * Provision a Hadoop Cluster
   
-    * Ambari provides a step-by-step wizard for installing Hadoop services across any number of hosts.
+    * Ambari provides an easy-to-use, step-by-step wizard for installing Hadoop services across any number of hosts.
     
     * Ambari handles configuration of Hadoop services for the cluster.
 
@@ -55,31 +61,52 @@ Introduction
     * Ambari leverages {{{http://ganglia.sourceforge.net/} Ganglia}} for metrics collection.
 
     * Ambari leverages {{{http://www.nagios.org/} Nagios}} for system alerting and will send emails when your attention is needed (e.g., a node goes down, remaining disk space is low, etc).
+  
+  []
+
+Ambari Source
+
+  Follow the  {{{./1.2.0/installing-hadoop-using-ambari/content/index.html} installation guide for 1.2.0 (stable)}} or check out the work going on in {{{./whats-new.html} trunk}} for the upcoming 1.2.1 release.
+
+Roadmap
+
+  * Support for additional Operating Systems
+
+    * Ambari currently supports 64-bit RHEL/CentOS 5 + 6 and SLES 11
 
   []
 
-  Ambari enables Application Developers and System Integrators to:
+  * RESTful API for integration
+  
+    * Ambari will expose a unified, RESTful API to enable third-party applications to integrate
+      Hadoop cluster management and monitoring capabilities (1.2.0)
+
+  []
 
-  * Easily integrate Hadoop provisioning, management, and monitoring capabilities to their own applications with the {{{https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md} Ambari REST APIs}}.
+  * Granular configurations
 
-Getting Started with Ambari
+      * Ambari currently applies service configurations at the cluster-level. For more
+        flexibility, Ambari will allow for configurations in a more granular manner, such as
+        applying a set of configurations to a specific group of hosts.
 
-  Follow the {{{./1.2.2/installing-hadoop-using-ambari/content/index.html} installation guide for Ambari 1.2.2}}.
+  []
 
-  Note: Ambari currently supports 64-bit RHEL/CentOS 5 + 6 and SLES 11.
+  * Security
 
-Get Involved
+      * Installation of secure Hadoop clusters (Kerberos-based)
 
-  Visit the {{{https://cwiki.apache.org/confluence/display/AMBARI/Ambari} Ambari Wiki}} for design documents, roadmap, development guidelines, etc.
+      * Role-based user authentication, authorization, and auditing
 
-  The first {{{http://www.meetup.com/Apache-Ambari-User-Group} Ambari User Group Meetup}} took place on April 2 in Palo Alto, California, USA.  {{{http://www.meetup.com/Apache-Ambari-User-Group/events/109316812/} See the slides and WebEx session from the Meetup}}.  
+      * Support for LDAP and Active Directory (1.2.0)
 
-What's New?
+  []
 
-  Check out the work going on for the {{{./whats-new.html} upcoming 1.3.0 release}}.
+  * Visualization
 
-Disclaimer
+      * Interactive visualization of the current state of the cluster for a number of key metrics (1.2.0)
 
-  Apache Ambari is an effort undergoing incubation at The Apache Software Foundation (ASF) sponsored by the Apache Incubator PMC. Incubation is required of all newly accepted projects until a further review indicates that the infrastructure, communications, and decision making process have stabilized in a manner consistent with other successful ASF projects. While incubation status is not necessarily a reflection of the completeness or stability of the code, it does indicate that the project has yet to be fully endorsed by the ASF.
+      * Interactive visualization of historical states of the cluster for a number of key metrics
 
+      * Interactive visualization of Pig, Hive, and MapReduce jobs (1.2.0)
 
+  []

Modified: incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/src/site/apt/whats-new.apt Sat May 11 17:25:52 2013
@@ -14,9 +14,28 @@
 ~~ limitations under the License.
 ~~
 
-What's New in Ambari 1.3.0?
+What's New?
 
- * Check back later for the new features in 1.3.0. 
+  There’s a lot of activity occurring in trunk. We’ve had a great response from the
+  community and received a lot of input into the direction. To date, we have added
+  many new features. To highlight a few...
+
+  * Added more options on provisioning so you can better target your cluster
+    setup for your needs
+
+  * More service controls, including the ability to run ad hoc smoke tests
+
+  * Exposed a RESTful API for cluster metrics (with management and
+    provisioning to follow)
+
+  * Moved to a JavaScript front-end for Ambari Web and refreshed the entire UI
+    experience
+
+  * Used that RESTful API and added more charts and heatmaps to the dashboard
+
+  * Reduced the number of dependencies on the Ambari Server + Ambari Agent
+
+  * Added support for SLES 11
 
 Getting Ambari Source
 
@@ -28,14 +47,14 @@ $ svn checkout http://svn.apache.org/rep
 
 JIRA
 
-  You can follow all the work going on by watching the Ambari JIRA project. You can see the
-  JIRA details for Ambari 1.3.0 here:
+  You can follow all the work going on by watching the Ambari JIRA project. The next release of Apache Ambari will be version 1.2.1. You can see the
+  JIRA details for this release here:
 
-  {{{https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.3.0%22%20AND%20project%20%3D%20AMBARI} https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.3.0%22%20AND%20project%20%3D%20AMBARI}}
+  {{{https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.2.1%22%20AND%20project%20%3D%20AMBARI} https://issues.apache.org/jira/issues/?jql=fixVersion%20%3D%20%221.2.1%22%20AND%20project%20%3D%20AMBARI}}
 
 User Guide
 
-  Take a look at {{{http://incubator.apache.org/ambari/1.2.2/installing-hadoop-using-ambari/content/index.html} how to install a Hadoop cluster using Ambari 1.2.2}}.
+  {{{http://incubator.apache.org/ambari/1.2.1/installing-hadoop-using-ambari/content/index.html} See how to install a Hadoop cluster using Ambari 1.2.1}}
 
 Stay Tuned
 

Modified: incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml
URL: http://svn.apache.org/viewvc/incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml?rev=1481380&r1=1481379&r2=1481380&view=diff
==============================================================================
--- incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml (original)
+++ incubator/ambari/branches/branch-1.2.3/docs/src/site/site.xml Sat May 11 17:25:52 2013
@@ -28,9 +28,7 @@
   <custom>
     <fluidoSkin>
       <topBarEnabled>true</topBarEnabled>
-      <sideBarEnabled>true</sideBarEnabled>
-      <leftColumnClass>span2</leftColumnClass>
-      <bodyColumnClass>span10</bodyColumnClass>
+      <sideBarEnabled>false</sideBarEnabled>
     </fluidoSkin>
   </custom>
   <bannerLeft>
@@ -61,62 +59,38 @@
           })();
        </script>
        <!-- End of Google analytics -->
-       <style>
-           .well.sidebar-nav {
-             background-color: #fff;
-           }
-           a.externalLink[href^=http] {
-             background-image: none;
-             padding-right: 0;
-           }
-           body.topBarEnabled {
-             padding-top: 40px;
-           }
-           #leftColumn .nav-list .active a {
-             background-color: #a0a0a0;
-           }
-           .nav-list .active a:hover {
-             background-color: #a0a0a0;
-           }
-       </style>
     </head>
 
     <links>
       <item name="JIRA" href="https://issues.apache.org/jira/browse/AMBARI" />
       <item name="SVN" href="https://svn.apache.org/repos/asf/incubator/ambari/" />
-      <item name="Wiki" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari" />
+      <item name="WIKI" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari" />
     </links>
 
     <breadcrumbs>
-      <item name="Incubator" href="../"/>
-      <item name="Ambari" href="http://incubator.apache.org/ambari/"/>
+      <item name="Ambari" href="index.html"/>
     </breadcrumbs>
 
     <menu name="Ambari">
-      <item name="Overview" href="index.html"/>
       <item name="What's New?" href="whats-new.html"/>
+      <item name="About" href="index.html"/>
+      <item name="Wiki" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari"/>
       <item name="Project Team" href="team-list.html"/>
-      <item name="IRC Channel" href="irc.html"/>
       <item name="Mailing Lists" href="mail-lists.html"/>
       <item name="Issue Tracking" href="issue-tracking.html"/>
-      <item name="User Group" href="http://www.meetup.com/Apache-Ambari-User-Group"/>
-      <item name="Project License" href="http://www.apache.org/licenses/"/>
+      <item name="Project License" href="license.html"/>
     </menu>
 
     <menu name="Releases">
-        <item name="1.2.2" href="http://www.apache.org/dist/incubator/ambari/ambari-1.2.2/"/>
-        <item name="1.2.1" href="http://www.apache.org/dist/incubator/ambari/ambari-1.2.1/"/>
+        <item name="1.2.1 (coming soon)" href="whats-new.html"/>
         <item name="1.2.0" href="http://www.apache.org/dist/incubator/ambari/ambari-1.2.0/"/>
         <item name="0.9" href="http://www.apache.org/dist/incubator/ambari/ambari-0.9-incubating/"/>
     </menu>
 
     <menu name="Documentation">
-        <item name="Install Guide for 1.2.2" href="1.2.2/installing-hadoop-using-ambari/content/index.html"/>
         <item name="Install Guide for 1.2.1" href="1.2.1/installing-hadoop-using-ambari/content/index.html"/>
         <item name="Install Guide for 1.2.0" href="1.2.0/installing-hadoop-using-ambari/content/index.html"/>
         <item name="Install Guide for 0.9" href="install-0.9.html"/>
-        <item name="API Reference" href="https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md"/>
-        <item name="Wiki" href="https://cwiki.apache.org/confluence/display/AMBARI/Ambari"/>
     </menu>
 
     <footer>