You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by ao...@apache.org on 2015/05/15 12:54:42 UTC

ambari git commit: AMBARI-11163. Some upgrade paths are not using su with correct parameters (aonishuk)

Repository: ambari
Updated Branches:
  refs/heads/trunk 8f7fb3d26 -> e10531e27


AMBARI-11163. Some upgrade paths are not using su with correct parameters (aonishuk)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/e10531e2
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/e10531e2
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/e10531e2

Branch: refs/heads/trunk
Commit: e10531e2720bb30f15468d19d4942c4269ee1771
Parents: 8f7fb3d
Author: Andrew Onishuk <ao...@hortonworks.com>
Authored: Fri May 15 13:54:21 2015 +0300
Committer: Andrew Onishuk <ao...@hortonworks.com>
Committed: Fri May 15 13:54:21 2015 +0300

----------------------------------------------------------------------
 .../HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py       | 8 ++++----
 1 file changed, 4 insertions(+), 4 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/e10531e2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
index ddd81fc..bc1e3ad 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
@@ -44,11 +44,11 @@ def reach_safemode_state(user, safemode_state, in_ha):
   original_state = SafeMode.UNKNOWN
 
   hostname = params.hostname
-  safemode_check = format("su - {user} -c 'hdfs dfsadmin -safemode get'")
+  safemode_check = format("hdfs dfsadmin -safemode get")
 
   grep_pattern = format("Safe mode is {safemode_state} in {hostname}") if in_ha else format("Safe mode is {safemode_state}")
-  safemode_check_with_grep = format("su - {user} -c 'hdfs dfsadmin -safemode get | grep \"{grep_pattern}\"'")
-  code, out = shell.call(safemode_check)
+  safemode_check_with_grep = format("hdfs dfsadmin -safemode get | grep '{grep_pattern}'")
+  code, out = shell.call(safemode_check, user=user)
   Logger.info("Command: %s\nCode: %d." % (safemode_check, code))
   if code == 0 and out is not None:
     Logger.info(out)
@@ -67,7 +67,7 @@ def reach_safemode_state(user, safemode_state, in_ha):
                 logoutput=True,
                 path=[params.hadoop_bin_dir])
 
-        code, out = call(safemode_check_with_grep)
+        code, out = call(safemode_check_with_grep, user=user)
         Logger.info("Command: %s\nCode: %d. Out: %s" % (safemode_check_with_grep, code, out))
         if code == 0:
           return (True, original_state)