You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by ao...@apache.org on 2015/05/15 12:54:42 UTC
ambari git commit: AMBARI-11163. Some upgrade paths are not using su
with correct parameters (aonishuk)
Repository: ambari
Updated Branches:
refs/heads/trunk 8f7fb3d26 -> e10531e27
AMBARI-11163. Some upgrade paths are not using su with correct parameters (aonishuk)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/e10531e2
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/e10531e2
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/e10531e2
Branch: refs/heads/trunk
Commit: e10531e2720bb30f15468d19d4942c4269ee1771
Parents: 8f7fb3d
Author: Andrew Onishuk <ao...@hortonworks.com>
Authored: Fri May 15 13:54:21 2015 +0300
Committer: Andrew Onishuk <ao...@hortonworks.com>
Committed: Fri May 15 13:54:21 2015 +0300
----------------------------------------------------------------------
.../HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py | 8 ++++----
1 file changed, 4 insertions(+), 4 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/e10531e2/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
index ddd81fc..bc1e3ad 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/namenode_upgrade.py
@@ -44,11 +44,11 @@ def reach_safemode_state(user, safemode_state, in_ha):
original_state = SafeMode.UNKNOWN
hostname = params.hostname
- safemode_check = format("su - {user} -c 'hdfs dfsadmin -safemode get'")
+ safemode_check = format("hdfs dfsadmin -safemode get")
grep_pattern = format("Safe mode is {safemode_state} in {hostname}") if in_ha else format("Safe mode is {safemode_state}")
- safemode_check_with_grep = format("su - {user} -c 'hdfs dfsadmin -safemode get | grep \"{grep_pattern}\"'")
- code, out = shell.call(safemode_check)
+ safemode_check_with_grep = format("hdfs dfsadmin -safemode get | grep '{grep_pattern}'")
+ code, out = shell.call(safemode_check, user=user)
Logger.info("Command: %s\nCode: %d." % (safemode_check, code))
if code == 0 and out is not None:
Logger.info(out)
@@ -67,7 +67,7 @@ def reach_safemode_state(user, safemode_state, in_ha):
logoutput=True,
path=[params.hadoop_bin_dir])
- code, out = call(safemode_check_with_grep)
+ code, out = call(safemode_check_with_grep, user=user)
Logger.info("Command: %s\nCode: %d. Out: %s" % (safemode_check_with_grep, code, out))
if code == 0:
return (True, original_state)