You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by ao...@apache.org on 2016/06/21 14:29:10 UTC

[1/2] ambari git commit: AMBARI-17337. Client installs failed on debian 7 (aonishuk)

Repository: ambari
Updated Branches:
  refs/heads/branch-2.4 3fac4f7e6 -> 3ce7d10b7
  refs/heads/trunk a6de04361 -> 4f78290a6


AMBARI-17337. Client installs failed on debian 7 (aonishuk)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/4f78290a
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/4f78290a
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/4f78290a

Branch: refs/heads/trunk
Commit: 4f78290a627f642b41e7ee1d75984f06eca90723
Parents: a6de043
Author: Andrew Onishuk <ao...@hortonworks.com>
Authored: Tue Jun 21 17:29:03 2016 +0300
Committer: Andrew Onishuk <ao...@hortonworks.com>
Committed: Tue Jun 21 17:29:03 2016 +0300

----------------------------------------------------------------------
 .../resource_management/libraries/functions/packages_analyzer.py | 2 +-
 .../common-services/HDFS/2.1.0.2.0/package/scripts/utils.py      | 4 ++--
 2 files changed, 3 insertions(+), 3 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/4f78290a/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
----------------------------------------------------------------------
diff --git a/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py b/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
index 860c1af..9c6bc1f 100644
--- a/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
+++ b/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
@@ -33,7 +33,7 @@ from resource_management.core.exceptions import Fail
 __all__ = ["installedPkgsByName", "allInstalledPackages", "allAvailablePackages", "nameMatch",
            "getInstalledRepos", "getInstalledPkgsByRepo", "getInstalledPkgsByNames", "getPackageDetails"]
 
-LIST_INSTALLED_PACKAGES_UBUNTU = "for i in $(dpkg -l |grep ^ii |awk -F' ' '{print $2}'); do      apt-cache showpkg \"$i\"|head -3|grep -v '^Versions'| tr -d '()' | awk '{ print $1\" \"$2 }'|sed -e 's/^Package: //;' | paste -d ' ' - -;  done"
+LIST_INSTALLED_PACKAGES_UBUNTU = "COLUMNS=9999 ; for i in $(dpkg -l |grep ^ii |awk -F' ' '{print $2}'); do      apt-cache showpkg \"$i\"|head -3|grep -v '^Versions'| tr -d '()' | awk '{ print $1\" \"$2 }'|sed -e 's/^Package: //;' | paste -d ' ' - -;  done"
 LIST_AVAILABLE_PACKAGES_UBUNTU = "packages=`for  i in $(ls -1 /var/lib/apt/lists  | grep -v \"ubuntu.com\") ; do grep ^Package: /var/lib/apt/lists/$i |  awk '{print $2}' ; done` ; for i in $packages; do      apt-cache showpkg \"$i\"|head -3|grep -v '^Versions'| tr -d '()' | awk '{ print $1\" \"$2 }'|sed -e 's/^Package: //;' | paste -d ' ' - -;  done"
 
 logger = logging.getLogger()

http://git-wip-us.apache.org/repos/asf/ambari/blob/4f78290a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
index e85fd7a..f6987b3 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
@@ -267,13 +267,13 @@ def service(action=None, name=None, user=None, options="", create_pid_dir=False,
     try:
       Execute(daemon_cmd, not_if=process_id_exists_command, environment=hadoop_env_exports)
     except:
-      show_logs(params.hdfs_log_dir, user)
+      show_logs(log_dir, user)
       raise
   elif action == "stop":
     try:
       Execute(daemon_cmd, only_if=process_id_exists_command, environment=hadoop_env_exports)
     except:
-      show_logs(params.hdfs_log_dir, user)
+      show_logs(log_dir, user)
       raise
     File(pid_file, action="delete")
 


[2/2] ambari git commit: AMBARI-17337. Client installs failed on debian 7 (aonishuk)

Posted by ao...@apache.org.
AMBARI-17337. Client installs failed on debian 7 (aonishuk)


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/3ce7d10b
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/3ce7d10b
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/3ce7d10b

Branch: refs/heads/branch-2.4
Commit: 3ce7d10b72605b900a8d70d402755149ba1cde79
Parents: 3fac4f7
Author: Andrew Onishuk <ao...@hortonworks.com>
Authored: Tue Jun 21 17:29:05 2016 +0300
Committer: Andrew Onishuk <ao...@hortonworks.com>
Committed: Tue Jun 21 17:29:05 2016 +0300

----------------------------------------------------------------------
 .../resource_management/libraries/functions/packages_analyzer.py | 2 +-
 .../common-services/HDFS/2.1.0.2.0/package/scripts/utils.py      | 4 ++--
 2 files changed, 3 insertions(+), 3 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/3ce7d10b/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
----------------------------------------------------------------------
diff --git a/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py b/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
index 860c1af..9c6bc1f 100644
--- a/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
+++ b/ambari-common/src/main/python/resource_management/libraries/functions/packages_analyzer.py
@@ -33,7 +33,7 @@ from resource_management.core.exceptions import Fail
 __all__ = ["installedPkgsByName", "allInstalledPackages", "allAvailablePackages", "nameMatch",
            "getInstalledRepos", "getInstalledPkgsByRepo", "getInstalledPkgsByNames", "getPackageDetails"]
 
-LIST_INSTALLED_PACKAGES_UBUNTU = "for i in $(dpkg -l |grep ^ii |awk -F' ' '{print $2}'); do      apt-cache showpkg \"$i\"|head -3|grep -v '^Versions'| tr -d '()' | awk '{ print $1\" \"$2 }'|sed -e 's/^Package: //;' | paste -d ' ' - -;  done"
+LIST_INSTALLED_PACKAGES_UBUNTU = "COLUMNS=9999 ; for i in $(dpkg -l |grep ^ii |awk -F' ' '{print $2}'); do      apt-cache showpkg \"$i\"|head -3|grep -v '^Versions'| tr -d '()' | awk '{ print $1\" \"$2 }'|sed -e 's/^Package: //;' | paste -d ' ' - -;  done"
 LIST_AVAILABLE_PACKAGES_UBUNTU = "packages=`for  i in $(ls -1 /var/lib/apt/lists  | grep -v \"ubuntu.com\") ; do grep ^Package: /var/lib/apt/lists/$i |  awk '{print $2}' ; done` ; for i in $packages; do      apt-cache showpkg \"$i\"|head -3|grep -v '^Versions'| tr -d '()' | awk '{ print $1\" \"$2 }'|sed -e 's/^Package: //;' | paste -d ' ' - -;  done"
 
 logger = logging.getLogger()

http://git-wip-us.apache.org/repos/asf/ambari/blob/3ce7d10b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
index dfba302..107ad7c 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py
@@ -268,13 +268,13 @@ def service(action=None, name=None, user=None, options="", create_pid_dir=False,
     try:
       Execute(daemon_cmd, not_if=process_id_exists_command, environment=hadoop_env_exports)
     except:
-      show_logs(params.hdfs_log_dir, user)
+      show_logs(log_dir, user)
       raise
   elif action == "stop":
     try:
       Execute(daemon_cmd, only_if=process_id_exists_command, environment=hadoop_env_exports)
     except:
-      show_logs(params.hdfs_log_dir, user)
+      show_logs(log_dir, user)
       raise
     File(pid_file, action="delete")