You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by dm...@apache.org on 2016/03/30 18:35:10 UTC
[1/2] ambari git commit: AMBARI-15627 Ambari is expecting hadoop
client package and configuration,
even if no HDFS components are installed on this host (dgrinenko via
dlysnichenko)
Repository: ambari
Updated Branches:
refs/heads/branch-2.2 a9e65decf -> 4b0802cac
refs/heads/trunk 9da28fec1 -> deb45e81a
AMBARI-15627 Ambari is expecting hadoop client package and configuration, even if no HDFS components are installed on this host (dgrinenko via dlysnichenko)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/deb45e81
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/deb45e81
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/deb45e81
Branch: refs/heads/trunk
Commit: deb45e81a3555de0430e65231e243919d1b9da32
Parents: 9da28fe
Author: Lisnichenko Dmitro <dl...@hortonworks.com>
Authored: Wed Mar 30 19:34:19 2016 +0300
Committer: Lisnichenko Dmitro <dl...@hortonworks.com>
Committed: Wed Mar 30 19:34:19 2016 +0300
----------------------------------------------------------------------
.../before-START/scripts/rack_awareness.py | 46 +++++---------------
.../hooks/before-START/test_before_start.py | 22 +++++-----
2 files changed, 21 insertions(+), 47 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/deb45e81/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
index 9579d18..7be1561 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
@@ -1,6 +1,6 @@
#!/usr/bin/env python
-'''
+"""
Licensed to the Apache Software Foundation (ASF) under one
or more contributor license agreements. See the NOTICE file
distributed with this work for additional information
@@ -16,55 +16,29 @@ distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
-'''
-from resource_management.core.resources import File,Directory
+"""
+
+from resource_management.core.resources import File
from resource_management.core.source import StaticFile, Template
from resource_management.libraries.functions import format
-import os
def create_topology_mapping():
import params
-
- path=params.net_topology_mapping_data_file_path
- parent_dir=os.path.dirname(path)
- # only create the parent directory and set its permission if it does not exist
- if not os.path.exists(parent_dir):
- Directory(parent_dir,
- create_parents = True,
- owner=params.hdfs_user,
- group=params.user_group)
-
- # placing the mappings file in the same folder where the topology script is located
- File(path,
+ File(params.net_topology_mapping_data_file_path,
content=Template("topology_mappings.data.j2"),
owner=params.hdfs_user,
group=params.user_group,
- # if there is no hadoop components, don't create the script
- only_if=format("test -d {net_topology_script_dir}"),
- )
+ only_if=format("test -d {net_topology_script_dir}"))
def create_topology_script():
import params
- path=params.net_topology_script_file_path
- parent_dir=os.path.dirname(path)
- # only create the parent directory and set its permission if it does not exist
- if not os.path.exists(parent_dir):
- Directory(parent_dir,
- create_parents = True,
- owner=params.hdfs_user,
- group=params.user_group)
-
- # installing the topology script to the specified location
- File(path,
+ File(params.net_topology_script_file_path,
content=StaticFile('topology_script.py'),
mode=0755,
- only_if=format("test -d {net_topology_script_dir}"),
- )
+ only_if=format("test -d {net_topology_script_dir}"))
def create_topology_script_and_mapping():
- import params
- if params.has_hadoop_env:
- create_topology_mapping()
- create_topology_script()
+ create_topology_mapping()
+ create_topology_script()
http://git-wip-us.apache.org/repos/asf/ambari/blob/deb45e81/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
index 90bd968..c0f279f 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
@@ -89,16 +89,16 @@ class TestHookBeforeStart(RMFTestCase):
group = 'hadoop',
)
self.assertResourceCalled('File', '/etc/hadoop/conf/topology_mappings.data',
- owner = 'hdfs',
- content = Template('topology_mappings.data.j2'),
- group = 'hadoop',
- only_if = 'test -d /etc/hadoop/conf',
- )
+ owner = 'hdfs',
+ content = Template('topology_mappings.data.j2'),
+ group = 'hadoop',
+ only_if = 'test -d /etc/hadoop/conf',
+ )
self.assertResourceCalled('File', '/etc/hadoop/conf/topology_script.py',
- content = StaticFile('topology_script.py'),
- mode = 0755,
- only_if = 'test -d /etc/hadoop/conf',
- )
+ content = StaticFile('topology_script.py'),
+ mode = 0755,
+ only_if = 'test -d /etc/hadoop/conf',
+ )
self.assertNoMoreResources()
def test_hook_secured(self):
@@ -326,12 +326,12 @@ class TestHookBeforeStart(RMFTestCase):
owner = 'hdfs',
content = Template('topology_mappings.data.j2'),
group = 'hadoop',
- only_if = 'test -d /etc/hadoop/conf'
+ only_if = 'test -d /etc/hadoop/conf',
)
self.assertResourceCalled('File', '/etc/hadoop/conf/topology_script.py',
content = StaticFile('topology_script.py'),
mode = 0755,
- only_if = 'test -d /etc/hadoop/conf'
+ only_if = 'test -d /etc/hadoop/conf',
)
self.assertNoMoreResources()
[2/2] ambari git commit: AMBARI-15627 Ambari is expecting hadoop
client package and configuration,
even if no HDFS components are installed on this host (dgrinenko via
dlysnichenko)
Posted by dm...@apache.org.
AMBARI-15627 Ambari is expecting hadoop client package and configuration, even if no HDFS components are installed on this host (dgrinenko via dlysnichenko)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/4b0802ca
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/4b0802ca
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/4b0802ca
Branch: refs/heads/branch-2.2
Commit: 4b0802cac478a93f03864e4bb69e387871a6a167
Parents: a9e65de
Author: Lisnichenko Dmitro <dl...@hortonworks.com>
Authored: Wed Mar 30 19:32:33 2016 +0300
Committer: Lisnichenko Dmitro <dl...@hortonworks.com>
Committed: Wed Mar 30 19:35:07 2016 +0300
----------------------------------------------------------------------
.../before-START/scripts/rack_awareness.py | 42 ++++----------------
.../hooks/before-START/test_before_start.py | 22 +++++-----
2 files changed, 19 insertions(+), 45 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/4b0802ca/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
index 453de63..a1d3ac3 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/rack_awareness.py
@@ -1,6 +1,6 @@
#!/usr/bin/env python
-'''
+"""
Licensed to the Apache Software Foundation (ASF) under one
or more contributor license agreements. See the NOTICE file
distributed with this work for additional information
@@ -16,55 +16,29 @@ distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
-'''
-from resource_management.core.resources import File,Directory
+"""
+
+from resource_management.core.resources import File
from resource_management.core.source import StaticFile, Template
from resource_management.libraries.functions import format
-import os
-
def create_topology_mapping():
import params
- path=params.net_topology_mapping_data_file_path
- parent_dir=os.path.dirname(path)
- # only create the parent directory and set its permission if it does not exist
- if not os.path.exists(parent_dir):
- Directory(parent_dir,
- recursive=True,
- owner=params.hdfs_user,
- group=params.user_group)
-
- # placing the mappings file in the same folder where the topology script is located
- File(path,
+ File(params.net_topology_mapping_data_file_path,
content=Template("topology_mappings.data.j2"),
owner=params.hdfs_user,
group=params.user_group,
- # if there is no hadoop components, don't create the script
- only_if=format("test -d {net_topology_script_dir}"),
- )
+ only_if=format("test -d {net_topology_script_dir}"))
def create_topology_script():
import params
- path=params.net_topology_script_file_path
- parent_dir=os.path.dirname(path)
- # only create the parent directory and set its permission if it does not exist
- if not os.path.exists(parent_dir):
- Directory(parent_dir,
- recursive=True,
- owner=params.hdfs_user,
- group=params.user_group)
-
- # installing the topology script to the specified location
- File(path,
+ File(params.net_topology_script_file_path,
content=StaticFile('topology_script.py'),
mode=0755,
- only_if=format("test -d {net_topology_script_dir}"),
- )
+ only_if=format("test -d {net_topology_script_dir}"))
def create_topology_script_and_mapping():
- import params
- if params.has_hadoop_env:
create_topology_mapping()
create_topology_script()
http://git-wip-us.apache.org/repos/asf/ambari/blob/4b0802ca/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
index 443c02e..ef429e9 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/hooks/before-START/test_before_start.py
@@ -89,16 +89,16 @@ class TestHookBeforeStart(RMFTestCase):
group = 'hadoop',
)
self.assertResourceCalled('File', '/etc/hadoop/conf/topology_mappings.data',
- owner = 'hdfs',
- content = Template('topology_mappings.data.j2'),
- group = 'hadoop',
- only_if = 'test -d /etc/hadoop/conf',
- )
+ owner = 'hdfs',
+ content = Template('topology_mappings.data.j2'),
+ group = 'hadoop',
+ only_if = 'test -d /etc/hadoop/conf',
+ )
self.assertResourceCalled('File', '/etc/hadoop/conf/topology_script.py',
- content = StaticFile('topology_script.py'),
- mode = 0755,
- only_if = 'test -d /etc/hadoop/conf',
- )
+ content = StaticFile('topology_script.py'),
+ mode = 0755,
+ only_if = 'test -d /etc/hadoop/conf',
+ )
self.assertNoMoreResources()
def test_hook_secured(self):
@@ -326,12 +326,12 @@ class TestHookBeforeStart(RMFTestCase):
owner = 'hdfs',
content = Template('topology_mappings.data.j2'),
group = 'hadoop',
- only_if = 'test -d /etc/hadoop/conf'
+ only_if = 'test -d /etc/hadoop/conf',
)
self.assertResourceCalled('File', '/etc/hadoop/conf/topology_script.py',
content = StaticFile('topology_script.py'),
mode = 0755,
- only_if = 'test -d /etc/hadoop/conf'
+ only_if = 'test -d /etc/hadoop/conf',
)
self.assertNoMoreResources()