You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by nc...@apache.org on 2015/10/19 15:54:14 UTC
[28/50] [abbrv] ambari git commit: AMBARI-13454. Make HDP 2.0 stack
inactive for Ambari 2.1.3 (aonishuk)
AMBARI-13454. Make HDP 2.0 stack inactive for Ambari 2.1.3 (aonishuk)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/e3958e2d
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/e3958e2d
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/e3958e2d
Branch: refs/heads/branch-dev-patch-upgrade
Commit: e3958e2da099f783ce637a57d752353fafd3f5b9
Parents: 8817569
Author: Andrew Onishuk <ao...@hortonworks.com>
Authored: Fri Oct 16 17:25:05 2015 +0300
Committer: Andrew Onishuk <ao...@hortonworks.com>
Committed: Fri Oct 16 17:25:05 2015 +0300
----------------------------------------------------------------------
.../2.1.0.2.0/package/scripts/params_linux.py | 7 +-----
.../0.12.0.2.0/package/scripts/params_linux.py | 20 +++++-----------
.../4.0.0.2.0/package/scripts/params_linux.py | 2 +-
.../2.0.6/hooks/after-INSTALL/scripts/params.py | 6 +----
.../2.0.6/hooks/before-ANY/scripts/params.py | 6 +----
.../2.0.6/hooks/before-START/scripts/params.py | 6 +----
.../main/resources/stacks/HDP/2.0/metainfo.xml | 2 +-
.../stacks/2.0.6/HIVE/test_hcat_client.py | 8 +++----
.../stacks/2.0.6/HIVE/test_hive_metastore.py | 8 +++++++
.../stacks/2.0.6/HIVE/test_webhcat_server.py | 24 ++++++++++----------
10 files changed, 36 insertions(+), 53 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
index 38eac2e..4cbce34 100644
--- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
+++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
@@ -106,12 +106,7 @@ hdfs_user_nofile_limit = default("/configurations/hadoop-env/hdfs_user_nofile_li
hdfs_user_nproc_limit = default("/configurations/hadoop-env/hdfs_user_nproc_limit", "65536")
create_lib_snappy_symlinks = not Script.is_hdp_stack_greater_or_equal("2.2")
-
-if Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.1") and not OSCheck.is_suse_family():
- # deprecated rhel jsvc_path
- jsvc_path = "/usr/libexec/bigtop-utils"
-else:
- jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
execute_path = os.environ['PATH'] + os.pathsep + hadoop_bin_dir
ulimit_cmd = "ulimit -c unlimited ; "
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
index e16df78..14cf1ac 100644
--- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
+++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
@@ -49,7 +49,7 @@ hostname = config["hostname"]
# This is expected to be of the form #.#.#.#
stack_version_unformatted = str(config['hostLevelParams']['stack_version'])
hdp_stack_version_major = format_hdp_stack_version(stack_version_unformatted)
-stack_is_hdp21 = Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.2")
+stack_is_hdp21 = Script.is_hdp_stack_less_than("2.2")
# this is not available on INSTALL action because hdp-select is not available
hdp_stack_version = functions.get_hdp_version('hive-server2')
@@ -83,10 +83,6 @@ sqoop_tar_file = '/usr/share/HDP-webhcat/sqoop*.tar.gz'
hive_specific_configs_supported = False
hive_etc_dir_prefix = "/etc/hive"
limits_conf_dir = "/etc/security/limits.d"
-hcat_conf_dir = '/etc/hcatalog/conf'
-config_dir = '/etc/hcatalog/conf'
-hcat_lib = '/usr/lib/hcatalog/share/hcatalog'
-webhcat_bin_dir = '/usr/lib/hcatalog/sbin'
hive_user_nofile_limit = default("/configurations/hive-env/hive_user_nofile_limit", "32000")
hive_user_nproc_limit = default("/configurations/hive-env/hive_user_nproc_limit", "16000")
@@ -101,11 +97,10 @@ hive_config_dir = status_params.hive_config_dir
hive_client_conf_dir = status_params.hive_client_conf_dir
hive_server_conf_dir = status_params.hive_server_conf_dir
-if Script.is_hdp_stack_greater_or_equal("2.1"):
- hcat_conf_dir = '/etc/hive-hcatalog/conf'
- config_dir = '/etc/hive-webhcat/conf'
- hcat_lib = '/usr/lib/hive-hcatalog/share/hcatalog'
- webhcat_bin_dir = '/usr/lib/hive-hcatalog/sbin'
+hcat_conf_dir = '/etc/hive-hcatalog/conf'
+config_dir = '/etc/hive-webhcat/conf'
+hcat_lib = '/usr/lib/hive-hcatalog/share/hcatalog'
+webhcat_bin_dir = '/usr/lib/hive-hcatalog/sbin'
# Starting from HDP2.3 drop should be executed with purge suffix
purge_tables = "false"
@@ -315,10 +310,7 @@ mysql_adduser_path = format("{tmp_dir}/addMysqlUser.sh")
mysql_deluser_path = format("{tmp_dir}/removeMysqlUser.sh")
######## Metastore Schema
-init_metastore_schema = False
-if Script.is_hdp_stack_greater_or_equal("2.1"):
- init_metastore_schema = True
-
+init_metastore_schema = True
########## HCAT
hcat_dbroot = hcat_lib
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
index d27d10a..fd9a462 100644
--- a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
+++ b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
@@ -183,7 +183,7 @@ if https_port is not None:
hdfs_site = config['configurations']['hdfs-site']
fs_root = config['configurations']['core-site']['fs.defaultFS']
-if Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.2"):
+if Script.is_hdp_stack_less_than("2.2"):
put_shared_lib_to_hdfs_cmd = format("hadoop --config {hadoop_conf_dir} dfs -put {oozie_shared_lib} {oozie_hdfs_user_dir}")
# for newer
else:
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
index 4aa5f8a..d7b18eb 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
@@ -62,11 +62,7 @@ hdfs_log_dir_prefix = config['configurations']['hadoop-env']['hdfs_log_dir_prefi
hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
hadoop_root_logger = config['configurations']['hadoop-env']['hadoop_root_logger']
-if Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.1") and not OSCheck.is_suse_family():
- # deprecated rhel jsvc_path
- jsvc_path = "/usr/libexec/bigtop-utils"
-else:
- jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize']
namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize']
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
index 2593ee4..e1f33d4 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
@@ -132,11 +132,7 @@ hdfs_log_dir_prefix = config['configurations']['hadoop-env']['hdfs_log_dir_prefi
hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
hadoop_root_logger = config['configurations']['hadoop-env']['hadoop_root_logger']
-if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.0') >= 0 and compare_versions(hdp_stack_version, '2.1') < 0 and not OSCheck.is_suse_family():
- # deprecated rhel jsvc_path
- jsvc_path = "/usr/libexec/bigtop-utils"
-else:
- jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize']
namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize']
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
index e771dd5..70c69fe 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
@@ -144,11 +144,7 @@ else:
#hadoop-env.sh
java_home = config['hostLevelParams']['java_home']
-if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.0') >= 0 and compare_versions(hdp_stack_version, '2.1') < 0 and not OSCheck.is_suse_family():
- # deprecated rhel jsvc_path
- jsvc_path = "/usr/libexec/bigtop-utils"
-else:
- jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize']
namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize']
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml b/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
index 0d4c27f..0659ca4 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
@@ -17,7 +17,7 @@
-->
<metainfo>
<versions>
- <active>true</active>
+ <active>false</active>
</versions>
<extends>2.0.6</extends>
<minJdk>1.7</minJdk>
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
index 6c9752a..e709693 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
@@ -37,7 +37,7 @@ class TestHcatClient(RMFTestCase):
group = 'hadoop',
recursive = True,
)
- self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+ self.assertResourceCalled('Directory', '/etc/hive-hcatalog/conf',
owner = 'hcat',
group = 'hadoop',
recursive = True,
@@ -54,7 +54,7 @@ class TestHcatClient(RMFTestCase):
configurations = self.getConfig()['configurations']['hive-site'],
configuration_attributes = self.getConfig()['configuration_attributes']['hive-site']
)
- self.assertResourceCalled('File', '/etc/hcatalog/conf/hcat-env.sh',
+ self.assertResourceCalled('File', '/etc/hive-hcatalog/conf/hcat-env.sh',
content = InlineTemplate(self.getConfig()['configurations']['hcat-env']['content']),
owner = 'hcat',
group = 'hadoop',
@@ -76,7 +76,7 @@ class TestHcatClient(RMFTestCase):
owner = 'hcat',
group = 'hadoop',
)
- self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+ self.assertResourceCalled('Directory', '/etc/hive-hcatalog/conf',
recursive = True,
owner = 'hcat',
group = 'hadoop',
@@ -93,7 +93,7 @@ class TestHcatClient(RMFTestCase):
configurations = self.getConfig()['configurations']['hive-site'],
configuration_attributes = self.getConfig()['configuration_attributes']['hive-site']
)
- self.assertResourceCalled('File', '/etc/hcatalog/conf/hcat-env.sh',
+ self.assertResourceCalled('File', '/etc/hive-hcatalog/conf/hcat-env.sh',
content = InlineTemplate(self.getConfig()['configurations']['hcat-env']['content']),
owner = 'hcat',
group = 'hadoop',
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
index 328b52d..7f3efcb 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
@@ -227,6 +227,10 @@ class TestHiveMetastore(RMFTestCase):
content = StaticFile('startMetastore.sh'),
mode = 0755,
)
+ self.assertResourceCalled('Execute', 'export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema -dbType mysql -userName hive -passWord \'!`"\'"\'"\' 1\'',
+ not_if = 'ambari-sudo.sh su hive -l -s /bin/bash -c \'[RMF_EXPORT_PLACEHOLDER]export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -info -dbType mysql -userName hive -passWord \'"\'"\'!`"\'"\'"\'"\'"\'"\'"\'"\'"\' 1\'"\'"\'\'',
+ user = 'hive',
+ )
self.assertResourceCalled('Directory', '/var/run/hive',
owner = 'hive',
group = 'hadoop',
@@ -331,6 +335,10 @@ class TestHiveMetastore(RMFTestCase):
content = StaticFile('startMetastore.sh'),
mode = 0755,
)
+ self.assertResourceCalled('Execute', 'export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema -dbType mysql -userName hive -passWord \'!`"\'"\'"\' 1\'',
+ not_if = 'ambari-sudo.sh su hive -l -s /bin/bash -c \'[RMF_EXPORT_PLACEHOLDER]export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -info -dbType mysql -userName hive -passWord \'"\'"\'!`"\'"\'"\'"\'"\'"\'"\'"\'"\' 1\'"\'"\'\'',
+ user = 'hive',
+ )
self.assertResourceCalled('Directory', '/var/run/hive',
owner = 'hive',
group = 'hadoop',
http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
index c11c2d0..b0b34bc 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
@@ -48,7 +48,7 @@ class TestWebHCatServer(RMFTestCase):
)
self.assert_configure_default()
- self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hcatalog/sbin/webhcat_server.sh start',
+ self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hive-hcatalog/sbin/webhcat_server.sh start',
environment = {'HADOOP_HOME': '/usr'},
not_if = "ambari-sudo.sh su hcat -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1'",
user = 'hcat',
@@ -64,7 +64,7 @@ class TestWebHCatServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Execute', '/usr/lib/hcatalog/sbin/webhcat_server.sh stop',
+ self.assertResourceCalled('Execute', '/usr/lib/hive-hcatalog/sbin/webhcat_server.sh stop',
user = 'hcat',
environment = {'HADOOP_HOME': '/usr' }
)
@@ -105,7 +105,7 @@ class TestWebHCatServer(RMFTestCase):
)
self.assert_configure_secured()
- self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hcatalog/sbin/webhcat_server.sh start',
+ self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hive-hcatalog/sbin/webhcat_server.sh start',
environment = {'HADOOP_HOME': '/usr'},
not_if = "ambari-sudo.sh su hcat -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1'",
user = 'hcat',
@@ -121,7 +121,7 @@ class TestWebHCatServer(RMFTestCase):
target = RMFTestCase.TARGET_COMMON_SERVICES
)
- self.assertResourceCalled('Execute', '/usr/lib/hcatalog/sbin/webhcat_server.sh stop',
+ self.assertResourceCalled('Execute', '/usr/lib/hive-hcatalog/sbin/webhcat_server.sh stop',
user = 'hcat',
environment = {'HADOOP_HOME': '/usr' }
)
@@ -152,7 +152,7 @@ class TestWebHCatServer(RMFTestCase):
recursive = True,
mode = 0755,
)
- self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+ self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf',
owner = 'hcat',
group = 'hadoop',
recursive = True,
@@ -161,11 +161,11 @@ class TestWebHCatServer(RMFTestCase):
self.assertResourceCalled('XmlConfig', 'webhcat-site.xml',
owner = 'hcat',
group = 'hadoop',
- conf_dir = '/etc/hcatalog/conf',
+ conf_dir = '/etc/hive-webhcat/conf',
configurations = self.getConfig()['configurations']['webhcat-site'],
configuration_attributes = self.getConfig()['configuration_attributes']['webhcat-site']
)
- self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-env.sh',
+ self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh',
content = InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']),
owner = 'hcat',
group = 'hadoop',
@@ -174,7 +174,7 @@ class TestWebHCatServer(RMFTestCase):
cd_access = 'a',
recursive=True
)
- self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-log4j.properties',
+ self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-log4j.properties',
content = 'log4jproperties\nline2',
owner = 'hcat',
group = 'hadoop',
@@ -194,7 +194,7 @@ class TestWebHCatServer(RMFTestCase):
recursive = True,
mode = 0755,
)
- self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+ self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf',
owner = 'hcat',
group = 'hadoop',
recursive = True,
@@ -207,11 +207,11 @@ class TestWebHCatServer(RMFTestCase):
self.assertResourceCalled('XmlConfig', 'webhcat-site.xml',
owner = 'hcat',
group = 'hadoop',
- conf_dir = '/etc/hcatalog/conf',
+ conf_dir = '/etc/hive-webhcat/conf',
configurations = self.getConfig()['configurations']['webhcat-site'],
configuration_attributes = self.getConfig()['configuration_attributes']['webhcat-site']
)
- self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-env.sh',
+ self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh',
content = InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']),
owner = 'hcat',
group = 'hadoop',
@@ -220,7 +220,7 @@ class TestWebHCatServer(RMFTestCase):
cd_access = 'a',
recursive=True
)
- self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-log4j.properties',
+ self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-log4j.properties',
content = 'log4jproperties\nline2',
owner = 'hcat',
group = 'hadoop',