You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hawq.apache.org by wl...@apache.org on 2015/12/17 05:48:14 UTC
[1/2] incubator-hawq git commit: HAWQ-186. Remove some useless files
Repository: incubator-hawq
Updated Branches:
refs/heads/master d9e1578c4 -> 10909db4a
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure-ha.xml b/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure-ha.xml
deleted file mode 100755
index 9433d31..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure-ha.xml
+++ /dev/null
@@ -1,224 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.encrypt.data.transfer</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.encrypt.data.transfer.algorithm</name>
- <value>rc4</value>
- <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
- </property>
- <property>
- <name>dfs.datanode.address</name>
- <value>0.0.0.0:1004</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:1006</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.journalnode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.journalnode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.web.authentication.kerberos.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.web.authentication.kerberos.keytab</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
-
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
- <property>
- <name>dfs.nameservices</name>
- <value>gphd-cluster</value>
- </property>
- <property>
- <name>dfs.ha.namenodes.gphd-cluster</name>
- <value>nn1,nn2</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.shared.edits.dir</name>
- <value>qjournal://%ZKSERVER1%:8485;%ZKSERVER2%:8485;%ZKSERVER3%:8485/gphd-cluster</value>
- </property>
- <property>
- <name>dfs.journalnode.edits.dir</name>
- <value>%JOURNAL_DATA_ROOT%</value>
- </property>
- <property>
- <name>dfs.client.failover.proxy.provider.gphd-cluster</name>
- <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
- </property>
- <property>
- <name>dfs.ha.fencing.methods</name>
- <value>shell(/bin/true)</value>
- </property>
- <property>
- <name>dfs.ha.fencing.ssh.connect-timeout</name>
- <value>20000</value>
- </property>
- <property>
- <name>dfs.ha.automatic-failover.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.namenode.fs-limits.min-block-size</name>
- <value>1024</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure.xml b/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure.xml
deleted file mode 100755
index 48c68e2..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-namenode-secure.xml
+++ /dev/null
@@ -1,181 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.encrypt.data.transfer</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.encrypt.data.transfer.algorithm</name>
- <value>rc4</value>
- <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
- </property>
- <property>
- <name>dfs.datanode.address</name>
- <value>0.0.0.0:1004</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:1006</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.principal</name>
- <value>hdfs/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.http.principal</name>
- <value>HTTP/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.secondary.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.principal</name>
- <value>hdfs/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.http.principal</name>
- <value>HTTP/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.web.authentication.kerberos.principal</name>
- <value>HTTP/%WEBAUTH_HOST%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.web.authentication.kerberos.keytab</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
-
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-secure-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-secure-ha.xml b/depends/libyarn/releng/bin/templates/hdfs-site-secure-ha.xml
deleted file mode 100755
index eb6d6ea..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-secure-ha.xml
+++ /dev/null
@@ -1,240 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.encrypt.data.transfer</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.encrypt.data.transfer.algorithm</name>
- <value>rc4</value>
- <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
- </property>
- <property>
- <name>dfs.datanode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.address</name>
- <value>0.0.0.0:1004</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:1006</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.journalnode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.journalnode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.journalnode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.web.authentication.kerberos.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.web.authentication.kerberos.keytab</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
-
- <property>
- <name>dfs.nameservices</name>
- <value>gphd-cluster</value>
- </property>
- <property>
- <name>dfs.ha.namenodes.gphd-cluster</name>
- <value>nn1,nn2</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.shared.edits.dir</name>
- <value>qjournal://%JOURNALNODE1%:8485;%JOURNALNODE2%:8485;%JOURNALNODE3%:8485/gphd-cluster</value>
- </property>
- <property>
- <name>dfs.journalnode.edits.dir</name>
- <value>%JOURNAL_DATA_ROOT%</value>
- </property>
- <property>
- <name>dfs.client.failover.proxy.provider.gphd-cluster</name>
- <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
- </property>
- <property>
- <name>dfs.ha.fencing.methods</name>
- <value>shell(/bin/true)</value>
- </property>
- <property>
- <name>dfs.ha.fencing.ssh.connect-timeout</name>
- <value>20000</value>
- </property>
- <property>
- <name>dfs.ha.automatic-failover.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.namenode.fs-limits.min-block-size</name>
- <value>1024</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-secure.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-secure.xml b/depends/libyarn/releng/bin/templates/hdfs-site-secure.xml
deleted file mode 100755
index 89e80bf..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-secure.xml
+++ /dev/null
@@ -1,193 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.encrypt.data.transfer</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.encrypt.data.transfer.algorithm</name>
- <value>rc4</value>
- <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
- </property>
- <property>
- <name>dfs.datanode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.address</name>
- <value>0.0.0.0:1004</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:1006</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.secondary.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.web.authentication.kerberos.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.web.authentication.kerberos.keytab</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
-
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site.xml b/depends/libyarn/releng/bin/templates/hdfs-site.xml
deleted file mode 100644
index 2137347..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site.xml
+++ /dev/null
@@ -1,123 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:59075</value>
- <description>
- The datanode http server address and port. If the port is 0 then the server will start on a free port.
- </description>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/log4j-ha.properties
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/log4j-ha.properties b/depends/libyarn/releng/bin/templates/log4j-ha.properties
deleted file mode 100644
index 88c4a75..0000000
--- a/depends/libyarn/releng/bin/templates/log4j-ha.properties
+++ /dev/null
@@ -1,75 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-# Define some default values that can be overridden by system properties
-zookeeper.root.logger=INFO, CONSOLE
-zookeeper.console.threshold=INFO
-zookeeper.log.dir=%ZOO_LOG_DIR%
-zookeeper.log.file=zookeeper.log
-zookeeper.log.threshold=DEBUG
-zookeeper.tracelog.dir=%ZOO_LOG_DIR%
-zookeeper.tracelog.file=zookeeper_trace.log
-
-#
-# ZooKeeper Logging Configuration
-#
-
-# Format is "<default threshold> (, <appender>)+
-
-# DEFAULT: console appender only
-log4j.rootLogger=${zookeeper.root.logger}
-
-# Example with rolling log file
-#log4j.rootLogger=DEBUG, CONSOLE, ROLLINGFILE
-
-# Example with rolling log file and tracing
-#log4j.rootLogger=TRACE, CONSOLE, ROLLINGFILE, TRACEFILE
-
-#
-# Log INFO level and above messages to the console
-#
-log4j.appender.CONSOLE=org.apache.log4j.ConsoleAppender
-log4j.appender.CONSOLE.Threshold=${zookeeper.console.threshold}
-log4j.appender.CONSOLE.layout=org.apache.log4j.PatternLayout
-log4j.appender.CONSOLE.layout.ConversionPattern=%d{ISO8601} [myid:%X{myid}] - %-5p [%t:%C{1}@%L] - %m%n
-
-#
-# Add ROLLINGFILE to rootLogger to get log file output
-# Log DEBUG level and above messages to a log file
-log4j.appender.ROLLINGFILE=org.apache.log4j.RollingFileAppender
-log4j.appender.ROLLINGFILE.Threshold=${zookeeper.log.threshold}
-log4j.appender.ROLLINGFILE.File=${zookeeper.log.dir}/${zookeeper.log.file}
-
-# Max log file size of 10MB
-log4j.appender.ROLLINGFILE.MaxFileSize=10MB
-# uncomment the next line to limit number of backup files
-#log4j.appender.ROLLINGFILE.MaxBackupIndex=10
-
-log4j.appender.ROLLINGFILE.layout=org.apache.log4j.PatternLayout
-log4j.appender.ROLLINGFILE.layout.ConversionPattern=%d{ISO8601} [myid:%X{myid}] - %-5p [%t:%C{1}@%L] - %m%n
-
-
-#
-# Add TRACEFILE to rootLogger to get log file output
-# Log DEBUG level and above messages to a log file
-log4j.appender.TRACEFILE=org.apache.log4j.FileAppender
-log4j.appender.TRACEFILE.Threshold=TRACE
-log4j.appender.TRACEFILE.File=${zookeeper.tracelog.dir}/${zookeeper.tracelog.file}
-
-log4j.appender.TRACEFILE.layout=org.apache.log4j.PatternLayout
-### Notice we are including log4j's NDC here (%x)
-log4j.appender.TRACEFILE.layout.ConversionPattern=%d{ISO8601} [myid:%X{myid}] - %-5p [%t:%C{1}@%L][%x] - %m%n
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/yarn-env.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/yarn-env.sh b/depends/libyarn/releng/bin/templates/yarn-env.sh
deleted file mode 100644
index 86c0ef9..0000000
--- a/depends/libyarn/releng/bin/templates/yarn-env.sh
+++ /dev/null
@@ -1,87 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements. See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-export HADOOP_HOME=/usr/lib/gphd/hadoop
-export HADOOP_COMMON_HOME=/usr/lib/gphd/hadoop
-export HADOOP_CONF_DIR=/etc/gphd/hadoop/conf
-export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
-export HADOOP_YARN_HOME=/usr/lib/gphd/hadoop-yarn
-export HADOOP_MAPRED_HOME=/usr/lib/gphd/hadoop-mapreduce
-export HADOOP_HDFS_HOME=/usr/lib/gphd/hadoop-hdfs
-export PATH=$HADOOP_YARN_HOME/bin:$HADOOP_YARN_HOME/sbin:$PATH
-export HADOOP_LIBEXEC_DIR=$HADOOP_YARN_HOME/libexec
-
-# User for YARN daemons
-export HADOOP_YARN_USER=${HADOOP_YARN_USER:-yarn}
-
-# resolve links - $0 may be a softlink
-export YARN_CONF_DIR="${YARN_CONF_DIR:-$HADOOP_YARN_HOME/conf}"
-
-# some Java parameters
-export JAVA_HOME=%JAVA_HOME%
-if [ "$JAVA_HOME" != "" ]; then
- echo "run java in $JAVA_HOME"
- JAVA_HOME=$JAVA_HOME
-fi
-
-if [ "$JAVA_HOME" = "" ]; then
- echo "Error: JAVA_HOME is not set."
- exit 1
-fi
-
-JAVA=$JAVA_HOME/bin/java
-JAVA_HEAP_MAX=-Xmx1000m
-
-# check envvars which might override default args
-if [ "$YARN_HEAPSIZE" != "" ]; then
- #echo "run with heapsize $YARN_HEAPSIZE"
-JAVA_HEAP_MAX="-Xmx""$YARN_HEAPSIZE""m"
- #echo $JAVA_HEAP_MAX
-fi
-
-# so that filenames w/ spaces are handled correctly in loops below
-IFS=
-
-
-# default log directory & file
-if [ "$YARN_LOG_DIR" = "" ]; then
- YARN_LOG_DIR="$HADOOP_YARN_HOME/logs"
-fi
-if [ "$YARN_LOGFILE" = "" ]; then
- YARN_LOGFILE='yarn.log'
-fi
-
-# default policy file for service-level authorization
-if [ "$YARN_POLICYFILE" = "" ]; then
- YARN_POLICYFILE="hadoop-policy.xml"
-fi
-
-# restore ordinary behaviour
-unset IFS
-
-
-YARN_OPTS="$YARN_OPTS -Dhadoop.log.dir=$YARN_LOG_DIR"
-YARN_OPTS="$YARN_OPTS -Dyarn.log.dir=$YARN_LOG_DIR"
-YARN_OPTS="$YARN_OPTS -Dhadoop.log.file=$YARN_LOGFILE"
-YARN_OPTS="$YARN_OPTS -Dyarn.log.file=$YARN_LOGFILE"
-YARN_OPTS="$YARN_OPTS -Dyarn.home.dir=$YARN_COMMON_HOME"
-YARN_OPTS="$YARN_OPTS -Dyarn.id.str=$YARN_IDENT_STRING"
-YARN_OPTS="$YARN_OPTS -Dhadoop.root.logger=${YARN_ROOT_LOGGER:-INFO,console}"
-YARN_OPTS="$YARN_OPTS -Dyarn.root.logger=${YARN_ROOT_LOGGER:-INFO,console}"
-if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
- YARN_OPTS="$YARN_OPTS -Djava.library.path=$JAVA_LIBRARY_PATH"
-fi
-YARN_OPTS="$YARN_OPTS -Dyarn.policy.file=$YARN_POLICYFILE"
-
-
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/yarn-site.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/yarn-site.xml b/depends/libyarn/releng/bin/templates/yarn-site.xml
deleted file mode 100644
index c871cd8..0000000
--- a/depends/libyarn/releng/bin/templates/yarn-site.xml
+++ /dev/null
@@ -1,112 +0,0 @@
-<?xml version="1.0"?>
-<!--
- Licensed to the Apache Software Foundation (ASF) under one or more
- contributor license agreements. See the NOTICE file distributed with
- this work for additional information regarding copyright ownership.
- The ASF licenses this file to You under the Apache License, Version 2.0
- (the "License"); you may not use this file except in compliance with
- the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
--->
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-
-<configuration>
-<property>
- <name>yarn.resourcemanager.address</name>
- <value>smdw:8032</value>
- </property>
-
-<property>
- <name>yarn.resourcemanager.resource-tracker.address</name>
- <value>smdw:8031</value>
- </property>
-
-<property>
- <name>yarn.resourcemanager.scheduler.address</name>
- <value>smdw:8030</value>
- </property>
-
-<property>
- <name>yarn.resourcemanager.admin.address</name>
- <value>smdw:8033</value>
- </property>
- <property>
- <name>yarn.resourcemanager.webapp.address</name>
- <value>smdw:8088</value>
- </property>
-
-<property>
- <name>yarn.nodemanager.vmem-pmem-ratio</name>
- <value>2.0</value>
-</property>
-
-<property>
- <name>yarn.scheduler.minimum-allocation-mb</name>
- <value>1024</value>
-</property>
-
-<property>
- <name>yarn.nodemanager.resource.memory-mb</name>
- <value>%RESOURCE_MEM%</value>
-</property>
-
- <property>
- <name>yarn.nodemanager.aux-services</name>
- <value>mapreduce_shuffle</value>
- </property>
-
- <property>
- <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
- <value>org.apache.hadoop.mapred.ShuffleHandler</value>
- </property>
-
- <property>
- <name>yarn.log-aggregation-enable</name>
- <value>true</value>
- </property>
-
- <property>
- <description>List of directories to store localized files in.</description>
- <name>yarn.nodemanager.local-dirs</name>
- <value>/var/lib/gphd/hadoop-yarn/cache/${user.name}/nm-local-dir</value>
- </property>
-
- <property>
- <description>Where to store container logs.</description>
- <name>yarn.nodemanager.log-dirs</name>
- <value>/var/log/gphd/hadoop-yarn/containers</value>
- </property>
-
- <property>
- <description>Where to aggregate logs to.</description>
- <name>yarn.nodemanager.remote-app-log-dir</name>
- <value>/var/log/gphd/hadoop-yarn/apps</value>
- </property>
- <property>
- <description>The address for the web proxy as HOST:PORT, if this is not
- given or if it matches yarn.resourcemanager.address then the proxy will
- run as part of the RM</description>
- <name>yarn.web-proxy.address</name>
- <value>0.0.0.0:9999</value>
- </property>
-
- <property>
- <description>Classpath for typical applications.</description>
- <name>yarn.application.classpath</name>
- <value>
- $HADOOP_CONF_DIR,
- /tmp/hawq-mapreduce-tool/*,
- $HADOOP_COMMON_HOME/*,$HADOOP_COMMON_HOME/lib/*,
- $HADOOP_HDFS_HOME/*,$HADOOP_HDFS_HOME/lib/*,
- $HADOOP_MAPRED_HOME/*,$HADOOP_MAPRED_HOME/lib/*,
- $HADOOP_YARN_HOME/*,$HADOOP_YARN_HOME/lib/*
- </value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/zoo.cfg
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/zoo.cfg b/depends/libyarn/releng/bin/templates/zoo.cfg
deleted file mode 100644
index ad1f666..0000000
--- a/depends/libyarn/releng/bin/templates/zoo.cfg
+++ /dev/null
@@ -1,8 +0,0 @@
-tickTime=2000
-initLimit=10
-syncLimit=5
-dataDir=/tmp/zookeeper
-clientPort=2181
-server.1=%ZKSERVER1%:2182:2183
-server.2=%ZKSERVER2%:2182:2183
-server.3=%ZKSERVER3%:2182:2183
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/etc/ivy.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/etc/ivy.xml b/depends/libyarn/releng/etc/ivy.xml
deleted file mode 100644
index 3356da4..0000000
--- a/depends/libyarn/releng/etc/ivy.xml
+++ /dev/null
@@ -1,38 +0,0 @@
-<?xml version='1.0'?>
-<!--
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing,
-software distributed under the License is distributed on an
-"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-KIND, either express or implied. See the License for the
-specific language governing permissions and limitations
-under the License.
--->
-
-<ivy-module version="2.0">
- <info organisation="Pivotal" module="libyarn" revision="<version>" status="release" publication="<time>">
- <license name="Pivotal"/>
- </info>
-
- <configurations>
- <conf name="osx106_x86" description="osx106_x86" visibility="public"/>
- <conf name="rhel5_x86_64" description="rhel5_x86_64" visibility="public"/>
- <conf name="suse10_x86_64" description="suse10_x86_64" visibility="public"/>
- </configurations>
-
- <publications>
- <artifact name="libyarn-osx106_x86" type="targz" ext="targz" conf="osx106_x86"/>
- <artifact name="libyarn-rhel5_x86_64" type="targz" ext="targz" conf="rhel5_x86_64"/>
- <artifact name="libyarn-suse10_x86_64" type="targz" ext="targz" conf="suse10_x86_64"/>
- </publications>
-
-</ivy-module>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/etc/local/.gitignore
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/etc/local/.gitignore b/depends/libyarn/releng/etc/local/.gitignore
deleted file mode 100644
index 1741722..0000000
--- a/depends/libyarn/releng/etc/local/.gitignore
+++ /dev/null
@@ -1 +0,0 @@
-hdfs-site.xml
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/etc/local/hdfs-site.xml.simple
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/etc/local/hdfs-site.xml.simple b/depends/libyarn/releng/etc/local/hdfs-site.xml.simple
deleted file mode 100644
index 37bdae8..0000000
--- a/depends/libyarn/releng/etc/local/hdfs-site.xml.simple
+++ /dev/null
@@ -1,65 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-
-<property>
- <name>fs.default.name</name>
- <value>hdfs://localhost:9000</value>
-</property>
-
-<property>
- <name>dfs.namenode.http-address</name>
- <value>hdfs://localhost:50070</value>
-</property>
-
-<property>
- <name>dfs.name.dir</name>
- <value>
- file:@data_path@/fs
- </value>
-</property>
-
-<property>
- <name>dfs.data.dir</name>
- <value>
- file:@data_path@/data
- </value>
-</property>
-
-<property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
-</property>
-
-<property>
- <name>hadoop.http.staticuser.user</name>
- <value>@user@</value>
-</property>
-
-<property>
- <name>dfs.block.local-path-access.user</name>
- <value>@user@</value>
-</property>
-
-<property>
- <name>dfs.namenode.fs-limits.min-block-size</name>
- <value>1024</value>
-</property>
-
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/.gitignore
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/.gitignore b/depends/libyarn/releng/make/dependencies/.gitignore
deleted file mode 100644
index 2d19fc7..0000000
--- a/depends/libyarn/releng/make/dependencies/.gitignore
+++ /dev/null
@@ -1 +0,0 @@
-*.html
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/build.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/build.xml b/depends/libyarn/releng/make/dependencies/build.xml
deleted file mode 100644
index 26236f0..0000000
--- a/depends/libyarn/releng/make/dependencies/build.xml
+++ /dev/null
@@ -1,84 +0,0 @@
-<?xml version='1.0'?>
-<!--
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing,
-software distributed under the License is distributed on an
-"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-KIND, either express or implied. See the License for the
-specific language governing permissions and limitations
-under the License.
--->
-
-<project name="cdb2" default="resolve" xmlns:ivy="antlib:org.apache.ivy.ant">
-
- <target name="resolve" if="BLD_ARCH">
- <ivy:resolve conf="${BLD_ARCH}"
- log="download-only"/>
- <ivy:report conf="${BLD_ARCH}"
- graph="false"/>
- </target>
-
- <target name="clean">
- <ivy:cleancache />
- </target>
-
-
- <target name="post-resolve-trigger">
- <echo>
- ======================================================================
- Post Resolve Trigger
- ======================================================================
-
- organisation=${dep.organisation}
- module=${dep.module}
- revision=${dep.revision}
- platform=${BLD_ARCH}
-
- </echo>
-
- <exec executable="./ivy.sh">
- <arg value="${dep.module}"/>
- <arg value="${dep.revision}"/>
- <arg value="${dep.organisation}"/>
- <arg value="${BLD_ARCH}"/>
- </exec>
-
- </target>
-
- <target name="download-untar-trigger">
- <echo>
-======================================================================
- Post download artifact
-======================================================================
- organisation=${dep.organisation}
- module=${dep.module}
- revision=${dep.revision}
- artifact=${dep.artifact}
- origin=${dep.origin}
- local=${dep.local}
- size=${dep.size}</echo>
-
- <exec executable="./ivy_util.sh">
- <arg value="${dep.file}"/>
- <arg value="${dep.revision}"/>
- </exec>
-
- <exec executable="./ivy.sh">
- <arg value="${dep.module}"/>
- <arg value="${dep.revision}"/>
- <arg value="${dep.organisation}"/>
- <arg value="${BLD_ARCH}"/>
- </exec>
-
- </target>
-
-</project>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/cacerts
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/cacerts b/depends/libyarn/releng/make/dependencies/cacerts
deleted file mode 100644
index 01e7700..0000000
Binary files a/depends/libyarn/releng/make/dependencies/cacerts and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/ivy-report.css
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/ivy-report.css b/depends/libyarn/releng/make/dependencies/ivy-report.css
deleted file mode 100644
index 4cb6091..0000000
--- a/depends/libyarn/releng/make/dependencies/ivy-report.css
+++ /dev/null
@@ -1,279 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- *
- */
-
-body {
- font-family:"Trebuchet MS",Verdana,Geneva,Arial,Helvetica,sans-serif;
- font-size:small;
-}
-
-div#logo {
- float: right;
- padding-left: 10px;
- padding-bottom: 10px;
- background: white;
- text-align: center;
-}
-
-#logo img {
- border: 0;
-}
-
-div#date {
- font-style: italic;
- padding-left: 60px;
- padding-bottom: 40px;
-}
-
-
-h1 {
- margin-bottom:2px;
-
- border-color:#7A9437;
- border-style:solid;
- border-width:0 0 3px 0;
-}
-
-span#module {
- color:#7A9437;
- text-decoration:none;
-}
-
-span#organisation {
- color:black;
- text-decoration:none;
-}
-
-#confmenu {
- color: #000;
- border-bottom: 2px solid black;
- margin: 12px 0px 0px 0px;
- padding: 0px;
- z-index: 1;
- padding-left: 10px
-}
-
-#confmenu li {
- display: inline;
- overflow: hidden;
- list-style-type: none;
-}
-
-#confmenu a, a.active {
- color: #DEDECF;
- background: #898B5E;
- font: bold 1em "Trebuchet MS", Arial, sans-serif;
- border: 2px solid black;
- padding: 2px 5px 0px 5px;
- text-decoration: none;
-}
-
-/*
-background: #ABAD85 #CED4BD
-background: #DEE4CD
- */
-
-#confmenu a.active {
- color: #7A9437;
- background: #DEE4CD;
- border-bottom: 3px solid #DEE4CD;
-}
-
-#confmenu a:hover {
- color: #fff;
- background: #ADC09F;
-}
-
-#confmenu a:visited {
- color: #DEDECF;
-}
-
-#confmenu a.active:visited {
- color: #7A9437;
-}
-
-#confmenu a.active:hover {
- background: #DEE4CD;
- color: #DEDECF;
-}
-
-#content {
- background: #DEE4CD;
- padding: 20px;
- border: 2px solid black;
- border-top: none;
- z-index: 2;
-}
-
-#content a {
- text-decoration: none;
- color: #E8E9BE;
-}
-
-#content a:hover {
- background: #898B5E;
-}
-
-
-h2 {
- margin-bottom:2px;
- font-size:medium;
-
- border-color:#7A9437;
- border-style:solid;
- border-width:0 0 2px 0;
-}
-
-h3 {
- margin-top:30px;
- margin-bottom:2px;
- padding: 5 5 5 0;
- font-size: 24px;
- border-style:solid;
- border-width:0 0 2px 0;
-}
-
-h4 {
- margin-bottom:2px;
- margin-top:2px;
- font-size:medium;
-
- border-color:#7A9437;
- border-style:dashed;
- border-width:0 0 1px 0;
-}
-
-h5 {
- margin-bottom:2px;
- margin-top:2px;
- margin-left:20px;
- font-size:medium;
-}
-
-span.resolved {
- padding-left: 15px;
- font-weight: 500;
- font-size: small;
-}
-
-
-#content table {
- border-collapse:collapse;
- width:90%;
- margin:auto;
- margin-top: 5px;
-}
-#content thead {
- background-color:#CED4BD;
- border:1px solid #7A9437;
-}
-#content tbody {
- border-collapse:collapse;
- background-color:#FFFFFF;
- border:1px solid #7A9437;
-}
-
-#content th {
- font-family:monospace;
- border:1px solid #7A9437;
- padding:5px;
-}
-
-#content td {
- border:1px dotted #7A9437;
- padding:0 3 0 3;
-}
-
-#content table a {
- color:#7A9437;
- text-decoration:none;
-}
-
-#content table a:hover {
- background-color:#CED4BD;
- color:#7A9437;
-}
-
-
-
-table.deps {
- border-collapse:collapse;
- width:90%;
- margin:auto;
- margin-top: 5px;
-}
-
-table.deps thead {
- background-color:#CED4BD;
- border:1px solid #7A9437;
-}
-table.deps tbody {
- border-collapse:collapse;
- background-color:#FFFFFF;
- border:1px solid #7A9437;
-}
-
-table.deps th {
- font-family:monospace;
- border:1px solid #7A9437;
- padding:2;
-}
-
-table.deps td {
- border:1px dotted #7A9437;
- padding:0 3 0 3;
-}
-
-
-
-
-
-table.header {
- border:0;
- width:90%;
- margin:auto;
- margin-top: 5px;
-}
-
-table.header thead {
- border:0;
-}
-table.header tbody {
- border:0;
-}
-table.header tr {
- padding:0px;
- border:0;
-}
-table.header td {
- padding:0 3 0 3;
- border:0;
-}
-
-td.title {
- width:150px;
- margin-right:15px;
-
- font-size:small;
- font-weight:700;
-}
-
-td.title:first-letter {
- color:#7A9437;
- background-color:transparent;
-}
-
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/ivy.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/ivy.sh b/depends/libyarn/releng/make/dependencies/ivy.sh
deleted file mode 100644
index 540e42e..0000000
--- a/depends/libyarn/releng/make/dependencies/ivy.sh
+++ /dev/null
@@ -1,70 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-## ======================================================================
-## Post download untar trigger : For EXT
-## ======================================================================
-
-MODULE=$1
-REVISION=$2
-ORG=$3
-BUILD_STAGE=$4
-
-# ##
-# ## Working directory is the source dir from where we do make
-# ## This will be used in creating ext dir
-# ##
-
-##WORKING_DIRECTORY=$(readlink -m ${PWD}/../../..)
-
-WORKING_DIRECTORY=$(cd ${PWD}/../../..; pwd -P)
-
-EXT_DIRECTORY=${WORKING_DIRECTORY}/ext/${BUILD_STAGE}
-
-if [ ! -d "${EXT_DIRECTORY}" ]; then
- mkdir -p ${EXT_DIRECTORY}
-fi
-
-TOOLS_DIRECTORY=/opt/releng/tools/${ORG}/${MODULE}/${REVISION}/${BUILD_STAGE}
-##
-## Copy all the files from /opt/tools to main/ext
-##
-
-if [ -d "${TOOLS_DIRECTORY}" ] && [ ! -f "${EXT_DIRECTORY}/checksums.${MODULE}-${REVISION}" ]; then
- cd $TOOLS_DIRECTORY
- tar cf - * | ( cd ${EXT_DIRECTORY}; tar xfp -)
-fi
-
-##
-## Softlink all the files for ext.
-##
-## This does require the dirs and sub-dirs to
-## to be created, which is done in the
-## first for loop
-##
-
-# for dir_entry in `find . -type d`
-# do
-# echo "$dir_entry"
-# mkdir -p ${EXT_DIRECTORY}/${BUILDSTAGE}$dir_entry
-# done
-
-# for entry in `find . -type f`
-# do
-# echo "$entry"
-# ln -s ${TOOLS_DIRECTORY}/$entry ${EXT_DIRECTORY}/${BUILDSTAGE}$entry
-# done
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/ivy.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/ivy.xml b/depends/libyarn/releng/make/dependencies/ivy.xml
deleted file mode 100644
index f3ba333..0000000
--- a/depends/libyarn/releng/make/dependencies/ivy.xml
+++ /dev/null
@@ -1,42 +0,0 @@
-<?xml version='1.0'?>
-<!--
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing,
-software distributed under the License is distributed on an
-"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-KIND, either express or implied. See the License for the
-specific language governing permissions and limitations
-under the License.
--->
-
-<ivy-module version="2.0">
- <info organisation="Pivotal"
- module="libyarn" />
-
- <configurations>
- <conf name="osx106_x86" description="osx106_x86" visibility="public"/>
- <conf name="osx106_x86_64" description="osx106_x86_64" visibility="public"/>
- <conf name="rhel5_x86_64" description="rhel5_x86_64" visibility="public"/>
- <conf name="suse10_x86_64" description="suse10_x86_64" visibility="public"/>
- </configurations>
-
- <dependencies>
- <dependency org="gnu" name="libstdc" rev="6.0.13" conf="osx106_x86->osx106_x86;rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- <dependency org="gnu" name="libuuid" rev="2.26" conf="rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- <dependency org="Boost.org" name="boost" rev="1.53.0" conf="osx106_x86->osx106_x86;rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- <dependency org="google" name="protobuf" rev="2.5.0" conf="osx106_x86->osx106_x86;osx106_x86_64->osx106_x86_64;rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- <dependency org="kitware" name="cmake" rev="2.8.11.2" conf="osx106_x86->osx106_x86;osx106_x86_64->osx106_x86_64;rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- <dependency org="mit" name="krb5" rev="1.11.3" conf="osx106_x86->osx106_x86;osx106_x86_64->osx106_x86_64;rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- <dependency org="gnu" name="libgsasl" rev="1.8.0" conf="osx106_x86->osx106_x86;osx106_x86_64->osx106_x86_64;rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- <dependency org="xmlsoft" name="libxml2" rev="2.7.8" conf="osx106_x86->osx106_x86;osx106_x86_64->osx106_x86_64;rhel5_x86_64->rhel5_x86_64;suse10_x86_64->suse10_x86_64" />
- </dependencies>
-</ivy-module>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/ivy_util.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/ivy_util.sh b/depends/libyarn/releng/make/dependencies/ivy_util.sh
deleted file mode 100644
index 37545a3..0000000
--- a/depends/libyarn/releng/make/dependencies/ivy_util.sh
+++ /dev/null
@@ -1,60 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-## ======================================================================
-## Post download untar trigger
-## ======================================================================
-
-FILE=$1
-REVISION=$2
-
-ARTIFACT_DIR=$( dirname ${FILE} )
-
-##
-## Expand tarball
-##
-
-if [ ! -f "${FILE}" ]; then
- echo "WARNING: tarball does not exist (${FILE})"
- exit 2
-else
- pushd ${ARTIFACT_DIR}/..
- gunzip -qc ${FILE} | tar xf -
- popd
-
- if [ $? != 0 ]; then
- echo "FATAL: Problem exapanding tarball (${FILE})"
- exit 1
- fi
-fi
-
-## ----------------------------------------------------------------------
-## Devel package processing
-## Restore object files into loation used for build
-## ----------------------------------------------------------------------
-
-echo "${FILE}" | grep devel
-if [ $? = 0 ]; then
- BASE_PULSE_PROJECT=`tar tvf ${FILE} | head -1 | sed -e 's|\(.* \./\)\(.*\)|\2|' | cut -d '/' -f1`
- BASE_BUILD_STAGE=`tar tvf ${FILE} | head -1 | sed -e 's|\(.* \./\)\(.*\)|\2|' | cut -d '/' -f2`
- rm -rf /opt/releng/build/${BASE_PULSE_PROJECT}/${BASE_BUILD_STAGE}
- cd ..
- [ ! -d /opt/releng/build/${BASE_PULSE_PROJECT} ] && mkdir -p /opt/releng/build/${BASE_PULSE_PROJECT}
- ln -s ${ARTIFACT_DIR}/../${BASE_PULSE_PROJECT}/${BASE_BUILD_STAGE} /opt/releng/build/${BASE_PULSE_PROJECT}/${BASE_BUILD_STAGE}
-fi
-
-exit 0
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/make/dependencies/ivysettings.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/make/dependencies/ivysettings.xml b/depends/libyarn/releng/make/dependencies/ivysettings.xml
deleted file mode 100644
index 236f5f7..0000000
--- a/depends/libyarn/releng/make/dependencies/ivysettings.xml
+++ /dev/null
@@ -1,68 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-
-<!--
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing,
-software distributed under the License is distributed on an
-"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-KIND, either express or implied. See the License for the
-specific language governing permissions and limitations
-under the License.
--->
-
-<ivysettings>
-
- <!-- ====================================================================== -->
-
- <caches defaultCacheDir="/opt/releng/tools"
- lockStrategy="artifact-lock"
- ivyPattern="[organisation]/[module]/[revision]/[type]s/ivy-[revision].xml"
- artifactPattern="[organisation]/[module]/[revision]/[type]s/[artifact]-[revision].[ext]"
- datafilePattern="[organisation]/[module]/ivydata-[revision].properties"
- />
-
- <!-- ====================================================================== -->
-
- <property name="repo.dir" value="/opt/repo"/>
- <settings defaultResolver="public"/>
-
- <!-- ====================================================================== -->
-
- <resolvers>
- <filesystem name="local">
- <ivy pattern="${repo.dir}/[organization]/[module]/[revision]/[type]s/ivy-[revision].xml" />
- <artifact pattern="${repo.dir}/[organization]/[module]/[revision]/[type]s/[artifact]-[revision].[ext]" />
- </filesystem>
- <url name="public">
- <ivy pattern="https://repo.eng.pivotal.io/artifactory/gpdb-ext-release-local/[organization]/[module]/[revision]/[type]s/ivy-[revision].xml" />
- <artifact pattern="https://repo.eng.pivotal.io/artifactory/gpdb-ext-release-local/[organization]/[module]/[revision]/[type]s/[artifact]-[revision].[ext]" />
- </url>
- </resolvers>
-
- <!-- ====================================================================== -->
-
- <triggers>
- <ant-call target="post-resolve-trigger"
- prefix="dep"
- event="post-resolve-dependency"/>
- </triggers>
-
- <triggers>
- <ant-call target="download-untar-trigger"
- prefix="dep"
- event="post-download-artifact"
- filter="type=targz AND status=successful"/>
- </triggers>
-
- <!-- ====================================================================== -->
-
-</ivysettings>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/publish/ivy.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/publish/ivy.xml b/depends/libyarn/releng/publish/ivy.xml
deleted file mode 100644
index 64a9e5d..0000000
--- a/depends/libyarn/releng/publish/ivy.xml
+++ /dev/null
@@ -1,36 +0,0 @@
-<?xml version='1.0'?>
-<!--
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing,
-software distributed under the License is distributed on an
-"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-KIND, either express or implied. See the License for the
-specific language governing permissions and limitations
-under the License.
--->
-
-<ivy-module version="2.0">
- <info organisation="Pivotal" module="libyarn"/>
-
- <configurations>
- <conf name="osx106_x86" description="osx106_x86" visibility="public"/>
- <conf name="rhel5_x86_64" description="rhel5_x86_64" visibility="public"/>
- <conf name="suse10_x86_64" description="suse10_x86_64" visibility="public"/>
- </configurations>
-
- <publications>
- <artifact name="libyarn-osx106_x86" type="targz" ext="targz" conf="osx106_x86"/>
- <artifact name="libyarn-rhel5_x86_64" type="targz" ext="targz" conf="rhel5_x86_64"/>
- <artifact name="libyarn-suse10_x86_64" type="targz" ext="targz" conf="suse10_x86_64"/>
- </publications>
-
-</ivy-module>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/publish/ivysettings.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/publish/ivysettings.xml b/depends/libyarn/releng/publish/ivysettings.xml
deleted file mode 100644
index f825c4e..0000000
--- a/depends/libyarn/releng/publish/ivysettings.xml
+++ /dev/null
@@ -1,56 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-
-<!--
-Licensed to the Apache Software Foundation (ASF) under one
-or more contributor license agreements. See the NOTICE file
-distributed with this work for additional information
-regarding copyright ownership. The ASF licenses this file
-to you under the Apache License, Version 2.0 (the
-"License"); you may not use this file except in compliance
-with the License. You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing,
-software distributed under the License is distributed on an
-"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-KIND, either express or implied. See the License for the
-specific language governing permissions and limitations
-under the License.
--->
-
-<ivysettings>
-
- <!-- ====================================================================== -->
-
- <caches lockStrategy="artifact-lock"
- ivyPattern="[organisation]/[module]/[revision]/[type]s/ivy-[revision].xml"
- artifactPattern="[organisation]/[module]/[revision]/[type]s/[artifact]-[revision].[ext]"
- datafilePattern="[organisation]/[module]/ivydata-[revision].properties"
- />
-
- <!-- ====================================================================== -->
-
- <property name="repo.dir" value="/opt/repo"/>
- <settings defaultResolver="local"/>
-
- <!-- ====================================================================== -->
- <!--Authentication required for publishing (deployment). 'Artifactory Realm' is the realm used by Artifactory so don't change it.-->
- <!-- ====================================================================== -->
-
- <credentials host="repo.eng.pivotal.io" realm="Artifactory Realm" username="build_gpdb" passwd="Piv@gpdb1" />
-
- <!-- ====================================================================== -->
-
- <resolvers>
- <filesystem name="local">
- <ivy pattern="${repo.dir}/[organization]/[module]/[revision]/[type]s/ivy-[revision].xml" />
- <artifact pattern="${repo.dir}/[organization]/[module]/[revision]/[type]s/[artifact]-[revision].[ext]" />
- </filesystem>
- <url name="public">
- <ivy pattern="https://repo.eng.pivotal.io/artifactory/gpdb-ext-release-local/[organization]/[module]/[revision]/[type]s/ivy-[revision].xml" />
- <artifact pattern="https://repo.eng.pivotal.io/artifactory/gpdb-ext-release-local/[organization]/[module]/[revision]/[type]s/[artifact]-[revision].[ext]" />
- </url>
- </resolvers>
-
-</ivysettings>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/set_bld_arch.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/set_bld_arch.sh b/depends/libyarn/releng/set_bld_arch.sh
deleted file mode 100644
index 5f59d4b..0000000
--- a/depends/libyarn/releng/set_bld_arch.sh
+++ /dev/null
@@ -1,63 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-# Define ARCH values
-#
-#
-case "`uname -s`" in
- Darwin)
- sw_vers | grep ProductVersion | grep 10.5
- if [ $? = 0 ]; then
- BLD_ARCH_HOST=osx105_x86
- else
- BLD_ARCH_HOST=osx106_x86
- fi
- ;;
-
- Linux)
- if [ -f /etc/redhat-release ]; then
- case "`cat /etc/redhat-release`" in
- "Red Hat Enterprise Linux WS release 3 (Taroon Update 4)")
- BLD_ARCH_HOST=rhel4_x86_32
- ;;
- *)
- BLD_ARCH_HOST="rhel`cat /etc/redhat-release | sed -e 's/Red Hat Enterprise Linux/RedHat/' -e 's/WS//' -e 's/Server//' -e 's/Client//' | awk '{print $3}' | awk -F. '{print $1}'`_`uname -p | sed -e s/i686/x86_32/`"
- ;;
- esac
- fi
- if [ -f /etc/SuSE-release ]; then
- case "`cat /etc/SuSE-release`" in
- *)
- BLD_ARCH_HOST="suse`head -1 /etc/SuSE-release | sed -e 's/SUSE Linux Enterprise Server/SuSE/' | awk '{print $2}'`_`uname -p | sed -e s/i686/x86_32/`"
- ;;
- esac
- fi
- ;;
-
- SunOS)
- BLD_ARCH_HOST="sol`uname -r | awk -F. ' {print $2}'`_`isainfo | awk ' {print $1}' | sed -e s/amd64/x86_64/ | sed -e s/sparcv9/sparc_64/`"
- ;;
-
- AIX)
- BLD_ARCH_HOST="aix`uname -v`_`prtconf | grep 'Processor Type' | awk '{print $3}' | perl -p -i -e 's,PowerPC_POWER5,ppc,'`_`prtconf -k | perl -p -i -e 's,Kernel Type: (\d+)-bit,\1,'`"
- ;;
-
- *)
- BLD_ARCH_HOST="BLD_ARCH_unknown"
- ;;
-esac
-
-echo ${BLD_ARCH_HOST}
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/src/port/win32.ico
----------------------------------------------------------------------
diff --git a/src/port/win32.ico b/src/port/win32.ico
deleted file mode 100755
index a58ee43..0000000
Binary files a/src/port/win32.ico and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/__init__.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/__init__.pyc b/tools/bin/ext/yaml/__init__.pyc
deleted file mode 100644
index 003ad21..0000000
Binary files a/tools/bin/ext/yaml/__init__.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/composer.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/composer.pyc b/tools/bin/ext/yaml/composer.pyc
deleted file mode 100644
index ea754d3..0000000
Binary files a/tools/bin/ext/yaml/composer.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/constructor.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/constructor.pyc b/tools/bin/ext/yaml/constructor.pyc
deleted file mode 100644
index c127d54..0000000
Binary files a/tools/bin/ext/yaml/constructor.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/cyaml.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/cyaml.pyc b/tools/bin/ext/yaml/cyaml.pyc
deleted file mode 100644
index 5829abe..0000000
Binary files a/tools/bin/ext/yaml/cyaml.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/dumper.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/dumper.pyc b/tools/bin/ext/yaml/dumper.pyc
deleted file mode 100644
index 611730d..0000000
Binary files a/tools/bin/ext/yaml/dumper.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/emitter.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/emitter.pyc b/tools/bin/ext/yaml/emitter.pyc
deleted file mode 100644
index 9dfa1b1..0000000
Binary files a/tools/bin/ext/yaml/emitter.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/error.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/error.pyc b/tools/bin/ext/yaml/error.pyc
deleted file mode 100644
index ead2e7d..0000000
Binary files a/tools/bin/ext/yaml/error.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/events.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/events.pyc b/tools/bin/ext/yaml/events.pyc
deleted file mode 100644
index 9fd228f..0000000
Binary files a/tools/bin/ext/yaml/events.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/loader.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/loader.pyc b/tools/bin/ext/yaml/loader.pyc
deleted file mode 100644
index cbd7503..0000000
Binary files a/tools/bin/ext/yaml/loader.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/nodes.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/nodes.pyc b/tools/bin/ext/yaml/nodes.pyc
deleted file mode 100644
index 337f71a..0000000
Binary files a/tools/bin/ext/yaml/nodes.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/parser.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/parser.pyc b/tools/bin/ext/yaml/parser.pyc
deleted file mode 100644
index fb23a0b..0000000
Binary files a/tools/bin/ext/yaml/parser.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/reader.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/reader.pyc b/tools/bin/ext/yaml/reader.pyc
deleted file mode 100644
index e41c788..0000000
Binary files a/tools/bin/ext/yaml/reader.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/representer.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/representer.pyc b/tools/bin/ext/yaml/representer.pyc
deleted file mode 100644
index f12198a..0000000
Binary files a/tools/bin/ext/yaml/representer.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/resolver.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/resolver.pyc b/tools/bin/ext/yaml/resolver.pyc
deleted file mode 100644
index 8f83512..0000000
Binary files a/tools/bin/ext/yaml/resolver.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/scanner.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/scanner.pyc b/tools/bin/ext/yaml/scanner.pyc
deleted file mode 100644
index 3ae15e8..0000000
Binary files a/tools/bin/ext/yaml/scanner.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/serializer.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/serializer.pyc b/tools/bin/ext/yaml/serializer.pyc
deleted file mode 100644
index 24fce03..0000000
Binary files a/tools/bin/ext/yaml/serializer.pyc and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/tools/bin/ext/yaml/tokens.pyc
----------------------------------------------------------------------
diff --git a/tools/bin/ext/yaml/tokens.pyc b/tools/bin/ext/yaml/tokens.pyc
deleted file mode 100644
index 47b5986..0000000
Binary files a/tools/bin/ext/yaml/tokens.pyc and /dev/null differ
[2/2] incubator-hawq git commit: HAWQ-186. Remove some useless files
Posted by wl...@apache.org.
HAWQ-186. Remove some useless files
Project: http://git-wip-us.apache.org/repos/asf/incubator-hawq/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-hawq/commit/10909db4
Tree: http://git-wip-us.apache.org/repos/asf/incubator-hawq/tree/10909db4
Diff: http://git-wip-us.apache.org/repos/asf/incubator-hawq/diff/10909db4
Branch: refs/heads/master
Commit: 10909db4a2f0a20121f14e666eb0e253547b9ecb
Parents: d9e1578
Author: Wen Lin <wl...@pivotal.io>
Authored: Thu Dec 17 12:55:30 2015 +0800
Committer: Wen Lin <wl...@pivotal.io>
Committed: Thu Dec 17 12:55:30 2015 +0800
----------------------------------------------------------------------
depends/libyarn/releng/Makefile | 235 --------
depends/libyarn/releng/bin/cleanup-phd.sh | 43 --
depends/libyarn/releng/bin/gphdinst.sh | 536 -------------------
depends/libyarn/releng/bin/ha_failover.sh | 36 --
depends/libyarn/releng/bin/install-phd.sh | 71 ---
depends/libyarn/releng/bin/setup_gphdinst.sh | 175 ------
.../bin/templates/.core-site-secure-ha.xml.swp | Bin 12288 -> 0 bytes
.../releng/bin/templates/core-site-ha.xml | 28 -
.../bin/templates/core-site-secure-ha.xml | 39 --
.../releng/bin/templates/core-site-secure.xml | 34 --
.../libyarn/releng/bin/templates/core-site.xml | 24 -
.../releng/bin/templates/download-config | 5 -
.../releng/bin/templates/hadoop-env-ha.sh | 72 ---
.../releng/bin/templates/hadoop-env-secure.sh | 73 ---
.../libyarn/releng/bin/templates/hadoop-env.sh | 72 ---
.../templates/hdfs-site-datanode-secure-ha.xml | 240 ---------
.../bin/templates/hdfs-site-datanode-secure.xml | 193 -------
.../releng/bin/templates/hdfs-site-ha.xml | 172 ------
.../templates/hdfs-site-namenode-secure-ha.xml | 224 --------
.../bin/templates/hdfs-site-namenode-secure.xml | 181 -------
.../bin/templates/hdfs-site-secure-ha.xml | 240 ---------
.../releng/bin/templates/hdfs-site-secure.xml | 193 -------
.../libyarn/releng/bin/templates/hdfs-site.xml | 123 -----
.../releng/bin/templates/log4j-ha.properties | 75 ---
.../libyarn/releng/bin/templates/yarn-env.sh | 87 ---
.../libyarn/releng/bin/templates/yarn-site.xml | 112 ----
depends/libyarn/releng/bin/templates/zoo.cfg | 8 -
depends/libyarn/releng/etc/ivy.xml | 38 --
depends/libyarn/releng/etc/local/.gitignore | 1 -
.../releng/etc/local/hdfs-site.xml.simple | 65 ---
.../libyarn/releng/make/dependencies/.gitignore | 1 -
.../libyarn/releng/make/dependencies/build.xml | 84 ---
.../libyarn/releng/make/dependencies/cacerts | Bin 83678 -> 0 bytes
.../releng/make/dependencies/ivy-report.css | 279 ----------
depends/libyarn/releng/make/dependencies/ivy.sh | 70 ---
.../libyarn/releng/make/dependencies/ivy.xml | 42 --
.../releng/make/dependencies/ivy_util.sh | 60 ---
.../releng/make/dependencies/ivysettings.xml | 68 ---
depends/libyarn/releng/publish/ivy.xml | 36 --
depends/libyarn/releng/publish/ivysettings.xml | 56 --
depends/libyarn/releng/set_bld_arch.sh | 63 ---
src/port/win32.ico | Bin 22486 -> 0 bytes
tools/bin/ext/yaml/__init__.pyc | Bin 11360 -> 0 bytes
tools/bin/ext/yaml/composer.pyc | Bin 4204 -> 0 bytes
tools/bin/ext/yaml/constructor.pyc | Bin 22333 -> 0 bytes
tools/bin/ext/yaml/cyaml.pyc | Bin 3846 -> 0 bytes
tools/bin/ext/yaml/dumper.pyc | Bin 2614 -> 0 bytes
tools/bin/ext/yaml/emitter.pyc | Bin 32743 -> 0 bytes
tools/bin/ext/yaml/error.pyc | Bin 3040 -> 0 bytes
tools/bin/ext/yaml/events.pyc | Bin 5094 -> 0 bytes
tools/bin/ext/yaml/loader.pyc | Bin 1925 -> 0 bytes
tools/bin/ext/yaml/nodes.pyc | Bin 2236 -> 0 bytes
tools/bin/ext/yaml/parser.pyc | Bin 14891 -> 0 bytes
tools/bin/ext/yaml/reader.pyc | Bin 6802 -> 0 bytes
tools/bin/ext/yaml/representer.pyc | Bin 15117 -> 0 bytes
tools/bin/ext/yaml/resolver.pyc | Bin 6782 -> 0 bytes
tools/bin/ext/yaml/scanner.pyc | Bin 34012 -> 0 bytes
tools/bin/ext/yaml/serializer.pyc | Bin 4430 -> 0 bytes
tools/bin/ext/yaml/tokens.pyc | Bin 6678 -> 0 bytes
59 files changed, 4154 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/Makefile
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/Makefile b/depends/libyarn/releng/Makefile
deleted file mode 100644
index 851f819..0000000
--- a/depends/libyarn/releng/Makefile
+++ /dev/null
@@ -1,235 +0,0 @@
-## ======================================================================
-##
-## ======================================================================
-
-BLD_TOP:=$(shell sh -c pwd)
-SHELL=/bin/bash
-
-##-------------------------------------------------------------------------------------
-## dependent modules
-##
-## NOTE: Dependent project module version is kept in $(BLD_TOP)/make/dependencies/ivy.xml
-##-------------------------------------------------------------------------------------
-
-GREP_SED_VAR = $(BLD_TOP)/make/dependencies/ivy.xml | sed -e 's|\(.*\)rev="\(.*\)"[ ]*conf\(.*\)|\2|'
-CMAKE_VER = $(shell grep "\"cmake\"" $(GREP_SED_VAR))
-
-TARGET_VERSION = ${LIBHDFS3_VERSION}
-
-ifeq "$(LIBHDFS3_VERSION)" ""
-TARGET_VERSION = `cat ${BLD_TOP}/../build/version`
-endif
-
-ifeq "$(LIBHDFS3_NOW)" ""
-TARGET_TIME = `date +%Y%m%d%H%M%S`
-endif
-
-ifeq "$(LIBHDFS3_TEST_URL)" ""
-LIBHDFS3_TEST_URL = localhost:9000
-endif
-
-ifneq "$(LIBYARN_PUBLISH_OVERWRITE)" ""
-LIBYARN_PUBLISH_OVERWRITE = -overwrite
-endif
-
-##-------------------------------------------------------------------------------------
-
-# Determine the default target architecture on this machine, if override not set
-ifeq "$(BLD_ARCH)" ""
-export BLD_ARCH:=$(shell $(BLD_TOP)/set_bld_arch.sh)
-endif
-
-export PATH:=${BLD_TOP}/../ext/${BLD_ARCH}/bin:${BLD_TOP}/../ext/${BLD_ARCH}/cmake-${CMAKE_VER}/bin:${PATH}
-
-# construct test directory prefix on hdfs.
-CUR_DATE=$(shell date +%s)
-TEST_HDFS_PREFIX="${BLD_ARCH}/${CUR_DATE}/"
-
-# prepare HDFS instance
-ifeq "${HDFS_CONFIG}" ""
-export HDFS_CONFIG:=local
-endif
-
-export HDFS_CONFIG_PATH:=${BLD_TOP}/etc/${HDFS_CONFIG}/hdfs-site.xml
-
-ifeq "${HDFS_USER}" ""
-export HDFS_USER:=${USER}
-endif
-
-ifeq "${DATA_PATH}" ""
-export DATA_PATH:=/tmp/libyarn-pulse
-endif
-
-ifneq "${LIBHDFS_HDFS_SUPERUSER}" ""
-export HDFS_SUPERUSER_OPTION:=-DHDFS_SUPERUSER=${LIBHDFS_HDFS_SUPERUSER}
-endif
-
-# include thirdparty infrastructure which depends on WHERE_THE...
-BLD_WHERE_THE_LIBRARY_THINGS_ARE=LD_LIBRARY_PATH
-osx106_x86_WHERE_THE_LIBRARY_THINGS_ARE=DYLD_LIBRARY_PATH
-ifneq "$($(BLD_ARCH)_WHERE_THE_LIBRARY_THINGS_ARE)" ""
-BLD_WHERE_THE_LIBRARY_THINGS_ARE=$($(BLD_ARCH)_WHERE_THE_LIBRARY_THINGS_ARE)
-endif
-
-SED_INLINE_EDIT_BAKUP_FILE=
-
-ifeq "$(BLD_ARCH)" "osx106_x86"
- SED_INLINE_EDIT_BAKUP_FILE=''
- export $(BLD_WHERE_THE_LIBRARY_THINGS_ARE):=/opt/gcc-4.4.2-osx106/lib:${BLD_TOP}/../ext/${BLD_ARCH}/lib:$($(BLD_WHERE_THE_LIBRARY_THINGS_ARE))
-else
- export $(BLD_WHERE_THE_LIBRARY_THINGS_ARE):=/opt/gcc-4.4.2/lib64:${BLD_TOP}/../ext/${BLD_ARCH}/lib:$($(BLD_WHERE_THE_LIBRARY_THINGS_ARE))
-endif
-
-##-------------------
-## set compiler
-##-------------------
-
-CCOMPILER=${CC}
-CXXCOMPILER=${CXX}
-
-ifeq "${CCOMPILER}" ""
-CCOMPILER=gcc
-endif
-
-ifeq "${CXXCOMPILER}" ""
-CXXCOMPILER=g++
-endif
-
-## ----------------------------------------------------------------------
-## libhdfs targets
-## ----------------------------------------------------------------------
-
-all: libhdfs-sync_tools libhdfs-clean libhdfs-configure libhdfs-build libhdfs-unittest
-
-${BLD_TOP}/../build:
- mkdir ${BLD_TOP}/../build
-
-libhdfs-sync_tools:
- make sync_tools
-
-libhdfs-clean:
- @rm -rf ${BLD_TOP}/../dist ${BLD_TOP}/../build
-
-libhdfs-configure: ${BLD_TOP}/../build
- (cd ${BLD_TOP}/../build && cmake ${BLD_TOP}/.. -DTEST_HDFS_PREFIX=${TEST_HDFS_PREFIX} $(LIBHDFS3_CONF_FLAGS) \
- -DCMAKE_C_COMPILER=${CCOMPILER} -DCMAKE_CXX_COMPILER=${CXXCOMPILER} \
- -DCMAKE_INSTALL_PREFIX=${BLD_TOP}/../dist/${BLD_ARCH} \
- -DCMAKE_PREFIX_PATH=${BLD_TOP}/../ext/${BLD_ARCH} \
- ${HDFS_SUPERUSER_OPTION})
-
-libhdfs-build:
- (cd ${BLD_TOP}/../build && make)
-
-libhdfs-install:
- rm -rf ${BLD_TOP}/../dist/${BLD_ARCH}
- (cd ${BLD_TOP}/../build && make install)
-
-libhdfs-test:
- (cd ${BLD_TOP}/../build && make test)
-
-libhdfs-showcov:
- (cd ${BLD_TOP}/../build && make ShowCoverage)
-
-libhdfs-functiontest:
- (export GTEST_OUTPUT="xml:functiontest.xml"; cd ${BLD_TOP}/../build && make functiontest)
-
-libhdfs-securetest:
- (export GTEST_OUTPUT="xml:securetest.xml"; cd ${BLD_TOP}/../build && make securetest)
-
-libhdfs-unittest:
- (export GTEST_OUTPUT="xml:unittest.xml"; cd ${BLD_TOP}/../build && make unittest)
-
-libhdfs-update-function-test.xml:
- (cd ${BLD_TOP}/../test/data && chmod a+w function-test.xml && sed -e 's|localhost:9000|${LIBHDFS3_TEST_URL}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} function-test.xml)
- (cd ${BLD_TOP}/../test/data && chmod a+w function-secure.xml && sed -e 's|localhost:9000|${LIBHDFS3_TEST_URL}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} function-secure.xml)
-
-libhdfs-package: libhdfs-install
- (cd ${BLD_TOP}/../dist/${BLD_ARCH} && find . -type f -or -type l | xargs ${MD5} > /tmp/checksums.libyarn && mv /tmp/checksums.libyarn ${BLD_TOP}/../dist/${BLD_ARCH} && \
- cd ${BLD_TOP}/../dist/ && tar -czvf libyarn-${BLD_ARCH}-${TARGET_VERSION}.targz ${BLD_ARCH} )
-
-libhdfs-update-ivys:
- (cd ${BLD_TOP}/../dist/ && mkdir -p ${TARGET_VERSION}/ivys && cp ${BLD_TOP}/etc/ivy.xml ${TARGET_VERSION}/ivys/ivy-${TARGET_VERSION}.xml \
- && sed -e 's|<version>|'${TARGET_VERSION}'|' -i ${SED_INLINE_EDIT_BAKUP_FILE} ${TARGET_VERSION}/ivys/ivy-${TARGET_VERSION}.xml \
- && sed -e 's|<time>|'${TARGET_TIME}'|' -i ${SED_INLINE_EDIT_BAKUP_FILE} ${TARGET_VERSION}/ivys/ivy-${TARGET_VERSION}.xml)
-
-libhdfs-ivy-package: libhdfs-package libhdfs-update-ivys
- (cd ${BLD_TOP}/../dist/ && mkdir -p ${TARGET_VERSION}/targzs && cp libyarn-${BLD_ARCH}-${TARGET_VERSION}.targz ${TARGET_VERSION}/targzs/ && tar -cjvf libyarn-${BLD_ARCH}-${TARGET_VERSION}.tar.bz2 ${TARGET_VERSION})
-
-libhdfs-update-hdfs-conf:
- (cd ${BLD_TOP}/etc/${HDFS_CONFIG} && cp -f hdfs-site.xml.simple hdfs-site.xml && sed -e 's|@user@|${HDFS_USER}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} hdfs-site.xml \
- && mkdir -p ${DATA_PATH} && sed -e 's|@data_path@|${DATA_PATH}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} hdfs-site.xml)
-
-libhdfs-cleanup-phd:
- (${BLD_TOP}/bin/cleanup-phd.sh)
-
-libhdfs-install-phd: libhdfs-update-hdfs-conf
- (${BLD_TOP}/bin/install-phd.sh)
-
-## ----------------------------------------------------------------------
-## Sync/Clean tools
-## ----------------------------------------------------------------------
-## Populate/clean up dependent releng supported tools. The projects are
-## downloaded and installed into /opt/releng/...
-##
-## Tool dependencies and platform config mappings are defined in:
-## * Apache Ivy dependency definition file
-## releng/make/dependencies/ivy.xml
-## ----------------------------------------------------------------------
-
-opt_write_test:
- @if [ ! -w /opt ]; then \
- echo ""; \
- echo "======================================================================"; \
- echo "ERROR: /opt is not writable."; \
- echo "----------------------------------------------------------------------"; \
- echo " Supporting tools are stored in /opt. Please ensure you have"; \
- echo " write access to /opt"; \
- echo "======================================================================"; \
- echo ""; \
- exit 1; \
- fi
-
-/opt/releng/apache-ant:
- ${MAKE} opt_write_test
- echo "Sync Ivy project dependency management framework ..."
- curl --silent http://build-prod.dh.greenplum.com/tools/apache-ant.1.8.1.tar.gz -o /tmp/apache-ant.1.8.1.tar.gz
- ( umask 002; [ ! -d /opt/releng ] && mkdir -p /opt/releng; \
- cd /opt/releng; \
- gunzip -qc /tmp/apache-ant.1.8.1.tar.gz | tar xf -; \
- rm -f /tmp/apache-ant.1.8.1.tar.gz; \
- chmod -R a+w /opt/releng/apache-ant )
-
-sync_tools: opt_write_test /opt/releng/apache-ant
- @cd make/dependencies; \
- (umask 002; /opt/releng/apache-ant/bin/ant -DBLD_ARCH=$(BLD_ARCH) resolve); \
- echo "Resolve finished"
-
-clean_tools: opt_write_test
- @cd releng/make/dependencies; \
- /opt/releng/apache-ant/bin/ant clean; \
- rm -rf /opt/releng/apache-ant; \
-
-# Publish Libhdfs
-local_publish: /opt/releng/apache-ant
- @if [ -f publish/ivy.xml ]; then \
- java -Xmx512M -jar /opt/releng/apache-ant/lib/ivy-2.2.0.jar \
- -settings publish/ivysettings.xml \
- -ivy publish/ivy.xml \
- -publish local \
- -publishpattern "../dist/[artifact]-[revision].[ext]" \
- -revision ${TARGET_VERSION} \
- -status release \
- -overwrite; \
- fi;
-public_publish: /opt/releng/apache-ant
- @if [ -f publish/ivy.xml ]; then \
- java -Xmx512M -Djavax.net.ssl.trustStore=$(BLD_TOP)/make/dependencies/cacerts \
- -jar /opt/releng/apache-ant/lib/ivy-2.2.0.jar \
- -settings publish/ivysettings.xml \
- -ivy publish/ivy.xml \
- -publish public \
- -publishpattern "../dist/[artifact]-[revision].[ext]" \
- -revision ${TARGET_VERSION} \
- -status release \
- $(LIBYARN_PUBLISH_OVERWRITE); \
- fi;
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/cleanup-phd.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/cleanup-phd.sh b/depends/libyarn/releng/bin/cleanup-phd.sh
deleted file mode 100644
index 7a9c2ff..0000000
--- a/depends/libyarn/releng/bin/cleanup-phd.sh
+++ /dev/null
@@ -1,43 +0,0 @@
-#!/bin/sh
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-if [ "x${DATA_PATH}" != "x" ]; then
- if [ -e ${DATA_PATH} ]; then
- echo "dump namenode log"
- cat ${DATA_PATH}/hadoop*/logs/*namenode*log*
- echo "dump datanode log"
- cat ${DATA_PATH}/hadoop*/logs/*datanode*log*
- fi
-
- if [ "x${KEEP_LOG}" != "xtrue" ]; then
- rm -rf ${DATA_PATH}/*
- fi
-
- rm -rf ${DATA_PATH}/fs
- rm -rf ${DATA_PATH}/data
-fi
-
-NAMENODE_PIDS=`ps aux | grep java | grep NameNode | awk '{print $2}'`
-if [ "x${NAMENODE_PIDS}" != "x" ]; then
- kill -9 ${NAMENODE_PIDS}
-fi
-
-DATANODE_PIDS=`ps aux | grep java | grep DataNode | awk '{print $2}'`
-if [ "x${DATANODE_PIDS}" != "x" ]; then
- kill -9 ${DATANODE_PIDS}
-fi
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/gphdinst.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/gphdinst.sh b/depends/libyarn/releng/bin/gphdinst.sh
deleted file mode 100644
index 99b7085..0000000
--- a/depends/libyarn/releng/bin/gphdinst.sh
+++ /dev/null
@@ -1,536 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-## ======================================================================
-## TODO
-## * make script work for single and muli-node installs
-## * should we be able to identify the usage of the hdfs instance
-## * need to determine how to setup/start httpd
-## * how to pass in server name values
-## ======================================================================
-
-# Do not run on gpdb9.rel.dh.greenplum.com as this script will delete its shared HDFS cluster
-if [[ $(hostname) =~ "gpdb9.rel" ]]; then
- echo "ABORT! Do not run this script on gpdb9.rel.dh.greenplum.com as this script will delete its shared cluster!"
- exit 1
-fi
-
-# Make sure only RHEL6.2 machines can run our script
-cat /etc/redhat-release | grep "6." > /dev/null
-if [ $? -ne 0 ]; then
- echo "This script must be run on a RHEL6.2 cluster"
- exit 1
-fi
-
-# Make sure only root can run our script
-if [[ $EUID -ne 0 ]]; then
- echo "This script must be run as root" 1>&2
- exit 1
-fi
-
-# Make sure if using HDFS HA that there are at least 3 datanodes not including a namenode
-if [ "${ENABLE_QUORUM}" = "true" ] && [ ! $( echo ${DATANODES} | wc -w ) -ge 3 ]; then
- echo "You must have at least three datanodes to have as the zkservers"
- exit 1
-fi
-
-##
-## Courtesy - http://stevemorin.blogspot.com/2007/10/bash-get-self-directory-trick.html
-## Updated to run on rhel3 (kite12)
-##
-
-script_path="$(cd $(dirname $0); pwd -P)/$(basename $0)"
-
-[[ ! -f "$script_path" ]] && script_path="$(cd $(/usr/bin/dirname "$0"); pwd -P)/$(basename "$0")"
-[[ ! -f "$script_path" ]] && script_path="" && echo 'No full path to running script found!' && exit 1
-
-script_dir="${script_path%/*}"
-
-##
-## Global variables
-##
-
-DCA_SETUP=${DCA_SETUP:=false}
-DOWNLOAD_DIR=${DOWNLOAD_DIR:=~/downloads}
-ARCHIVE_URL=${ARCHIVE_URL:=http://hdsh132.lss.emc.com:8080/view/Ongoing-sprint-release-build/job/HudsonHD2_0_1_0ReleaseBuild_ongoing_sprint_release/lastSuccessfulBuild/artifact/*zip*/archive.zip}
-ARCHIVE=$( basename ${ARCHIVE_URL} )
-HADOOP_CONF_DIR=/etc/gphd/hadoop/conf
-ZK_CONF_DIR=/etc/gphd/zookeeper/conf
-NAMENODE=${NAMENODE:=smdw}
-SECONDARY_NAMENODE=${SECONDARY_NAMENODE:=mdw}
-DATANODES=${DATANODES:=sdw1 sdw2}
-RESOURCEMANAGER=${RESOURCEMANAGER:=smdw}
-NODEMANAGERS=${DATANODES:=sdw1 sdw2}
-HDFS_DATA_DIRECTORY_ROOT=${HDFS_DATA_DIRECTORY_ROOT:=/data}
-JOURNAL_DATA_ROOT=${JOURNAL_DATA_ROOT:=/data/journal}
-export HADOOP_HOME=${HADOOP_HOME:=/usr/lib/gphd/hadoop}
-export HADOOP_YARN_HOME=${HADOOP_YARN_HOME:=/usr/lib/gphd/hadoop-yarn}
-export ZK_HOME=${ZK_HOME:=/usr/lib/gphd/zookeeper}
-export ZOO_LOG_DIR=${ZOO_LOG_DIR:=/tmp/zookeeper/logs}
-export JAVA_HOME=${JAVA_HOME:=/opt/jdk1.7.0_15}
-
-## ======================================================================
-## Functions
-## ======================================================================
-
-## ----------------------------------------------------------------------
-## gphd_stop_remove()
-## o Stop any existing namenode and datanode processes
-## o Display running java processes with "jps" command
-## o Remove hadoop*, bigtop* and zookeeper rpms
-## o Remove /usr/lib/gphd
-## ----------------------------------------------------------------------
-
-display_jps(){
- echo "======================================================================"
- for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
- echo $node
- ssh $node "${JAVA_HOME}/bin/jps"
- echo
- done
- echo "======================================================================"
-}
-
-gphd_stop_remove(){
-
- ## Sanity Stop request for NN, 2NN, and DN only (Actual stop is when run finishes in sys_mgmt_utils.sh)
- ssh ${NAMENODE} "[ -f ${HADOOP_HOME}/sbin/hadoop-daemon.sh ] && sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh stop namenode"
- ssh ${SECONDARY_NAMENODE} "[ -f ${HADOOP_HOME}/sbin/hadoop-daemon.sh ] && sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh stop secondarynamenode"
-
- for node in ${DATANODES}; do
- ssh $node "[ -f ${HADOOP_HOME}/sbin/hadoop-daemon.sh ] && sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh stop datanode"
- done
-
- display_jps
-
- for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
- ssh $node "rm -rf /data*/hdfs /tmp/hadoop-*.pid /tmp/hsperfdata_*;"
- done
-
- # Remove hadoop, bigtop, and zookeeper rpms in each node
- for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
- ssh $node '\
- hostname; \
- for rpm in hadoop bigtop zookeeper; do \
- echo "----------------------------------------------------------------------"; \
- echo "Processing $rpm"; \
- rpm -qa | grep $rpm; \
- if [ $? = 0 ]; then \
- yum remove -y $( rpm -qa | grep $rpm ); \
- rpm -qa | grep $rpm; \
- fi; \
- rm -rf /etc/gphd/hadoop* /etc/gphd/zookeeper* /usr/lib/gphd/zookeeper* /var/lib/gphd/hadoop* /var/log/gphd/hadoop* /var/lib/gphd/zookeeper*; \
- echo "----------------------------------------------------------------------"; \
- done'
- done
-}
-
-retrieve_hdfs_archive(){
- echo "======================================================================"
-
- echo "Retrieving ${ARCHIVE_URL}"
-
- if [ ! -d "${DOWNLOAD_DIR}" ]; then
- mkdir ${DOWNLOAD_DIR}
- fi
-
- rm -rf ${DOWNLOAD_DIR}/*
-
- wget --no-verbose ${ARCHIVE_URL} -O ${DOWNLOAD_DIR}/${ARCHIVE}
- if [ $? != 0 ]; then
- echo "FATAL: retrieval failed (${ARCHIVE_URL})."
- exit 2
- fi
-
- echo "======================================================================"
-}
-
-## ######################################################################
-## MAIN Script
-## ######################################################################
-
-##
-## If necessary, stop hadoop services and remove previously installed
-## rpms (and files).
-##
-
-gphd_stop_remove
-
-##
-## Retrieve latest hdfs archive from jenkins server
-##
-
-retrieve_hdfs_archive
-
-##
-## Process retrieved archive
-##
-
-cd ${DOWNLOAD_DIR}
-
-if [ ! -f "${ARCHIVE}" ]; then
- echo "FATAL: archive does note exist (${ARCHIVE})"
- exit 2
-fi
-
-ls | grep archive.zip >> /dev/null
-if [ $? == 0 ]; then
- unzip ${ARCHIVE}
- if [ $? != 0 ]; then
- echo "FATAL: unzip of archive failed (${ARCHIVE})."
- exit 2
- fi
-
- cd archive/dist
-fi
-
-GPHD_TARBALL=$( ls )
-if [ ! -f "${GPHD_TARBALL}" ]; then
- echo "FATAL: gphd tarball does not exist."
- exit 2
-fi
-
-GPHD_DIR=$( basename * .tar.gz )
-
-tar xf ${GPHD_TARBALL}
-if [ $? != 0 ]; then
- echo "FATAL: tar expansion failed (${GPHD_TARBALL})."
- exit 2
-fi
-
-##
-## Prepare file based yum repository
-##
-
-echo "======================================================================"
-
-cd ${GPHD_DIR}
-REPO_DIR=$( pwd )
-createrepo ${REPO_DIR}
-
-echo "======================================================================"
-
-rm -f /etc/yum.repos.d/gphd.repo
-
-cat > /etc/yum.repos.d/gphd.repo <<-EOF
- [gphd]
- name=Greenplum HD repository
- baseurl=http://${SECONDARY_NAMENODE}/yum/base
- enabled=1
- gpgcheck=0
-EOF
-
-if [ ! -d /var/www/html/yum ] ; then
- mkdir -p /var/www/html/yum
- chmod 777 /var/www/html/yum
-fi
-rm -rf /var/www/html/yum/base
-
-rsync -auv ${REPO_DIR}/* /var/www/html/yum/base
-chmod -R 777 /var/www/html/yum/base
-cd ${DOWNLOAD_DIR}
-rm -rf ${DOWNLOAD_DIR}/*
-
-service httpd start
-
-##
-## Configure hdfs installation
-##
-
-if [ "${ENABLE_QUORUM}" = "true" ] && [ -n "${KDC}" ] ; then
- CORE_SITE=core-site-secure-ha.xml
- HADOOP_ENV=hadoop-env-secure.sh
- HDFS_SITE=hdfs-site-secure-ha.xml
-elif [ "${ENABLE_QUORUM}" = "true" ] ; then
- CORE_SITE=core-site-ha.xml
- HADOOP_ENV=hadoop-env-ha.sh
- HDFS_SITE=hdfs-site-ha.xml
-elif [ -n "${KDC}" ] ; then
- CORE_SITE=core-site-secure.xml
- HADOOP_ENV=hadoop-env-secure.sh
- HDFS_SITE=hdfs-site-secure.xml
-else
- CORE_SITE=core-site.xml
- HADOOP_ENV=hadoop-env.sh
- HDFS_SITE=hdfs-site.xml
-fi
-
-
-if [ "${ENABLE_QUORUM}" = "true" ] ; then
- ZOO_LOG=log4j-ha.properties
- ZOO_CFG=zoo.cfg
-
- # Deploy zookeeper and jounalnode depend on datanodes number.
- datanode_num=0
-
- for node in ${DATANODES/%${SECONDARY_NAMENODE}/}; do
- datanode_num=`expr $datanode_num + 1`
- done
-
- for num in {1..3}; do
- declare ZKSERVER_$num=$( echo $DATANODES | cut -d" " -f ${num} )
-
- if [ $datanode_num -ge 6 ]; then
- num_jn=`expr $num + 3`
- elif [ $datanode_num -gt 3 ]; then
- num_jn=`expr $datanode_num - 3 + $num`
- else
- num_jn=$num
- fi
-
- declare JOURNALNODE_$num=$( echo $DATANODES | cut -d " " -f ${num_jn} )
- done
-fi
-
-sed -e "s|%HDFS_HOST%|${NAMENODE}:9000|" --in-place=.orig ${script_dir}/templates/${CORE_SITE}
-
-if [ "${DCA_SETUP}" = true ]; then
- sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" \
- -e "s|%NAMENODE_MEMORY%|-Xmx10240m|" \
- -e "s|%DATANODE_MEMORY%|-Xmx8192m|" \
- -e "s|%SECONDARYNAMENODE_MEMORY%|-Xmx10240m|" \
- -e "s|%HADOOP_HOME%|${HADOOP_HOME}|" \
- --in-place=.orig ${script_dir}/templates/${HADOOP_ENV}
-elif [ "${DCA_SETUP}" = "half" ]; then
- sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" \
- -e "s|%NAMENODE_MEMORY%|-Xmx10240m|" \
- -e "s|%DATANODE_MEMORY%|-Xmx8192m|" \
- -e "s|%SECONDARYNAMENODE_MEMORY%|-Xmx10240m|" \
- -e "s|%HADOOP_HOME%|${HADOOP_HOME}|" \
- --in-place=.orig ${script_dir}/templates/${HADOOP_ENV}
-else
- sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" \
- -e "s|%NAMENODE_MEMORY%|-Xmx4096m|" \
- -e "s|%DATANODE_MEMORY%|-Xmx2048m|" \
- -e "s|%SECONDARYNAMENODE_MEMORY%|-Xmx2048m|" \
- -e "s|%HADOOP_HOME%|${HADOOP_HOME}|" \
- --in-place=.orig ${script_dir}/templates/${HADOOP_ENV}
-fi
-
-if [ "${TWO_NAMENODE_DIR}" = true ]; then
- sed -e "s|%DATA_DIR_ROOT%/hdfs/name<|${HDFS_DATA_DIRECTORY_ROOT}/hdfs/name,file:${HDFS_DATA_DIRECTORY_ROOT}/hdfs/name2<|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-fi
-
-sed -e "s|%DATA_DIR_ROOT%|${HDFS_DATA_DIRECTORY_ROOT}|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-
-DATA_FILESYSTEMS=$( ssh $( echo ${DATANODES} | cut -d" " -f 1 ) "df -kh 2> /dev/null | grep data | wc -l" )
-
-if [ $DATA_FILESYSTEMS -ge 2 ] ; then
- echo "/data1 and /data2 found... using both!"
- sed -e "s|%HDFS_DATANODE_DIR%|file:${HDFS_DATA_DIRECTORY_ROOT}1/hdfs/data,file:${HDFS_DATA_DIRECTORY_ROOT}2/hdfs/data|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-else
- sed -e "s|%HDFS_DATANODE_DIR%|${HDFS_DATA_DIRECTORY_ROOT}/hdfs/data|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-fi
-
-if [ "${ENABLE_QUORUM}" = "true" ] ; then
- sed -e "s|%ZKSERVER1%|${ZKSERVER_1}|" -e "s|%ZKSERVER2%|${ZKSERVER_2}|" -e "s|%ZKSERVER3%|${ZKSERVER_3}|" --in-place=.orig ${script_dir}/templates/${ZOO_CFG}
- sed -e "s|%ZKSERVER1%|${ZKSERVER_1}|" -e "s|%ZKSERVER2%|${ZKSERVER_2}|" -e "s|%ZKSERVER3%|${ZKSERVER_3}|" --in-place=.orig ${script_dir}/templates/${CORE_SITE}
-
- sed -e "s|%NAMENODE%|${NAMENODE}|" -e "s|%SECONDARY_NAMENODE%|${SECONDARY_NAMENODE}|" -e "s|%JOURNALNODE1%|${JOURNALNODE_1}|" -e "s|%JOURNALNODE2%|${JOURNALNODE_2}|" \
- -e "s|%JOURNALNODE3%|${JOURNALNODE_3}|" -e "s|%JOURNAL_DATA_ROOT%|${JOURNAL_DATA_ROOT}|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-
- sed -e "s|%ZOO_LOG_DIR%|${ZOO_LOG_DIR}|" --in-place=.orig ${script_dir}/templates/${ZOO_LOG}
-fi
-
-if [ "${YARN}" = "true" ] ; then
- if [ "${DCA_SETUP}" = true ]; then
- RESOURCE_MEM="8192"
- elif [ "${DCA_SETUP}" = "half" ]; then
- RESOURCE_MEM="8192"
- else
- RESOURCE_MEM="2048"
- fi
- sed -e "s|%RESOURCE_MEM%|${RESOURCE_MEM}|" --in-place=.orig ${script_dir}/templates/yarn-site.xml
- sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" --in-place=.orig ${script_dir}/templates/yarn-env.sh
-fi
-
-rm ${script_dir}/templates/slaves
-for datanode in ${DATANODES}; do
- echo "${datanode}" >> ${script_dir}/templates/slaves
-done
-
-rm ${script_dir}/templates/masters
-echo "${NAMENODE}" >> ${script_dir}/templates/masters
-echo "${SECONDARY_NAMENODE}" >> ${script_dir}/templates/masters
-
-##
-## Install hdfs rpms
-##
-
-for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
- echo ""
- echo "Remove old gphd directories in $node"
- ssh $node "rm -rf /var/log/gphd/* /usr/lib/gphd/*"
-done
-
-for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
- echo ""
- echo "Passing YUM repo file to $node"
- scp /etc/yum.repos.d/gphd.repo $node:/etc/yum.repos.d/gphd.repo
-
- echo "Installing hdfs in $node"
- ssh $node 'yum --disablerepo "*" --enablerepo "gphd" list available; \
- yum install -y hadoop-conf-pseudo; \
- yum install -y zookeeper-server;'
-
- ##
- ## Post rpm installation processing
- ##
-
- echo "cp libexec files in $node"
- ssh $node "cp ${HADOOP_HOME}/libexec/* ${HADOOP_HOME}-hdfs/libexec"
- ssh $node "cp /usr/lib/gphd/hadoop/libexec/* /usr/lib/gphd/hadoop-yarn/libexec/"
-
- echo "scp slaves, masters, and core-site.xml file to $node"
- scp ${script_dir}/templates/slaves $node:${HADOOP_CONF_DIR}/slaves
- scp ${script_dir}/templates/masters $node:${HADOOP_CONF_DIR}/masters
- scp ${script_dir}/templates/${CORE_SITE} $node:${HADOOP_CONF_DIR}/core-site.xml
- scp ${script_dir}/templates/${HADOOP_ENV} $node:${HADOOP_CONF_DIR}/hadoop-env.sh
-
- if [ "${YARN}" = "true" ] ; then
- scp ${script_dir}/templates/yarn-env.sh $node:${HADOOP_CONF_DIR}/yarn-env.sh
- scp ${script_dir}/templates/yarn-site.xml $node:${HADOOP_CONF_DIR}/yarn-site.xml
-
- ssh $node mkdir ${HADOOP_YARN_HOME}/logs
- ssh $node chmod 777 ${HADOOP_YARN_HOME}/logs
- ssh $node chmod 777 /var/log/gphd/hadoop-yarn
- fi
-
- if [ "${ENABLE_QUORUM}" = "true" ] ; then
- ssh $node "mkdir /usr/lib/gphd/zookeeper/etc; \
- ln -s /etc/gphd/zookeeper/conf /usr/lib/gphd/zookeeper/etc/zookeeper; \
- echo ZOO_LOG_DIR=/tmp/zookeeper/logs/ >> ${ZK_HOME}/bin/zkEnv.sh"
- scp ${script_dir}/templates/${ZOO_CFG} $node:${ZK_CONF_DIR}/zoo.cfg
- scp ${script_dir}/templates/${ZOO_LOG} $node:${ZK_CONF_DIR}/log4j.properties
-
- if [ $node == $ZKSERVER_1 ] ; then
- ZK_NUM=1
- elif [ $node == $ZKSERVER_2 ] ; then
- ZK_NUM=2
- elif [ $node == $ZKSERVER_3 ] ; then
- ZK_NUM=3
- fi
-
- if [ $node == $ZKSERVER_1 ] || [ $node == $ZKSERVER_2 ] || [ $node == $ZKSERVER_3 ] ; then
- ssh $node "mkdir -p /tmp/zookeeper/logs && chown -R zookeeper.hadoop /tmp/zookeeper/ && echo ${ZK_NUM} > /tmp/zookeeper/myid; \
- grep ZK_HOME ~/.bash_profile >> /dev/null; \
- if [ $? != 0 ] ; then \
- echo 'export ZK_HOME=/usr/lib/gphd/zookeeper' >> ~/.bash_profile; \
- echo 'export ZOO_LOG_DIR=/tmp/zookeeper/logs' >> ~/.bash_profile; \
- fi"
- fi
- fi
-
- scp ${script_dir}/templates/${HDFS_SITE} $node:${HADOOP_CONF_DIR}/hdfs-site.xml
-
- ssh $node mkdir ${HADOOP_HOME}/logs
- ssh $node chmod 777 ${HADOOP_HOME}/logs
- echo ""
-done
-
-if [ "${YARN}" = "true" ] ; then
- for node in ${RESOURCEMANAGER} ${NODEMANAGERS}; do
- ssh ${node} cp /etc/gphd/hadoop/conf.empty/capacity-scheduler.xml /usr/lib/gphd/hadoop/etc/hadoop/
- done
-fi
-
-if [ "${ENABLE_QUORUM}" = "true" ] ; then
- if [ -n "${KDC}" ] ; then
- HDFS_DN_SSH_COMMAND=""
- else
- HDFS_DN_SSH_COMMAND="sudo -u hdfs"
- fi
-
- for zkserver in ${ZKSERVER_1} ${ZKSERVER_2} ${ZKSERVER_3}; do
- ssh $zkserver "sudo JAVA_HOME=${JAVA_HOME} /etc/init.d/zookeeper-server start"
- done
-
- for journalnode in ${JOURNALNODE_1} ${JOURNALNODE_2} ${JOURNALNODE_3}; do
- ssh $journalnode "rm -rf ${JOURNAL_DATA_ROOT}; mkdir ${JOURNAL_DATA_ROOT}; chown hdfs.hdfs ${JOURNAL_DATA_ROOT};\
- JAVA_HOME=${JAVA_HOME} ${HDFS_DN_SSH_COMMAND} ${HADOOP_HOME}/sbin/hadoop-daemon.sh start journalnode"
- done
-
- ssh ${NAMENODE} "\
- JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs namenode -format -force 2>&1 | tee ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
- grep 'Exiting with status 0' ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
- if [ \$? != 0 ]; then \
- echo 'FATAL: format failed.'; \
- exit 1; \
- fi;
- sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh start namenode"
-
- ssh ${SECONDARY_NAMENODE} "\
- JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs namenode -bootstrapStandby 2>&1 | tee ${HADOOP_HOME}/logs/hdfs-namenode-bootstrapStandby.out; \
- grep 'Exiting with status 0' ${HADOOP_HOME}/logs/hdfs-namenode-bootstrapStandby.out; \
- if [ \$? != 0 ]; then \
- echo 'FATAL: bootstrapStandby failed.'; \
- exit 1; \
- fi;
- sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh start namenode"
-
- ssh ${NAMENODE} "JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs zkfc -formatZK -force"
- for namenode in ${NAMENODE} ${SECONDARY_NAMENODE}; do
- ssh $namenode "JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh start zkfc"
- done
-
- for datanode in ${DATANODES}; do
- ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
- done
-else
- ##
- ## Format hdfs file system
- ##
-
- ssh ${NAMENODE} "\
- sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs namenode -format -force 2>&1 | tee ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
- grep 'Exiting with status 0' ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
- if [ \$? != 0 ]; then \
- echo 'FATAL: format failed.'; \
- exit 1; \
- fi"
-
- ##
- ## Startup hadoop namenode and datanode services
- ##
-
- ssh ${NAMENODE} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
-
- ssh ${SECONDARY_NAMENODE} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start secondarynamenode
-
- for datanode in ${DATANODES}; do
- if [ -n "${KDC}" ] ; then
- ssh $datanode /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
- else
- ssh $datanode sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
- fi
- done
-
- if [ "${YARN}" = "true" ] ; then
- ssh ${PHD_USER}@${RESOURCEMANAGER} JAVA_HOME=${JAVA_HOME} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start resourcemanager
- for node in ${NODEMANAGERS}; do
- ssh ${PHD_USER}@${node} JAVA_HOME=${JAVA_HOME} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start nodemanager
- done
- fi
-fi
-
-##
-## Before exiting, display running java processes with "jps" command
-##
-
-display_jps
-
-exit 0
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/ha_failover.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/ha_failover.sh b/depends/libyarn/releng/bin/ha_failover.sh
deleted file mode 100644
index f596041..0000000
--- a/depends/libyarn/releng/bin/ha_failover.sh
+++ /dev/null
@@ -1,36 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-nn1_status=`/usr/bin/hdfs haadmin -getServiceState nn1`
-echo nn1 status is: $nn1_status
-nn2_status=`/usr/bin/hdfs haadmin -getServiceState nn2`
-echo nn2 status is: $nn2_status
-if [ "${nn1_status}" = "active" ]; then
- echo "hdfs haadmin -failover nn1 nn2"
- hdfs haadmin -failover nn1 nn2
-elif [ "${nn1_status}" = "standby" ]; then
- echo "hdfs haadmin -failover nn2 nn1"
- hdfs haadmin -failover nn2 nn1
-else
- echo "Can't get valid status of nn1, exit."
- exit 1
-fi
-nn1_status=`/usr/bin/hdfs haadmin -getServiceState nn1`
-echo nn1 status now is: $nn1_status
-nn2_status=`/usr/bin/hdfs haadmin -getServiceState nn2`
-echo nn2 status now is: $nn2_status
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/install-phd.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/install-phd.sh b/depends/libyarn/releng/bin/install-phd.sh
deleted file mode 100644
index 4d63191..0000000
--- a/depends/libyarn/releng/bin/install-phd.sh
+++ /dev/null
@@ -1,71 +0,0 @@
-#!/bin/sh
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-if [ -z ${DATA_PATH} ]; then
- echo "DATA_PATH not set"
- exit 1
-fi
-
-if [ -d ${DATA_PATH} ]; then
- cd ${DATA_PATH}
-else
- echo "DATA_PATH not exist"
- exit 1
-fi
-
-HADOOP_TARBALL=`curl --silent --no-buffer "http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/latest/?C=M;O=D" | grep -o 'PHD-2.[0-9].[0-9].[0-9]-bin-[0-9]*.tar.gz' | head -n 1`
-echo "HADOOP tarball: " ${HADOOP_TARBALL}
-
-HADOOP_URL="http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/latest/"${HADOOP_TARBALL}
-echo "Download HADOOP from " ${HADOOP_URL}
-
-curl --silent -o ${DATA_PATH}/${HADOOP_TARBALL} -L ${HADOOP_URL}
-
-tar -xzf ${DATA_PATH}/${HADOOP_TARBALL}
-
-HADOOP_PACKAGE=`echo ${HADOOP_TARBALL} | grep -o 'PHD-[0-9].[0-9].[0-9].[0-9]-bin-[0-9]*'`
-HADOOP_VERSION=`ls ${DATA_PATH}/${HADOOP_PACKAGE}/hadoop/tar/*.tar.gz | grep -o 'hadoop-[0-9].[0-9].[0-9]-[A-Za-z0-9\-]*-[0-9].[0-9].[0-9].[0-9]'`
-echo "HADOOP version: " ${HADOOP_VERSION}
-
-if [ -z ${HADOOP_VERSION} ]; then
- echo "cannot get HADOOP version"
- exit 1
-fi
-
-tar -xzf ${HADOOP_PACKAGE}/hadoop/tar/${HADOOP_VERSION}.tar.gz
-
-if [ -z ${HDFS_CONFIG_PATH} ]; then
- echo "HDFS_CONFIG_PATH not set"
- exit 1
-fi
-
-if [ -f ${HDFS_CONFIG_PATH} ]; then
- cp -f ${HDFS_CONFIG_PATH} ${DATA_PATH}/${HADOOP_VERSION}/etc/hadoop/
-else
- echo "HDFS_CONFIG_PATH not a file"
- exit 1
-fi
-
-HADOOP_BIN=${DATA_PATH}/${HADOOP_VERSION}/bin
-HADOOP_SBIN=${DATA_PATH}/${HADOOP_VERSION}/sbin
-
-${HADOOP_BIN}/hdfs namenode -format
-${HADOOP_SBIN}/hadoop-daemon.sh start namenode
-${HADOOP_SBIN}/hadoop-daemon.sh start datanode
-${HADOOP_BIN}/hdfs dfs -mkdir hdfs://localhost:9000/user
-${HADOOP_BIN}/hdfs dfs -chmod 777 hdfs://localhost:9000/user
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/setup_gphdinst.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/setup_gphdinst.sh b/depends/libyarn/releng/bin/setup_gphdinst.sh
deleted file mode 100644
index 64a82c1..0000000
--- a/depends/libyarn/releng/bin/setup_gphdinst.sh
+++ /dev/null
@@ -1,175 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-#
-# Usage: ./setup_gphdinst.sh setup / start HK / stop HK
-# HK means this is for HA/Kerberos
-#
-
-JAVA_HOME=${JAVA_HOME:="/opt/jdk1.7.0_15"}
-HADOOP_HOME="/usr/lib/gphd/hadoop"
-LOCAL_HDFS_RPMINSTALL="true"
-DCA_SETUP=${DCA_SETUP:=false}
-TWO_NAMENODE_DIR=${TWO_NAMENODE_DIR:=false}
-SEGMENT_SYSTEM_LIST=${SEGMENT_SYSTEM_LIST:="sdw1 sdw2 sdw3"}
-MASTERHOST=${MASTERHOST:=mdw}
-STANDBYHOST=${STANDBYHOST:=smdw}
-DATANODES_LIST=${SEGMENT_SYSTEM_LIST}
-USER="gpadmin"
-STEP="$1"
-ENABLE_QUORUM=${ENABLE_QUORUM:=true}
-
-if [ "${ENABLE_QUORUM}" = "true" ]; then
- ZKSERVER_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f1-3 )
-
- # Deploy zookeeper and jounalnode depend on datanodes number.
- datanode_num=0
-
- for node in ${SEGMENT_SYSTEM_LIST}; do
- datanode_num=`expr $datanode_num + 1`
- done
-
- if [ $datanode_num -ge 6 ]; then
- JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f4-6 )
- elif [ $datanode_num -eq 5 ]; then
- JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f3-5 )
- elif [ $datanode_num -eq 4 ]; then
- JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f2-4 )
- else
- JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f1-3 )
- fi
- echo "ZKSERVER_LIST is ${ZKSERVER_LIST}"
- echo "JOURNALNODE_LIST is ${JOURNALNODE_LIST}"
-fi
-
-if [ "${STEP}" = "setup" ]; then
- PHD_PACKAGE=$( curl --silent --no-buffer "http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/testing/?C=N;O=D" | grep -o 'PHD-[0-9].[0-9].[0-9].[0-9]-[0-9]*.tar.gz' | head -n 1 )
- HDFS_DOWNLOAD_URL="http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/testing/${PHD_PACKAGE}"
-
- echo "Hadoop Package Download URL: ${HDFS_DOWNLOAD_URL}"
- echo "Setting up latest HDFS..."
-
- sudo DATANODES="${DATANODES_LIST}" \
- NAMENODE=${STANDBYHOST} \
- SECONDARY_NAMENODE=${MASTERHOST} \
- RESOURCEMANAGER=${STANDBYHOST} \
- NODEMANAGERS="${DATANODES_LIST}" \
- ARCHIVE_URL=${HDFS_DOWNLOAD_URL} \
- JAVA_HOME=${JAVA_HOME} \
- DCA_SETUP=${DCA_SETUP} \
- TWO_NAMENODE_DIR=${TWO_NAMENODE_DIR} \
- KDC=${KDC} \
- ENABLE_QUORUM=${ENABLE_QUORUM} \
- YARN=${YARN} \
- PHD_USER=${USER} \
- bash gphdinst.sh
-fi
-
-if [ -n "${KDC}" ] ; then
- HDFS_DN_SSH_COMMAND="sudo"
-elif [ "$2" = "HK" ]; then
- HDFS_DN_SSH_COMMAND="sudo"
-else
- HDFS_DN_SSH_COMMAND="sudo -u hdfs"
-fi
-
-if [ "${STEP}" = "stop" ]; then
- if [ -f "${HADOOP_HOME}/bin/stop-all.sh" ]; then
- ${HADOOP_HOME}/bin/stop-all.sh
- fi
-
- if [ "${LOCAL_HDFS_RPMINSTALL}" = "true" ]; then
- if [ "${USE_MASTER_AS_DATANODE}" = "true" ]; then
- DATANODES_LIST="${SEGMENT_SYSTEM_LIST} ${MASTERHOST}"
- else
- DATANODES_LIST=${SEGMENT_SYSTEM_LIST}
- fi
-
- if [ "${ENABLE_QUORUM}" = "true" ]; then
- echo "Stopping local HDFS HA cluster..."
- for datanode in ${DATANODES_LIST}; do
- ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop datanode
- done
-
- for node in ${JOURNALNODE_LIST}; do
- ssh $node ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop journalnode
- done
-
- for node in ${ZKSERVER_LIST}; do
- ssh $node sudo JAVA_HOME=${JAVA_HOME} /etc/init.d/zookeeper-server stop
- done
-
- sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop zkfc
- ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop zkfc
- sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop namenode
- ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop namenode
- else
- echo "Stopping local HDFS cluster..."
-
- sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop secondarynamenode
-
- for datanode in ${DATANODES_LIST}; do
- ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop datanode
- if [ "${YARN}" = "true" ] ; then
- ssh $datanode "/usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh stop nodemanager"
- fi
- done
-
- ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop namenode
- if [ "${YARN}" = "true" ] ; then
- ssh ${STANDBYHOST} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh stop resourcemanager
- fi
- fi
- fi
-fi
-
-if [ "${STEP}" = "start" ]; then
-
- if [ "${ENABLE_QUORUM}" = "true" ]; then
- echo "Starting up local HDFS HA"
-
- for node in ${JOURNALNODE_LIST}; do
- ssh $node ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start journalnode
- done
-
- for node in ${ZKSERVER_LIST}; do
- ssh $node sudo JAVA_HOME=${JAVA_HOME} /etc/init.d/zookeeper-server start
- done
-
- ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start zkfc
- sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start zkfc
- ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
- sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
-
- for datanode in ${DATANODES_LIST}; do
- ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
- done
- else
- echo "Starting up local HDFS"
- ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
- if [ "${YARN}" = "true" ]; then
- ssh ${STANDBYHOST} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start resourcemanager
- fi
- for datanode in ${DATANODES_LIST}; do
- ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
- if [ "${YARN}" = "true" ]; then
- ssh $datanode /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start nodemanager
- fi
- done
- sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start secondarynamenode
- fi
-fi
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp b/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp
deleted file mode 100644
index 43e78f7..0000000
Binary files a/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp and /dev/null differ
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site-ha.xml b/depends/libyarn/releng/bin/templates/core-site-ha.xml
deleted file mode 100644
index ef31461..0000000
--- a/depends/libyarn/releng/bin/templates/core-site-ha.xml
+++ /dev/null
@@ -1,28 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>fs.defaultFS</name>
- <value>hdfs://gphd-cluster</value>
- </property>
- <property>
- <name>ha.zookeeper.quorum</name>
- <value>%ZKSERVER1%:2181,%ZKSERVER2%:2181,%ZKSERVER3%:2181</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml b/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml
deleted file mode 100755
index cf96ce1..0000000
--- a/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml
+++ /dev/null
@@ -1,39 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>hadoop.security.authentication</name>
- <value>kerberos</value>
- </property>
-
- <property>
- <name>hadoop.security.authorization</name>
- <value>true</value>
- </property>
-
- <property>
- <name>fs.defaultFS</name>
- <value>hdfs://gphd-cluster</value>
- </property>
-
- <property>
- <name>ha.zookeeper.quorum</name>
- <value>%ZKSERVER1%:2181,%ZKSERVER2%:2181,%ZKSERVER3%:2181</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site-secure.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site-secure.xml b/depends/libyarn/releng/bin/templates/core-site-secure.xml
deleted file mode 100755
index 1424ebb..0000000
--- a/depends/libyarn/releng/bin/templates/core-site-secure.xml
+++ /dev/null
@@ -1,34 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>fs.default.name</name>
- <value>hdfs://%HDFS_HOST%</value>
- </property>
-
- <property>
- <name>hadoop.security.authentication</name>
- <value>kerberos</value>
- </property>
-
- <property>
- <name>hadoop.security.authorization</name>
- <value>true</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site.xml b/depends/libyarn/releng/bin/templates/core-site.xml
deleted file mode 100644
index 120e1d4..0000000
--- a/depends/libyarn/releng/bin/templates/core-site.xml
+++ /dev/null
@@ -1,24 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>fs.default.name</name>
- <value>hdfs://%HDFS_HOST%</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/download-config
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/download-config b/depends/libyarn/releng/bin/templates/download-config
deleted file mode 100644
index e517b49..0000000
--- a/depends/libyarn/releng/bin/templates/download-config
+++ /dev/null
@@ -1,5 +0,0 @@
-hawq-1.1.0=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.0.1/PHD-1.0.1.0-19.tar.gz
-hawq-1.1.1=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.0.2/PHD-1.0.2.0-7.tar.gz
-hawq-1.1.2=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.0.3/PHD-1.0.3.0-66.tar.gz
-hawq-1.1.3=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.1.0/PHD-1.1.0.0-76.tar.gz
-hawq-1.1.4=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/testing/PHD-1.1.1.0-82.tar.gz
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh b/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh
deleted file mode 100644
index ede6744..0000000
--- a/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh
+++ /dev/null
@@ -1,72 +0,0 @@
-# Copyright 2011 The Apache Software Foundation
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME. All others are
-# optional. When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=%JAVA_HOME%
-export HADOOP_HOME=%HADOOP_HOME%
-export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
-export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements. Automatically insert capacity-scheduler.
-for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
- if [ "$HADOOP_CLASSPATH" ]; then
- export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
- else
- export HADOOP_CLASSPATH=$f
- fi
-done
-
-# Extra Java runtime options. Empty by default.
-export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
-
-# Command specific options appended to HADOOP_OPTS when specified
-export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %NAMENODE_MEMORY% $HADOOP_NAMENODE_OPTS"
-export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS %DATANODE_MEMORY% -Xss256k $HADOOP_DATANODE_OPTS"
-
-export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %SECONDARYNAMENODE_MEMORY% $HADOOP_SECONDARYNAMENODE_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-export HADOOP_CLIENT_OPTS="-Xmx128m $HADOOP_CLIENT_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges
-export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}
-
-# Where log files are stored. $HADOOP_HOME/logs by default.
-#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-# The directory where pid files are stored. /tmp by default.
-# NOTE: this should be set to a directory that can only be written to by
-# the user that will run the hadoop daemons. Otherwise there is the
-# potential for a symlink attack.
-export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-export HADOOP_IDENT_STRING=$USER
-
-umask 022
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh b/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh
deleted file mode 100755
index 03f9900..0000000
--- a/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh
+++ /dev/null
@@ -1,73 +0,0 @@
-# Copyright 2011 The Apache Software Foundation
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME. All others are
-# optional. When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=%JAVA_HOME%
-export HADOOP_HOME=%HADOOP_HOME%
-export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
-export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements. Automatically insert capacity-scheduler.
-for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
- if [ "$HADOOP_CLASSPATH" ]; then
- export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
- else
- export HADOOP_CLASSPATH=$f
- fi
-done
-
-# Extra Java runtime options. Empty by default.
-export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
-
-# Command specific options appended to HADOOP_OPTS when specified
-export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %NAMENODE_MEMORY% $HADOOP_NAMENODE_OPTS"
-export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS %DATANODE_MEMORY% -Xss256k $HADOOP_DATANODE_OPTS"
-
-export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %SECONDARYNAMENODE_MEMORY% $HADOOP_SECONDARYNAMENODE_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-export HADOOP_CLIENT_OPTS="-Xmx128m $HADOOP_CLIENT_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges
-export HADOOP_SECURE_DN_USER=hdfs
-export JSVC_HOME=/usr/bin
-
-# Where log files are stored. $HADOOP_HOME/logs by default.
-#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-# The directory where pid files are stored. /tmp by default.
-# NOTE: this should be set to a directory that can only be written to by
-# the user that will run the hadoop daemons. Otherwise there is the
-# potential for a symlink attack.
-export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-export HADOOP_IDENT_STRING=$USER
-
-umask 022
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hadoop-env.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hadoop-env.sh b/depends/libyarn/releng/bin/templates/hadoop-env.sh
deleted file mode 100644
index ede6744..0000000
--- a/depends/libyarn/releng/bin/templates/hadoop-env.sh
+++ /dev/null
@@ -1,72 +0,0 @@
-# Copyright 2011 The Apache Software Foundation
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME. All others are
-# optional. When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=%JAVA_HOME%
-export HADOOP_HOME=%HADOOP_HOME%
-export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
-export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements. Automatically insert capacity-scheduler.
-for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
- if [ "$HADOOP_CLASSPATH" ]; then
- export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
- else
- export HADOOP_CLASSPATH=$f
- fi
-done
-
-# Extra Java runtime options. Empty by default.
-export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
-
-# Command specific options appended to HADOOP_OPTS when specified
-export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %NAMENODE_MEMORY% $HADOOP_NAMENODE_OPTS"
-export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS %DATANODE_MEMORY% -Xss256k $HADOOP_DATANODE_OPTS"
-
-export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %SECONDARYNAMENODE_MEMORY% $HADOOP_SECONDARYNAMENODE_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-export HADOOP_CLIENT_OPTS="-Xmx128m $HADOOP_CLIENT_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges
-export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}
-
-# Where log files are stored. $HADOOP_HOME/logs by default.
-#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-# The directory where pid files are stored. /tmp by default.
-# NOTE: this should be set to a directory that can only be written to by
-# the user that will run the hadoop daemons. Otherwise there is the
-# potential for a symlink attack.
-export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-export HADOOP_IDENT_STRING=$USER
-
-umask 022
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml b/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml
deleted file mode 100755
index 966234d..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml
+++ /dev/null
@@ -1,240 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.encrypt.data.transfer</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.encrypt.data.transfer.algorithm</name>
- <value>rc4</value>
- <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
- </property>
- <property>
- <name>dfs.datanode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.address</name>
- <value>0.0.0.0:1004</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:1006</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.journalnode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.journalnode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.journalnode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.principal</name>
- <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.http.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.web.authentication.kerberos.principal</name>
- <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.web.authentication.kerberos.keytab</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
-
- <property>
- <name>dfs.nameservices</name>
- <value>gphd-cluster</value>
- </property>
- <property>
- <name>dfs.ha.namenodes.gphd-cluster</name>
- <value>nn1,nn2</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.shared.edits.dir</name>
- <value>qjournal://%ZKSERVER1%:8485;%ZKSERVER2%:8485;%ZKSERVER3%:8485/gphd-cluster</value>
- </property>
- <property>
- <name>dfs.journalnode.edits.dir</name>
- <value>%JOURNAL_DATA_ROOT%</value>
- </property>
- <property>
- <name>dfs.client.failover.proxy.provider.gphd-cluster</name>
- <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
- </property>
- <property>
- <name>dfs.ha.fencing.methods</name>
- <value>shell(/bin/true)</value>
- </property>
- <property>
- <name>dfs.ha.fencing.ssh.connect-timeout</name>
- <value>20000</value>
- </property>
- <property>
- <name>dfs.ha.automatic-failover.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.namenode.fs-limits.min-block-size</name>
- <value>1024</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml b/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml
deleted file mode 100755
index 89b33cc..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml
+++ /dev/null
@@ -1,193 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.encrypt.data.transfer</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.encrypt.data.transfer.algorithm</name>
- <value>rc4</value>
- <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
- </property>
- <property>
- <name>dfs.datanode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.principal</name>
- <value>hdfs/%DATANODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.kerberos.http.principal</name>
- <value>HTTP/%DATANODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.datanode.address</name>
- <value>0.0.0.0:1004</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:1006</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.principal</name>
- <value>hdfs/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.http.principal</name>
- <value>HTTP/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.secondary.namenode.keytab.file</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.principal</name>
- <value>hdfs/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.http.principal</name>
- <value>HTTP/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.secondary.namenode.kerberos.internal.spnego.principal</name>
- <value>HTTP/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
- </property>
-
- <property>
- <name>dfs.block.access.token.enable</name>
- <value>true</value>
- </property>
-
- <property>
- <name>dfs.web.authentication.kerberos.principal</name>
- <value>HTTP/%WEBAUTH_HOST%@HAWQ.PIVOTAL.COM</value>
- </property>
- <property>
- <name>dfs.web.authentication.kerberos.keytab</name>
- <value>/etc/gphd/hawq-krb5.keytab</value>
- </property>
-
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
-</configuration>
http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml b/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml
deleted file mode 100644
index 7fb7c1d..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml
+++ /dev/null
@@ -1,172 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- <property>
- <name>dfs.web.ugi</name>
- <value>hdfs,hdfs</value>
- <description>The user account used by the web interface.</description>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.webhdfs.enabled</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.support.append</name>
- <value>true</value>
- </property>
- <property>
- <name>dfs.client.read.shortcircuit</name>
- <value>false</value>
- </property>
- <property>
- <name>dfs.block.local-path-access.user</name>
- <value>gpadmin</value>
- <description>
- Specify the user allowed to do short circuit read
- </description>
- </property>
- <property>
- <name>dfs.safemode.extension</name>
- <value>0</value>
- </property>
- <property>
- <name>dfs.safemode.min.datanodes</name>
- <value>1</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>%DATA_DIR_ROOT%</value>
- </property>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.dir</name>
- <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
- </property>
- <property>
- <name>dfs.namenode.checkpoint.period</name>
- <value>3600</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>%HDFS_DATANODE_DIR%</value>
- </property>
- <property>
- <name>dfs.datanode.data.dir.perm</name>
- <value>755</value>
- </property>
- <property>
- <name>dfs.replication</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.datanode.max.transfer.threads</name>
- <value>40960</value>
- </property>
- <property>
- <name>dfs.client.socket-timeout</name>
- <value>300000000</value>
- </property>
- <property>
- <name>dfs.datanode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>dfs.namenode.handler.count</name>
- <value>60</value>
- </property>
- <property>
- <name>ipc.client.connection.maxidletime</name>
- <value>3600000</value>
- </property>
- <property>
- <name>ipc.server.handler.queue.size</name>
- <value>3300</value>
- </property>
- <property>
- <name>ipc.client.connection</name>
- <value>3</value>
- </property>
- <property>
- <name>dfs.namenode.accesstime.precision</name>
- <value>-1</value>
- </property>
- <property>
- <name>dfs.datanode.http.address</name>
- <value>0.0.0.0:59075</value>
- <description>
- The datanode http server address and port. If the port is 0 then the server will start on a free port.
- </description>
- </property>
-
- <property>
- <name>dfs.nameservices</name>
- <value>gphd-cluster</value>
- </property>
- <property>
- <name>dfs.ha.namenodes.gphd-cluster</name>
- <value>nn1,nn2</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.rpc-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:9000</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn1</name>
- <value>%NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.http-address.gphd-cluster.nn2</name>
- <value>%SECONDARY_NAMENODE%:50070</value>
- </property>
- <property>
- <name>dfs.namenode.shared.edits.dir</name>
- <value>qjournal://%JOURNALNODE1%:8485;%JOURNALNODE2%:8485;%JOURNALNODE3%:8485/gphd-cluster</value>
- </property>
- <property>
- <name>dfs.journalnode.edits.dir</name>
- <value>%JOURNAL_DATA_ROOT%</value>
- </property>
- <property>
- <name>dfs.client.failover.proxy.provider.gphd-cluster</name>
- <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
- </property>
- <property>
- <name>dfs.ha.fencing.methods</name>
- <value>shell(/bin/true)</value>
- </property>
- <property>
- <name>dfs.ha.fencing.ssh.connect-timeout</name>
- <value>20000</value>
- </property>
- <property>
- <name>dfs.ha.automatic-failover.enabled</name>
- <value>true</value>
- </property>
-</configuration>